There continues to be research and proposed solutions to grammar checking and word usage errors. What these solutions have in common is the targeted design for specific types of errors. For example, in a word processor, grammar checking is one of the processes the word processor was designed to do effectively. Where there is a subject-word agreement problem where the number on the verb does not agree with the number of the subject, the checker can handle this with a high degree of success. These grammar checkers are designed to target a specific set of errors, such as subject-verb disagreement.
Similarly, with respect to learning the English language and grammar, a daunting task for non-English speaking users, typical systems for automatic error correction and detection are designed by identifying the typical errors made, and then developing specialized subsystems that target a specific error. For example, prepositions are difficult for learners because prepositional systems across languages are very different. Thus, a typical system design for automatic correction of learner language specifically looks at preposition errors. To do so, it may consider the assigned part of speech tag for each of the words in a sentence, and then identify noun phrases based on the tags. Further analysis checks whether a preposition precedes the noun phrase and determines which preposition would be the best choice, given the words and parts-of-speech in the context.
These more focused designs employ different models from which to check if the preposition entered by the user was correct, and if no preposition precedes the noun phrase, another model from which to determine if a preposition should be present in this context. The more focused designs are costly in terms of performance and maintenance, since each error type amongst a potentially large set requires its own targeted sub-component.
The following presents a simplified summary in order to provide a basic understanding of some novel embodiments described herein. This summary is not an extensive overview, and it is not intended to identify key/critical elements or to delineate the scope thereof. Its sole purpose is to present some concepts in a simplified form as a prelude to the more detailed description that is presented later.
The disclosed architecture is a general-purpose framework for the generation of fluent alternatives to dysfluent or ungrammatical word sequences (e.g., in a written sample), such as the correction of grammatical and word usage errors, whether the errors are related to fluency, prepositions, articles, form of verb, or other word choices. The architecture detects and corrects writing errors in a human language based on the utilization of three different stages: error detection, correction candidate generation, and correction candidate ranking in terms of fluency.
Error detection is addressed by a suite of language model related scores and possibly other scores that can identify a particularly unlikely sequence of words. Correction candidate generation is addressed by a lookup in a very large corpus of “correct” English that looks for alternative arrangements of the same or similar words or subsequences of these words in the same context. Correction candidate ranking is addressed by a language model ranker.
To the accomplishment of the foregoing and related ends, certain illustrative aspects are described herein in connection with the following description and the annexed drawings. These aspects are indicative of the various ways in which the principles disclosed herein can be practiced and all aspects and equivalents thereof are intended to be within the scope of the claimed subject matter. Other advantages and novel features will become apparent from the following detailed description when considered in conjunction with the drawings.
The disclosed architecture is a system that addresses grammatical and word usage correction in a general way (without targeting specific error types with specific mechanisms). The architecture detects and corrects any type of error that manifests itself in the choice or ordering of words, and thus, is capable of achieving higher coverage than current error detection systems or proofing tools.
Reference is now made to the drawings, wherein like reference numerals are used to refer to like elements throughout. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding thereof. It may be evident, however, that the novel embodiments can be practiced without these specific details. In other instances, well known structures and devices are shown in block diagram form in order to facilitate a description thereof. The intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the claimed subject matter.
The error detection component 102 detects the erroneous subsequence of words 108 based on scores from at least one of multiple language models or phrase count tables that detect an unlikely sequence of words. The error detection component 102 operates as the sliding window 110 over sequences of words of the sequence 104. The candidate selection component 118 selects a candidate string based on a score of the potential candidate strings 114 from an associated language model. The scores for error detection can be used as features in a classification or sequence model that learns how scores of the language models indicate presence of the erroneous subsequence of words 108. The corpus 116 used by the potential candidate generation component 112 can be represented as a suffix array to increase efficiency of lookup.
The potential candidate generation component 112 generates the potential candidate strings 114 based on context. The potential candidate generation component 112 generates the potential candidate strings 114 based on comparison of the number of words in the potential candidate strings 114 to the number of words in the erroneous subsequence of words 108. The potential candidate generation component 112 generates the potential candidate strings 114 based on the number of words in the potential candidate strings 114 that match words in the erroneous subsequence of words 108.
Note that although a single signal is described as one implementation of use, multiple signals can be generated and then combined for an overall measure of error detection. For example, one or more signals from linguistic analysis can be employed in addition to language model scores, such as signals based on parse trees and dependency tuples. Machine learning can then be employed in a classifier or sequence modeling approach that operates on a given word of the input sequence 104. Moreover, the resources accessible for error detection are not limited to language models and linguistic analysis, but can include many different sources of information, such as web-based frequency information.
The general-purpose error detection system 200 employs one or more large language models 204 (and/or word and phrase count tables) as resources, and calculates one or more scores computed from these resources in the sliding window 110 (over the subsequence of words 108) over the input sentence (sequence of words 104).
Generally, a sequence of reasonably common words, where the sequence itself is highly unlikely can indicate an area in the input string (e.g., sentence) where there could be one or more errors (e.g., word ordering). The scores can also be used as features in a classification or sequence modeling approach, which can be trained on annotated error data. Such an algorithm learns how the different language model scores provide a signal for the presence of an error. Output of the language model(s) 204 or, the optional classifier or sequence model 206, is to an error detection process 208 (of the error detection component 102). The error detection process 208 detects an error in the word sequence defined by the sliding window 110, generates and sends a score for presence of an error to a potential candidate generation phase 210.
The potential candidate generation phase 210 includes a potential candidate generation process 212 (of the potential candidate generation component 112) that generates the potential candidate strings 114 from the corpus 116. The algorithm for generating the potential candidate strings 114 can comprise a lookup in a very large corpus (e.g., the corpus 116). In one particular instantiation the corpus 116 can be represented as a suffix array, for efficient lookup.
Once the error detection phase 202 has determined that a sequence of words (defined by the sliding window 110) might contain one or more errors, the potential candidate strings 114 are generated. The potential candidate generation process 210 finds all potential candidate strings 114 in the corpus 116 based on one or more of the following criteria, for example (note that other criteria may be employed):
(i) the strings occur in the same context, where context can be defined as “preceded and followed by the same word or words”;
(ii) the strings are of the same length or within no more than m words compared to the original string;
(iii) the strings contain at least n words from the original string;
(iv) the strings contain words that are semantically similar to the words in the original string; and
(v) the strings contain words that are morphologically similar to the words in original string (i.e., contain the same stem, but a different suffix)
These criteria can be heuristically determined or can be optimized (i.e. optimal values for m can be found, optimal range of context can be defined) by maximizing the number of valid correction candidate strings 114 on an annotated corpus.
Once a set of potential correction candidate strings 114 has been identified, the candidate generation process 216 (of the potential candidate generation component 112) in a final candidate generation phase 214 selects and presents one or more of the potential candidate strings 114 as the candidate strings 120 for correction of the error(s) in the sequence 104. Selection can be by ranking the potential candidate strings 114 so that only the top t candidates are shown to the user. This phase 214 can be implemented as one or more language models 218 that rank the potential candidate strings 114 in the context of the complete user input. Only potential candidate strings that rank higher than the original input (e.g., word sequence defined by the window 110) are considered to be valid options, for example, and of those, only a subset may be presented to the user.
A language model assigns a probability to a sequence of m words.
Included herein is a set of flow charts representative of exemplary methodologies for performing novel aspects of the disclosed architecture. While, for purposes of simplicity of explanation, the one or more methodologies shown herein, for example, in the form of a flow chart or flow diagram, are shown and described as a series of acts, it is to be understood and appreciated that the methodologies are not limited by the order of acts, as some acts may, in accordance therewith, occur in a different order and/or concurrently with other acts from that shown and described herein. For example, those skilled in the art will understand and appreciate that a methodology could alternatively be represented as a series of interrelated states or events, such as in a state diagram. Moreover, not all acts illustrated in a methodology may be required for a novel implementation.
As used in this application, the terms “component” and “system” are intended to refer to a computer-related entity, either hardware, a combination of software and tangible hardware, software, or software in execution. For example, a component can be, but is not limited to, tangible components such as a processor, chip memory, mass storage devices (e.g., optical drives, solid state drives, and/or magnetic storage media drives), and computers, and software components such as a process running on a processor, an object, an executable, a data structure (stored in volatile or non-volatile storage media), a module, a thread of execution, and/or a program. By way of illustration, both an application running on a server and the server can be a component. One or more components can reside within a process and/or thread of execution, and a component can be localized on one computer and/or distributed between two or more computers. The word “exemplary” may be used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs.
Referring now to
The computing system 900 for implementing various aspects includes the computer 902 having processing unit(s) 904, a computer-readable storage such as a system memory 906, and a system bus 908. The processing unit(s) 904 can be any of various commercially available processors such as single-processor, multi-processor, single-core units and multi-core units. Moreover, those skilled in the art will appreciate that the novel methods can be practiced with other computer system configurations, including minicomputers, mainframe computers, as well as personal computers (e.g., desktop, laptop, etc.), hand-held computing devices, microprocessor-based or programmable consumer electronics, and the like, each of which can be operatively coupled to one or more associated devices.
The system memory 906 can include computer-readable storage (physical storage media) such as a volatile (VOL) memory 910 (e.g., random access memory (RAM)) and non-volatile memory (NON-VOL) 912 (e.g., ROM, EPROM, EEPROM, etc.). A basic input/output system (BIOS) can be stored in the non-volatile memory 912, and includes the basic routines that facilitate the communication of data and signals between components within the computer 902, such as during startup. The volatile memory 910 can also include a high-speed RAM such as static RAM for caching data.
The system bus 908 provides an interface for system components including, but not limited to, the system memory 906 to the processing unit(s) 904. The system bus 908 can be any of several types of bus structure that can further interconnect to a memory bus (with or without a memory controller), and a peripheral bus (e.g., PCI, PCIe, AGP, LPC, etc.), using any of a variety of commercially available bus architectures.
The computer 902 further includes machine readable storage subsystem(s) 914 and storage interface(s) 916 for interfacing the storage subsystem(s) 914 to the system bus 908 and other desired computer components. The storage subsystem(s) 914 (physical storage media) can include one or more of a hard disk drive (HDD), a magnetic floppy disk drive (FDD), and/or optical disk storage drive (e.g., a CD-ROM drive DVD drive), for example. The storage interface(s) 916 can include interface technologies such as EIDE, ATA, SATA, and IEEE 1394, for example.
One or more programs and data can be stored in the memory subsystem 906, a machine readable and removable memory subsystem 918 (e.g., flash drive form factor technology), and/or the storage subsystem(s) 914 (e.g., optical, magnetic, solid state), including an operating system 920, one or more application programs 922, other program modules 924, and program data 926.
The one or more application programs 922, other program modules 924, and program data 926 can include the entities and components of the system 100 of
Generally, programs include routines, methods, data structures, other software components, etc., that perform particular tasks or implement particular abstract data types. All or portions of the operating system 920, applications 922, modules 924, and/or data 926 can also be cached in memory such as the volatile memory 910, for example. It is to be appreciated that the disclosed architecture can be implemented with various commercially available operating systems or combinations of operating systems (e.g., as virtual machines).
The storage subsystem(s) 914 and memory subsystems (906 and 918) serve as computer readable media for volatile and non-volatile storage of data, data structures, computer-executable instructions, and so forth. Such instructions, when executed by a computer or other machine, can cause the computer or other machine to perform one or more acts of a method. The instructions to perform the acts can be stored on one medium, or could be stored across multiple media, so that the instructions appear collectively on the one or more computer-readable storage media, regardless of whether all of the instructions are on the same media.
Computer readable media can be any available media that can be accessed by the computer 902 and includes volatile and non-volatile internal and/or external media that is removable or non-removable. For the computer 902, the media accommodate the storage of data in any suitable digital format. It should be appreciated by those skilled in the art that other types of computer readable media can be employed such as zip drives, magnetic tape, flash memory cards, flash drives, cartridges, and the like, for storing computer executable instructions for performing the novel methods of the disclosed architecture.
A user can interact with the computer 902, programs, and data using external user input devices 928 such as a keyboard and a mouse. Other external user input devices 928 can include a microphone, an IR (infrared) remote control, a joystick, a game pad, camera recognition systems, a stylus pen, touch screen, gesture systems (e.g., eye movement, head movement, etc.), and/or the like. The user can interact with the computer 902, programs, and data using onboard user input devices 930 such a touchpad, microphone, keyboard, etc., where the computer 902 is a portable computer, for example. These and other input devices are connected to the processing unit(s) 904 through input/output (I/O) device interface(s) 932 via the system bus 908, but can be connected by other interfaces such as a parallel port, IEEE 1394 serial port, a game port, a USB port, an IR interface, short-range wireless (e.g., Bluetooth) and other personal area network (PAN) technologies, etc. The I/O device interface(s) 932 also facilitate the use of output peripherals 934 such as printers, audio devices, camera devices, and so on, such as a sound card and/or onboard audio processing capability.
One or more graphics interface(s) 936 (also commonly referred to as a graphics processing unit (GPU)) provide graphics and video signals between the computer 902 and external display(s) 938 (e.g., LCD, plasma) and/or onboard displays 940 (e.g., for portable computer). The graphics interface(s) 936 can also be manufactured as part of the computer system board.
The computer 902 can operate in a networked environment (e.g., IP-based) using logical connections via a wired/wireless communications subsystem 942 to one or more networks and/or other computers. The other computers can include workstations, servers, routers, personal computers, microprocessor-based entertainment appliances, peer devices or other common network nodes, and typically include many or all of the elements described relative to the computer 902. The logical connections can include wired/wireless connectivity to a local area network (LAN), a wide area network (WAN), hotspot, and so on. LAN and WAN networking environments are commonplace in offices and companies and facilitate enterprise-wide computer networks, such as intranets, all of which may connect to a global communications network such as the Internet.
When used in a networking environment the computer 902 connects to the network via a wired/wireless communication subsystem 942 (e.g., a network interface adapter, onboard transceiver subsystem, etc.) to communicate with wired/wireless networks, wired/wireless printers, wired/wireless input devices 944, and so on. The computer 902 can include a modem or other means for establishing communications over the network. In a networked environment, programs and data relative to the computer 902 can be stored in the remote memory/storage device, as is associated with a distributed system. It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between the computers can be used.
The computer 902 is operable to communicate with wired/wireless devices or entities using the radio technologies such as the IEEE 802.xx family of standards, such as wireless devices operatively disposed in wireless communication (e.g., IEEE 802.11 over-the-air modulation techniques) with, for example, a printer, scanner, desktop and/or portable computer, personal digital assistant (PDA), communications satellite, any piece of equipment or location associated with a wirelessly detectable tag (e.g., a kiosk, news stand, restroom), and telephone. This includes at least Wi-Fi (or Wireless Fidelity) for hotspots, WiMax, and Bluetooth™ wireless technologies. Thus, the communications can be a predefined structure as with a conventional network or simply an ad hoc communication between at least two devices. Wi-Fi networks use radio technologies called IEEE 802.11x (a, b, g, etc.) to provide secure, reliable, fast wireless connectivity. A Wi-Fi network can be used to connect computers to each other, to the Internet, and to wire networks (which use IEEE 802.3-related media and functions).
What has been described above includes examples of the disclosed architecture. It is, of course, not possible to describe every conceivable combination of components and/or methodologies, but one of ordinary skill in the art may recognize that many further combinations and permutations are possible. Accordingly, the novel architecture is intended to embrace all such alterations, modifications and variations that fall within the spirit and scope of the appended claims. Furthermore, to the extent that the term “includes” is used in either the detailed description or the claims, such term is intended to be inclusive in a manner similar to the term “comprising” as “comprising” is interpreted when employed as a transitional word in a claim.
This application claims the benefit of U.S. Provisional Patent application Ser. No. 61/391,197 entitled “GENERAL PURPOSE CORRECTION OF GRAMMATICAL AND WORD USAGE ERRORS” and filed Oct. 8, 2010, the entirety of which is incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
4674065 | Lange | Jun 1987 | A |
4730269 | Kucera | Mar 1988 | A |
4771401 | Kaufman et al. | Sep 1988 | A |
4868750 | Kucera et al. | Sep 1989 | A |
5477448 | Golding et al. | Dec 1995 | A |
5485372 | Golding et al. | Jan 1996 | A |
5490061 | Tolin et al. | Feb 1996 | A |
5537317 | Schabes et al. | Jul 1996 | A |
5577249 | Califano | Nov 1996 | A |
5799269 | Schabes | Aug 1998 | A |
5845306 | Schabes et al. | Dec 1998 | A |
5855000 | Waibel et al. | Dec 1998 | A |
6064959 | Young et al. | May 2000 | A |
6173252 | Qiu et al. | Jan 2001 | B1 |
6205261 | Goldberg | Mar 2001 | B1 |
6401060 | Critchlow et al. | Jun 2002 | B1 |
6424983 | Schabes et al. | Jul 2002 | B1 |
6618697 | Kantrowitz et al. | Sep 2003 | B1 |
6691088 | Blasig | Feb 2004 | B1 |
6816830 | Kempe | Nov 2004 | B1 |
7003446 | Trower et al. | Feb 2006 | B2 |
7043422 | Gao | May 2006 | B2 |
7120582 | Young et al. | Oct 2006 | B1 |
7165019 | Lee | Jan 2007 | B1 |
7366983 | Brill | Apr 2008 | B2 |
7447627 | Jessee et al. | Nov 2008 | B2 |
7539619 | Seligman | May 2009 | B1 |
7574348 | Hon et al. | Aug 2009 | B2 |
7587308 | Kasravi | Sep 2009 | B2 |
7680649 | Park | Mar 2010 | B2 |
7702512 | Gopinath et al. | Apr 2010 | B2 |
7774193 | Gao et al. | Aug 2010 | B2 |
7809744 | Nevidomski | Oct 2010 | B2 |
7835902 | Gamon et al. | Nov 2010 | B2 |
8086453 | Detlef et al. | Dec 2011 | B2 |
8170868 | Gamon et al. | May 2012 | B2 |
8386234 | Uchimoto et al. | Feb 2013 | B2 |
8661012 | Baker | Feb 2014 | B1 |
20020128821 | Ehsani | Sep 2002 | A1 |
20030036900 | Weise | Feb 2003 | A1 |
20030120481 | Murata et al. | Jun 2003 | A1 |
20030229497 | Wilson | Dec 2003 | A1 |
20040024601 | Gopinath et al. | Feb 2004 | A1 |
20040098263 | Hwang et al. | May 2004 | A1 |
20040122656 | Abir | Jun 2004 | A1 |
20040138881 | Divay et al. | Jul 2004 | A1 |
20050075877 | Minamino et al. | Apr 2005 | A1 |
20060241944 | Potter et al. | Oct 2006 | A1 |
20060247912 | Suzuki et al. | Nov 2006 | A1 |
20060277029 | Green et al. | Dec 2006 | A1 |
20060277031 | Ramsey | Dec 2006 | A1 |
20070219776 | Gamon et al. | Sep 2007 | A1 |
20080109209 | Fraser et al. | May 2008 | A1 |
20080162132 | Doulton | Jul 2008 | A1 |
20080208567 | Brockett et al. | Aug 2008 | A1 |
20080270118 | Kuo | Oct 2008 | A1 |
20090048833 | Fritsch et al. | Feb 2009 | A1 |
20100049498 | Cao et al. | Feb 2010 | A1 |
20100082348 | Silverman et al. | Apr 2010 | A1 |
20100138210 | Seo | Jun 2010 | A1 |
20100180198 | Iakobashvili | Jul 2010 | A1 |
20100185435 | Deshmukh et al. | Jul 2010 | A1 |
20110087961 | Fitusi | Apr 2011 | A1 |
20110161072 | Terao et al. | Jun 2011 | A1 |
20110184723 | Huang et al. | Jul 2011 | A1 |
20110202876 | Badger | Aug 2011 | A1 |
20110282667 | Hernandez-Abrego | Nov 2011 | A1 |
20120029910 | Medlock | Feb 2012 | A1 |
20120089387 | Gamon et al. | Apr 2012 | A1 |
20120101804 | Roth et al. | Apr 2012 | A1 |
20120239379 | Gershnik | Sep 2012 | A1 |
20130304453 | Fritch et al. | Nov 2013 | A9 |
Entry |
---|
Jager et al., “Learning Lexical Collocations with Concordancing and Scaffolding”, Retrieved at << http://clic.cimec.unitn.it/marco/publications/It—Journ—Lings.submitted.pdf >>,Jul. 7, 2010, pp. 95. |
Yi, et al., “AWeb-based English Proofing System for English as a Second Language Users”, Retrieved at << http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.68.9984&rep=rep1&type=pdf >>, Retrieved date: Sep. 28, 2010, pp. 6. |
Onnis, et al., “Generalizable distributional regularities aid fluent language processing: The case of semantic valence tendencies”, Retrieved at http://clic.cimec.unitn.it/marco/publications/It—Journ—Lings.submitted.pdf >>, Retrieved Date: Sep. 28, 2010, pp. 50. |
Athanaselis, T., S. Bakamidis, I. Dologlou, A novel technique for words reordering based on N-grams, 9th International Symposium on Signal Processing and Its Applications, ISSPA 2007, Feb. 12-15, 2007, pp. 1-4, Sharjah, United Arab Emirates. |
Dagan, I., L. Lee, F. C. N. Pereira, Similarity-based models of word cooccurrence probabilities, Machine Learning, Feb. 1999, vol. 34, No. 1-3, pp. 43-69. |
Wang, K., C. Thrasher, E. Viegas, X. Li, P. Hsu, An overview of Microsoft Web N-gram corpus and applications, Proc. of the NAACL HLT 2010: Demonstration Session, Jun. 2010, pp. 45-48, Los Angeles, California. |
Spooner, L., Notice of Allowance, U.S. Appl. No. 13/193,248, May 29, 2014, pp. 1-10. |
Number | Date | Country | |
---|---|---|---|
20120089387 A1 | Apr 2012 | US |
Number | Date | Country | |
---|---|---|---|
61391197 | Oct 2010 | US |