The present invention generally relates to the field of mark recognition. The present invention particularly relates to scanning an image of a document and performing mark recognition on the scanned image. In an embodiment, a response mark may be indicative of an intended response to a question and/or may be representative of coded information. In a preferred embodiment, the present invention provides a hardware independent methodology for evaluating the presence or absence of one or more marks on a document, which may be configured to higher or lower resolutions as needed.
Conventional mark reading systems have been employed to extract information, such as pencil marks, from a document. These systems typically include a scanner configured to read marks from the document. The scanner creates a digitized representation of the document. The digitized document is then processed by a separate application to evaluate the document. In an assessment environment, this system can be used to score examination responses. For example, many students and other examinees are familiar with the typical response format for answering multiple choice questions which requires the student to fill-in a space (e.g. a circle, square, etc.) associated with the question. Mark reading systems can grade such answer sheets significantly faster and more accurately than a human grader. However, conventional mark reading systems are limited by the accuracy with which the document is digitized. Increased accuracy is available with more sophisticated and more expensive systems. When a high degree of accuracy is required, expensive scanning equipment can be purchased which will provide a digitized representation of the document within most accepted error rates. However, in addition to the expense associated with such systems, they may not be easily configured to alternate environments where differing accuracy levels are required. When accuracy in the digitizing process is sacrificed, such systems often improperly grade test answers. For example, the systems may not recognize properly marked target areas that are incompletely filled or in which the pencil marking is lighter than a threshold value. Moreover, the systems may recognize stray marks and erasures as incorrect answers when a test taker did not intend such an answer.
A particularly egregious error may result when a test taker properly marks a correct answer for a question, but a mark reading system recognizes a stray mark or erasure as an incorrect answer for the same question. In this case, although the test taker correctly answered the question, the test taker would not receive credit because the system would report that the test taker provided two answers. As such, the test taker may be penalized because of the medium in which the test is presented rather than the substantive material on which the test is based.
What is needed is an improved image mark recognition system that can accurately evaluate the marks on a document to be digitized which is both cost effective and independent of the hardware being used to scan the document.
Before the present methods, systems, and materials are described, it is to be understood that this invention is not limited to the particular methodologies, systems and materials described, as these may vary. It is also to be understood that the terminology used in the description is for the purpose of describing the particular versions or embodiments only, and is not intended to limit the scope of the present invention which will be limited only by the appended claims.
It must also be noted that as used herein and in the appended claims, the singular forms “a,” “an,” and “the” include plural references unless the context clearly dictates otherwise. Thus, for example, reference to a “mark” is a reference to one or more marks and equivalents thereof known to those skilled in the art, and so forth. Unless defined otherwise, all technical and scientific terms used herein have the same meanings as commonly understood by one of ordinary skill in the art. Although any methods, materials, and devices similar or equivalent to those described herein can be used in the practice or testing of embodiments of the present invention, the preferred methods, materials, and devices are now described. All publications mentioned herein are incorporated by reference. Nothing herein is to be construed as an admission that the invention is not entitled to antedate such disclosure by virtue of prior invention.
The present invention relates to enhancing a scanned image of a document or more accurately performing mark recognition of the scanned image to determine the presence or absence of a mark. The methodology of the present invention begins after a document has been digitized. The digitized image is then mapped to a coordinate system. The mapping is performed on the digitized image by using trigger rows, which are determined using existing marks on one margin of the document. The trigger rows are then logically extended across the document. Next, the present invention completes the mapping process by logically creating columns orthogonal to the extended trigger rows. The number of columns created can vary and may be configured to the degree of accuracy required. Alternately, the mapping process may be performed by using corner marks (also known as anchor marks or crop marks), which are marks printed at each corner of the document. A coordinate system is then created by determining the number of rows desired, which may vary, and equally dividing the space into horizontal rows, and by determining the number of columns desired, which may vary, and equally dividing the space into vertical columns. Document deskew may also be performed. Those skilled in the art will understand that other methods for mapping the digitized image to a coordinate system may also be used.
Next, a resolver engine evaluates each cell on a pixel-by-pixel basis. The pixel-by-pixel analysis results in a value for each cell. The value may represent the opacity of a response mark present in the cell. The result is that the digitized image has been converted into a series of numeric values each representing a potential response mark on the original document.
In a preferred embodiment, a method of performing image mark recognition includes accessing a scanned image of a document containing a plurality of image marks, performing reference image mark recognition on the scanned image, performing coordinate mapping on the scanned image to determine coordinates for a plurality of cells within the scanned image, analyzing one or more pixels in each cell, and determining a cell color value for each cell based on values assigned to the one or more pixels. In an embodiment, performing reference image mark recognition includes determining a location for each of one or more alignment marks and may further include verifying image mark alignment and spacing based on a resolution of the scanned image and the locations of the one or more alignment marks. In an embodiment, determining the location of one or more alignment marks includes examining each corner of the scanned image. In an embodiment, determining the location of one or more alignment marks includes searching for trigger rows on at least one margin of the scanned image.
In an embodiment, performing coordinate mapping includes computing logical row coordinates for a number of rows in the scanned image, computing logical column coordinates for a number of columns in the scanned image, creating a mapping within the scanned image, and resolving the mapping into coordinate values for each of a plurality of cells. In an embodiment, computing logical row coordinates and logical column coordinates includes using one or more trigger marks as reference points. In an embodiment, computing logical row coordinates and logical column coordinates includes using one or more corner marks as reference points. The number of rows and the number of columns may be automatically determined based on the step of performing reference image mark recognition. Alternately, the number of rows and the number of columns may be pre-determined. In an embodiment, creating a mapping includes determining absolute pixel positions for each cell within the scanned document as defined by an intersection of a row and a column.
In an embodiment, analyzing one or more pixels includes incrementing a total color value by a color value for each pixel in a cell. Determining a cell color value may include setting the cell color value to the ratio of the total color value to a maximum color value.
In a preferred embodiment, a system for performing image mark recognition includes a processor and one or more computer readable mediums operatively coupled to the processor. At least one of the one or more computer readable mediums contains a scanned image of a document. At least one of the one or more computer readable mediums contains computer program instructions for performing a method of performing image mark recognition.
The accompanying drawings, which are incorporated in and form a part of the specification, illustrate preferred embodiments of the present invention and, together with the description serve to explain the principles of the invention. The embodiments illustrated in the drawings should not be read to constitute limiting requirements, but instead are intended to assist the reader in understanding the invention.
The present invention relates to resolving a scanner image of a document to perform mark recognition on the scanned image. The present invention provides a system and method for resolving a digitized image of a document, independent of the hardware used to create the digitized image. The digitized image is enlisted by the present invention and a sequence of values representing the presence or absence of response marks on the document is created. This sequence of values is then available for use by any application which may be created to enhance the document, such as a scoring application which may be used to return a score for an examination.
An aspect of the present invention involves mapping the digitized image of a document to a coordinate system. This coordinate mapping may be performed by detecting trigger row marks and/or corner/crop marks to create a logical document mapping. A trigger row mark is a mark on an edge of a document that denotes boundaries of a row logically extending across the document. In an assessment environment, responses may be entered within the logical rows. Each row for which a test taker can potentially enter one or more responses may have a separate trigger row mark. Alternatively, a trigger row mark may be used to denote every second row, third row, or any other pattern that may be used to logically subdivide a document. An alternate method of coordinate mapping includes use of corner/crop marks placed in each corner of the document. The crop marks are located and used to define the coordinate system by sub-dividing the space between the crop marks into a logical coordinate system. For example, adjacent crop marks along the vertical margin of the document may be subdivided into a predetermined number of equally sized rows. Similarly, adjacent crop marks along the horizontal margin of the document may be subdivided into a predetermined number of equally sized columns intersecting with the rows. The pre-determined number of rows and columns may be verified by analyzing the location of the response marks. Alternatively, a user may enter information regarding the document, such as a document number, from which the row and column placement is determined.
As shown in
Returning to
Two exemplary computations of a cell color value are shown in
Similar computations are used to compute the total color value, maximum color value and cell color value in
Preferably, a value of 0 may represent the lightest possible mark (e.g., white), and a value of 100 may represent the darkest possible mark (e.g., black). The machine-readable data values for each cell may optionally be stored 125 to an output file.
A disk controller 404 interfaces with one or more optional disk drives to the system bus 428. These disk drives may be external or internal floppy disk drives such as 410, CD ROM drives 406, or external or internal hard drives 408. As indicated previously, these various disk drives and disk controllers are optional devices.
Program instructions may be stored in the ROM 418 and/or the RAM 420. Optionally, program instructions may be stored on a computer readable medium such as a floppy disk or a digital disk or other recording medium, a communications signal or a carrier wave.
An optional display interface 422 may permit information from the bus 428 to be displayed on the display 424 in audio, graphic or alphanumeric format. Communication with external devices may optionally occur using various communication ports 426. An exemplary communication port 426 may be attached to a scanner 430 used to scan an image of a document into one or of the memory devices or disk drives.
In addition to the standard computer-type components, the hardware may also include an interface 412 which allows for receipt of data from input devices such as a keyboard 414 or other input device 416 such as a remote control, pointer and/or joystick.
An embedded system may optionally be used to perform one, some or all of the operations of the present invention. Likewise, a multiprocessor system may optionally be used to perform one, some or all of the operations of the present invention.
Although the invention has been described with reference to the preferred embodiments, it will be apparent to one skilled in the art that variations and modifications are contemplated within the spirit and scope of the invention. The drawings and description of the preferred embodiments are made by way of example rather than to limit the scope of the invention, and it is intended to cover within the spirit and scope of the invention all such changes and modifications.
This is a continuation application of U.S. Non-Provisional Patent Application No. 10/760,850, filed Jan. 20, 2004, titled “Method and System for Image Mark Recognition,” which was pending at the time the present application was filed and was issued as U.S. Pat. No. 7,298,902B2 on Nov. 20, 2007 and is hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
3601906 | Roche | Aug 1971 | A |
3870865 | Schneiderhan et al. | Mar 1975 | A |
4153895 | Weisbrod et al. | May 1979 | A |
4217487 | Kjeer | Aug 1980 | A |
4300123 | McMillin et al. | Nov 1981 | A |
4817179 | Buck | Mar 1989 | A |
4979136 | Weiman et al. | Dec 1990 | A |
5038393 | Nanba | Aug 1991 | A |
5101448 | Kawachiya et al. | Mar 1992 | A |
5102341 | Koslin | Apr 1992 | A |
5103490 | McMillin | Apr 1992 | A |
5129014 | Bloomberg | Jul 1992 | A |
5134669 | Keogh et al. | Jul 1992 | A |
5253304 | LeCun et al. | Oct 1993 | A |
5452379 | Poor | Sep 1995 | A |
5568571 | Willis et al. | Oct 1996 | A |
5572601 | Bloomberg | Nov 1996 | A |
5600732 | Ott et al. | Feb 1997 | A |
5625770 | Nomura | Apr 1997 | A |
5649026 | Heins | Jul 1997 | A |
5672060 | Poor | Sep 1997 | A |
5675671 | Hayduchok et al. | Oct 1997 | A |
5754674 | Ott et al. | May 1998 | A |
5818976 | Pasco et al. | Oct 1998 | A |
5825947 | Sasaki et al. | Oct 1998 | A |
5870488 | Rush et al. | Feb 1999 | A |
5956422 | Alam | Sep 1999 | A |
6079624 | Apperson et al. | Jun 2000 | A |
6129278 | Wang et al. | Oct 2000 | A |
6173154 | Kucinski et al. | Jan 2001 | B1 |
6188787 | Ohmae et al. | Feb 2001 | B1 |
6213664 | Kondo | Apr 2001 | B1 |
6259814 | Krtolica et al. | Jul 2001 | B1 |
6299066 | Howland et al. | Oct 2001 | B1 |
6311040 | Kucinski et al. | Oct 2001 | B1 |
6325286 | Howland et al. | Dec 2001 | B1 |
6408106 | Tatsuta et al. | Jun 2002 | B1 |
6542629 | Wu | Apr 2003 | B1 |
6575367 | Longacre | Jun 2003 | B1 |
6580820 | Fan | Jun 2003 | B1 |
6684052 | Kucinski et al. | Jan 2004 | B2 |
6741738 | Taylor | May 2004 | B2 |
6927872 | Currans | Aug 2005 | B2 |
6970267 | Scanlon | Nov 2005 | B1 |
7004393 | Schum et al. | Feb 2006 | B2 |
7164797 | Simard et al. | Jan 2007 | B2 |
7236653 | Constantin et al. | Jun 2007 | B2 |
7242806 | Johnson et al. | Jul 2007 | B2 |
20010033688 | Taylor | Oct 2001 | A1 |
20020168090 | Bruce et al. | Nov 2002 | A1 |
20020181805 | Loeb et al. | Dec 2002 | A1 |
20030072019 | Haines et al. | Apr 2003 | A1 |
20030173404 | Chung et al. | Sep 2003 | A1 |
20040139391 | Stumbo et al. | Jul 2004 | A1 |
20040185424 | Kucinski et al. | Sep 2004 | A1 |
20040202992 | Moulthrop et al. | Oct 2004 | A1 |
20050157930 | Cichielo et al. | Jul 2005 | A1 |
20050201639 | Wu | Sep 2005 | A1 |
20060062453 | Schacht | Mar 2006 | A1 |
20060250660 | Cui | Nov 2006 | A1 |
20060252023 | Cui | Nov 2006 | A1 |
20060255145 | Chung et al. | Nov 2006 | A1 |
20070116358 | Klotz et al. | May 2007 | A1 |
20070192176 | Onischuk | Aug 2007 | A1 |
Number | Date | Country |
---|---|---|
62263584 | Nov 1987 | JP |
6176192 | Jun 1994 | JP |
Number | Date | Country | |
---|---|---|---|
20080253658 A1 | Oct 2008 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10760850 | Jan 2004 | US |
Child | 11869527 | US |