Image sensors are commonly used to obtain an image of an object, e.g., a physical document or a form, for processing markings of the object. For example, circuitry of the image sensor or in communication with the image sensor converts markings of the obtained image into data that can be machine-processed. Example forms of which an image is obtained for processing of markings in the image are lottery playslips. An example playslip is a card, ticket, or other printed media that indicates a player's lottery number selections. It may be used by the player to purchase a lottery ticket having the indicated player lottery number selections. Markings of the lottery playslips are processed, e.g., for determining numbers selected by a player for an upcoming lottery drawing.
It is conventional to use a linear image sensor to obtain the image for processing of markings of the image. For example a linear barcode scanner is used to obtain an image of a portion of a barcode, i.e., a representative line of the barcode. Individual sub-markings of the obtained line are converted into processable data in order for a processor to “read” the barcode. Barcode scanners are commonly used in lottery applications.
Where lines to be processed are not uniform, so that one line of markings is not representative of all other lines of markings, it is necessary for the linear image sensor to progressively obtain a plurality of images, for example, one image per line. To do so, it is conventional for a user to tediously pass a hand-held linear image sensor over the object until all of the lines of the marking are obtained by the linear image sensor. The use of such linear image sensors to obtain an image that includes a number of non-uniform lines of markings to be processed takes up much time and effort by the user.
It is also conventional for a linear image sensor to include a mechanical device either for passing the linear image sensor over the object or for passing the object by the linear image sensor. Such sensors are bulky, and the use of such linear image sensors takes up much time.
Furthermore, some forms include markings, the processing of which is in accordance with a position of the markings with respect to the form. To process such markings, it is conventional to provide a linear image sensor with an electro-mechanical device to precisely position the form in a predetermined position with respect to the linear image sensor and to pass the form so positioned over a readhead of the linear image sensor. Since the form is positioned in a predetermined manner, a processor can process markings of the form according to the location of the markings in the precisely positioned form. For example, some lottery tickets, e.g., lottery playslips often include geometric shapes to be filled in by a user, e.g. fill-in squares, circles, etc. Each filled-in shape is processed as a particular number based on a position of the mark with respect to the ticket. It is conventional for an electro-mechanical device to pass such playslips over a readhead of a linear image sensor in order to obtain an image of the playslip for processing of the fill-in markings, for example as number selections for a lottery drawing. However such electro-mechanical devices are expensive.
It is conventional to use an area-array image sensor to simultaneously obtain an image of numerous lines of a form, e.g., to simultaneously obtain all fill-in markings of a lottery playslip. However, an image of the same form sensed by the area-array sensor may vary depending on the form's position with respect to the area-array sensor. For example, for a particular sensing area sensed by the area-array sensor, the sensed image of the form consumes much or all of the sensing area when the form is positioned near the area-array sensor, and consumes less of the sensing area when the form is placed farther from the area-array sensor. Differences in the degree of tilt of the form with respect to the area-array sensor also result in differences in the image sensed by the area-array sensor. For example, if a top side of a rectangular form is tilted away from the area-array sensor, the sensor senses the form as though it is of a trapezoidal shape, such as that of the form image 100 illustrated in
There is a need in the art for a system and method for processing markings of an image of a form sensed with an area-array sensor, without the use of a stand for positioning the form in a predetermined position with respect to the sensor. Some embodiments of the present invention generally relate to a system and method for obtaining an image of a form in order to process information in the form with an area-array image sensor, while the form is in free-space, e.g., without the use of a stand for precisely positioning the form with respect to the image sensor. The embodiments particularly, though not exclusively, relate to sensing an image of a lottery playslip for processing markings of the playslip as a lottery entry. Some of the embodiments provide for aiding a user to correctly position a form for input by aligning light patterns at indicated positions on the form.
In an alternative example embodiment of the present invention, any predetermined light-emitted pattern may be used to indicate placement of a form to be processed with respect to the image recording device. For example, a circle, triangle, or even a single line may be used. For example, a form may include a narrow rectangular box, instead of the four markings 302. The image recording device may include lights that emit light in the form of a single line. A user may position the form so that the light-emitted line falls within the narrow rectangular box of the form. However, use of a light-emitted pattern that encompasses all four corners of the form may optimize minimization of the degree of the form's tilt during image sensing.
In an example embodiment of the present invention, the image sensing device 202 may sense an image of the entire surface of a form to be processed, e.g., form 300, simultaneously. Alternatively, where only a particular portion of the form 300 is to be processed, the image sensing device 202 may record an image of the entire surface of the particular portion of the form 300 simultaneously. Accordingly, the form 300 need not be moved with respect to the image sensing device 202 during the image recordation in order to continuously sense light reflected by sub-portions of the form 300. Both the image sensing device 202 and the form 300 may remain stationary, e.g., with respect one another, during image sensing.
For example, the image sensing device 202 may include an area-array image sensor 206, e.g., that includes an array of light sensitive diodes to simultaneously sense light reflected by the entire surface of the form 300. The image sensor 206 may convert the sensed light into electrical charges, and may convert the electrical charges into corresponding binary values that represent an image of the form 300 sensed by the image sensor 206. For example, the image sensor 206 may be a charge coupled device (CCD) or a complementary metal oxide semiconductor (CMOS). Example image sensing devices 202 may be the SX5000 from Symagery, the SI-1300 from Silicon Imaging, and the IBIS406600 from Fill Factory, all of which are CMOS imagers.
In an example embodiment of the present invention, a form 300 may be illuminated by ambient light. The form 300 may reflect the ambient light towards the image sensing device 202. The image sensor 206 of the image sensing device 202 may sense the reflected ambient light. The image sensing device 202 may sense an image of the form 300 based solely on the sensed reflected ambient light.
In an alternative example embodiment of the present invention, the image sensing device 202 may include lights, e.g., LEDs 208, to illuminate the form 300. For example, the LEDs 208 may supplement any ambient light that may be illuminating the form 300. According to this embodiment, the image sensing device 202 may sense an image of the form 300 even in an environment in which no ambient light illuminates the form 300.
In an example embodiment of the present invention, the image sensing device 202 may selectively sense particular portions of the form 300, such that the sensed image of the form 300 does not include particular markings of the form 300. For example, the form 300 may include markings that instruct a user, e.g., a player, how to complete the form, and that are not intended for processing. The image sensing device 202 may record an image of the form without these instruction markings.
According to this embodiment, the image sensing device 202 may sense an image of a form 300 excluding all markings of a particular color. Any markings of the form 300 not intended for processing may be printed in the particular color. For example, the image sensing device 202 may exclude from the sensed image any marking in the form 300 that is red. To exclude markings of a particular color from the sensed image, the image sensing device may include a color filter 210 that causes all markings of the particular color, e.g., red, to be sensed by the image sensor 206 as a blank area of the form.
In an example embodiment of the present invention, the terminal 201 may include a processor 212 to process markings in a sensed image of a form 300, e.g., binary values that represent the image. Alternatively, or in addition, a processor may be located at a location external to the terminal 201, e.g., at a server 214. According to this embodiment, the server 214 and the terminal 201 may be connected via a network 215. Any conventional network may be used. Accordingly, processing may be performed locally at the terminal 201 and/or at the server 214. For example, the types of processing performed at the server 214 may be that which relates to an overall system, e.g., a lottery system, while processing performed at the terminal 201 may be that which relates to the execution of individual transactions within the system. The image sensing device 202 may transmit towards the processor 212 the sensed image of the form 300. The processor 212 may process a marking within the image based on the location of the marking within the sensed image. For example, if a user filled in a circle located at an upper-left corner of a lottery playslip, the processor 212 may process a representation of the mark obtained from the image sensor 206 as an indication that a lottery player has chosen the number 1 as the first of a set of lottery numbers. The light emitted by the LEDs 204 may indicate to a user to place the playslip in a particular position with respect to the image sensing device 202 so that the circle representing the number 1 of the first of a set of lottery numbers will be located at the upper-left corner of the sensed image. The processor 212 may accordingly correctly process markings within the playslip.
According to an example embodiment of the present invention, the processor 212 may correctly process markings of the form 300 if the form 300 is placed with respect to the image sensing device 202 in any position that is within a range of proximally located predetermined positions. For example, the LEDs 204 may indicate to a user to place the playslip in a particular position with respect to the image recording device. The user may attempt to position the playslip in the indicated position but may mistakenly slightly shift and/or tilt the form from the indicated position. The processor 212 may be able to process markings of the form despite the shift or tilt, as long as the shift or tilt is not a very drastic shift or tilt by which the playslip is positioned out of a predetermined range of positions.
The LEDs 204 may indicate an optimum position into which the user is to place the playslip. The processor 212 may be configured to recognize the markings of the playslip even if the top of the playslip is tilted forward or backward, e.g., up to 200 relative to the axis that is perpendicular to the playslip when positioned in the optimum position. For example, the processor 212 may be configured to recognize the tilt and correct the image before processing the markings so that the processor 212 may process markings of the image of the playslip as though the playslip was placed at the optimum position. Alternatively, the processor 212 may be configured to correctly process markings of an image of a playslip even without correcting a misplaced form 300. For example, the form 300 may include markings that may be positioned at a predetermined distance from one another, such that when the form 300 is placed within any of a range of positions, each marking of the form 300 may always fall within a single corresponding predetermined area. The processor 212 may accordingly process the markings of the form 300 if the form 300 is placed within any of the predetermined range of positions. In a further alternative example embodiment, the processor 212 may convert the image to grayscale. The processor 212 may then determine the extent of the shift or tilt angle. According to the determined position, the processor 212 may process the markings. For example, for each position within a predetermined range of positions, the processor 212 may process the image in a different manner.
In an example embodiment of the present invention, the processor 212 may be able to correctly process markings of an image of a form if the form is placed in a position that is within a distance range of six to eight inches from the image sensing device 202.
According to an example embodiment of the present invention, the processor 212 may be configured for optical character recognition (OCR), and may process alpha-numeric characters, whether machine-printed or hand-printed, that are within a form 300, e.g., a sensed image of the form 300. For example, a lottery playslip may include a series of blank lines above which a player may print selected numbers for a lottery drawing. The selected numbers may vary from playslip to playslip. The processor 212 may process the printed numbers of the image of the playslip sensed by the image sensing device 202.
In an example embodiment of the present invention, the light-emitted pattern of the image sensing device 202 may indicate for the form 300 a predetermined placement position with respect to the abscissa position of the form 300, the ordinate position of the form 300, the distance of the form 300 from the image sensing device 202, and/or the orientation of the form 300. For example, if a playslip is positioned 12″ from the image sensing device 202, a light-emitted square pattern of the image sensing device 202 may form towards the center of the playslip. For the pattern to properly form for example at the edges of the playslip, it may be required to move the playslip closer to the image sensing device 202.
In an example embodiment of the present invention, a predetermined position with respect to the rotational orientation of a form 300 to be processed may not be required. According to this embodiment, an orientation of a sensed image may be predetermined. The processor 212 may be configured to rotate a sensed image of the form 300 to the predetermined orientation. The processor 212 may then process the rotated image.
According to this embodiment, a form 300 may include markings 304 that indicate to the processor 212 a manner in which to rotate the recorded image. For example, the form 300 may include markings 304 along two sides, e.g., adjacent edges, of the form 300. The processor may be configured to rotate a recorded image of the form until the two markings are located at a predetermined position, e.g., one marking 304 at a lower portion of a left side of the image, and the other marking 304 at a left portion of a bottom side of the image.
The image recording device may rely solely on ambient light to illuminate the lottery playslip. Alternatively, in step 400, LEDs may be provided to emit light to supplement ambient light in illuminating the lottery playslip.
In step 402, an image sensor, e.g. an area-array sensor, such as a CCD or CMOS image sensor, or any conventional area-array image sensor, may sense light reflected by a lottery playslip positioned with respect to the image sensing device according to the light-emitted pattern formed by the LEDs in step 400.
The playslip may include instructions to a lottery player. It may not be required for the instructions portion of the lottery playslip to be processed, since the instructions may be intended only to convey information to the lottery player. These instructions may include instructions with respect to how to fill out the playslip, how to claim prize money, rules of the lottery game, and any instructions conventionally printed on a lottery playslip. The included instructions may be printed on the lottery playslip in red. A color may be provided in order to exclude the red instruction markings of the lottery playslip during sensing of an image of the lottery playslip.
In step 404, the image sensing device may convert the sensed light into electrical charges. In step 406, the image sensing device may convert the electrical charges into binary values. The binary values may represent an image of the lottery playslip. The represented image may exclude red instruction markings of the lottery playslip. In step 408, the image sensing device may transmit the binary values to a processor. The processor may be located in a lottery POS terminal. Alternatively, the processor may be located at an external location, for example, at a central lottery processing location, connected to the POS terminal, e.g., via a network. In an alternative embodiment, the processor may be integrated in the image sensing device. In one embodiment, a plurality of processors may be provided. The various processes to be performed may be divided amongst the plurality of processors. According to this embodiment, a processor may be located in the image recording device, at a location that is external to the image recording device and within the POS terminal, and at a location external to the POS terminal.
In step 410, the processor may decode playslip markings. The processor may interpret the transmitted binary values as an image of the lottery playslip. The processor may interpret individual markings of the represented image. For example, the processor may determine from markings within the playslip the type of lottery game played with the lottery playslip, and may determine player selected numbers for the lottery game, e.g., for a future lottery drawing, a time of the entry, a date of a particular drawing, a bet amount, etc.
In an example embodiment of the present invention, after the processor processes the markings, the processor may transmit a message to a user that indicates the information the processor determined from the markings. In response, the user may indicate to the processor whether the processor correctly processed the markings. For example, the processor may display in a window within a display area of a graphical user interface. The processor may include the message and two buttons in the window. The user may click one button to indicate that the processor correctly processed the markings and may click the other button to indicate that the processor incorrectly processed the markings. If the user indicates that the markings were correctly processed, the processor may record the entry. Otherwise, the processor may return to 410 and process the markings again. Alternatively, the processor may indicate to the user that the playslip is to be input a second time. This procedure may be repeated until the user indicates that the processor has correctly processed the playslip markings.
The processor may output a receipt associated with the information obtained from the markings of the playslip. For example, in step 412, the processor may associate the player selected numbers with a particular barcode or other identifier. The barcode may be printed on the receipt. The receipt may be printed at the time of playslip purchase. For example, the barcode may be printed at the time of playslip purchase. Alternatively, the barcode may be pre-printed on a receipt portion of the lottery playslip. According to this alternative, the image sensing device may sense an image of both the number selection portion of a lottery playslip and the receipt portion of the lottery playslip. A lottery player may remove the receipt portion of the lottery playslip after the image of the playslip is sensed. In step 412, the processor may associate the player selected numbers to the barcode of the playslip receipt. If the playslip does not include a receipt portion, the processor may additionally, in step 412, generate a unique barcode to be associated with the player selected numbers of the lottery playslip. After processing of the playslip, the generated barcode may be printed on a receipt. The processor may store, e.g., in a central database, an identification of the barcode and its association with the player selected numbers.
After a drawing associated with the playslip, the lottery player may present the receipt. The processor may access the central database to ascertain whether the lottery player is a lottery drawing winner.
In an example embodiment of the present invention, a processor 212 of
In one example embodiment of the present invention, the barcode 306 may indicate to the processor 212 a way in which to process other markings 308 within the form. See, e.g., “Method and Apparatus for Providing and Processing Active Barcodes,” Thomas K. Oram, U.S. patent application Ser. No. 10/698,152. For example, the barcode 306 may indicate that the processor 212 is to process a marking 308 positioned five inches above and a half an inch to the left of the barcode 306 as the number 1 and as the first position of a set of selected numbers for a lottery drawing. The processor 212 may process the barcode 306 and subsequently process markings 308 within the image in accordance with the processed data of the barcode 306. Alternatively, the barcode 306 may indicate, with respect to the position of the barcode 306, positions within the sensed image of particular edges of the form. The processor 212 may accordingly ascertain the positions of each coordinate along the entire surface of the form 300 with respect to the sensed image. The barcode 306 may also indicate to the processor 212 how to process markings at various positions with respect to the ascertained representation of the form surface. Alternatively, the processor 212 may be configured to process a marking according to the marking's position with respect to the ascertained representation of the form surface without further instructions of the barcode 306.
In an alternative example embodiment of the present invention, the barcode 306 may omit processing instructions. According to this embodiment, the processor 212 may locate and recognize the barcode 306 from within the image, and may process markings 308 within the sensed image based on the position of the markings 308 with respect to the barcode 306.
According to an example embodiment of the present invention, an image sensing device may be provided that may sense an image of markings that are as narrow as one-hundredth of an inch. According to this embodiment, the barcode 306 may include a bar as narrow as one-hundredth of an inch.
According to the embodiment in which a form includes an identifier, such as a barcode, to aid in the processing of form markings, in step 400 of
In one example embodiment, the processor may process markings of a number of different types of forms. For example, different forms may be used for different lottery games. Different barcodes may be included for the different form types. The processor may determine how to process markings based on the particular barcode included in the form. For example, before proceeding to 410, the processor may identify the barcode included in the form.
Those skilled in the art can appreciate from the foregoing description that the present invention may be implemented in a variety of forms. Therefore, while the embodiments of this invention have been described in connection with particular examples thereof, the true scope of the embodiments of the invention should not be so limited since other modifications will become apparent to the skilled practitioner upon a study of the drawings, specification, and following claims.
This application is a continuation of application Ser. No. 11/376,052, filed Mar. 14, 2006, now U.S. Pat. No. 7,920,299 which claims the benefit of priority of provisional Application No. 60/661,698, filed Mar. 14, 2005. application Ser. No. 11/376,052 is hereby incorporated by reference in its entirety. A portion of the disclosure of this patent document contains material that is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or patent disclosure as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever.
Number | Name | Date | Kind |
---|---|---|---|
4141036 | Diehl | Feb 1979 | A |
4240748 | Blanc et al. | Dec 1980 | A |
4724307 | Dutton et al. | Feb 1988 | A |
4736109 | Dvorzsak | Apr 1988 | A |
4760247 | Keane et al. | Jul 1988 | A |
5039847 | Morii et al. | Aug 1991 | A |
5216595 | Protheroe | Jun 1993 | A |
5239165 | Novak | Aug 1993 | A |
5262624 | Koch | Nov 1993 | A |
5362105 | Scott | Nov 1994 | A |
5417424 | Snowden et al. | May 1995 | A |
5452379 | Poor | Sep 1995 | A |
5673125 | Merecki et al. | Sep 1997 | A |
5689104 | Suzuki et al. | Nov 1997 | A |
5734153 | Swartz et al. | Mar 1998 | A |
5748783 | Rhoads | May 1998 | A |
5818026 | Melling et al. | Oct 1998 | A |
5821518 | Sussmeier et al. | Oct 1998 | A |
5834749 | Durbin | Nov 1998 | A |
5945655 | Gilgeous et al. | Aug 1999 | A |
6150930 | Cooper | Nov 2000 | A |
6157435 | Slater et al. | Dec 2000 | A |
6176429 | Reddersen et al. | Jan 2001 | B1 |
6199044 | Ackley et al. | Mar 2001 | B1 |
6223986 | Bobba et al. | May 2001 | B1 |
6234899 | Nulph | May 2001 | B1 |
6262670 | Ballou | Jul 2001 | B1 |
6340114 | Correa et al. | Jan 2002 | B1 |
6357658 | Garczynski et al. | Mar 2002 | B1 |
6366696 | Hertz et al. | Apr 2002 | B1 |
6494372 | Bruchlos et al. | Dec 2002 | B2 |
6553129 | Rhoads | Apr 2003 | B1 |
6650427 | Brooks et al. | Nov 2003 | B2 |
6722569 | Ehrhart et al. | Apr 2004 | B2 |
6736319 | Swanson et al. | May 2004 | B2 |
6760476 | Meder | Jul 2004 | B2 |
6776337 | Irwin, Jr. et al. | Aug 2004 | B2 |
6875015 | Tiernan | Apr 2005 | B1 |
6875105 | Behm et al. | Apr 2005 | B1 |
6887153 | Walker et al. | May 2005 | B2 |
6954290 | Braudaway et al. | Oct 2005 | B1 |
7060968 | Leviton | Jun 2006 | B1 |
7090131 | Natsuno | Aug 2006 | B2 |
7124945 | Kim | Oct 2006 | B2 |
7221487 | Sesek et al. | May 2007 | B2 |
7247095 | Nulph | Jul 2007 | B2 |
7474760 | Hertzman et al. | Jan 2009 | B2 |
7635088 | Chung et al. | Dec 2009 | B2 |
7674171 | Walker et al. | Mar 2010 | B2 |
7688467 | Enmei | Mar 2010 | B2 |
7764408 | Ohama et al. | Jul 2010 | B2 |
20010025880 | Bruchlos et al. | Oct 2001 | A1 |
20010029513 | Kuwano et al. | Oct 2001 | A1 |
20020010022 | Vincent | Jan 2002 | A1 |
20020048403 | Guerreri | Apr 2002 | A1 |
20020109866 | Yang et al. | Aug 2002 | A1 |
20030102376 | Meier et al. | Jun 2003 | A1 |
20030173404 | Chung et al. | Sep 2003 | A1 |
20030186734 | LeMay et al. | Oct 2003 | A1 |
20040029630 | Walker et al. | Feb 2004 | A1 |
20040106652 | Hamilton et al. | Jun 2004 | A1 |
20040209665 | Walker et al. | Oct 2004 | A1 |
20040218835 | Loew | Nov 2004 | A1 |
20050098633 | Poloniewicz et al. | May 2005 | A1 |
20050161511 | Parker et al. | Jul 2005 | A1 |
20050233797 | Gilmore et al. | Oct 2005 | A1 |
20050237580 | Coleman et al. | Oct 2005 | A1 |
20050238260 | Coleman et al. | Oct 2005 | A1 |
20060063589 | Chong | Mar 2006 | A1 |
20060079311 | Nulph | Apr 2006 | A1 |
20060122910 | Chau et al. | Jun 2006 | A1 |
20060152595 | Ryu et al. | Jul 2006 | A1 |
20060221354 | Slaten et al. | Oct 2006 | A1 |
20060255145 | Chung et al. | Nov 2006 | A1 |
20060290886 | Santos | Dec 2006 | A1 |
20070109511 | Kelly et al. | May 2007 | A1 |
20080240614 | Garcia | Oct 2008 | A1 |
20090020606 | Chung et al. | Jan 2009 | A1 |
20090167960 | Miyasato | Jul 2009 | A1 |
20100060943 | Monga et al. | Mar 2010 | A1 |
Number | Date | Country |
---|---|---|
0910032 | Apr 1999 | EP |
1 172 756 | Jan 2002 | EP |
1 441 315 | Jul 2004 | EP |
05-242279 | Sep 1993 | JP |
06-266876 | Sep 1994 | JP |
07037018 | Feb 1995 | JP |
07-105133 | Apr 1995 | JP |
07-239920 | Sep 1995 | JP |
11-220588 | Aug 1999 | JP |
2000149081 | May 2000 | JP |
2002056348 | Feb 2002 | JP |
2004-219255 | Aug 2004 | JP |
WO 0004487 | Jan 2000 | WO |
WO 0104571 | Jan 2001 | WO |
WO 2007114241 | Oct 2007 | WO |
Number | Date | Country | |
---|---|---|---|
20110176180 A1 | Jul 2011 | US |
Number | Date | Country | |
---|---|---|---|
60661698 | Mar 2005 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11376052 | Mar 2006 | US |
Child | 13026712 | US |