Browser-based mobile image capture

Information

  • Patent Grant
  • 12130882
  • Patent Number
    12,130,882
  • Date Filed
    Friday, July 7, 2023
    a year ago
  • Date Issued
    Tuesday, October 29, 2024
    2 months ago
Abstract
Systems and methods for integrating mobile image capture with an internet browser application on a mobile device are provided, wherein a user viewing a web page via the internet browser application can initiate a mobile image capture process on the mobile device through the web page, capture images with a mobile image capture user interface, and then provide secure access to the captured images for the web page in order to utilize the content of the captured images. The mobile image capture process is accessible through any internet browser application via the mobile device and allows any service running on a web page to benefit from the ability to capture an image on the mobile device and obtain the content of the image through mobile image processing. The browser-based mobile image capture allows mobile devices which do not run certain third party applications to utilize mobile image capture and processing.
Description
BACKGROUND
1. Field of the Invention

Various embodiments described herein relate generally to the field of image capture using a mobile device. More particularly, various embodiments are directed to integrating mobile image capture with an internet-browser application on the mobile device.


2. Related Art

Mobile phone adoption continues to escalate, including ever-growing smart phone adoption and tablet usage. Mobile imaging is a discipline where a consumer takes a picture of a document, and that document is processed, extracting and extending the data contained within it for selected purposes. The convenience of this technique is powerful and is currently driving a desire for this technology throughout financial services and other industries.


Numerous applications utilizing mobile image capture have been developed for various financial services such as check deposits, bill pay, credit card balance transfers, and insurance account services, to name a few. These applications rely on the use of a third party application running on the mobile device that is managed by the service provider, such as a bank, utility or insurance company. The mobile image capture process is built into the third party application and is often customized for the particular third party service through the use of specific user interfaces and image capture parameters that are designed to capture certain types of documents with known properties and content. The mobile image capture process is therefore limited to use only within the specific third party application and cannot be utilized for another purpose for which it has not already been integrated. This effectively prevents the use of mobile image capture for mobile devices which do not run these third party applications or for services which are accessed through a mobile web browser instead of the third party application that is resident on the mobile device.


Therefore, it is desirable to broaden the use of mobile image capture for use with additional services and platforms available on the mobile device.


SUMMARY

Disclosed herein are systems and methods for integrating mobile image capture with an internet browser application on a mobile device. Various embodiments provide for a user viewing a web page via the internet browser application to initiate a mobile image capture process on the mobile device through the web page, capture one or more images with a mobile image capture user interface, and then provide secure access to the one or more captured images for the web page to utilize the content of the one or more captured images. The mobile image capture process is accessible through any internet browser application via the mobile device and allows any service running on a web page to benefit from the ability to capture an image on the mobile device and obtain the content of the image through mobile image processing.


Other features and advantages should become apparent from the following description of the preferred embodiments, taken in conjunction with the accompanying drawings.





BRIEF DESCRIPTION OF THE DRAWINGS

Various embodiments disclosed herein are described in detail with reference to the following figures. The drawings are provided for purposes of illustration only and merely depict typical or exemplary embodiments. These drawings are provided to facilitate the reader's understanding and shall not be considered limiting of the breadth, scope, or applicability of the embodiments. It should be noted that for clarity and ease of illustration these drawings are not necessarily made to scale.



FIG. 1A is a workflow diagram illustrating an exemplary network topology that may be used for browser-based mobile image capture, according to embodiments.



FIG. 1B is a workflow diagram illustrating an alternative exemplary network topology that may be used for browser-based mobile image capture, according to embodiments.



FIG. 2 is a flow diagram illustrating an exemplary method of performing browser-based mobile image capture according to embodiments.



FIG. 3 is a flow diagram illustrating a further exemplary method of performing browser-based mobile image capture according to embodiments.



FIG. 4 illustrates a block diagram of an embodiment of a computer/server system upon which an embodiment of the inventive methodology may be implemented





The various embodiments mentioned above are described in further detail with reference to the aforementioned figured and the following detailed description of exemplary embodiments.


DETAILED DESCRIPTION

The embodiments described herein provide for an internet browser application running on a mobile device to interface with a mobile image capture application running on the mobile device to provide for mobile image capture and processing for web-based services accessible through web pages on the internet browser application. This browser-based mobile image capture allows services which do not provide downloadable applications that run on the mobile device to utilize mobile image capture and processing features through the web browser. Additionally, mobile devices which do not provide or have access to third party applications can also utilize mobile image capture and processing. Furthermore, a mobile image capture application may be run on the mobile device as a standalone image capture application by the user of the mobile device in order to utilize the image capture and processing features apart from a related application.


System Overview



FIG. 1A is a workflow diagram illustrating one embodiment of a system and method for browser-based mobile image capture. As illustrated in FIG. 1A, the system utilizes a mobile device for accessing the web-based service and the mobile image capture application and a server for storing and providing secure access to the captured images to the web-based service. The mobile device is configured to run at least one internet browser application which allows a user to access the Internet, and specifically web pages for services which require capturing one or more images with content required for providing the services. The mobile device may run more than one internet browser application which the user can select from when accessing a web page; however, the user can select any internet browser application when utilization of the mobile image capture process is needed.


The mobile device is also configured to run a micro application (“MicroApp”) which interfaces with the browser to receive a request for mobile image capture, initiates a mobile image capture application and coordinates the transmission of the captured images to a remote server. The micro application also coordinates the transmission of schema from the browser to the mobile image capture application and the server which may pertain to the destination URL for the captured images or image capture settings to use when capturing the at least one image using the mobile device. In one embodiment, the schema may be transmitted using a JSON (JavaScript Object Notation) format. The mobile image capture application utilizes a camera on the mobile device and provides a user interface that allows a user to capture one or more high quality images with content that is needed for the web-based services. The features of the mobile image capture application provide for the user to easily and quickly capture high quality images of content with the mobile device, such as documents, financial papers, identification cards, license plates, vehicle identification numbers (VIN) and other words, numbers and symbols that may be needed by a web-based service and which would otherwise require manual entry of the content through more cumbersome entry methods such as typing or touch-based inputs.


The server may be configured to receive the one or more captured images from the micro application and store the captured images in a connected database for access by the web-based service. The server may also receive the schema from the micro application with the data on the destination for the captured images. The captured images may be securely stored and provide limited access using a token-based system or other secure protocol to ensure that the captured images and any content extracted therefrom are utilized only for the specific service that the user selects. The server may then send a response to the micro application confirming that the image and related data has been received and stored, which may prompt the micro application to close and return the user to the browser application. As will be described in further detail below, the captured images may then be requested (securely via the token) by the web-based service that the user is accessing with the internet browser on the mobile device, after which the captured images are transmitted to servers for the web-based service (not illustrated herein) at the destination URL provided in the schema. In one embodiment, the servers for the web-based service may perform image correction on the received images to obtain images which meet certain standards and requirements for content extraction such that the content can be extracted with a high degree of confidence. The servers for the web-based service may then perform content extraction steps on the corrected images in order to produce a content file of the extracted content for use by the web-based service.


In an alternate embodiment illustrated in FIG. 1B, the captured images may be stored directly on the mobile device without needing to store the images at a remote server, and the captured images can then be transmitted directly to the servers of the web-based service directly from the mobile device. In this embodiment, the micro application (also known as the “Collaborative Application”) provides for storing the one or more captured images in the RAM in the collaborative app on the mobile device, where it can then be retrieved within a short time frame using the same session GUID to ensure the security of the captured images.


It should be noted that although the micro application and mobile image capture (or “MiSnap”) application are illustrated as separate applications, they may be integrated into a single application that is executable on the mobile device rather than separate applications which require separate executions. The configurability of the applications may depend upon the operating system of the mobile device, but will nonetheless provide the same functionality regardless of the integration or separation of these applications.


Exemplary Method


One embodiment of a method for browser-based mobile image capture is illustrated in FIG. 2, beginning with the user accessing the internet browser application (step 202). When the user arrives at a web page which requests a mobile image of a document or other content, the user may select a hyperlink that invokes the micro application on the mobile device (step 204). The hyperlink may also contain information on the type of image, document or content to be captured and provide settings to the mobile image capture application that will improve the process and quality for capturing of the mobile image. In step 206, the mobile image capture application is invoked and one or more images are captured, after which the captured images are stored (step 208) on the remote server. An access token is created in step 210 which provides information specific to the capture of the images that will be used to verify that the captured and stored images are not requested by an unauthorized user. In step 212, the user is returned to the internet browser application where the user can then request the captured images using the token data (step 214) from the server. Once the token is authenticated, the captured images are obtained (step 216) by transmitting them from the stored location on the remote server to the server of the service providing the web page. The service will then extract the content from the captured images to utilize in providing the user with a particular service.



FIG. 3 illustrates a detailed flowchart of the process for browser-based mobile image capture, which illustrates the steps taken by the browser, the micro application and the mobile image capture application to provide the browser-based interface and image capture on the mobile device. In this embodiment, at the first step where the user selects a hyperlink on the browser, the hyperlink acts to invoke the micro application and first determines whether the micro application was invoked by the browser hyperlink or by another invocation means (such as a user launching the micro application directly on the mobile device for generic image capture applications). In one embodiment, the micro application may be integrated with the browser application as a plugin application which is automatically invoked by the web page at a certain step or which is invoked by the user selecting an icon for the plugin on the browser menu. If the micro application was invoked by the browser (Invoked—Yes), several schema will be passed along from the browser hyperlink, including settings for image capture, a server URL where the captured images and a token should be stored, and a return URL where the browser may be pointed to once the images have been captured. If the return URL is included, the micro application will direct the browser application to the return URL. If no return URL is supplied, the user will be returned to the calling URL that initiated the micro application. Thus, the micro application may store the destination information for the captured image so that the captured image can be automatically sent back to the destination specified by the web service without requiring the user to perform a separate step. If the micro application was invoked by other means (Invoked—No), a generic mode of the micro application and mobile image capture application are initiated to provide for a general image capture application using the camera of the mobile device and allowing the user to send any captured images to a destination of their choice via numerous transmission modes (text message, e-mail, upload to a social media page, cloud storage, etc.).


Once the mobile image capture application is initiated, the camera on the mobile device is utilized along with a user interface which provides for automatic capture of images of a document or other content based on the schema passed in through the hyperlink. When the image or images have been captured, the mobile image capture application terminates. The micro application then determines whether the captured images were requested by the browser (Invoke—Yes) and transmits the captured images to the server. If the captured images were requested by other means (Invoke—No), the captured images may be saved within the memory of the mobile device or sent to a destination selected by the user using a user interface provided by the micro application.


When the captured images are transmitted to the server, a token is also created and stored at the server along with the captured images. The token is designed to prevent unauthorized access to the captured image, and may be created with data from one or more criteria which is associated with the captured images, such as a time stamp of the time the images were created, a device identifier which uniquely identified the mobile device which captured the images, or a user identifier for the user of the mobile device that the user has set up with the web page service provider, phone carrier or through the micro application. The token may also be instance specific in that it can only be used once, or it must be used within a certain period of time before it expires. The captured images can only be transmitted from the server when a proper request is made which identifies the address where the images are stored on the server (server URL) and provides the token which matches the token stored at the server. Once the token and captured images are transmitted to the server, the micro application terminates and the mobile device returns the user to the browser application. The browser application is still displaying the browser session with the web page that requested the captured images, and the browser can either automatically request the captured images or wait for the user to select a link to then request that the captured images be transmitted to the server for the web page.


Additional Applications


In another embodiment, the internet browser may be configured to operate within the micro application such that the user uses the micro application to access the web page for the web service that requires mobile image capture. The embedded browser functionality within the micro application provides for streamlined functionality between the browser, micro application and mobile image capture application since the entire user interface and browser functionality takes place within a single application on the mobile device.


In another embodiment, the micro application may also provide independent functionality as a generic mobile image capture application which can be invoked separately from the internet browser where the user would like to capture one or more images of content for their own use or for use with an application that provides for uploading the image through another application or other communication means. For example, a service may require that an image be captured and then sent via a text message or e-mail message to an address for image processing and content extraction. The user could then initiate the micro application in the Generic Mode shown in FIG. 3 in order to capture one or more images using the mobile image capture application and then select an option to send the captured one or more images via text or e-mail message to a destination address for the service. The mobile image capture application on its own would provide benefits over a default image capture application on the mobile device by providing functionality for automatic capture of a high quality image based on a real-time evaluation of mobile device settings and the image settings.


Computer-Implemented Embodiment



FIG. 4 is a block diagram that illustrates an embodiment of a computer/server system 400 upon which an embodiment of the inventive methodology may be implemented. The system 400 includes a computer/server platform 401 including a processor 402 and memory 403 which operate to execute instructions, as known to one of skill in the art. The term “computer-readable storage medium” as used herein refers to any tangible medium, such as a disk or semiconductor memory, that participates in providing instructions to processor 402 for execution. Additionally, the computer platform 401 receives input from a plurality of input devices 404, such as a keyboard, mouse, touch device or verbal command. The computer platform 401 may additionally be connected to a removable storage device 405, such as a portable hard drive, optical media (CD or DVD), disk media or any other tangible medium from which a computer can read executable code. The computer platform may further be connected to network resources 406 which connect to the Internet or other components of a local public or private network. The network resources 406 may provide instructions and data to the computer platform from a remote location on a network 407. The connections to the network resources 406 may be via wireless protocols, such as the 802.11 standards, Bluetooth® or cellular protocols, or via physical transmission media, such as cables or fiber optics. The network resources may include storage devices for storing data and executable instructions at a location separate from the computer platform 401. The computer interacts with a display 408 to output data and other information to a user, as well as to request additional instructions and input from the user. The display 408 may therefore further act as an input device 404 for interacting with a user


While various embodiments have been described above, it should be understood that they have been presented by way of example only, and not of limitation. The breadth and scope should not be limited by any of the above-described exemplary embodiments. Where this document refers to technologies that would be apparent or known to one of ordinary skill in the art, such technologies encompass those apparent or known to the skilled artisan now or at any time in the future. In addition, the described embodiments are not restricted to the illustrated example architectures or configurations, but the desired features can be implemented using a variety of alternative architectures and configurations. As will become apparent to one of ordinary skill in the art after reading this document, the illustrated embodiments and their various alternatives can be implemented without confinement to the illustrated example. One of ordinary skill in the art would also understand how alternative functional, logical or physical partitioning and configurations could be utilized to implement the desired features of the described embodiments.


Furthermore, although items, elements or components may be described or claimed in the singular, the plural is contemplated to be within the scope thereof unless limitation to the singular is explicitly stated. The presence of broadening words and phrases such as “one or more,” “at least,” “but not limited to” or other like phrases in some instances shall not be read to mean that the narrower case is intended or required in instances where such broadening phrases may be absent.

Claims
  • 1. A method comprising using at least one computer processor of a mobile device to: within a browser application on the mobile device, retrieve a web page from a web server over at least one network, wherein the web page comprises a hyperlink to a micro application on the mobile device, and wherein the hyperlink contains a schema comprising one or more settings for image capture;in response to selection of the hyperlink in the web page within the browser application, invoke the micro application on the mobile device; and,by the micro application, initiate an image capture of one or more images using the mobile device in accordance with the schema contained within the selected hyperlink, andsend the captured one or more images to a remote server over the at least one network.
  • 2. The method of claim 1, wherein the remote server is the web server.
  • 3. The method of claim 1, wherein the remote server is different from the web server, and wherein the captured one or more images are transmitted to the remote server with a token.
  • 4. The method of claim 3, further comprising using at least one computer processor of the remote server to: receive the captured one or more images and the token from the mobile device via the at least one network;store the captured one or more images in association with the token;receive a request for the captured one or more images from the web server, wherein the request comprises a token; and,when the token in the request matches the token stored in association with the captured one or more images, send the captured one or more images to the web server.
  • 5. The method of claim 4, further comprising using the at least one computer processor of the remote server to: temporarily store the captured one or more images in a memory of the remote server; anddelete the captured one or more images from the memory in response to expiration of the token stored in association with the captured one or more images.
  • 6. The method of claim 4, further comprising using the at least one computer processor of the remote server to perform image correction on the captured one or more images prior to sending the captured one or more images to the web server.
  • 7. The method of claim 4, wherein the request further comprises a server Uniform Resource Locator (URL) identifying a location on the remote server at which the captured one or more images are stored.
  • 8. The method of claim 1, wherein the captured one or more images are transmitted to the remote server with a destination Uniform Resource Locator (URL).
  • 9. The method of claim 8, further comprising using at least one computer processor of the remote server to: receive the captured one or more images and the destination URL from the mobile device via the at least one network; andsend the captured one or more images to the destination URL.
  • 10. The method of claim 9, further comprising the at least one computer processor of the remote server to perform image correction on the captured one or more images prior to sending the captured one or more images to the destination URL.
  • 11. The method of claim 1, further comprising using the at least one computer processor of the mobile device to: temporarily store the captured one or more images in a memory of the mobile device; anddelete the captured one or more images from the memory in response to a session ending between the browser application and the web server.
  • 12. The method of claim 1, further comprising using the at least one computer processor of the mobile device to close the micro application in response to a session ending between the browser application and the web server.
  • 13. The method of claim 1, wherein the micro application is a plug-in of the browser application.
  • 14. The method of claim 1, further comprising using the at least one computer processor of the mobile device to: receive an acknowledgement, from the remote server, that the captured one or more images were received by the remote server; and,in response to receiving the acknowledgement, close the micro application and return to the browser application.
  • 15. The method of claim 1, further comprising using the at least one computer processor of the mobile device to, by the micro application: determine whether or not a return Uniform Resource Locator (URL) was provided during invocation of the micro application;when determining that the return URL was provided during invocation of the micro application, direct the browser application to the return URL; and,when determining that no return URL was provided during invocation of the micro application, return to the web page.
  • 16. The method of claim 1, wherein the remote server is different from the web server, wherein the web page further comprises a request hyperlink, and wherein the method further comprises using the at least one computer processor of the mobile device to, in response to selection of the request hyperlink, send a request to the remote server that the captured one or more images be sent to the web server.
  • 17. The method of claim 1, wherein the at least one network comprises the Internet.
  • 18. The method of claim 1, wherein the one or more settings comprise one or more of a type of image to be captured, a document to be captured by the one or more images, or content to be captured by the one or more images, and wherein the schema is sent to the remote server with the captured one or more images.
  • 19. A system comprising at least one computer processor configured to: within a browser application on the mobile device, retrieve a web page from a web server over at least one network, wherein the web page comprises a hyperlink to a micro application on the mobile device, and wherein the hyperlink contains a schema comprising one or more settings for image capture;in response to selection of the hyperlink in the web page within the browser application, invoke the micro application on the mobile device; and,by the micro application, initiate an image capture of one or more images using the mobile device in accordance with the schema contained within the selected hyperlink, andsend the captured one or more images to a remote server over the at least one network.
  • 20. A non-transitory computer-readable medium having instructions stored thereon, wherein the instructions, when executed by a processor, cause the processor to: within a browser application on the mobile device, retrieve a web page from a web server over at least one network, wherein the web page comprises a hyperlink to a micro application on the mobile device, and wherein the hyperlink contains a schema comprising one or more settings for image capture;in response to selection of the hyperlink in the web page within the browser application, invoke the micro application on the mobile device; and,by the micro application, initiate an image capture of one or more images using the mobile device in accordance with the schema contained within the selected hyperlink, andsend the captured one or more images to a remote server over the at least one network.
RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 17/189,085, filed on Mar. 1, 2021 and issued as U.S. Pat. No. 11,741,181 on Aug. 29, 2023, which is a continuation of U.S. patent application Ser. No. 14/568,085, filed on Dec. 11, 2014 and issued as U.S. Pat. No. 10,963,535 on Mar. 30, 2021, which claims priority to U.S. Provisional Patent App. No. 61/914,847, filed on Dec. 11, 2013, and is a continuation in part of U.S. patent application Ser. No. 14/042,677, filed on Sep. 30, 2013 and issued as U.S. Pat. No. 9,324,073 on Apr. 26, 2016, U.S. patent application Ser. No. 14/042,719, filed on Sep. 30, 2013 and issued as U.S. Pat. No. 9,672,510 on Jun. 6, 2017, and U.S. patent application Ser. No. 13/769,833, filed on Feb. 19, 2013 and now abandoned, which are all hereby incorporated herein by reference as if set forth in full.

US Referenced Citations (384)
Number Name Date Kind
4311914 Huber Jan 1982 A
5326959 Perazza Jul 1994 A
5600732 Ott et al. Feb 1997 A
5751841 Leong et al. May 1998 A
5761686 Bloomberg Jun 1998 A
5920847 Kolling et al. Jul 1999 A
5966473 Takahashi et al. Oct 1999 A
5999636 Juang Dec 1999 A
6038351 Rigakos Mar 2000 A
6038553 Hyde, Jr. Mar 2000 A
6070150 Remington et al. May 2000 A
6125362 Elworthy Sep 2000 A
6282326 Lee et al. Aug 2001 B1
6304684 Niczyporuk et al. Oct 2001 B1
6345130 Dahl Feb 2002 B1
6408094 Mirzaoff et al. Jun 2002 B1
6516078 Yang et al. Feb 2003 B1
6621919 Mennie et al. Sep 2003 B2
6735341 Horie et al. May 2004 B1
6807294 Yamazaki Oct 2004 B2
6947610 Sun Sep 2005 B2
6985631 Zhang Jan 2006 B2
6993205 Lorie et al. Jan 2006 B1
7020320 Filatov Mar 2006 B2
7072862 Wilson Jul 2006 B1
7133558 Ohara et al. Nov 2006 B1
7245765 Myers et al. Jul 2007 B2
7283656 Blake et al. Oct 2007 B2
7301564 Fan Nov 2007 B2
7331523 Meier et al. Feb 2008 B2
7376258 Klein et al. May 2008 B2
7377425 Ma et al. May 2008 B1
7426316 Vehvilinen Sep 2008 B2
7433098 Klein et al. Oct 2008 B2
7478066 Remington et al. Jan 2009 B2
7548641 Gilson et al. Jun 2009 B2
7558418 Verma et al. Jul 2009 B2
7584128 Mason et al. Sep 2009 B2
7593595 Heaney, Jr. et al. Sep 2009 B2
7606741 King et al. Oct 2009 B2
7636483 Yamaguchi et al. Dec 2009 B2
7735721 Ma et al. Jun 2010 B1
7778457 Nepomniachtchi et al. Aug 2010 B2
7793831 Beskitt Sep 2010 B2
7793835 Coggeshall et al. Sep 2010 B1
7817854 Taylor Oct 2010 B2
7869098 Corso et al. Jan 2011 B2
7873200 Oakes, III et al. Jan 2011 B1
7876949 Oakes, III et al. Jan 2011 B1
7949176 Nepomniachtchi May 2011 B2
7950698 Popadic et al. May 2011 B2
7953268 Nepomniachtchi May 2011 B2
7982770 Kahn et al. May 2011 B1
7974899 Prasad et al. Jul 2011 B1
7978900 Nepomniachtchi et al. Jul 2011 B2
7983468 Ibikunle et al. Jul 2011 B2
7986346 Kaneda et al. Jul 2011 B2
7995196 Fraser Aug 2011 B1
7996317 Gurz Aug 2011 B1
8000514 Nepomniachtchi et al. Aug 2011 B2
8023155 Jiang Sep 2011 B2
8025226 Hopkins et al. Sep 2011 B1
8109436 Hopkins, III Feb 2012 B1
8118216 Hoch et al. Feb 2012 B2
8121948 Gustin et al. Feb 2012 B2
8126252 Abernethy et al. Feb 2012 B2
8160149 Demos Apr 2012 B2
8180137 Faulkner et al. May 2012 B2
8233714 Zuev et al. Jul 2012 B2
8238638 Mueller et al. Aug 2012 B2
8290237 Burks et al. Oct 2012 B1
8300917 Borgia et al. Oct 2012 B2
8320657 Burks et al. Nov 2012 B1
8326015 Nepomniachtchi Dec 2012 B2
8339642 Ono Dec 2012 B2
8340452 Marchesotti Dec 2012 B2
8358826 Medina et al. Jan 2013 B1
8370254 Hopkins et al. Feb 2013 B1
8374383 Long et al. Feb 2013 B2
8379914 Nepomniachtchi et al. Feb 2013 B2
8442844 Trandal et al. May 2013 B1
8532419 Coleman Sep 2013 B2
8538124 Harpel et al. Sep 2013 B1
8540158 Lei et al. Sep 2013 B2
8542921 Medina Sep 2013 B1
8559766 Tilt et al. Oct 2013 B2
8582862 Nepomniachtchi et al. Nov 2013 B2
8688579 Ethington et al. Apr 2014 B1
8699779 Prasad et al. Apr 2014 B1
8837833 Wang et al. Sep 2014 B1
8861883 Tanaka Oct 2014 B2
8879783 Wang et al. Nov 2014 B1
8959033 Oakes, III et al. Feb 2015 B1
8977571 Bueche, Jr. et al. Mar 2015 B1
9058512 Medina, III Jun 2015 B1
9208393 Kotovich et al. Dec 2015 B2
9460141 Coman Oct 2016 B1
9613258 Chen et al. Apr 2017 B2
9679214 Kotovich et al. Jun 2017 B2
9710702 Nepomniachtchi et al. Jul 2017 B2
9773186 Nepomniachtchi et al. Sep 2017 B2
9786011 Engelhorn et al. Oct 2017 B1
9842331 Nepomniachtchi et al. Dec 2017 B2
10095947 Nepomniachtchi et al. Oct 2018 B2
10102583 Strange Oct 2018 B2
10275673 Kotovich et al. Apr 2019 B2
10360447 Nepomniachtchi et al. Jul 2019 B2
10373136 Pollack et al. Aug 2019 B1
10452908 Ramanathan et al. Oct 2019 B1
10546206 Nepomniachtchi et al. Jan 2020 B2
10621660 Medina et al. Apr 2020 B1
10789496 Kotovich et al. Sep 2020 B2
10789501 Nepomniachtchi et al. Sep 2020 B2
10891475 Nepomniachtchi et al. Jan 2021 B2
10909362 Nepomniachtchi et al. Feb 2021 B2
11157731 Nepomniachtchi et al. Oct 2021 B2
11380113 Nepomniachtchi et al. Jul 2022 B2
11393272 Kriegsfeld et al. Jul 2022 B2
20010014183 Sansom-Wai et al. Aug 2001 A1
20010016084 Pollard et al. Aug 2001 A1
20010019334 Carrai et al. Sep 2001 A1
20010019664 Pilu Sep 2001 A1
20010044899 Levy Nov 2001 A1
20020003896 Yamazaki Jan 2002 A1
20020012462 Fujiwara Jan 2002 A1
20020023055 Antognini et al. Feb 2002 A1
20020037097 Hoyos et al. Mar 2002 A1
20020041717 Murata et al. Apr 2002 A1
20020044689 Roustaei et al. Apr 2002 A1
20020046341 Kazaks et al. Apr 2002 A1
20020067846 Foley Jun 2002 A1
20020073044 Singhal Jun 2002 A1
20020077976 Meyer et al. Jun 2002 A1
20020080013 Anderson, III et al. Jun 2002 A1
20020085745 Jones et al. Jul 2002 A1
20020120846 Stewart et al. Aug 2002 A1
20020128967 Meyer et al. Sep 2002 A1
20020138351 Houvener et al. Sep 2002 A1
20020143804 Dowdy Oct 2002 A1
20020150279 Scott et al. Oct 2002 A1
20030009420 Jones Jan 2003 A1
20030072568 Lin et al. Apr 2003 A1
20030086615 Dance et al. May 2003 A1
20030099379 Monk et al. May 2003 A1
20030099401 Driggs et al. May 2003 A1
20030156201 Zhang Aug 2003 A1
20030161523 Moon et al. Aug 2003 A1
20030177100 Filatov Sep 2003 A1
20040012679 Fan Jan 2004 A1
20040017947 Yang Jan 2004 A1
20040024769 Forman et al. Feb 2004 A1
20040037448 Brundage Feb 2004 A1
20040081332 Tuttle et al. Apr 2004 A1
20040109597 Lugg Jun 2004 A1
20040205474 Eskin et al. Oct 2004 A1
20040213434 Emerson et al. Oct 2004 A1
20040213437 Howard et al. Oct 2004 A1
20040218799 Mastie et al. Nov 2004 A1
20040236688 Bozeman Nov 2004 A1
20040236690 Bogosian et al. Nov 2004 A1
20040247168 Pintsov et al. Dec 2004 A1
20050011957 Attia et al. Jan 2005 A1
20050065893 Josephson Mar 2005 A1
20050071283 Randle et al. Mar 2005 A1
20050080698 Perg et al. Apr 2005 A1
20050091161 Gustin et al. Apr 2005 A1
20050097046 Singfield May 2005 A1
20050100216 Myers et al. May 2005 A1
20050125295 Tidwell et al. Jun 2005 A1
20050129300 Sandison et al. Jun 2005 A1
20050141028 Koppich Jun 2005 A1
20050143136 Lev et al. Jun 2005 A1
20050163362 Jones et al. Jul 2005 A1
20050180661 El Bernoussi et al. Aug 2005 A1
20050192897 Rogers et al. Sep 2005 A1
20050196069 Yonaha Sep 2005 A1
20050196071 Prakash et al. Sep 2005 A1
20050213805 Blake et al. Sep 2005 A1
20050219367 Kanda et al. Oct 2005 A1
20050220324 Klein et al. Oct 2005 A1
20050229010 Monk et al. Oct 2005 A1
20050242186 Ohbuchi Nov 2005 A1
20050261990 Gocht et al. Nov 2005 A1
20060008167 Yu et al. Jan 2006 A1
20060008267 Kim Jan 2006 A1
20060012699 Miki Jan 2006 A1
20060039629 Li et al. Feb 2006 A1
20060045322 Clarke et al. Mar 2006 A1
20060045342 Kim et al. Mar 2006 A1
20060045344 Paxton et al. Mar 2006 A1
20060045379 Heaney et al. Mar 2006 A1
20060071950 Kurzweil et al. Apr 2006 A1
20060072822 Hatzav et al. Apr 2006 A1
20060088214 Handley et al. Apr 2006 A1
20060106717 Randle et al. May 2006 A1
20060140504 Fujimoto et al. Jun 2006 A1
20060164682 Lev Jul 2006 A1
20060177118 Ibikunle et al. Aug 2006 A1
20060182331 Gilson et al. Aug 2006 A1
20060186194 Richardson et al. Aug 2006 A1
20060210192 Orhun Sep 2006 A1
20060221415 Kawamoto Oct 2006 A1
20060242063 Peterson et al. Oct 2006 A1
20060280354 Murray Dec 2006 A1
20060291727 Bargeron Dec 2006 A1
20070009155 Potts et al. Jan 2007 A1
20070053574 Verma et al. Mar 2007 A1
20070058851 Quine et al. Mar 2007 A1
20070064991 Douglas et al. Mar 2007 A1
20070071324 Thakur Mar 2007 A1
20070076940 Goodall et al. Apr 2007 A1
20070081796 Fredlund et al. Apr 2007 A1
20070084911 Crowell Apr 2007 A1
20070086642 Foth et al. Apr 2007 A1
20070086643 Spier et al. Apr 2007 A1
20070110277 Hayduchok et al. May 2007 A1
20070114785 Porter May 2007 A1
20070118391 Malaney et al. May 2007 A1
20070131759 Cox et al. Jun 2007 A1
20070140678 Yost et al. Jun 2007 A1
20070154071 Lin et al. Jul 2007 A1
20070156438 Popadic et al. Jul 2007 A1
20070168382 Tillberg et al. Jul 2007 A1
20070171288 Inoue et al. Jul 2007 A1
20070174214 Welsh et al. Jul 2007 A1
20070195174 Oren Aug 2007 A1
20070206877 Wu et al. Sep 2007 A1
20070211964 Agam et al. Sep 2007 A1
20070214078 Coppinger Sep 2007 A1
20070244782 Chimento Oct 2007 A1
20070265887 Mclaughlin et al. Nov 2007 A1
20070288382 Narayanan et al. Dec 2007 A1
20070297664 Blaikie Dec 2007 A1
20080010215 Rackley, III et al. Jan 2008 A1
20080031543 Nakajima et al. Feb 2008 A1
20080040259 Snow et al. Feb 2008 A1
20080040280 Davis et al. Feb 2008 A1
20080062437 Rizzo Mar 2008 A1
20080086420 Gilder et al. Apr 2008 A1
20080089573 Mori et al. Apr 2008 A1
20080128505 Challa et al. Jun 2008 A1
20080152238 Sarkar Jun 2008 A1
20080174815 Komaki Jul 2008 A1
20080183576 Kim et al. Jul 2008 A1
20080192129 Walker et al. Aug 2008 A1
20080193020 Sibiryakov et al. Aug 2008 A1
20080212901 Castiglia et al. Sep 2008 A1
20080231714 Estevez et al. Sep 2008 A1
20080235263 Riaz et al. Sep 2008 A1
20080247629 Gilder et al. Oct 2008 A1
20080249931 Gilder et al. Oct 2008 A1
20080249936 Miller et al. Oct 2008 A1
20080267510 Paul et al. Oct 2008 A1
20080306787 Hamilton et al. Dec 2008 A1
20090041377 Edgar Feb 2009 A1
20090063431 Erol et al. Mar 2009 A1
20090092322 Erol et al. Apr 2009 A1
20090108080 Meyer et al. Apr 2009 A1
20090114716 Ramachandran May 2009 A1
20090125510 Graham et al. May 2009 A1
20090141962 Borgia et al. Jun 2009 A1
20090159659 Norris et al. Jun 2009 A1
20090185241 Nepomniachtchi Jul 2009 A1
20090185736 Nepomniachtchi Jul 2009 A1
20090185737 Nepomniachtchi Jul 2009 A1
20090185738 Nepomniachtchi Jul 2009 A1
20090185752 Dwivedula et al. Jul 2009 A1
20090190830 Hasegawa Jul 2009 A1
20090196485 Mueller et al. Aug 2009 A1
20090198493 Hakkani-Tur et al. Aug 2009 A1
20090201541 Neogi et al. Aug 2009 A1
20090216672 Zulf Aug 2009 A1
20090261158 Lawson Oct 2009 A1
20090265134 Sambasivan et al. Oct 2009 A1
20090271287 Halpern Oct 2009 A1
20090285444 Erol et al. Nov 2009 A1
20100030524 Warren Feb 2010 A1
20100037059 Sun et al. Feb 2010 A1
20100038839 Dewitt et al. Feb 2010 A1
20100073735 Hunt et al. Mar 2010 A1
20100074547 Yu et al. Mar 2010 A1
20100080471 Haas et al. Apr 2010 A1
20100082470 Walach et al. Apr 2010 A1
20100102119 Gustin et al. Apr 2010 A1
20100104171 Faulkner et al. Apr 2010 A1
20100114765 Gustin et al. May 2010 A1
20100114766 Gustin et al. May 2010 A1
20100114771 Gustin et al. May 2010 A1
20100114772 Gustin et al. May 2010 A1
20100150424 Nepomniachtchi et al. Jun 2010 A1
20100161466 Gilder Jun 2010 A1
20100200660 Moed et al. Aug 2010 A1
20100208282 Isaev Aug 2010 A1
20100239160 Enomoto et al. Sep 2010 A1
20100246972 Koyama et al. Sep 2010 A1
20100253787 Grant Oct 2010 A1
20100254604 Prabhakara et al. Oct 2010 A1
20100284611 Lee et al. Nov 2010 A1
20110013822 Blackson et al. Jan 2011 A1
20110026810 Hu Feb 2011 A1
20110052065 Nepomniachtchi et al. Mar 2011 A1
20110075936 Deaver Mar 2011 A1
20110081051 Tayal et al. Apr 2011 A1
20110091092 Nepomniachtchi et al. Apr 2011 A1
20110134248 Heit et al. Jun 2011 A1
20110170740 Coleman Jul 2011 A1
20110188759 Filimonova et al. Aug 2011 A1
20110194750 Nepomniachtchi Aug 2011 A1
20110249905 Singh et al. Oct 2011 A1
20110255795 Nakamura Oct 2011 A1
20110280450 Nepomniachtchi et al. Nov 2011 A1
20110289028 Sato Nov 2011 A1
20120010885 Hakkani-Tr et al. Jan 2012 A1
20120023567 Hammad Jan 2012 A1
20120030104 Huff et al. Feb 2012 A1
20120033892 Blenkhorn et al. Feb 2012 A1
20120051649 Saund et al. Mar 2012 A1
20120070062 Houle et al. Mar 2012 A1
20120072859 Wang et al. Mar 2012 A1
20120086989 Collins et al. Apr 2012 A1
20120106802 Hsieh et al. May 2012 A1
20120109792 Eftekhari et al. May 2012 A1
20120113489 Heit et al. May 2012 A1
20120150773 Dicorpo et al. Jun 2012 A1
20120197640 Hakkani-Tr et al. Aug 2012 A1
20120201416 Dewitt et al. Aug 2012 A1
20120226600 Dolev Sep 2012 A1
20120230577 Calman et al. Sep 2012 A1
20120265655 Stroh Oct 2012 A1
20120278336 Malik et al. Nov 2012 A1
20120308139 Dhir Dec 2012 A1
20130004076 Koo et al. Jan 2013 A1
20130022231 Nepomniachtchi et al. Jan 2013 A1
20130051610 Roach et al. Feb 2013 A1
20130058531 Hedley et al. Mar 2013 A1
20130085935 Nepomniachtchi et al. Apr 2013 A1
20130120595 Roach et al. May 2013 A1
20130148862 Roach et al. Jun 2013 A1
20130155474 Roach et al. Jun 2013 A1
20130181054 Durham et al. Jul 2013 A1
20130182002 Macciola et al. Jul 2013 A1
20130182951 Shustorovich et al. Jul 2013 A1
20130182973 Macciola et al. Jul 2013 A1
20130202185 Irwin, Jr. et al. Aug 2013 A1
20130204777 Irwin, Jr. et al. Aug 2013 A1
20130223721 Nepomniachtchi et al. Aug 2013 A1
20130272607 Chattopadhyay et al. Oct 2013 A1
20130297353 Strange et al. Nov 2013 A1
20130311362 Milam et al. Nov 2013 A1
20130317865 Tofte et al. Nov 2013 A1
20130325706 Wilson et al. Dec 2013 A1
20140032406 Roach et al. Jan 2014 A1
20140037183 Gorski et al. Feb 2014 A1
20140040141 Gauvin et al. Feb 2014 A1
20140044303 Chakraborti Feb 2014 A1
20140046841 Gauvin et al. Feb 2014 A1
20140064621 Reese et al. Mar 2014 A1
20140108456 Ramachandrula et al. Apr 2014 A1
20140126790 Duchesne et al. May 2014 A1
20140133767 Lund et al. May 2014 A1
20140172467 He et al. Jun 2014 A1
20140188715 Barlok et al. Jul 2014 A1
20140233837 Sandoz et al. Aug 2014 A1
20140254887 Amtrup et al. Sep 2014 A1
20140258838 Evers et al. Sep 2014 A1
20140270540 Spector et al. Sep 2014 A1
20140281871 Brunner et al. Sep 2014 A1
20140307959 Filimonova et al. Oct 2014 A1
20150012382 Ceribelli et al. Jan 2015 A1
20150012442 Ceribelli et al. Jan 2015 A1
20150040001 Kannan et al. Feb 2015 A1
20150142545 Ceribelli et al. May 2015 A1
20150142643 Ceribelli et al. May 2015 A1
20150334184 Liverance Nov 2015 A1
20160092730 Smirnov et al. Mar 2016 A1
20170185972 Bozeman Jun 2017 A1
20170316263 Nepomniachtchi et al. Nov 2017 A1
20180101751 Ghosh et al. Apr 2018 A1
20180101836 Nepomniachtchi et al. Apr 2018 A1
20180240081 Doyle et al. Aug 2018 A1
20200304650 Roach et al. Sep 2020 A1
20200342248 Nepomniachtchi et al. Oct 2020 A1
20210090372 Kriegsfeld et al. Mar 2021 A1
20220351161 Roach et al. Nov 2022 A1
Foreign Referenced Citations (7)
Number Date Country
2773730 Apr 2012 CA
1020040076131 Aug 2004 KR
1020070115834 Dec 2007 KR
03069425 Aug 2003 WO
2006075967 Jul 2006 WO
2006136958 Dec 2006 WO
2012144957 Oct 2012 WO
Non-Patent Literature Citations (35)
Entry
PDF417, Wikipedia: the free encyclopedia, Oct. 21, 2008, https://en.wikipedia.org/w/index.php?title=PDF417&oldid=246681430 (Year: 2008); 3 pages.
“OCR: The Most Important Scanning Feature You Never Knew You Needed.” hp (blog), Feb. 24, 2012. Accessed May 13, 2015. http://h71036.www7.hp.com/hho/cache/608037-0-0-39-121.html.
Abdulkader et al. “Low Cost Correction of OCR Errors Using Learning in a Multi-Engine Environment.” Proceedings of the 10th International Conference on Document Analysis and Recognition (ICDAR '09). IEEE Computer Society, Washington, D.C., USA. pp. 576-580. http://dx.doi.org/10.1109/ICDAR.2009.24.
Bassil, Youssef. “Ocr Post-Processing Error Correction Algorithm Using Google's Online Spelling Suggestion.” Journal of Emergin Trends in Computing and Information Sciences 3, No. 1 (Jan. 2012): 1. Accessed May 13, 2015. http://arxiv.org/ftp/arxiv/papers/1204/1204.0191.pdf.
Bieniecki et al. “Image preprocessing for improving OCR accuracy.” Perspective Technologies and Methods in MEMS Design, 2007. International Conference on MEMSTECH 2007. IEEE, 2007.
Chattopadhyay et al. “On the Enhancement and Binarization of Mobile Captured Vehicle Identification No. for an Embedded Solution.” 10th IAPR International Workshop on Document Analysis Systems (DAS), 2012. Pages 235-239. Mar. 27-29, 2012.
Cook, John. “Three Algorithms for Converting Color to Grayscale.” Singular Value Consulting. Aug. 24, 2009. Accessed May 13, 2015. http://www.johndcook.com/blog/2009/08/24/algorithms-convert-color-grayscale/.
Gatos et al. “Improved Document Image Binarization by Using a Combination of Multiple Binarization Techniques and Adapted Edge Information.” 19th International Conference on Pattern Recognition, 2008. IEEE.
He et al, “Comer deterctor Based on Global and Local Curvature Properties” Optical Engineering 47(5), 0570008 (2008).
International Search Report and Written Opinion received in PCT/US2011/056593, mailed May 30, 2012, 9 pages.
International Search Report issued in related International Application No. PCT/US2011/056593 dated May 30, 2012 (3 pages).
Notice of Allowance dated Feb. 22, 2023 received in U.S. Appl. No. 17/236,373 in 30 pages.
Notice of Allowance for related U.S. Appl. No. 16/160,796, mailed on Jan. 22, 2021, in 17 pages.
Notice of Allowance for related U.S. Appl. No. 16/579,625, mailed on Jan. 13, 2020 in 27 pages.
Notice of Allowance for related U.S. Appl. No. 16/742,439, mailed on Sep. 18, 2020, in 39 pages.
Notice of Allowance for related U.S. Appl. No. 16/987,782, mailed May 17, 2023, in 9 pages.
Notice of Allowance for related U.S. Appl. No. 17/189,085, mailed on Apr. 12, 2023, in 8 pages.
Notice of Allowance for related U.S. Appl. No. 17/829,025, mailed on Apr. 11, 2023, in 13 pages.
Notice of Allowance for related U.S. Appl. No. 17/829,025, mailed on Apr. 19, 2023, in 4 pages.
Office Action dated Feb. 1, 2023 in related U.S. Appl. No. 16/987,782, in 104 pages.
Office Action dated Jan. 12, 2023 in related U.S. Appl. No. 17/479,904, in 34 pages.
Office Action dated Jan. 9, 2020 for related U.S. Appl. No. 16/397,728 in 56 pages.
Office Action dated Jul. 11, 2019 for related U.S. Appl. No. 15/614,456 in 45 pages.
Office Action dated Jul. 26, 2019 for related U.S. Appl. No. 16/282,250 in 21 pages.
Office Action dated Mar. 20, 2020 in related U.S. Appl. No. 16/282,250, in 20 pages.
Office Action dated May 27, 2020 for related U.S. Appl. No. 16/282,250 in 18 pages.
Office Action dated Sep. 25, 2019 for related U.S. Appl. No. 16/518,815, in 10 pages.
Office Action for related CA Patent Application No. 2,773,730, dated Aug. 21, 2017, in 4 pages. .
Office Action for related U.S. Appl. No. 16/259,896, mailed on Dec. 12, 2019, in 22 pages.
Office Action for related U.S. Appl. No. 17/983,785, mailed on Mar. 30, 2023, in 46 pages.
Relativity. “Searching Manual.” Aug. 27, 2010. Accessed May 13, 2015. http://www.inventus.com/wp-content/uploads/2010/09/Relativity-Searching-Manual-6.6.pdf.
Shah et al. “OCR-Based chassis-nunber recognition using artificial neural networks.” 2009 IEEE Conference on Vehicular Electronics and Safety. pp. 31-34. Nov. 11-12, 2009.
Stevens. “Advanced Programming in the UNIX Environment.” Addison-Wesley Publishing Company, pp. 195-196 (1992).
“Tokenworks Introduces IDWedge ID Scanner Solution.” 2008.
Junker et al. “Evaluating OCR and Non-OCR Text Representation for Learning Document Classifiers.” Proceedings of the 4th International Conference on Document Analysis and Recognition. Ulm, Germany. Aug. 18-20, 1997. pp. 1060-1066 (1997). Accessed http://citeseerxist.psu.eduviewdoc/download?doi=10.1.1.6.6732&rep=rep1-&type=pdf.
Related Publications (1)
Number Date Country
20230359687 A1 Nov 2023 US
Provisional Applications (1)
Number Date Country
61914847 Dec 2013 US
Continuations (2)
Number Date Country
Parent 17189085 Mar 2021 US
Child 18219309 US
Parent 14568085 Dec 2014 US
Child 17189085 US
Continuation in Parts (3)
Number Date Country
Parent 14042677 Sep 2013 US
Child 14568085 US
Parent 14042719 Sep 2013 US
Child 14568085 US
Parent 13769833 Feb 2013 US
Child 14568085 US