System and method for selectively reading code symbols

Information

  • Patent Grant
  • 9773142
  • Patent Number
    9,773,142
  • Date Filed
    Monday, July 22, 2013
    12 years ago
  • Date Issued
    Tuesday, September 26, 2017
    8 years ago
Abstract
A system for selectively reading code symbols includes a code-symbol-capturing subsystem for acquiring information about code symbols within the code-symbol-capturing subsystem's field of view. The system also includes a code-symbol-decoding processor that detects a marked region of interest within the code-symbol-capturing subsystem's field of view. After initialization, the code-symbol-decoding processor decodes only those code symbols falling within the marked region of interest.
Description
FIELD OF THE INVENTION

The present invention relates to code symbol readers. More specifically, the present invention relates to a system and method for selectively reading code symbols.


BACKGROUND

There are instances when more than one code symbol (e.g., barcode, QR code, etc.) may be present within the field of view of a code symbol reader (e.g., barcode scanner). A barcode menu, for example, may be a page with several barcodes in each row. In these cases, when the user scans an entire row of barcodes and the reader returns the information relating to a decoded barcode (e.g., product identification number), the user may not know which barcode the reader actually decoded. In the case of handheld scanners, for example, the user may attempt to solve this problem by adjusting the field of view of the scanner (e.g., by bringing the scanner closer to the barcode that is desired to be decoded). This solution is not always practical, however, because it takes time for the user to reposition the scanner to achieve the desired field of view, and because the scanner may not be capable of repositioning (e.g., a fixed-position scanner). Particularly in situations requiring repetitive scanning of barcodes in the same position within the scanner's field of view, requiring the user to manually reposition the scanner to adjust the field of view to that one area results in needless use of the operator's time and focus.


A need therefore exists for a system for reading code symbols that can be initialized to read only those code symbols that fall within a region of interest that is within the system's field of view. For example, a need exists for a system for reading code symbols that can be initialized to decode only the barcodes that fall within the second column on a barcode menu having four columns.


SUMMARY

Accordingly, in one aspect, the present invention embraces a system for reading code symbols. The system includes a code-symbol-capturing subsystem for acquiring information about code symbols within the code-symbol-capturing subsystem's field of view. The system also includes a code-symbol-decoding processor. The code-symbol-decoding processor is configured for detecting a marked region of interest within the code-symbol-capturing subsystem's field of view. The code-symbol-decoding processor is also configured for decoding the code-symbol information inside the marked region of interest within the code-symbol-capturing subsystem's field of view. The code-symbol-decoding processor is also configured to disregard code-symbol information that is outside the marked region of interest.


In an exemplary embodiment, the code-symbol-capturing subsystem is an imaging subsystem for capturing images within the imaging subsystem's field of view.


In another exemplary embodiment, the code-symbol-capturing subsystem is a laser scanning subsystem for scanning code symbols within the laser scanning subsystem's field of view.


In yet another exemplary embodiment, the code-symbol-decoding processor is configured for detecting the region of interest in response to a user-positioned marker.


In yet another exemplary embodiment, the code-symbol-decoding processor is configured for detecting the region of interest in response to a user-positioned marker, which user-positioned marker is a code symbol.


In yet another exemplary embodiment, the code-symbol-decoding processor is configured for detecting the region of interest in response to a user-positioned marker, which user-positioned marker is a signal-enhancing marker.


In yet another exemplary embodiment, the code-symbol-decoding processor is configured for detecting the region of interest in response to a user-positioned marker, which user-positioned marker is a signal-decreasing marker.


In another aspect, the present invention embraces a system for reading code symbols that includes a code-symbol-capturing subsystem configured to detect a marked region of interest within the code-symbol-capturing subsystem's initial field of view. The code-symbol-capturing subsystem is also configured to adjust the code-symbol-capturing subsystem's initial field of view to an adjusted field of view that corresponds to the region of interest. The code-symbol-capturing subsystem is also configured to acquire information about code symbols within the code-symbol-capturing subsystem's adjusted field of view that corresponds to the region of interest. The system for reading code symbols also includes a code-symbol-decoding processor for processing code-symbol information acquired by the code-symbol-capturing subsystem.


In an exemplary embodiment, the code-symbol-capturing subsystem is an imaging subsystem for capturing images of code symbols.


In another exemplary embodiment, the code-symbol-capturing subsystem is a laser scanning subsystem for scanning code symbols.


In yet another exemplary embodiment, the code-symbol-capturing subsystem is configured for detecting the region of interest in response to a user-positioned marker.


In yet another exemplary embodiment, the code-symbol-capturing subsystem is configured for detecting the region of interest in response to a user-positioned marker, which user-positioned marker is a code symbol.


In yet another exemplary embodiment, the code-symbol-capturing subsystem is configured for detecting the region of interest in response to a user-positioned marker, which user-positioned marker is a signal-enhancing marker.


In yet another exemplary embodiment, the code-symbol-capturing subsystem is configured for detecting the region of interest in response to a user-positioned marker, which user-positioned marker is a signal-decreasing marker.


In another aspect, the present invention embraces a method for selectively reading code symbols. According to the method of the present invention, a code symbol reader is provided. The code symbol reader has a code-symbol-capturing subsystem for acquiring information about code symbols within the code-symbol-capturing subsystem's field of view. The code symbol reader also has a code-symbol-decoding processor for decoding code symbols captured by the code-symbol-capturing subsystem. A marked region of interest within the code-symbol-capturing subsystem's field of view is detected by the code symbol reader. Only those code symbols that are within the marked region of interest are decoded by the code symbol reader.


In an exemplary embodiment, the code-symbol-decoding processor of the method according to the present invention is configured to disregard code-symbol information that is outside the marked region of interest.


In another exemplary embodiment of the method according to the present invention, the code-symbol-capturing subsystem's initial field of view is adjusted to an adjusted field of view. The adjusted field of view corresponds to the region of interest. Information about code symbols within the code-symbol-capturing subsystem's adjusted field of view is acquired.


In yet another exemplary embodiment of the method according to the present invention, a region of interest is marked with a user-position marker, which user-positioned marker is a finger.


In yet another exemplary embodiment of the method according to the present invention, a region of interest is marked with a user-position marker, which user-positioned marker is a signal-enhancing marker.


In yet another exemplary embodiment of the method according to the present invention, the marked region of interest is indicated to a user of the code symbol reader by a region-of-interest indicator.


The foregoing illustrative summary, as well as other exemplary objectives and/or advantages of the invention, and the manner in which the same are accomplished, are further explained within the following detailed description and its accompanying drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram of an exemplary system for reading code symbols according to the present invention.



FIG. 2 is a block diagram of a first alternative embodiment of an exemplary system for reading code symbols according to the present invention.



FIG. 3 is a block diagram of a second alternative embodiment of an exemplary system for reading code symbols according to the present invention.



FIG. 4 is an illustration of an exemplary system being used to selectively read a barcode menu.



FIG. 5 is an illustration of an exemplary system being used to selectively read a barcode menu.



FIG. 6 depicts the adjusted field of view of an exemplary system for reading code symbols according to the present invention.





DETAILED DESCRIPTION

The present invention embraces a system for reading code symbols. The term “code symbol” is intended broadly to refer to any machine-readable indicia or device used to store information about an object, including barcodes, linear barcodes, 1D barcodes, matrix barcodes, 2D barcodes, QR codes, RFID tags, and optical character recognition (OCR) symbols. A barcode, for example, is a machine-readable representation of information in a graphic format. The most familiar of these graphic symbols is a series of parallel bars and spaces of varying widths, which format gave rise to the term “barcode.”


In some instances, an object may bear more than one code symbol. A code symbol menu (e.g., barcode menu), for example, may display a series of different code symbols arranged in columns and rows. When there are a plurality of code symbols within the field of view of the code symbol reader, the user must determine which code symbol the user wants the code symbol reader to read. The system for reading code symbols according to the present invention can be initialized to read only those code symbols that appear in a selected portion (e.g., the region of interest) within its field of view. Typically, the system is initialized by a marker that indicates region of interest. After the initialization, the system only reads those barcodes which appear within the region of interest.


Referring now to FIG. 1 through FIG. 5, the system 100 for reading code symbols according to the present invention includes a code-symbol-capturing subsystem 110. The code-symbol-capturing subsystem 110 acquires information about code symbols (e.g., code-symbol information) within the code-symbol-capturing subsystem's field of view 116. The form that the code-symbol information takes depends upon the configuration of the code-symbol-capturing subsystem 110. As will be discussed below, the code-symbol-capturing subsystem 110 may be an imaging subsystem 110A, or it may be a laser scanning subsystem 110B. Where an imaging subsystem 110A is employed, the code-symbol information will be a digital image of the code symbol(s), which digital image may be decoded via image processing techniques (e.g., optical character recognition (OCR) techniques). On the other hand, a laser scanning subsystem 110B collects code-symbol information in the form of a reflected optical signal, which signal may be decoded based upon an analysis of the signal's intensity.


The system 100 for reading code symbols according to the present invention also includes a code-symbol-decoding processor 115. The code-symbol-decoding processor 115 processes code-symbol information acquired by the code-symbol-capturing subsystem 110. The code-symbol-decoding processor 115 of the present system 100 is configured to detect a marked region of interest 120 within the code-symbol-capturing subsystem's 110 field of view 116. Unlike traditional code symbol scanners, the code-symbol-decoding processor 115 decodes only the code-symbol information positioned inside the marked region of interest 120. The code-symbol-decoding processor 115 disregards code-symbol information that is outside the marked region of interest 120.


As shown in FIG. 2, in a first alternative exemplary embodiment, the code-symbol-capturing subsystem 110 is an imaging subsystem 110A. The imaging subsystem 110A acquires code-symbol information by gathering light from an object 112 within the imaging subsystem's 110A field of view. Typically, the light is gathered through a lens, where it is focused onto an image sensor (e.g., a CCD, CMOS sensor, or any other sensor typically used in a digital imaging device). For example, the camera component of a smartphone device may serve as the imaging subsystem 110A. The image sensor converts the gathered light into a digital image depicting a two dimensional representation of the object 112 within the imaging subsystem's 110a field of view 116. The digital image represents the code-symbol information that is passed on to the code-symbol-decoding processor 115.


The code-symbol-decoding processor 115 is configured (e.g., through software) to analyze the code-symbol information (e.g., digital image) and detect a marked region of interest 120. For example, the where the object 112 within the imaging subsystem's 110A field of view 116 is a box with four code symbols positioned horizontally in a row, the code-symbol information will be a digital image that includes the four code symbols. The code-symbol-decoding processor 115 will analyze the code-symbol information to determine whether a region of interest 120 has been identified. For example, if the code-symbol-decoding processor determines that the leftmost code-symbol in the row has been marked as the region of interest 120, it will only read (e.g., decode) the leftmost code symbol. The other code symbols in the digital image will be disregarded. In this way, the code-symbol-decoding processor 115 determines that, in this instance, the leftmost code symbol is the region of interest 120. Having been initialized to regard the leftmost code symbol as the region of interest 120, for example, the code-symbol-decoding processor 115 will continue to only decode the leftmost code symbol in all subsequent digital images captured by the imaging subsystem 110A (at least until the system 100 is re-initialized to recognize a different region of interest). Consequently, once the system 100 has been initialized to decode only code symbols within the region of interest 120, the user knows that each time the user scans an object 112 having a row of code symbols, the system 100 will decode only the leftmost code symbol. This advantageously allows a user to continuously scan objects 112 bearing multiple code symbols without uncertainty as to which code symbol will be decoded. In this example, it will always be the leftmost code symbol. Another advantage of this technique is that it reduces the time to process code-symbol information because the code-symbol-decoding processor 115 only has to decode one of the code symbols.


As depicted in FIG. 3, a second alternative embodiment of the system 100 according to the present invention has a code-symbol-capturing subsystem 110 that is a laser scanning subsystem 110B. The laser scanning subsystem 110B gathers code-symbol information by projecting a beam (e.g., laser beam) from a light source (e.g., laser source, LED source) onto an object 112 within the field of view of the laser scanning subsystem 110B. Typically, the laser scanning subsystem 110B then detects the level of reflected light as the light beam sweeps across the code symbol(s) within the field of view. Using this technique, the laser scanning subsystem 110B is able to distinguish between dark areas and light areas on the code symbol (e.g., distinguish between black lines and white lines that form a barcode). More light is reflected from the light areas on the code symbol than the dark areas, so the optical energy reflected back to the laser scanning subsystem 110B will be a signal containing a series of peaks corresponding to the light areas and valleys corresponding to the dark areas. The laser scanning subsystem 110B typically uses a photoreceptor (e.g. photodiode) to receive the optical signal reflected from the code symbol, which optical signal represents code-symbol information that can be decoded by the code-symbol-decoding processor 115.


The region of interest 120 may be designated in various ways. As shown in FIG. 4, the region of interest 120 may be designated by a user-positioned marker 135. The marker 135 indicates which code symbol within the field of view 116 of the code-symbol-capturing subsystem 110 should be decoded, thereby initializing to thereafter decode only those code symbols appearing in the same region of interest 120 (e.g., the same position relative to the field of view 116) as the marked code symbol. In the example shown in FIG. 4, the marker 135 is positioned to the left of the second code symbol in the first row of code symbols on a barcode menu 112A. In this example, the system 100 would be pre-configured to recognize that the area in the field of view 116 that contains the code symbol to the immediate right of the marker 135 represents that portion of the field of view 116 that is the region of interest 120.


The type of marker 135 that is used will typically vary depending upon whether the code-symbol-capturing subsystem 110 is an imaging subsystem 110A or a laser scanning subsystem 110B. Where an imaging subsystem 110A is used, the marker 135 could be any symbol that the code-symbol-decoding processor 115 could recognize (e.g., through OCR techniques) as a marker 135 distinct from the code symbol(s) within the field of view 116. Because a laser scanning subsystem 110B typically cannot be configured to recognize different symbols (e.g., shapes, images) in the way that an imaging subsystem 110A can, a different type of marker is typically used for applications using a laser scanning subsystem 110B. For example, the marker 135 may be an area of enhanced reflectivity (e.g., a reflective sticker) that results in the reflected optical signal being greater than the reflected optical signal from the white spaces in the code-symbol. By using a signal-enhancing marker 135, the system 100 recognizes that the marker 135 is not part of the background of the object 112 (e.g., the white space on the barcode menu 112A page). Similarly, a signal-decreasing marker 135 could be used to create a reflected optical signal that is substantially below the other optical signal valleys. For example, the marker 135 could be a sticker that absorbed more of the light beam than the other areas of the barcode menu 112A on which the sticker was placed. Utilizing a sticker (e.g., a signal-enhancing sticker, a signal-decreasing sticker, or a sticker bearing a symbol) as a marker 135 would advantageously allow a user to define the region of interest 120 at any point in time during the scanning process, and to likewise change the region of interest simply by placing a sticker in a new position relative to the code-symbols within the field of view 116.



FIG. 5 illustrates an alternative technique for identifying a region of interest 120. The system 100 could be configured to recognize a finger as a marker 135. To initialize the system 100 to recognize a desired region of interest 120, the user points a finger to an area adjacent to the code symbol within the desired region of interest 120. A system 100 configured to recognize a finger as a marker 135 advantageously allows the user to indicate the region of interest 120 without the need for additional materials (e.g., stickers). Additionally, using a finger as a marker 135 results in quick identification of the region of interest 120 because there is no need to take time to retrieve a sticker, for example, and position the sticker in the appropriate location.


As another example, a code symbol may be used as a marker 135. It will be appreciated by persons of ordinary skill in the art that the settings on code symbol readers are often changed by scanning code symbols containing configuration information recognized by the code reader. Bar code readers, for example, are often configured to desired settings by scanning a barcode that will prompt the barcode reader to change its settings in the desired manner. In the same way, a code symbol could be used as a marker 135 that, when decoded by the system 100, could initialize the system 100 as to the desired region of interest 120. For instance, the code symbol marker 135 could contain information instructing the system 100 to only decode subsequently-scanned code symbols that are in the same region of interest 120 as the code symbol marker 135.


Referring now to FIG. 6, in another alternative embodiment, the code-symbol-capturing subsystem 110, in response to the detection of a marker 135 within its field of view 116, actually narrows its field of view to an adjusted field of view 116A that includes the region of interest 120 but excludes code symbols outside the region of interest 120. In this way, when the code-symbol-capturing subsystem 110 detects a marker 135, it narrows its field of view 116 such that it will subsequently (at least until re-initialized) scan only the region of interest 120 instead of a broader scan of the entire original field of view (e.g., an entire row of a barcode menu). By initializing the system 100 to adopt an adjusted field of view 116A that only scans the region of interest 120, the system 100 can achieve scans more quickly because it is scanning less of an area.


In a system 100 that utilizes a laser scanning subsystem 110B, for example, the adjusted field of view 116A can be achieved by resetting the sweep angle of the laser to a sweep angle that results in a scan of only the region of interest 120. It will be appreciated by one of ordinary skill in the art that, typically, the sweep angle is adjusted by modifying the rotation of the mirror(s) that reflects the laser substantially linearly across the field of view 116.


It will also be appreciated by a person of ordinary skill in the art that the code-symbol-capturing subsystem 110 will typically require a processor (e.g., a microprocessor, computer processor) to recognize a marker 135 and the associated marked region of interest, and to adjust the initial field of view 116 to an adjusted field of view 116A. This processor may be part of the same module that comprises the code-symbol-decoding processor 115, or this processor may be a separate processor.


By employing the technique of modifying the field of view 116 to an adjusted field of view 116A, only code-symbol information associated with code symbols within the region of interest 120 is transmitted to the code-symbol-decoding processor 115. Because the code-symbol-decoding processor 115 does not have to decipher multiple code symbols and/or determine which code symbol is marked, the decoding process is expedited.


In another alternative embodiment of the system 100 for reading code symbols according to the present disclosure, the adjusted field of view 116A that corresponds to the region of interest 120 is indicated (e.g., visually identified) to the user of the system 100 by a region of interest indicator. Typically, the region of interest indicator will be a light beam that is projected from the system 100 onto the region of interest 120. For example, if the region of interest 120 includes the third of four columns in a barcode menu, the system 100 will project a region of interest indicator (e.g., a blinking dot, a steady dot) onto the barcode appearing in the third column. This region of interest indicator identifies to the user that the system 100 has been initialized to scan only those barcodes that lie within the region of interest, which is presently set to correspond to the third barcode in the row of four barcodes.


To supplement the present disclosure, this application incorporates entirely by reference the following patents, patent application publications, and patent applications: U.S. Pat. No. 6,832,725; U.S. Pat. No. 7,159,783; U.S. Pat. No. 7,128,266; U.S. Pat. No. 7,413,127; U.S. Pat. No. 7,726,575; U.S. Pat. No. 8,390,909; U.S. Pat. No. 8,294,969; U.S. Pat. No. 8,408,469; U.S. Pat. No. 8,408,468; U.S. Pat. No. 8,381,979; U.S. Pat. No. 8,408,464; U.S. Pat. No. 8,317,105; U.S. Pat. No. 8,366,005; U.S. Pat. No. 8,424,768; U.S. Pat. No. 8,322,622; U.S. Pat. No. 8,371,507; U.S. Pat. No. 8,376,233; U.S. Pat. No. 8,457,013; U.S. Pat. No. 8,448,863; U.S. Pat. No. 8,459,557; U.S. Pat. No. 8,469,272; U.S. Pat. No. 8,474,712; U.S. Pat. No. 8,479,992; U.S. Pat. No. 8,490,877; U.S. Patent Application Publication No. 2012/0111946; U.S. Patent Application Publication No. 2012/0223141; U.S. Patent Application Publication No. 2012/0193423; U.S. Patent Application Publication No. 2012/0203647; U.S. Patent Application Publication No. 2012/0248188; U.S. Patent Application Publication No. 2012/0228382; U.S. Patent Application Publication No. 2012/0193407; U.S. Patent Application Publication No. 2012/0168511; U.S. Patent Application Publication No. 2012/0168512; U.S. Patent Application Publication No. 2010/0177749; U.S. Patent Application Publication No. 2010/0177080; U.S. Patent Application Publication No. 2010/0177707; U.S. Patent Application Publication No. 2010/0177076; U.S. Patent Application Publication No. 2009/0134221; U.S. Patent Application Publication No. 2012/0318869; U.S. Patent Application Publication No. 2013/0043312; U.S. Patent Application Publication No. 2013/0068840; U.S. Patent Application Publication No. 2013/0070322; U.S. Patent Application Publication No. 2013/0075168; U.S. Patent Application Publication No. 2013/0056285; U.S. Patent Application Publication No. 2013/0075464; U.S. Patent Application Publication No. 2013/0082104; U.S. Patent Application Publication No. 2010/0225757; U.S. patent application Ser. No. 13/347,219 for an OMNIDIRECTIONAL LASER SCANNING BAR CODE SYMBOL READER GENERATING A LASER SCANNING PATTERN WITH A HIGHLY NON-UNIFORM SCAN DENSITY WITH RESPECT TO LINE ORIENTATION, filed Jan. 10, 2012 (Good); U.S. patent application Ser. No. 13/347,193 for a HYBRID-TYPE BIOPTICAL LASER SCANNING AND DIGITAL IMAGING SYSTEM EMPLOYING DIGITAL IMAGER WITH FIELD OF VIEW OVERLAPPING FIELD OF FIELD OF LASER SCANNING SUBSYSTEM, filed Jan. 10, 2012 (Kearney et al.); U.S. patent application Ser. No. 13/367,047 for LASER SCANNING MODULES EMBODYING SILICONE SCAN ELEMENT WITH TORSIONAL HINGES, filed Feb. 6, 2012 (Feng et al.); U.S. patent application Ser. No. 13/400,748 for a LASER SCANNING BAR CODE SYMBOL READING SYSTEM HAVING INTELLIGENT SCAN SWEEP ANGLE ADJUSTMENT CAPABILITIES OVER THE WORKING RANGE OF THE SYSTEM FOR OPTIMIZED BAR CODE SYMBOL READING PERFORMANCE, filed Feb. 21, 2012 (Wilz); U.S. patent application Ser. No. 13/432,197 for a LASER SCANNING SYSTEM USING LASER BEAM SOURCES FOR PRODUCING LONG AND SHORT WAVELENGTHS IN COMBINATION WITH BEAM-WAIST EXTENDING OPTICS TO EXTEND THE DEPTH OF FIELD THEREOF WHILE RESOLVING HIGH RESOLUTION BAR CODE SYMBOLS HAVING MINIMUM CODE ELEMENT WIDTHS, filed Mar. 28, 2012 (Havens et al.); U.S. patent application Ser. No. 13/492,883 for a LASER SCANNING MODULE WITH ROTATABLY ADJUSTABLE LASER SCANNING ASSEMBLY, filed Jun. 10, 2012 (Hennick et al.); U.S. patent application Ser. No. 13/367,978 for a LASER SCANNING MODULE EMPLOYING AN ELASTOMERIC U-HINGE BASED LASER SCANNING ASSEMBLY, filed Feb. 7, 2012 (Feng et al.); U.S. patent application Ser. No. 13/852,097 for a System and Method for Capturing and Preserving Vehicle Event Data, filed Mar. 28, 2013 (Barker et al.); U.S. patent application Ser. No. 13/780,356 for a Mobile Device Having Object-Identification Interface, filed Feb. 28, 2013 (Samek et al.); U.S. patent application Ser. No. 13/780,158 for a Distraction Avoidance System, filed Feb. 28, 2013 (Sauerwein); U.S. patent application Ser. No. 13/784,933 for an Integrated Dimensioning and Weighing System, filed Mar. 5, 2013 (McCloskey et al.); U.S. patent application Ser. No. 13/785,177 for a Dimensioning System, filed Mar. 5, 2013 (McCloskey et al.); U.S. patent application Ser. No. 13/780,196 for Android Bound Service Camera Initialization, filed Feb. 28, 2013 (Todeschini et al.); U.S. patent application Ser. No. 13/792,322 for a Replaceable Connector, filed Mar. 11, 2013 (Skvoretz); U.S. patent application Ser. No. 13/780,271 for a Vehicle Computer System with Transparent Display, filed Feb. 28, 2013 (Fitch et al.); U.S. patent application Ser. No. 13/736,139 for an Electronic Device Enclosure, filed Jan. 8, 2013 (Chaney); U.S. patent application Ser. No. 13/771,508 for an Optical Redirection Adapter, filed Feb. 20, 2013 (Anderson); U.S. patent application Ser. No. 13/750,304 for Measuring Object Dimensions Using Mobile Computer, filed Jan. 25, 2013; U.S. patent application Ser. No. 13/471,973 for Terminals and Methods for Dimensioning Objects, filed May 15, 2012; U.S. patent application Ser. No. 13/895,846 for a Method of Programming a Symbol Reading System, filed Apr. 10, 2013 (Corcoran); U.S. patent application Ser. No. 13/867,386 for a Point of Sale (POS) Based Checkout System Supporting a Customer-Transparent Two-Factor Authentication Process During Product Checkout Operations, filed Apr. 22, 2013 (Cunningham et al.); U.S. patent application Ser. No. 13/888,884 for an Indicia Reading System Employing Digital Gain Control, filed May 7, 2013 (Xian et al.); U.S. patent application Ser. No. 13/895,616 for a Laser Scanning Code Symbol Reading System Employing Multi-Channel Scan Data Signal Processing with Synchronized Digital Gain Control (SDGC) for Full Range Scanning, filed May 16, 2013 (Xian et al.); U.S. patent application Ser. No. 13/897,512 for a Laser Scanning Code Symbol Reading System Providing Improved Control over the Length and Intensity Characteristics of a Laser Scan Line Projected Therefrom Using Laser Source Blanking Control, filed May 20, 2013 (Brady et al.); U.S. patent application Ser. No. 13/897,634 for a Laser Scanning Code Symbol Reading System Employing Programmable Decode Time-Window Filtering, filed May 20, 2013 (Wilz, Sr. et al.); U.S. patent application Ser. No. 13/902,242 for a System For Providing A Continuous Communication Link With A Symbol Reading Device, filed May 24, 2013 (Smith et al.); U.S. patent application Ser. No. 13/902,144, for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Chamberlin); U.S. patent application Ser. No. 13/902,110 for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Hollifield); U.S. patent application Ser. No. 13/912,262 for a Method of Error Correction for 3D Imaging Device, filed Jun. 7, 2013 (Jovanovski et al.); U.S. patent application Ser. No. 13/912,702 for a System and Method for Reading Code Symbols at Long Range Using Source Power Control, filed Jun. 7, 2013 (Xian et al.); U.S. patent application Ser. No. 13/922,339 for a System and Method for Reading Code Symbols Using a Variable Field of View, filed Jun. 20, 2013 (Xian et al.); U.S. patent application Ser. No. 13/927,398 for a Code Symbol Reading System Having Adaptive Autofocus, filed Jun. 26, 2013 (Todeschini); U.S. patent application Ser. No. 13/930,913 for a Mobile Device Having an Improved User Interface for Reading Code Symbols, filed Jun. 28, 2013 (Gelay et al.); and U.S. patent application Ser. No. 13/933,415 for an Electronic Device Case, filed Jul. 2, 2013 (London et al.).


In the specification and/or figures, typical embodiments of the invention have been disclosed. The present invention is not limited to such exemplary embodiments. The use of the term “and/or” includes any and all combinations of one or more of the associated listed items. The figures are schematic representations and so are not necessarily drawn to scale. Unless otherwise noted, specific terms have been used in a generic and descriptive sense and not for purposes of limitation.

Claims
  • 1. A system for reading code symbols on an object, comprising: a code-symbol-capturing subsystem configured to acquire information about code symbols within a full region of the code-symbol-capturing subsystem's field of view; anda code-symbol-decoding processor configured to: identify a marker within the full region designating a smaller region within the full region as a region of interest,detect code-symbol information only within the region of interest,disregard code-symbol information outside the region of interest, anddetect and decode code-symbols in the region of interest during all subsequent information acquisitions until a new region of interest is designated by the marker.
  • 2. The system for reading code symbols according to claim 1, wherein the code-symbol-capturing subsystem is an imaging subsystem for capturing images within the imaging subsystem's field of view.
  • 3. The system for reading code symbols according to claim 1, wherein the code-symbol-capturing subsystem is a laser scanning subsystem for scanning code symbols within the laser scanning subsystem's field of view.
  • 4. The system for reading code symbols according to claim 1, wherein the code-symbol-decoding processor is configured for identifying the region of interest in response to a user-positioned marker.
  • 5. The system for reading code symbols according to claim 4, wherein the user-positioned marker is a code symbol.
  • 6. The system for reading code symbols according to claim 4, wherein the user-positioned marker is a signal-enhancing marker.
  • 7. The system for reading code symbols according to claim 4, wherein the user-positioned marker is a signal-diminishing marker.
  • 8. A system for reading code symbols on an object, comprising: a code-symbol-capturing subsystem configured for: identifying a marker within a full region of the code-symbol-capturing subsystem's field of view, the marker designating a smaller region within the full region as a region of interest; andadjusting the code-symbol-capturing subsystem's initial full region of the field of view to an adjusted field of view that corresponds to the region of interest, andacquiring information only about code symbols within the region of interest while disregarding code-symbol information outside the region of interest, anddetecting and decoding code-symbols in the region of interest during all subsequent information acquisitions until a new region of interest is designated by the marker; anda code-symbol-decoding processor for processing code-symbol information acquired by the code-symbol-capturing subsystem.
  • 9. The system for reading code symbols according to claim 8, wherein the code-symbol-capturing subsystem is an imaging subsystem for capturing images of code symbols.
  • 10. The system for reading code symbols according to claim 8, wherein the code-symbol-capturing subsystem is a laser scanning subsystem for scanning code symbols.
  • 11. The system for reading code symbols according to claim 8, wherein the code-symbol-capturing subsystem is configured for identifying the region of interest in response to a user-positioned marker.
  • 12. The system for reading code symbols according to claim 8, wherein the user-positioned marker is a code symbol.
  • 13. The system for reading code symbols according to claim 8, wherein the user-positioned marker is a signal-enhancing marker.
  • 14. The system for reading code symbols according to claim 8, wherein the user-positioned marker is a signal-diminishing marker.
  • 15. A method for selectively reading code symbols on an object, comprising: providing a code symbol reader having a code-symbol-capturing subsystem for acquiring information about code symbols within the code-symbol-capturing subsystem's field of view, and a code-symbol-decoding processor for decoding code-symbols captured by the code-symbol-capturing subsystem;identifying a marker within a full region of the field of view, the marker designating a smaller region within the full region as a marked region of interest;decoding only those code symbols that are within the region of interest; anddetecting and decoding code-symbols in the marked region of interest during all subsequent information acquisitions until a new region of interest is designated by the marker.
  • 16. The method for selectively reading code symbols of claim 15, wherein the code-symbol-decoding processor is configured to disregard code-symbol information that is outside the marked region of interest.
  • 17. The method for selectively reading code symbols of claim 15, comprising: adjusting the code-symbol-capturing subsystem's initial field of view to an adjusted field of view that corresponds to the marked region of interest; andacquiring information about code symbols within the code-symbol-capturing subsystem's adjusted field of view that corresponds to the marked region of interest.
  • 18. The method for selectively reading code symbols of claim 15, comprising: marking the region of interest with a user-positioned marker;wherein the user-positioned marker is a finger.
  • 19. The method for selectively reading code symbols of claim 15, comprising: marking the region of interest with a user-positioned marker;wherein the user-positioned marker is a signal-enhancing marker.
  • 20. The method for selectively reading code symbols of claim 15, comprising: indicating the marked region of interest to a user of the code symbol reader by a region-of-interest indicator.
US Referenced Citations (211)
Number Name Date Kind
6832725 Gardiner et al. Dec 2004 B2
7128266 Zhu et al. Oct 2006 B2
7159783 Walczyk et al. Jan 2007 B2
7413127 Ehrhart et al. Aug 2008 B2
7726575 Wang et al. Jun 2010 B2
8294969 Plesko Oct 2012 B2
8317105 Kotlarsky et al. Nov 2012 B2
8322622 Liu Dec 2012 B2
8366005 Kotlarsky et al. Feb 2013 B2
8371507 Haggerty et al. Feb 2013 B2
8376233 Van Horn et al. Feb 2013 B2
8381979 Franz Feb 2013 B2
8390909 Plesko Mar 2013 B2
8408464 Zhu et al. Apr 2013 B2
8408468 Horn et al. Apr 2013 B2
8408469 Good Apr 2013 B2
8424768 Rueblinger et al. Apr 2013 B2
8448863 Xian et al. May 2013 B2
8457013 Essinger et al. Jun 2013 B2
8459557 Havens et al. Jun 2013 B2
8469272 Kearney Jun 2013 B2
8474712 Kearney et al. Jul 2013 B2
8479992 Kotlarsky et al. Jul 2013 B2
8490877 Kearney Jul 2013 B2
8517271 Kotlarsky et al. Aug 2013 B2
8523076 Good Sep 2013 B2
8528818 Ehrhart et al. Sep 2013 B2
8544737 Gomez et al. Oct 2013 B2
8548420 Grunow et al. Oct 2013 B2
8550335 Samek et al. Oct 2013 B2
8550354 Gannon et al. Oct 2013 B2
8550357 Kearney Oct 2013 B2
8556174 Kosecki et al. Oct 2013 B2
8556176 Van Horn et al. Oct 2013 B2
8556177 Hussey et al. Oct 2013 B2
8559767 Barber et al. Oct 2013 B2
8561895 Gomez et al. Oct 2013 B2
8561903 Sauerwein Oct 2013 B2
8561905 Edmonds et al. Oct 2013 B2
8565107 Pease et al. Oct 2013 B2
8571307 Li et al. Oct 2013 B2
8579200 Samek et al. Nov 2013 B2
8583924 Caballero et al. Nov 2013 B2
8584945 Wang et al. Nov 2013 B2
8587595 Wang Nov 2013 B2
8587697 Hussey et al. Nov 2013 B2
8588869 Sauerwein et al. Nov 2013 B2
8590789 Nahill et al. Nov 2013 B2
8596539 Havens et al. Dec 2013 B2
8596542 Havens et al. Dec 2013 B2
8596543 Havens et al. Dec 2013 B2
8599271 Havens et al. Dec 2013 B2
8599957 Peake et al. Dec 2013 B2
8600158 Li et al. Dec 2013 B2
8600167 Showering Dec 2013 B2
8602309 Longacre et al. Dec 2013 B2
8608053 Meier et al. Dec 2013 B2
8608071 Liu et al. Dec 2013 B2
8611309 Wang et al. Dec 2013 B2
8615487 Gomez et al. Dec 2013 B2
8621123 Caballero Dec 2013 B2
8622303 Meier et al. Jan 2014 B2
8628013 Ding Jan 2014 B2
8628015 Wang et al. Jan 2014 B2
8628016 Winegar Jan 2014 B2
8629926 Wang Jan 2014 B2
8630491 Longacre et al. Jan 2014 B2
8635309 Berthiaume et al. Jan 2014 B2
8636200 Kearney Jan 2014 B2
8636212 Nahill et al. Jan 2014 B2
8636215 Ding et al. Jan 2014 B2
8636224 Wang Jan 2014 B2
8638806 Wang et al. Jan 2014 B2
8640958 Lu et al. Feb 2014 B2
8640960 Wang et al. Feb 2014 B2
8643717 Li et al. Feb 2014 B2
8646692 Meier et al. Feb 2014 B2
8646694 Wang et al. Feb 2014 B2
8657200 Ren et al. Feb 2014 B2
8659397 Vargo et al. Feb 2014 B2
8668149 Good Mar 2014 B2
8678285 Kearney Mar 2014 B2
8678286 Smith et al. Mar 2014 B2
8682077 Longacre Mar 2014 B1
D702237 Oberpriller et al. Apr 2014 S
8687282 Feng et al. Apr 2014 B2
8692927 Pease et al. Apr 2014 B2
8695880 Bremer et al. Apr 2014 B2
8698949 Grunow et al. Apr 2014 B2
8702000 Barber et al. Apr 2014 B2
8717494 Gannon May 2014 B2
8720783 Biss et al. May 2014 B2
8723804 Fletcher et al. May 2014 B2
8723904 Marty et al. May 2014 B2
8727223 Wang May 2014 B2
20050279832 Kobayashi et al. Dec 2005 A1
20070063048 Havens et al. Mar 2007 A1
20070119939 Epshteyn et al. May 2007 A1
20080185432 Caballero et al. Aug 2008 A1
20090108071 Carlson Apr 2009 A1
20090134221 Zhu et al. May 2009 A1
20100177076 Essinger et al. Jul 2010 A1
20100177080 Essinger et al. Jul 2010 A1
20100177707 Essinger et al. Jul 2010 A1
20100177749 Essinger et al. Jul 2010 A1
20110075931 Chiou Mar 2011 A1
20110169999 Grunow et al. Jul 2011 A1
20110202554 Powilleit et al. Aug 2011 A1
20120111945 Kearney May 2012 A1
20120111946 Golant May 2012 A1
20120138685 Qu et al. Jun 2012 A1
20120168511 Kotlarsky et al. Jul 2012 A1
20120168512 Kotlarsky et al. Jul 2012 A1
20120193407 Barten Aug 2012 A1
20120193423 Samek Aug 2012 A1
20120203647 Smith Aug 2012 A1
20120223141 Good et al. Sep 2012 A1
20120228382 Havens et al. Sep 2012 A1
20120248188 Kearney Oct 2012 A1
20130043312 Van Horn Feb 2013 A1
20130056285 Meagher Mar 2013 A1
20130070322 Fritz et al. Mar 2013 A1
20130075168 Amundsen et al. Mar 2013 A1
20130082104 Kearney et al. Apr 2013 A1
20130175341 Kearney et al. Jul 2013 A1
20130175343 Good Jul 2013 A1
20130200158 Feng et al. Aug 2013 A1
20130214048 Wilz Aug 2013 A1
20130256418 Havens et al. Oct 2013 A1
20130257744 Daghigh et al. Oct 2013 A1
20130257759 Daghigh Oct 2013 A1
20130270346 Xian et al. Oct 2013 A1
20130278425 Cunningham et al. Oct 2013 A1
20130287258 Kearney Oct 2013 A1
20130292474 Xian et al. Nov 2013 A1
20130292475 Kotlarsky et al. Nov 2013 A1
20130292477 Hennick et al. Nov 2013 A1
20130293539 Hunt et al. Nov 2013 A1
20130293540 Laffargue et al. Nov 2013 A1
20130306728 Thuries et al. Nov 2013 A1
20130306730 Brady et al. Nov 2013 A1
20130306731 Pedraro Nov 2013 A1
20130306734 Xian et al. Nov 2013 A1
20130307964 Bremer et al. Nov 2013 A1
20130308625 Park et al. Nov 2013 A1
20130313324 Koziol et al. Nov 2013 A1
20130313325 Wilz et al. Nov 2013 A1
20130313326 Ehrhart Nov 2013 A1
20130327834 Hennick et al. Dec 2013 A1
20130341399 Xian et al. Dec 2013 A1
20130342717 Havens et al. Dec 2013 A1
20140001267 Giordano et al. Jan 2014 A1
20140002828 Laffargue et al. Jan 2014 A1
20140008430 Soule et al. Jan 2014 A1
20140008439 Wang Jan 2014 A1
20140021256 Qu et al. Jan 2014 A1
20140025584 Liu et al. Jan 2014 A1
20140027518 Edmonds et al. Jan 2014 A1
20140034723 Van Horn et al. Feb 2014 A1
20140034734 Sauerwein Feb 2014 A1
20140036848 Pease et al. Feb 2014 A1
20140039693 Havens et al. Feb 2014 A1
20140042814 Kather et al. Feb 2014 A1
20140049120 Kohtz et al. Feb 2014 A1
20140049635 Laffargue et al. Feb 2014 A1
20140061305 Nahill et al. Mar 2014 A1
20140061306 Wu et al. Mar 2014 A1
20140061307 Wang et al. Mar 2014 A1
20140063289 Hussey et al. Mar 2014 A1
20140066136 Sauerwein et al. Mar 2014 A1
20140067692 Ye et al. Mar 2014 A1
20140070005 Nahill et al. Mar 2014 A1
20140071840 Venancio Mar 2014 A1
20140074746 Wang Mar 2014 A1
20140075846 Woodburn Mar 2014 A1
20140076974 Havens et al. Mar 2014 A1
20140078341 Havens et al. Mar 2014 A1
20140078342 Li et al. Mar 2014 A1
20140078345 Showering Mar 2014 A1
20140084068 Gillet et al. Mar 2014 A1
20140086348 Peake et al. Mar 2014 A1
20140097249 Gomez et al. Apr 2014 A1
20140098284 Oberpriller et al. Apr 2014 A1
20140098792 Wang et al. Apr 2014 A1
20140100774 Showering Apr 2014 A1
20140100813 Showering Apr 2014 A1
20140103115 Meier et al. Apr 2014 A1
20140104413 McCloskey et al. Apr 2014 A1
20140104414 McCloskey et al. Apr 2014 A1
20140104416 Giordano et al. Apr 2014 A1
20140104451 Todeschini et al. Apr 2014 A1
20140106594 Skvoretz Apr 2014 A1
20140106725 Sauerwein Apr 2014 A1
20140108010 Maltseff et al. Apr 2014 A1
20140108402 Gomez et al. Apr 2014 A1
20140108682 Caballero Apr 2014 A1
20140110485 Toa et al. Apr 2014 A1
20140114530 Fitch et al. Apr 2014 A1
20140124577 Wang et al. May 2014 A1
20140124579 Ding May 2014 A1
20140125842 Winegar May 2014 A1
20140125853 Wang May 2014 A1
20140125999 Longacre et al. May 2014 A1
20140129378 Richardson May 2014 A1
20140131438 Kearney May 2014 A1
20140131441 Nahill et al. May 2014 A1
20140131445 Ding et al. May 2014 A1
20140133379 Wang et al. May 2014 A1
20140140585 Wang May 2014 A1
20140151453 Meier et al. Jun 2014 A1
20140160329 Ren et al. Jun 2014 A1
Foreign Referenced Citations (6)
Number Date Country
2093695 Aug 2009 EP
2450828 May 2012 EP
2518284 Mar 2015 GB
2013163789 Nov 2013 WO
2013173985 Nov 2013 WO
2014019130 Feb 2014 WO
Non-Patent Literature Citations (55)
Entry
U.S. Appl. No. 13/367,978, filed Feb. 7, 2012, (Feng et al.); now abandoned.
U.S. Appl. No. 13/736,139 for an Electronic Device Enclosure, filed Jan. 8, 2013 (Chaney); 40 pages.
U.S. Appl. No. 13/771,508 for an Optical Redirection Adapter, filed Feb. 20, 2013 (Anderson); 26 pages.
U.S. Appl. No. 13/780,356 for a Mobile Device Having Object Identification Interface, filed Feb. 28, 2013 (Samek et al.); 21 pages.
U.S. Appl. No. 13/852,097 for a System and Method for Capturing and Preserving Vehicle Event Data, filed Mar. 28, 2013 (Barker et al.); 20 pages.
U.S. Appl. No. 13/902,110 for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Hollifield); 29 pages.
U.S. Appl. No. 13/902,144, for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Chamberlin); 23 pages.
U.S. Appl. No. 13/902,242 for a System for Providing a Continuous Communication Link With a Symbol Reading Device, filed May 24, 2013 (Smith et al.); 24 pages.
U.S. Appl. No. 13/912,262 for a Method of Error Correction for 3D Imaging Device, filed Jun. 7, 2013 (Jovanovski et al.); 33 pages.
U.S. Appl. No. 13/912,702 for a System and Method for Reading Code Symbols at Long Range Using Source Power Control, filed Jun. 7, 2013 (Xian et al.); 24 pages.
U.S. Appl. No. 13/922,339 for a System and Method for Reading Code Symbols Using a Variable Field of View, filed Jun. 20, 2013 (Xian et al.); 23 pages.
U.S. Appl. No. 13/927,398 for a Code Symbol Reading System Having Adaptive Autofocus, filed Jun. 26, 2013 (Todeschini); 24 pages.
U.S. Appl. No. 13/930,913 for a Mobile Device Having an Improved User Interface for Reading Code Symbols, filed Jun. 28, 2013 (Gelay et al.); 24 pages.
U.S. Appl. No. 13/933,415 for an Electronic Device Case, filed Jul. 2, 2013 (London et al.); 47 pages.
U.S. Appl. No. 13/947,296 for a System and Method for Selectively Reading Code Symbols, filed Jul. 22, 2013 (Rueblinger et al.); 29 pages.
U.S. Appl. No. 13/950,544 for a Code Symbol Reading System Having Adjustable Object Detection, filed Jul. 25, 2013 (Jiang); 28 pages.
U.S. Appl. No. 13/961,408 for a Method for Manufacturing Laser Scanners, filed Aug. 7, 2013 (Saber et al.); 26 pages.
U.S. Appl. No. 14/018,729 for a Method for Operating a Laser Scanner, filed Sep. 5, 2013 (Feng et al.); 24 pages.
U.S. Appl. No. 14/019,616 for a Device Having Light Source to Reduce Surface Pathogens, filed Sep. 6, 2013 (Todeschini); 23 pages.
U.S. Appl. No. 14/023,762 for a Handheld Indicia Reader Having Locking Endcap, filed Sep. 11, 2013 (Gannon); 31 pages.
U.S. Appl. No. 14/035,474 for Augmented-Reality Signature Capture, filed Sep. 24, 2013 (Todeschini); 33 pages.
U.S. Appl. No. 14/047,896 for Terminal Having Illumination and Exposure Control filed Oct. 7, 2013 (Jovanovski et al.); 32 pages.
U.S. Appl. No. 14/053,175 for Imaging Apparatus Having Imaging Assembly, filed Oct. 14, 2013 (Barber); 39 pages.
U.S. Appl. No. 14/055,234 for Dimensioning System, filed Oct. 16, 2013 (Fletcher); 26 pages.
U.S. Appl. No. 14/053,314 for Indicia Reader, filed Oct. 14, 2013 (Huck); 29 pages.
U.S. Appl. No. 14/065,768 for Hybrid System and Method for Reading Indicia, filed Oct. 29, 2013 (Meier et al.); 22 pages.
U.S. Appl. No. 14/074,746 for Self-Checkout Shopping System, filed Nov. 8, 2013 (Hejl et al.); 26 pages.
U.S. Appl. No. 14/074,787 for Method and System for Configuring Mobile Devices via NFC Technology, filed Nov. 8, 2013 (Smith et al.); 28 pages.
U.S. Appl. No. 14/087,190 for Optimal Range Indicators for Bar Code Validation, filed Nov. 22, 2013 (Hejl); 27 pages.
U.S. Appl. No. 14/345,735 for Optical Indicia Reading Terminal with Combined Illumination filed Mar. 19, 2014 (Ouyang); 19 pages.
U.S. Appl. No. 14/101,965 for High Dynamic-Range Indicia Reading System, filed Dec. 10, 2013 (Xian); 28 pages.
U.S. Appl. No. 14/118,400 for Indicia Decoding Device with Security Lock, filed Nov. 18, 2013 (Liu); 28 pages.
U.S. Appl. No. 14/150,393 for Incicia-reader Having Unitary Construction Scanner, filed Jan. 8, 2014 (Colavito et al.); 28 pages.
U.S. Appl. No. 14/154,207 for Laser Barcode Scanner, filed Jan. 14, 2014 (Hou et al.); 26 pages.
U.S. Appl. No. 14/154,915 for Laser Scanning Module Employing a Laser Scanning Assembly having Elastomeric Wheel Hinges, filed Jan. 14, 2014 (Havens et al.); 24 pages.
U.S. Appl. No. 14/158,126 for Methods and Apparatus to Change a Feature Set on Data Collection Devices, filed Jan. 17, 2014 (Berthiaume et al.); 53 pages.
U.S. Appl. No. 14/342,551 for Terminal Having Image Data Format Conversion filed Mar. 4, 2014 (Lui et al.); 25 pages.
U.S. Appl. No. 14/342,544 for Imaging Based Barcode Scanner Engine with Multiple Elements Supported on a Common Printed Circuit Board filed Mar. 4, 2014 (Liu et al.); 27 pages.
U.S. Appl. No. 14/257,174 for Reading Apparatus Having Partial Frame Operating Mode filed Apr. 21, 2014, (Barber et al.), 67 pages.
U.S. Appl. No. 14/200,405 for Indicia Reader for Size-Limited Applications filed Mar. 7, 2014 (Feng et al.); 42 pages.
U.S. Appl. No. 14/166,103 for Indicia Reading Terminal Including Optical Filter filed Jan. 28, 2014 (Lu et al.); 29 pages.
U.S. Appl. 14/274,858 for Mobile Printer With Optional Battery Accessory, filed May 12, 2014, (Marty et al.), 26 pages.
U.S. Appl. No. 14/264,173 for Autofocus Lens System for Indicia Readers filed Apr. 29, 2014, (Ackley et al.), 39 pages.
U.S. Appl. No. 14/230,322 for Focus Module and Components with Actuator filed Mar. 31, 2014 (Feng et al.); 92 pages.
U.S. Appl. No. 14/222,994 for Method and Apparatus for Reading Optical Indicia Using a Plurality of Data filed Mar. 24, 2014 (Smith et al.); 30 pages.
U.S. Appl. No. 14/231,898 for Hand-Mounted Indicia-Reading Device with Finger Motion Triggering filed Apr. 1, 2014 (Van Horn et al.); 36 pages.
U.S. Appl. No. 29/486,759 for an Imaging Terminal, filed Apr. 2, 2014 (Oberpriller et al.); 8 pages.
U.S. Appl. No. 29/436,337 for an Electronic Device, filed Nov. 5, 2012 (Fitch et al.); 19 pages.
U.S. Appl. No. 29/458,405 for an Electronic Device, filed Jun. 19, 2013 (Fitch et al.); 22 pages.
U.S. Appl. No. 29/459,620 for an Electronic Device Enclosure, filed Jul. 2, 2013 (London et al.); 21 pages.
U.S. Appl. No. 29/459,681 for an Electronic Device Enclosure, filed Jul. 2, 2013 (Chaney et al.); 14 pages.
U.S. Appl. No. 29/459,785 for a Scanner and Charging Base, filed Jul. 3, 2013 (Fitch et al.); 21 pages.
U.S. Appl. No. 29/459,823 for a Scanner, filed Jul. 3, 2013 (Zhou et al.); 13 pages.
U.S. Appl. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.); 44 pages.
Combined Search and Examination Report, Application No. GB1412470.5, Related to Current Application, Dated Jan. 13, 2015, 6 pages.
Related Publications (1)
Number Date Country
20150021397 A1 Jan 2015 US