High ambient light electronic screen communication method

Information

  • Patent Grant
  • 10896361
  • Patent Number
    10,896,361
  • Date Filed
    Tuesday, January 22, 2019
    5 years ago
  • Date Issued
    Tuesday, January 19, 2021
    3 years ago
Abstract
The present invention embraces a method to obtain barcoded information off a computer screen under high ambient lighting conditions, including a direct sunlight environment. The method includes a user prompting a computer to present an active window on all or a portion of its screen. The computer generates a communication sequence, comprising information commonly included on a barcode, and displays the communication sequence on the computer screen. After activating the scanner into an alternative or screen mode, the user places the scanner on the active window of the computer screen and the scanner asynchronously receives the communication sequence. The scanner decodes the received communication sequence and obtains encoded information of the computer. The encoded information may include a unique identifier of the computer, such as a MAC address. With the MAC address, the scanner may wirelessly communicate with the computer to exchange other information.
Description
FIELD OF THE INVENTION

The present invention relates to methods of communicating with a computer, and in particularly methods of transferring information via screen communication to a bar code scanner in a high ambient light environment without the limitation of an in-focus optics system.


BACKGROUND

Generally speaking, the proliferation of 2D Imagers and their ease of operation have increased the demand for reading barcode symbols off a computer screen. Reading on-screen barcodes allows customers to take advantage of their inherent ability to be dynamically-created and therefore accommodate variable data, such as PC-specific Bluetooth MAC addresses. These techniques may support mission-critical applications such as package delivery. However, this process may break down in high ambient light environments. Often, the computer screen is located where significant and uncontrollable amounts of ambient light are flooding the screen, making the on-screen barcodes unusable.


Therefore, a need exists for a method to obtain the barcoded information off a computer screen under high ambient lighting conditions.


SUMMARY

Accordingly, in one aspect, the present invention embraces a method to transmit data between a computer screen and a decoding device utilizing an asynchronous communication method and an out-of focus condition for the decoding device. The data may be information which may have been encoded in a barcode and the decoding device may be a barcode scanner. The method may operate in a high ambient lighting environment, such as direct sunlight.


In an exemplary embodiment, the method of communicating with a computer may comprise activating a scanner to operate in a screen mode (or an alternative mode); and placing the scanner in contact with an active window displayed on a screen of the computer. The computer receives a prompt to cause the active window to be displayed on the screen, and the computer displays a communication sequence on the active window. The method continues with the scanner scanning the active window to asynchronously receive the communication sequence that comprises encoded information. The scanner operates in a defocused focus condition of the scanner and can operate successfully in a direct sunlight environment. The method further continues with the scanner converting the received encoded information to a unique identifier of the computer; and connecting the scanner to the computer using the unique identifier via an electronic communication method. The communication sequence comprises at least three optical states.


The active window is displayed on a portion of the screen of the computer. The unique identifier of the computer may be, but is not limited to, a MAC address. The electronic communication method may utilize, but is not limited to, a Bluetooth Low Energy (LE) technology.


In another exemplary embodiment, a method of communicating with a computer with a screen may comprise scanning, with a scanner, to detect a magnitude of an output of the screen, converting, by the scanner, the output of the screen into encoded information; and coupling the scanner to the computer using the encoded information. The computer causes all or part of the screen to display the output. The encoded information varies at least at a frame rate of the screen. The scanner operates independent of any focus condition. The output of the screen may be a multi-state sequence.


The scanner detects the magnitude of the output by averaging the pixels of the screen. The scanner may average all of the pixels of the screen. The scanner may use its internal CMOS sensor for this detection. The changing screen states enable asynchronous communications. The encoded information may comprise one or more instructions intended for the scanner. For example, the encoded information may include a MAC address. The scanner operates in a direct sunlight environment. The communication sequence may comprise a sequence of more than three optical states.


In yet another exemplary embodiment, a method of communicating with a computer may comprise receiving a prompt, at the computer, to cause a window to be displayed on a screen of the computer; displaying, by the computer, on the window a communication sequence comprising encoded information, wherein a scanner performs the steps of: (i) activating the scanner to operate in a screen mode) (or an alternative mode), (ii) placing the scanner in contact with the window, (iii) scanning the window with the scanner to asynchronously receive encoded information, and (iv) converting the received encoded information from the computer to a unique identifier of the computer. The method continues with the computer connecting the scanner to the computer using the unique identifier via an electronic communication method. The scanner operates independent of any focus condition of the scanner and operates independent of a direct sunlight environment. The communication sequence comprises a sequence of three optical states.


The aforementioned exemplary embodiments may include the following elements: The communication sequence may comprise a sequence of three optical states including an “on” state, an “off” state, and an intermediate state. On the active window, the “on” state is displayed as a white color, the “off” state is displayed as a black color, and the intermediate state is displayed as a gray color. At the “on” state or the “off” state, the communication sequence or multi-state sequence transitions to and from the intermediate state. At the intermediate state, the communication sequence or multi-state sequence transitions to and from either the “on” state or the “off” state. The “on” state, the “off” state or the intermediate state have at least a duration of a scanner frame rate. A transition from the intermediate state to the “on” or “off” state and back to a transition state comprises one bit of information. A transmission period for one bit is equal to two times a frame duration plus two times a sum of a rise time plus a fall time. The communication sequence or output is displayed by the computer onto the computer screen, or a portion of the computer screen.


In yet another exemplary embodiment, a method of communicating with a computer comprises activating a scanner to operate in a barcode mode. The scanner then attempts to decode a barcode displayed on a computer screen. If the scanner does not successfully decode the barcode: (i) the operation of the scanner changes to a screen mode, and (ii) the scanner proceeds to process a plurality of frames in a communication sequence displayed on the computer screen. The scanner then determines if the communication sequence comprises multiple states in the frames of the communication sequence.


If the communication sequence comprises multiple states, the scanner continues to decode the plurality of frames comprising bits of encoded information. After receiving and decoding a full bit stream, the operation of the scanner returns to the barcode mode. The scanner converts the encoded information to a unique ID of the computer, and proceeds to wirelessly connect to the computer using the unique ID.


In the barcode mode, if the scanner successfully decodes the barcode, the decoded barcode may be transmitted to a host computer. Then, the scanner repeats the operation in the barcode mode by attempting to decode another barcode. In the screen mode, if the communication sequence does not comprise multiple states, the operation of the scanner may be returned to the barcode mode.


In yet another exemplary embodiment, a method of obtaining a machine-readable code information with a barcode scanner from a digital display comprising the steps of: converting the machine-readable code information into a communication sequence comprised of sequential images, wherein each sequential image is defined by at least three colors or patterns; configuring the barcode scanner to operate in a manner that analyzes a plurality of images sequentially; displaying by the digital display the communication sequence; receiving with the barcode scanner the communication sequence, wherein the barcode scanner captures the sequential images; converting the captured sequential images back into the machine-readable code information; and connecting the barcode scanner to a computer using a unique identifier via an electronic communication method.


The communication sequence comprises a sequence of three optical states including an “on” state, an “off” state, and an intermediate state. Each optical state represents a different color. On the active window, the “on” state is displayed as a white color, the “off” state is displayed as a black color, and the intermediate state is displayed as a gray color. At the “on” state or the “off” state, the communication sequence transitions to and from the intermediate state; and at the intermediate state, the communication sequence transitions to and from either the “on” state or the “off” state. The “on” state, the “off” state or the intermediate state have at least a duration of a barcode scanner frame rate.


In yet another exemplary embodiment, a system for obtaining a machine-readable code information with a barcode scanner from a digital display, the system comprising: a computer capable of generating a multi-state sequence based on the machine-readable code information and presenting the multi-state sequence on the digital display; the barcode scanner capable of scanning the digital display and receiving the multi-state sequence using out-of-focus asynchronous data transmission, wherein, the barcode scanner converts the multi-state sequence back into the machine-readable code information; and the digital display for displaying the multi-state sequence.


The foregoing illustrative summary, as well as other exemplary objectives and/or advantages of the invention, and the manner in which the same are accomplished, are further explained within the following detailed description and its accompanying drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A illustrates a computer screen flooded with ambient light.



FIG. 1B illustrates a system for obtaining barcoded information off a computer screen under high ambient lighting conditions.



FIG. 2 illustrates an exemplary embodiment of a communication sequence that is displayed on the computer screen and subsequently scanned by a scanner in a high ambient light environment such as illustrated in FIG. 1A and FIG. 1B.



FIG. 3 illustrates another exemplary embodiment of the communication sequence that is produced by the scanner, after the communication sequence has been displayed on the computer screen.



FIG. 4 illustrates yet another exemplary embodiment of the communication sequence that is displayed on the computer screen including the white-gray-black pattern displayed on the computer screen.



FIG. 5 is a flow diagram illustrating a method of receiving at a scanner the communication sequence that includes encoded information.



FIG. 6 is a flow diagram illustrating another method of receiving at a scanner the communication sequence that includes encoded information.





DETAILED DESCRIPTION

The present invention embraces a method to obtain barcoded information off a computer screen under high ambient lighting conditions, including a direct sunlight environment. The method includes a user prompting a computer to present an active window on all or a portion of its screen. The computer generates a communication sequence, comprising information normally included in a barcode, and displays the communication sequence on the computer screen. After activating the scanner into a screen mode (or an alternative mode), the user places the scanner on the active window of the computer screen and the scanner asynchronously receives the communication sequence. The scanner decodes the received communication sequence and obtains encoded information from the computer. The encoded information may include a unique identifier of the computer, such as a MAC address. With the MAC address, the scanner may wirelessly communicate with the computer to exchange other information. The other information may include tracking the location of product inventory in a facility.


The present invention may be useful in an industrial environment where a user is picking up a product from inventory and delivering it to a delivery location in the facility. To track the user's activity, the user first connects to a specific computer. The connection includes reading three bar codes. The first barcode resets the scanner; the second barcode reconfigures the scanner or operation; and the third barcode includes a unique identifier of the computer. The unique identifier may be a MAC address.


After connecting with the specific computer, the user may walk in the facility and pick up a product and deliver it to a delivery location in the facility. At each stop, the user reads the barcode of the product, which is immediately transmitted to the specific computer. Accordingly, the location of the product is tracked by the specific computer. The specific computer is often in a high ambient light environment. A work around for this situation may include posting the barcodes on a piece of paper and attaching the paper to the specific computer. The scanner may be able to read the barcode on the paper in a high ambient light environment, but the paper may become lost or damaged. The present invention provides a method to avoid such a work around.


As described, herein, a communication sequence is equivalent to a communication signal. An optical state is equivalent to a color. A scanner is equivalent to a barcode scanner. A communication sequence may be a multi-state sequence.



FIG. 1A illustrates an embodiment 100 of a computer screen flooded with ambient light 102. The high ambient light environment may inhibit a barcode scanner from reading a barcode on the computer screen. The present invention may solve this problem by creating a small “window” on the computer screen that changes light intensity uniformly, allowing the scanner to be in contact with the screen and to operate in any focus condition including out-of-focus.



FIG. 1B illustrates a system 120 for obtaining barcoded information off a computer screen under high ambient lighting conditions, as depicted in FIG. 1A. System 120 comprises barcode scanner 122 that includes a scanner optical input/output (I/O) port 123, display 124 that includes a visual portion of the display 125, computer 126, and high ambient light 127. As illustrated, barcode scanner 122 attempts to reads a barcode off display 124 via the scanner optical I/O port 123 and visual portion of the display 125. Inasmuch as a successful reading of a barcode may not be possible, the user of barcode scanner 122 changes from a barcode reading mode to a screen scanning mode (or screen mode) and requests the computer 126 to display a communication sequence representing the barcode information. As illustrated, barcode scanner 122 can wirelessly communication with computer 126. Although not shown, barcode scanner 122 can communication with computer 126 on a non-wireless basis. A “screen mode” is sometimes referred to as an “alternative mode.”


The present invention may also solve the problem of how to communicate information without using a method to synchronize to the computer screen. Other methods have relied on using the time-dependent rastering scans for each frame of a CRT display to transmit a byte and a focused photodiode to receive the information in a synchronized fashion. This CRT method may not be useful with current screen technology and may be very sensitive to high ambient light conditions. Also, the CRT method may use the entire frame to transmit a bit and the receiving device may need to be focused. In the present invention, the method of communicating information may not need to be synchronized and may not need to rely on time dependent frame generation. Moreover, the receiving device may be a standard imaging barcode scanner.



FIG. 2 illustrates an exemplary embodiment of a communication sequence 200 that is displayed, i.e., projected, on the computer screen and subsequently scanned by a scanner in a high ambient light environment such as illustrated in FIG. 1. The display or projection of communication sequence 200 may be on all or a designated area of the computer screen such as in the lower right corner of the computer screen.


The scanner, such as a Honeywell SF61, may run at 60 frames per second, meaning the scanner may capture a new image approximately every 17 ms (i.e., the frame duration, or Ts is 17 ms). This situation may limit the fundamental information transfer since a “bit” has to be captured during this time period. A “bit” may be defined when the screen is either all on or all off. An LCD screen, Honeywell CV61, may have a “rise” time of 36 ms and a “fall” time of 20 ms between the two optical states that comprise a bit. The problem becomes how to make sure that the scanner sees the “on” state without missing it, while at the same time being sure that the scanner does not capture the bit twice. One solution is to utilize an intermediate state where half the screen illumination is on and half off creating a gray level that may serve as a “catch up” zone both to be sure that at least one frame is caught and to know when two frames are caught so as to avoid a double count.


The definition of a bit is therefore a process where the screen transitions from gray to, all-white (i.e., “on”) and stays “on” for approximately 17 ms, then transitions back to gray for approximately 17 ms. This method may allow for two “on states” of the same value to be adjacent, since they are separated by a gray period. This method may allow for two “on states” to be adjacent, but not separated by a gray value, thereby allowing the scanner to ignore one of the “on states”. This method may also allow for the accounting of any intermediate point of the rise and fall of the screen.



FIG. 2 shows the method of the present invention diagrammatically, and specifically the encodation of two bits, 01, on the CV61 display screen as a function of time. The encodation starts on the left with a gray frame that may be long enough to be sure the scanner does not miss a frame. The frame duration may be 17 ms because in the worst case, the scanner captures a frame every 17 ms. Hence, the method may not miss the gray frame. The 10 ms is half the “fall” time (Tf) to the first “off state” that needs to stay at zero level (off=0) again for 17 ms. The 18 ms is half the screen's inherent “rise” time (Tr) to return to gray that allows the scanner to become ready to measure the next “on or off” state, which in this example may be 256 (on=1). An “on” state (white) means that all pixels (256) are turned on. An “off” state (black) means that no pixels (zero) are turned on. A gray level means that half the pixels (128) are turned on or all the pixels are set at an intermediate color, such as light blue.



FIG. 2 also shows the minimum times for display output on the CV61 display screen to generate two bits dependably. The “on” or white state must be of at least the duration of the scanner frame rate to be sure to capture it. There may be an intermediate state, gray, to distinguish between two adjacent “on” or “off” states. The ½ rise time (Tr) and ½ fall time (Tf) are inherent in the display and are 18 and 10 ms respectively in this exemplary embodiment.


The intermediate state (where half the screen illumination is on and half off) creates a gray level that serves as a neutral or “rest” condition, and provides a mechanism for bit synchronization.


The total time for transmission of a bit is 2(Ts)+2(Tr+Tf). For 2 bits=4(17 ms)+2 (10 ms+18 ms)=124 ms. For 40 bits=124×20=2.5 s. For 48 bits=124×24=3.0 s (“s”=second, “ms”=milli-second). Ts is the frame duration.


When the scanner receives the MAC address of the computer, the scanner may wirelessly communicate with the computer. As it turns out, there may be some repeating data in the MAC address for the CV61 that may be the same 6-digit prefix and may be 001040, which may basically identify the chipset vendor. For example, valid CV61 MAC addresses may be: 00104057a64c, 001040b6afe3, and 00104042d4a0. Consequently, six hex characters must be represented, which is three bytes or 24 bits, in order to be useful to the application. Check bits could be added as well as a gray prefix and still be within the practical range of about two seconds of contact time with the display screen.


In summary, FIG. 2 illustrates a communication sequence comprising a sequence of three optical states. The communication sequence includes 2 bits, a “01” bit pattern. The time to transmit 2 bits=(2GF+BF+WF)+2(Tr)+2(Tf), where GF=gray frame; BF=black frame; WF=white frame; Tr=½ rise time; Tf=½ fall time. For one embodiment: GF=BF=WF=17 ms; Tf=10 ms; Tr=18 ms. A rise time equals the rise from the black level to the white level; similarly for the fall time. Tr may be a different value than Tf. The number of pixels displayed is the bit state. In another exemplary embodiment, the communication sequence may comprise a sequence of more than three optical states.



FIG. 3 illustrates another exemplary embodiment of the communication sequence that is produced by the scanner, after the communication sequence has been displayed on the computer screen. The communication sequence 300 of the present invention may be simulated by turning the screen on, off, or to gray. The scanner can detect the value of the constant state with the following algorithm:














int MatrixPluginDecode(unsigned char *pBuffer, int width, int height){


int iRow, iCol;


unsigned long ulSum, ulAvrg;


int MaxRow = 640; //# of vertical pixels


int MaxCol = 844; //# of horizontal pixels


ulAvrg = 0;


for (iRow = 0; iRow < MaxRow; iRow++){


    ulSum = 0;


    for (iCol = 0; iCol < MaxCol; iCol++){


        ulSum += pBuffer[iCol + iRow * MaxRow];


        }


    ulSum /= MaxCol;


    ulAvrg += ulSum;


    }


ulAvrg /= MaxRow;


printf(“%u\r\n”, ulAvrg); //outputs pixel average value to host computer


}









In this code, snippet “ulAvrg” is the variable of interest; it may represent the pixel average values returned by an Xenon scanner of its entire sensor area, for a single acquisition. The resulting output of the scanner (i.e. all successive acquisitions) at the end of the experiment is then fed into a spreadsheet and displayed graphically, as shown FIG. 3. The pixel average values are shown on the Y-axis, and taken together in sequence, illustrate a “0110” bit pattern. The pixel average values are represented by a bit state measured on a scale from 0 to 255, where 0 represents “perfect black” and 255 represents “perfect white”. This measurement is essentially the average value for all pixels of the sensor, where each individual pixel is encoded as an 8-bit value. In the aforementioned experiment, a “perfect white” was not achieved. Rather, a white frame was measured at approximately 100 bits state.



FIG. 4 illustrates yet another exemplary embodiment 400 of communication sequence 402 that is displayed on the computer screen including the white-gray-black pattern displayed on the computer screen. The present invention may be demonstrated utilizing the following devices: a standard Android tablet (e.g., Galaxy Tab 2 by Samsung) and a standard Xenon 1900 barcode scanner by Honeywell, with a black & white CMOS sensor. An Android application named AndroidScreenFlicker was created for the purpose of this experiment. The result of the experiment is illustrated in exemplary embodiment 400 of FIG. 4. The time-distributed pattern of communication sequence 402 is displayed, i.e. projected, on the LCD screen and alternates on the LCD screen of the tablet between the 3 defined color states i.e., the white-gray-black pattern 404. If each frame is approximately 17 ms, the user may not be able to observe the change in the white-gray-black pattern 404, and the LCD screen may appear gray during the projection of the communication sequence 402. The projection results in communication sequence 402 comprising the bits 0110 which may be the same as the bit pattern of FIG. 3. In this experiment, each time slot has a fixed 1-second duration, although other time periods may be selected for the experiment. The aforementioned Android application may be implemented on another OS, i.e., the application is OS agnostic.


In the experiment, special attention was taken to ensure that gray levels are effectively converted into a median value, and in this particular experiment the best performing color was Light Steel Blue. An example color may be html color code #B0C4DE, as defined at W3schools.com.


This experiment also highlights the importance of having an intermediate state (where all the pixels are Light Steel Blue or other intermediate color or half the screen illumination is on and half off creating a gray level to the scanner) since the scanner may capture an image at any given time including during the “rise” and the “fall” time of the LCD screen, as illustrated by indicator 302 in FIG. 3. The intermediate state creates a gray level that serves as a neutral or “rest” condition, and provides a mechanism for bit synchronization.



FIG. 5 is a flow diagram 500 illustrating a method of receiving at a scanner a communication sequence that includes encoded information. The method may comprises the steps of:


User prompts a computer to display an active window on the computer screen. (step 502) The prompting may include the user touching the “touch screen” of the computer. An active window on the computer screen changes light intensity uniformly to allow the scanner to be in contact with the screen and to operate in any focus condition, including out-of focus


User activates the scanner in a screen mode. (step 504)


User physically places the opening of the scanner in contact with the active window of the computer screen (step 506


Scanner begins to receive bits of encoded information. (step 508) Reception of bits may begin essentially immediately after the scanner in placed in contact with the active window. The bits may be received via asynchronous communication.


After receiving a full bit stream, the scanner converts the encoded information to a unique identifier of the computer. The unique identifier may be a MAC address of the computer (step 510). After a successful reception and conversion of the bit stream, the scanner provides a positive notification to the user, such as turning on a light, generating a sound or causing a vibration.


Scanner connects with computer utilizing ad hoc communication and the MAC address. (step 512)


Scanners may have several embodiments of operation. In one embodiment, the scanner continuously scans for barcodes after the scanner is turned on. In another embodiment, the scanner only scans when a scan button (trigger) is depressed. Generally, for the scan button case, after a barcode is read, the scanner turns off. However, there may be modes where the scanning continues as long as the scan button is activated.


In another embodiment, if the barcode scan is not successful, the scanner may automatically switch to a screen mode where the scanner scans a communication sequence that is displayed on the screen.



FIG. 6 is a flow diagram 600 illustrating another method of receiving at a scanner the communication sequence that includes encoded information. The method may comprises the steps of:


Activate a scanner to operate in barcode mode. (step 601)


In the barcode mode, the scanner attempts to decode a barcode displayed on computer screen. (step 602) For one embodiment of decoding, the scanner scans for contrast areas and then determines edges between high and low contrast.


Does the scanner successfully decode the barcode? (step 604) If the scanner does not successfully decode the barcode: (i) change the operation of the scanner to a screen mode and process a plurality of frames in a communication sequence displayed on the computer screen. (step 608) In one embodiment, several frames (e.g., six) are processed by averaging pixels.


The scanner determines if the communication sequence comprises multiple states in the frames of the communication sequence. (step 610) In one embodiment, there may be three states corresponding to black, gray, and white. In another embodiment there may be more than three states.


If the communication sequence comprises multiple states, the scanner continues to decode the plurality of frames comprising bits of encoded information. (step 612) After receiving and decoding a full bit stream, the scanner converts the encoded information to a unique ID of the computer, and then wirelessly connects to the computer using the unique ID. (step 614) The wireless connection may utilize Bluetooth technology. Subsequently, the operation of the scanner returns to the barcode mode. A full bit stream comprises all the bits contained between prolonged gray states that last for multiple frames. (step 616) In the barcode mode, if the scanner successfully decodes the barcode, the scanner transmits the decoded barcode to a host computer, and then the scanner repeats the operation in the barcode mode by attempting to decode another barcode. (steps 602, 604, 606)


In the screen mode, if the communication sequence does not comprise multiple states, the operation of the scanner returns to barcode mode. (steps 610, 616)


In summary, the computer may transmit its communication sequence with a unique identifier continuously in a loop in a designated area such as in the lower right corner of the computer screen. The user places the scanner in contact with the screen and makes an average reading of all or a portion of its sensor pixels during each frame of its scan time. The scanner records the bits and the long gray segment that indicates the start of the data until a full data transmission is obtained. The scanner may then unambiguously connect with the computer terminal, such as a CV61, using the acquired data. The present invention in a broad sense is a generic method of communication from a computer screen without the limitation of an in-focus optics system and in the presence of high environmental challenges such as direct sunlight or rain. The invention was inspired by observing the difficulties of a user in a package carrier environment when trying to read a bar code off the screen in the presence of direct sunlight. The present invention, therefore, is not limited to a specific application such as communicating a computer MAC address, but can be a method of communicating data for any purpose, including data collection.


The present invention allows the scanners already deployed to be used with only software modifications. With the present invention, a user may connect a common scanning device to a host computer in the presence of direct sunlight with a high confidence level of success.


To supplement the present disclosure, this application incorporates entirely by reference the following commonly assigned patents, patent application publications, and patent applications:

  • U.S. Pat. Nos. 6,832,725; 7,128,266;
  • 7,159,783; 7,413,127;
  • 7,726,575; 8,294,969;
  • 8,317,105; 8,322,622;
  • 8,366,005; 8,371,507;
  • 8,376,233; 8,381,979;
  • 8,390,909; 8,408,464;
  • 8,408,468; 8,408,469;
  • 8,424,768; 8,448,863;
  • 8,457,013; 8,459,557;
  • 8,469,272; 8,474,712;
  • 8,479,992; 8,490,877;
  • 8,517,271; 8,523,076;
  • 8,528,818; 8,544,737;
  • 8,548,242; 8,548,420;
  • 8,550,335; 8,550,354;
  • 8,550,357; 8,556,174;
  • 8,556,176; 8,556,177;
  • 8,559,767; 8,599,957;
  • 8,561,895; 8,561,903;
  • 8,561,905; 8,565,107;
  • 8,571,307; 8,579,200;
  • 8,583,924; 8,584,945;
  • 8,587,595; 8,587,697;
  • 8,588,869; 8,590,789;
  • 8,596,539; 8,596,542;
  • 8,596,543; 8,599,271;
  • 8,599,957; 8,600,158;
  • 8,600,167; 8,602,309;
  • 8,608,053; 8,608,071;
  • 8,611,309; 8,615,487;
  • 8,616,454; 8,621,123;
  • 8,622,303; 8,628,013;
  • 8,628,015; 8,628,016;
  • 8,629,926; 8,630,491;
  • 8,635,309; 8,636,200;
  • 8,636,212; 8,636,215;
  • 8,636,224; 8,638,806;
  • 8,640,958; 8,640,960;
  • 8,643,717; 8,646,692;
  • 8,646,694; 8,657,200;
  • 8,659,397; 8,668,149;
  • 8,678,285; 8,678,286;
  • 8,682,077; 8,687,282;
  • 8,692,927; 8,695,880;
  • 8,698,949; 8,717,494;
  • 8,717,494; 8,720,783;
  • 8,723,804; 8,723,904;
  • 8,727,223; U.S. Pat. No. D702,237;
  • U.S. Pat. Nos. 8,740,082; 8,740,085;
  • 8,746,563; 8,750,445;
  • 8,752,766; 8,756,059;
  • 8,757,495; 8,760,563;
  • 8,763,909; 8,777,108;
  • 8,777,109; 8,779,898;
  • 8,781,520; 8,783,573;
  • 8,789,757; 8,789,758;
  • 8,789,759; 8,794,520;
  • 8,794,522; 8,794,525;
  • 8,794,526; 8,798,367;
  • 8,807,431; 8,807,432;
  • 8,820,630; 8,822,848;
  • 8,824,692; 8,824,696;
  • 8,842,849; 8,844,822;
  • 8,844,823; 8,849,019;
  • 8,851,383; 8,854,633;
  • 8,866,963; 8,868,421;
  • 8,868,519; 8,868,802;
  • 8,868,803; 8,870,074;
  • 8,879,639; 8,880,426;
  • 8,881,983; 8,881,987;
  • 8,903,172; 8,908,995;
  • 8,910,870; 8,910,875;
  • 8,914,290; 8,914,788;
  • 8,915,439; 8,915,444;
  • 8,916,789; 8,918,250;
  • 8,918,564; 8,925,818;
  • 8,939,374; 8,942,480;
  • 8,944,313; 8,944,327;
  • 8,944,332; 8,950,678;
  • 8,967,468; 8,971,346;
  • 8,976,030; 8,976,368;
  • 8,978,981; 8,978,983;
  • 8,978,984; 8,985,456;
  • 8,985,457; 8,985,459;
  • 8,985,461; 8,988,578;
  • 8,988,590; 8,991,704;
  • 8,996,194; 8,996,384;
  • 9,002,641; 9,007,368;
  • 9,010,641; 9,015,513;
  • 9,016,576; 9,022,288;
  • 9,030,964; 9,033,240;
  • 9,033,242; 9,036,054;
  • 9,037,344; 9,038,911;
  • 9,038,915; 9,047,098;
  • 9,047,359; 9,047,420;
  • 9,047,525; 9,047,531;
  • 9,053,055; 9,053,378;
  • 9,053,380; 9,058,526;
  • 9,064,165; 9,064,167;
  • 9,064,168; 9,064,254;
  • 9,066,032; 9,070,032;
  • U.S. Design Pat. No. D716,285;
  • U.S. Design Pat. No. D723,560;
  • U.S. Design Pat. No. D730,357;
  • U.S. Design Pat. No. D730,901;
  • U.S. Design Pat. No. D730,902;
  • U.S. Design Pat. No. D733,112;
  • U.S. Design Pat. No. D734,339;
  • International Publication No. 2013/163789;
  • International Publication No. 2013/173985;
  • International Publication No. 2014/019130;
  • International Publication No. 2014/110495;
  • U.S. Patent Application Publication No. 2008/0185432;
  • U.S. Patent Application Publication No. 2009/0134221;
  • U.S. Patent Application Publication No. 2010/0177080;
  • U.S. Patent Application Publication No. 2010/0177076;
  • U.S. Patent Application Publication No. 2010/0177707;
  • U.S. Patent Application Publication No. 2010/0177749;
  • U.S. Patent Application Publication No. 2010/0265880;
  • U.S. Patent Application Publication No. 2011/0202554;
  • U.S. Patent Application Publication No. 2012/0111946;
  • U.S. Patent Application Publication No. 2012/0168511;
  • U.S. Patent Application Publication No. 2012/0168512;
  • U.S. Patent Application Publication No. 2012/0193423;
  • U.S. Patent Application Publication No. 2012/0203647;
  • U.S. Patent Application Publication No. 2012/0223141;
  • U.S. Patent Application Publication No. 2012/0228382;
  • U.S. Patent Application Publication No. 2012/0248188;
  • U.S. Patent Application Publication No. 2013/0043312;
  • U.S. Patent Application Publication No. 2013/0082104;
  • U.S. Patent Application Publication No. 2013/0175341;
  • U.S. Patent Application Publication No. 2013/0175343;
  • U.S. Patent Application Publication No. 2013/0257744;
  • U.S. Patent Application Publication No. 2013/0257759;
  • U.S. Patent Application Publication No. 2013/0270346;
  • U.S. Patent Application Publication No. 2013/0287258;
  • U.S. Patent Application Publication No. 2013/0292475;
  • U.S. Patent Application Publication No. 2013/0292477;
  • U.S. Patent Application Publication No. 2013/0293539;
  • U.S. Patent Application Publication No. 2013/0293540;
  • U.S. Patent Application Publication No. 2013/0306728;
  • U.S. Patent Application Publication No. 2013/0306731;
  • U.S. Patent Application Publication No. 2013/0307964;
  • U.S. Patent Application Publication No. 2013/0308625;
  • U.S. Patent Application Publication No. 2013/0313324;
  • U.S. Patent Application Publication No. 2013/0313325;
  • U.S. Patent Application Publication No. 2013/0342717;
  • U.S. Patent Application Publication No. 2014/0001267;
  • U.S. Patent Application Publication No. 2014/0008439;
  • U.S. Patent Application Publication No. 2014/0025584;
  • U.S. Patent Application Publication No. 2014/0034734;
  • U.S. Patent Application Publication No. 2014/0036848;
  • U.S. Patent Application Publication No. 2014/0039693;
  • U.S. Patent Application Publication No. 2014/0042814;
  • U.S. Patent Application Publication No. 2014/0049120;
  • U.S. Patent Application Publication No. 2014/0049635;
  • U.S. Patent Application Publication No. 2014/0061306;
  • U.S. Patent Application Publication No. 2014/0063289;
  • U.S. Patent Application Publication No. 2014/0066136;
  • U.S. Patent Application Publication No. 2014/0067692;
  • U.S. Patent Application Publication No. 2014/0070005;
  • U.S. Patent Application Publication No. 2014/0071840;
  • U.S. Patent Application Publication No. 2014/0074746;
  • U.S. Patent Application Publication No. 2014/0076974;
  • U.S. Patent Application Publication No. 2014/0078341;
  • U.S. Patent Application Publication No. 2014/0078345;
  • U.S. Patent Application Publication No. 2014/0097249;
  • U.S. Patent Application Publication No. 2014/0098792;
  • U.S. Patent Application Publication No. 2014/0100813;
  • U.S. Patent Application Publication No. 2014/0103115;
  • U.S. Patent Application Publication No. 2014/0104413;
  • U.S. Patent Application Publication No. 2014/0104414;
  • U.S. Patent Application Publication No. 2014/0104416;
  • U.S. Patent Application Publication No. 2014/0104451;
  • U.S. Patent Application Publication No. 2014/0106594;
  • U.S. Patent Application Publication No. 2014/0106725;
  • U.S. Patent Application Publication No. 2014/0108010;
  • U.S. Patent Application Publication No. 2014/0108402;
  • U.S. Patent Application Publication No. 2014/0110485;
  • U.S. Patent Application Publication No. 2014/0114530;
  • U.S. Patent Application Publication No. 2014/0124577;
  • U.S. Patent Application Publication No. 2014/0124579;
  • U.S. Patent Application Publication No. 2014/0125842;
  • U.S. Patent Application Publication No. 2014/0125853;
  • U.S. Patent Application Publication No. 2014/0125999;
  • U.S. Patent Application Publication No. 2014/0129378;
  • U.S. Patent Application Publication No. 2014/0131438;
  • U.S. Patent Application Publication No. 2014/0131441;
  • U.S. Patent Application Publication No. 2014/0131443;
  • U.S. Patent Application Publication No. 2014/0131444;
  • U.S. Patent Application Publication No. 2014/0131445;
  • U.S. Patent Application Publication No. 2014/0131448;
  • U.S. Patent Application Publication No. 2014/0133379;
  • U.S. Patent Application Publication No. 2014/0136208;
  • U.S. Patent Application Publication No. 2014/0140585;
  • U.S. Patent Application Publication No. 2014/0151453;
  • U.S. Patent Application Publication No. 2014/0152882;
  • U.S. Patent Application Publication No. 2014/0158770;
  • U.S. Patent Application Publication No. 2014/0159869;
  • U.S. Patent Application Publication No. 2014/0166755;
  • U.S. Patent Application Publication No. 2014/0166759;
  • U.S. Patent Application Publication No. 2014/0168787;
  • U.S. Patent Application Publication No. 2014/0175165;
  • U.S. Patent Application Publication No. 2014/0175172;
  • U.S. Patent Application Publication No. 2014/0191644;
  • U.S. Patent Application Publication No. 2014/0191913;
  • U.S. Patent Application Publication No. 2014/0197238;
  • U.S. Patent Application Publication No. 2014/0197239;
  • U.S. Patent Application Publication No. 2014/0197304;
  • U.S. Patent Application Publication No. 2014/0214631;
  • U.S. Patent Application Publication No. 2014/0217166;
  • U.S. Patent Application Publication No. 2014/0217180;
  • U.S. Patent Application Publication No. 2014/0231500;
  • U.S. Patent Application Publication No. 2014/0232930;
  • U.S. Patent Application Publication No. 2014/0247315;
  • U.S. Patent Application Publication No. 2014/0263493;
  • U.S. Patent Application Publication No. 2014/0263645;
  • U.S. Patent Application Publication No. 2014/0267609;
  • U.S. Patent Application Publication No. 2014/0270196;
  • U.S. Patent Application Publication No. 2014/0270229;
  • U.S. Patent Application Publication No. 2014/0278387;
  • U.S. Patent Application Publication No. 2014/0278391;
  • U.S. Patent Application Publication No. 2014/0282210;
  • U.S. Patent Application Publication No. 2014/0284384;
  • U.S. Patent Application Publication No. 2014/0288933;
  • U.S. Patent Application Publication No. 2014/0297058;
  • U.S. Patent Application Publication No. 2014/0299665;
  • U.S. Patent Application Publication No. 2014/0312121;
  • U.S. Patent Application Publication No. 2014/0319220;
  • U.S. Patent Application Publication No. 2014/0319221;
  • U.S. Patent Application Publication No. 2014/0326787;
  • U.S. Patent Application Publication No. 2014/0332590;
  • U.S. Patent Application Publication No. 2014/0344943;
  • U.S. Patent Application Publication No. 2014/0346233;
  • U.S. Patent Application Publication No. 2014/0351317;
  • U.S. Patent Application Publication No. 2014/0353373;
  • U.S. Patent Application Publication No. 2014/0361073;
  • U.S. Patent Application Publication No. 2014/0361082;
  • U.S. Patent Application Publication No. 2014/0362184;
  • U.S. Patent Application Publication No. 2014/0363015;
  • U.S. Patent Application Publication No. 2014/0369511;
  • U.S. Patent Application Publication No. 2014/0374483;
  • U.S. Patent Application Publication No. 2014/0374485;
  • U.S. Patent Application Publication No. 2015/0001301;
  • U.S. Patent Application Publication No. 2015/0001304;
  • U.S. Patent Application Publication No. 2015/0003673;
  • U.S. Patent Application Publication No. 2015/0009338;
  • U.S. Patent Application Publication No. 2015/0009610;
  • U.S. Patent Application Publication No. 2015/0014416;
  • U.S. Patent Application Publication No. 2015/0021397;
  • U.S. Patent Application Publication No. 2015/0028102;
  • U.S. Patent Application Publication No. 2015/0028103;
  • U.S. Patent Application Publication No. 2015/0028104;
  • U.S. Patent Application Publication No. 2015/0029002;
  • U.S. Patent Application Publication No. 2015/0032709;
  • U.S. Patent Application Publication No. 2015/0039309;
  • U.S. Patent Application Publication No. 2015/0039878;
  • U.S. Patent Application Publication No. 2015/0040378;
  • U.S. Patent Application Publication No. 2015/0048168;
  • U.S. Patent Application Publication No. 2015/0049347;
  • U.S. Patent Application Publication No. 2015/0051992;
  • U.S. Patent Application Publication No. 2015/0053766;
  • U.S. Patent Application Publication No. 2015/0053768;
  • U.S. Patent Application Publication No. 2015/0053769;
  • U.S. Patent Application Publication No. 2015/0060544;
  • U.S. Patent Application Publication No. 2015/0062366;
  • U.S. Patent Application Publication No. 2015/0063215;
  • U.S. Patent Application Publication No. 2015/0063676;
  • U.S. Patent Application Publication No. 2015/0069130;
  • U.S. Patent Application Publication No. 2015/0071819;
  • U.S. Patent Application Publication No. 2015/0083800;
  • U.S. Patent Application Publication No. 2015/0086114;
  • U.S. Patent Application Publication No. 2015/0088522;
  • U.S. Patent Application Publication No. 2015/0096872;
  • U.S. Patent Application Publication No. 2015/0099557;
  • U.S. Patent Application Publication No. 2015/0100196;
  • U.S. Patent Application Publication No. 2015/0102109;
  • U.S. Patent Application Publication No. 2015/0115035;
  • U.S. Patent Application Publication No. 2015/0127791;
  • U.S. Patent Application Publication No. 2015/0128116;
  • U.S. Patent Application Publication No. 2015/0129659;
  • U.S. Patent Application Publication No. 2015/0133047;
  • U.S. Patent Application Publication No. 2015/0134470;
  • U.S. Patent Application Publication No. 2015/0136851;
  • U.S. Patent Application Publication No. 2015/0136854;
  • U.S. Patent Application Publication No. 2015/0142492;
  • U.S. Patent Application Publication No. 2015/0144692;
  • U.S. Patent Application Publication No. 2015/0144698;
  • U.S. Patent Application Publication No. 2015/0144701;
  • U.S. Patent Application Publication No. 2015/0149946;
  • U.S. Patent Application Publication No. 2015/0161429;
  • U.S. Patent Application Publication No. 2015/0169925;
  • U.S. Patent Application Publication No. 2015/0169929;
  • U.S. Patent Application Publication No. 2015/0178523;
  • U.S. Patent Application Publication No. 2015/0178534;
  • U.S. Patent Application Publication No. 2015/0178535;
  • U.S. Patent Application Publication No. 2015/0178536;
  • U.S. Patent Application Publication No. 2015/0178537;
  • U.S. Patent Application Publication No. 2015/0181093;
  • U.S. Patent Application Publication No. 2015/0181109;
  • U.S. patent application Ser. No. 13/367,978 for a Laser Scanning Module Employing an Elastomeric U-Hinge Based Laser Scanning Assembly, filed Feb. 7, 2012 (Feng et al.);
  • U.S. patent application Ser. No. 29/458,405 for an Electronic Device, filed Jun. 19, 2013 (Fitch et al.);
  • U.S. patent application Ser. No. 29/459,620 for an Electronic Device Enclosure, filed Jul. 2, 2013 (London et al.);
  • U.S. patent application Ser. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/150,393 for Indicia-reader Having Unitary Construction Scanner, filed Jan. 8, 2014 (Colavito et al.);
  • U.S. patent application Ser. No. 14/200,405 for Indicia Reader for Size-Limited Applications filed Mar. 7, 2014 (Feng et al.);
  • U.S. patent application Ser. No. 14/231,898 for Hand-Mounted Indicia-Reading Device with Finger Motion Triggering filed Apr. 1, 2014 (Van Horn et al.);
  • U.S. patent application Ser. No. 29/486,759 for an Imaging Terminal, filed Apr. 2, 2014 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/257,364 for Docking System and Method Using Near Field Communication filed Apr. 21, 2014 (Showering);
  • U.S. patent application Ser. No. 14/264,173 for Autofocus Lens System for Indicia Readers filed Apr. 29, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/277,337 for MULTIPURPOSE OPTICAL READER, filed May 14, 2014 (Jovanovski et al.);
  • U.S. patent application Ser. No. 14/283,282 for TERMINAL HAVING ILLUMINATION AND FOCUS CONTROL filed May 21, 2014 (Liu et al.);
  • U.S. patent application Ser. No. 14/327,827 for a MOBILE-PHONE ADAPTER FOR ELECTRONIC TRANSACTIONS, filed Jul. 10, 2014 (Hejl);
  • U.S. patent application Ser. No. 14/334,934 for a SYSTEM AND METHOD FOR INDICIA VERIFICATION, filed Jul. 18, 2014 (Hejl);
  • U.S. patent application Ser. No. 14/339,708 for LASER SCANNING CODE SYMBOL READING SYSTEM, filed Jul. 24, 2014 (Xian et al.);
  • U.S. patent application Ser. No. 14/340,627 for an AXIALLY REINFORCED FLEXIBLE SCAN ELEMENT, filed Jul. 25, 2014 (Rueblinger et al.);
  • U.S. patent application Ser. No. 14/446,391 for MULTIFUNCTION POINT OF SALE APPARATUS WITH OPTICAL SIGNATURE CAPTURE filed Jul. 30, 2014 (Good et al.);
  • U.S. patent application Ser. No. 14/452,697 for INTERACTIVE INDICIA READER, filed Aug. 6, 2014 (Todeschini);
  • U.S. patent application Ser. No. 14/453,019 for DIMENSIONING SYSTEM WITH GUIDED ALIGNMENT, filed Aug. 6, 2014 (Li et al.);
  • U.S. patent application Ser. No. 14/462,801 for MOBILE COMPUTING DEVICE WITH DATA COGNITION SOFTWARE, filed on Aug. 19, 2014 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/483,056 for VARIABLE DEPTH OF FIELD BARCODE SCANNER filed Sep. 10, 2014 (McCloskey et al.);
  • U.S. patent application Ser. No. 14/513,808 for IDENTIFYING INVENTORY ITEMS IN A STORAGE FACILITY filed Oct. 14, 2014 (Singel et al.);
  • U.S. patent application Ser. No. 14/519,195 for HANDHELD DIMENSIONING SYSTEM WITH FEEDBACK filed Oct. 21, 2014 (Laffargue et al.);
  • U.S. patent application Ser. No. 14/519,179 for DIMENSIONING SYSTEM WITH MULTIPATH INTERFERENCE MITIGATION filed Oct. 21, 2014 (Thuries et al.);
  • U.S. patent application Ser. No. 14/519,211 for SYSTEM AND METHOD FOR DIMENSIONING filed Oct. 21, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/519,233 for HANDHELD DIMENSIONER WITH DATA-QUALITY INDICATION filed Oct. 21, 2014 (Laffargue et al.);
  • U.S. patent application Ser. No. 14/519,249 for HANDHELD DIMENSIONING SYSTEM WITH MEASUREMENT-CONFORMANCE FEEDBACK filed Oct. 21, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/527,191 for METHOD AND SYSTEM FOR RECOGNIZING SPEECH USING WILDCARDS IN AN EXPECTED RESPONSE filed Oct. 29, 2014 (Braho et al.);
  • U.S. patent application Ser. No. 14/529,563 for ADAPTABLE INTERFACE FOR A MOBILE COMPUTING DEVICE filed Oct. 31, 2014 (Schoon et al.);
  • U.S. patent application Ser. No. 14/529,857 for BARCODE READER WITH SECURITY FEATURES filed Oct. 31, 2014 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/398,542 for PORTABLE ELECTRONIC DEVICES HAVING A SEPARATE LOCATION TRIGGER UNIT FOR USE IN CONTROLLING AN APPLICATION UNIT filed Nov. 3, 2014 (Bian et al.);
  • U.S. patent application Ser. No. 14/531,154 for DIRECTING AN INSPECTOR THROUGH AN INSPECTION filed Nov. 3, 2014 (Miller et al.);
  • U.S. patent application Ser. No. 14/533,319 for BARCODE SCANNING SYSTEM USING WEARABLE DEVICE WITH EMBEDDED CAMERA filed Nov. 5, 2014 (Todeschini);
  • U.S. patent application Ser. No. 14/535,764 for CONCATENATED EXPECTED RESPONSES FOR SPEECH RECOGNITION filed Nov. 7, 2014 (Braho et al.);
  • U.S. patent application Ser. No. 14/568,305 for AUTO-CONTRAST VIEWFINDER FOR AN INDICIA READER filed Dec. 12, 2014 (Todeschini);
  • U.S. patent application Ser. No. 14/573,022 for DYNAMIC DIAGNOSTIC INDICATOR GENERATION filed Dec. 17, 2014 (Goldsmith);
  • U.S. patent application Ser. No. 14/578,627 for SAFETY SYSTEM AND METHOD filed Dec. 22, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/580,262 for MEDIA GATE FOR THERMAL TRANSFER PRINTERS filed Dec. 23, 2014 (Bowles);
  • U.S. patent application Ser. No. 14/590,024 for SHELVING AND PACKAGE LOCATING SYSTEMS FOR DELIVERY VEHICLES filed Jan. 6, 2015 (Payne);
  • U.S. patent application Ser. No. 14/596,757 for SYSTEM AND METHOD FOR DETECTING BARCODE PRINTING ERRORS filed Jan. 14, 2015 (Ackley);
  • U.S. patent application Ser. No. 14/416,147 for OPTICAL READING APPARATUS HAVING VARIABLE SETTINGS filed Jan. 21, 2015 (Chen et al.);
  • U.S. patent application Ser. No. 14/614,706 for DEVICE FOR SUPPORTING AN ELECTRONIC TOOL ON A USER'S HAND filed Feb. 5, 2015 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/614,796 for CARGO APPORTIONMENT TECHNIQUES filed Feb. 5, 2015 (Morton et al.);
  • U.S. patent application Ser. No. 29/516,892 for TABLE COMPUTER filed Feb. 6, 2015 (Bidwell et al.);
  • U.S. patent application Ser. No. 14/619,093 for METHODS FOR TRAINING A SPEECH RECOGNITION SYSTEM filed Feb. 11, 2015 (Pecorari);
  • U.S. patent application Ser. No. 14/628,708 for DEVICE, SYSTEM, AND METHOD FOR DETERMINING THE STATUS OF CHECKOUT LANES filed Feb. 23, 2015 (Todeschini);
  • U.S. patent application Ser. No. 14/630,841 for TERMINAL INCLUDING IMAGING ASSEMBLY filed Feb. 25, 2015 (Gomez et al.);
  • U.S. patent application Ser. No. 14/635,346 for SYSTEM AND METHOD FOR RELIABLE STORE-AND-FORWARD DATA HANDLING BY ENCODED INFORMATION READING TERMINALS filed Mar. 2, 2015 (Sevier);
  • U.S. patent application Ser. No. 29/519,017 for SCANNER filed Mar. 2, 2015 (Zhou et al.);
  • U.S. patent application Ser. No. 14/405,278 for DESIGN PATTERN FOR SECURE STORE filed Mar. 9, 2015 (Zhu et al.);
  • U.S. patent application Ser. No. 14/660,970 for DECODABLE INDICIA READING TERMINAL WITH COMBINED ILLUMINATION filed Mar. 18, 2015 (Kearney et al.);
  • U.S. patent application Ser. No. 14/661,013 for REPROGRAMMING SYSTEM AND METHOD FOR DEVICES INCLUDING PROGRAMMING SYMBOL filed Mar. 18, 2015 (Soule et al.);
  • U.S. patent application Ser. No. 14/662,922 for MULTIFUNCTION POINT OF SALE SYSTEM filed Mar. 19, 2015 (Van Horn et al.);
  • U.S. patent application Ser. No. 14/663,638 for VEHICLE MOUNT COMPUTER WITH CONFIGURABLE IGNITION SWITCH BEHAVIOR filed Mar. 20, 2015 (Davis et al.);
  • U.S. patent application Ser. No. 14/664,063 for METHOD AND APPLICATION FOR SCANNING A BARCODE WITH A SMART DEVICE WHILE CONTINUOUSLY RUNNING AND DISPLAYING AN APPLICATION ON THE SMART DEVICE DISPLAY filed Mar. 20, 2015 (Todeschini);
  • U.S. patent application Ser. No. 14/669,280 for TRANSFORMING COMPONENTS OF A WEB PAGE TO VOICE PROMPTS filed Mar. 26, 2015 (Funyak et al.);
  • U.S. patent application Ser. No. 14/674,329 for AIMER FOR BARCODE SCANNING filed Mar. 31, 2015 (Bidwell);
  • U.S. patent application Ser. No. 14/676,109 for INDICIA READER filed Apr. 1, 2015 (Huck);
  • U.S. patent application Ser. No. 14/676,327 for DEVICE MANAGEMENT PROXY FOR SECURE DEVICES filed Apr. 1, 2015 (Yeakley et al.);
  • U.S. patent application Ser. No. 14/676,898 for NAVIGATION SYSTEM CONFIGURED TO INTEGRATE MOTION SENSING DEVICE INPUTS filed Apr. 2, 2015 (Showering);
  • U.S. patent application Ser. No. 14/679,275 for DIMENSIONING SYSTEM CALIBRATION SYSTEMS AND METHODS filed Apr. 6, 2015 (Laffargue et al.);
  • U.S. patent application Ser. No. 29/523,098 for HANDLE FOR A TABLET COMPUTER filed Apr. 7, 2015 (Bidwell et al.);
  • U.S. patent application Ser. No. 14/682,615 for SYSTEM AND METHOD FOR POWER MANAGEMENT OF MOBILE DEVICES filed Apr. 9, 2015 (Murawski et al.);
  • U.S. patent application Ser. No. 14/686,822 for MULTIPLE PLATFORM SUPPORT SYSTEM AND METHOD filed Apr. 15, 2015 (Qu et al.);
  • U.S. patent application Ser. No. 14/687,289 for SYSTEM FOR COMMUNICATION VIA A PERIPHERAL HUB filed Apr. 15, 2015 (Kohtz et al.);
  • U.S. patent application Ser. No. 29/524,186 for SCANNER filed Apr. 17, 2015 (Zhou et al.);
  • U.S. patent application Ser. No. 14/695,364 for MEDICATION MANAGEMENT SYSTEM filed Apr. 24, 2015 (Sewell et al.);
  • U.S. patent application Ser. No. 14/695,923 for SECURE UNATTENDED NETWORK AUTHENTICATION filed Apr. 24, 2015 (Kubler et al.);
  • U.S. patent application Ser. No. 29/525,068 for TABLET COMPUTER WITH REMOVABLE SCANNING DEVICE filed Apr. 27, 2015 (Schulte et al.);
  • U.S. patent application Ser. No. 14/699,436 for SYMBOL READING SYSTEM HAVING PREDICTIVE DIAGNOSTICS filed Apr. 29, 2015 (Nahill et al.);
  • U.S. patent application Ser. No. 14/702,110 for SYSTEM AND METHOD FOR REGULATING BARCODE DATA INJECTION INTO A RUNNING APPLICATION ON A SMART DEVICE filed May 1, 2015 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/702,979 for TRACKING BATTERY CONDITIONS filed May 4, 2015 (Young et al.);
  • U.S. patent application Ser. No. 14/704,050 for INTERMEDIATE LINEAR POSITIONING filed May 5, 2015 (Charpentier et al.);
  • U.S. patent application Ser. No. 14/705,012 for HANDS-FREE HUMAN MACHINE INTERFACE RESPONSIVE TO A DRIVER OF A VEHICLE filed May 6, 2015 (Fitch et al.);
  • U.S. patent application Ser. No. 14/705,407 for METHOD AND SYSTEM TO PROTECT SOFTWARE-BASED NETWORK-CONNECTED DEVICES FROM ADVANCED PERSISTENT THREAT filed May 6, 2015 (Hussey et al.);
  • U.S. patent application Ser. No. 14/707,037 for SYSTEM AND METHOD FOR DISPLAY OF INFORMATION USING A VEHICLE-MOUNT COMPUTER filed May 8, 2015 (Chamberlin);
  • U.S. patent application Ser. No. 14/707,123 for APPLICATION INDEPENDENT DEX/UCS INTERFACE filed May 8, 2015 (Pape);
  • U.S. patent application Ser. No. 14/707,492 for METHOD AND APPARATUS FOR READING OPTICAL INDICIA USING A PLURALITY OF DATA SOURCES filed May 8, 2015 (Smith et al.);
  • U.S. patent application Ser. No. 14/710,666 for PRE-PAID USAGE SYSTEM FOR ENCODED INFORMATION READING TERMINALS filed May 13, 2015 (Smith);
  • U.S. patent application Ser. No. 29/526,918 for CHARGING BASE filed May 14, 2015 (Fitch et al.);
  • U.S. patent application Ser. No. 14/715,672 for AUGUMENTED REALITY ENABLED HAZARD DISPLAY filed May 19, 2015 (Venkatesha et al.);
  • U.S. patent application Ser. No. 14/715,916 for EVALUATING IMAGE VALUES filed May 19, 2015 (Ackley);
  • U.S. patent application Ser. No. 14/722,608 for INTERACTIVE USER INTERFACE FOR CAPTURING A DOCUMENT IN AN IMAGE SIGNAL filed May 27, 2015 (Showering et al.);
  • U.S. patent application Ser. No. 29/528,165 for IN-COUNTER BARCODE SCANNER filed May 27, 2015 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/724,134 for ELECTRONIC DEVICE WITH WIRELESS PATH SELECTION CAPABILITY filed May 28, 2015 (Wang et al.);
  • U.S. patent application Ser. No. 14/724,849 for METHOD OF PROGRAMMING THE DEFAULT CABLE INTERFACE SOFTWARE IN AN INDICIA READING DEVICE filed May 29, 2015 (Barten);
  • U.S. patent application Ser. No. 14/724,908 for IMAGING APPARATUS HAVING IMAGING ASSEMBLY filed May 29, 2015 (Barber et al.);
  • U.S. patent application Ser. No. 14/725,352 for APPARATUS AND METHODS FOR MONITORING ONE OR MORE PORTABLE DATA TERMINALS (Caballero et al.);
  • U.S. patent application Ser. No. 29/528,590 for ELECTRONIC DEVICE filed May 29, 2015 (Fitch et al.);
  • U.S. patent application Ser. No. 29/528,890 for MOBILE COMPUTER HOUSING filed Jun. 2, 2015 (Fitch et al.);
  • U.S. patent application Ser. No. 14/728,397 for DEVICE MANAGEMENT USING VIRTUAL INTERFACES CROSS-REFERENCE TO RELATED APPLICATIONS filed Jun. 2, 2015 (Caballero);
  • U.S. patent application Ser. No. 14/732,870 for DATA COLLECTION MODULE AND SYSTEM filed Jun. 8, 2015 (Powilleit);
  • U.S. patent application Ser. No. 29/529,441 for INDICIA READING DEVICE filed Jun. 8, 2015 (Zhou et al.);
  • U.S. patent application Ser. No. 14/735,717 for INDICIA-READING SYSTEMS HAVING AN INTERFACE WITH A USER'S NERVOUS SYSTEM filed Jun. 10, 2015 (Todeschini);
  • U.S. patent application Ser. No. 14/738,038 for METHOD OF AND SYSTEM FOR DETECTING OBJECT WEIGHING INTERFERENCES filed Jun. 12, 2015 (Amundsen et al.);
  • U.S. patent application Ser. No. 14/740,320 for TACTILE SWITCH FOR A MOBILE ELECTRONIC DEVICE filed Jun. 16, 2015 (Bandringa);
  • U.S. patent application Ser. No. 14/740,373 for CALIBRATING A VOLUME DIMENSIONER filed Jun. 16, 2015 (Ackley et al.);
  • U.S. patent application Ser. No. 14/742,818 for INDICIA READING SYSTEM EMPLOYING DIGITAL GAIN CONTROL filed Jun. 18, 2015 (Xian et al.);
  • U.S. patent application Ser. No. 14/743,257 for WIRELESS MESH POINT PORTABLE DATA TERMINAL filed Jun. 18, 2015 (Wang et al.);
  • U.S. patent application Ser. No. 29/530,600 for CYCLONE filed Jun. 18, 2015 (Vargo et al);
  • U.S. patent application Ser. No. 14/744,633 for IMAGING APPARATUS COMPRISING IMAGE SENSOR ARRAY HAVING SHARED GLOBAL SHUTTER CIRCUITRY filed Jun. 19, 2015 (Wang);
  • U.S. patent application Ser. No. 14/744,836 for CLOUD-BASED SYSTEM FOR READING OF DECODABLE INDICIA filed Jun. 19, 2015 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/745,006 for SELECTIVE OUTPUT OF DECODED MESSAGE DATA filed Jun. 19, 2015 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/747,197 for OPTICAL PATTERN PROJECTOR filed Jun. 23, 2015 (Thuries et al.);
  • U.S. patent application Ser. No. 14/747,490 for DUAL-PROJECTOR THREE-DIMENSIONAL SCANNER filed Jun. 23, 2015 (Jovanovski et al.); and
  • U.S. patent application Ser. No. 14/748,446 for CORDLESS INDICIA READER WITH A MULTIFUNCTION COIL FOR WIRELESS CHARGING AND EAS DEACTIVATION, filed Jun. 24, 2015 (Xie et al.).


In the specification and/or figures, typical embodiments of the invention have been disclosed. The present invention is not limited to such exemplary embodiments. The use of the term “and/or” includes any and all combinations of one or more of the associated listed items. The figures are schematic representations and so are not necessarily drawn to scale. Unless otherwise noted, specific terms have been used in a generic and descriptive sense and not for purposes of limitation.

Claims
  • 1. A method of communicating with an electronic device, the method comprising the steps of: in response to being in a display screen reading mode and being within imaging range of a digital display communicatively coupled to the electronic device, capturing, by a barcode scanner, a sequence of multi-state frames displayed on the digital display, wherein the sequence of multi-state frames comprises a bit-wise information stream;detecting, by the barcode scanner, a magnitude of an output of the sequence of multi-state frames for a complete transmission of the sequence of multi-state frames from the digital display;in response to detecting the magnitude of the output of the sequence of multi-state frames, generating a notification at the barcode scanner upon reception of the complete transmission of the sequence of multi-state frames from the digital display, wherein the notification comprises at least one of light, a sound, and a vibration; andentering, by the barcode scanner, a bar code reading mode.
  • 2. The method according to claim 1, wherein, the sequence of multi-state frames comprises three optical states including an “on” state, an “off” state, and an intermediate state, wherein, the three optical states are represented by respective colors;at the “on” state or the “off” state, the sequence of multi-state frames transitions to and from the intermediate state; andat the intermediate state, the sequence of multi-state frames transitions to and from either the “on” state or the “off” state.
  • 3. The method of claim 2, wherein the intermediate state is displayed as a light steel blue, wherein the light steel blue is represented with an html color code of #B0C4DE.
  • 4. The method according to claim 2, wherein the “on” state, the “off” state or the intermediate state have a duration that is at least as long as a frame duration associated with a barcode scanner frame rate.
  • 5. The method according to claim 2, wherein, a transition from the intermediate state to the “on” state or the “off” state and back to a transition state comprises one bit of information; anda transmission period for one bit is equal to two times a frame duration plus two times a sum of a rise time plus a fall time.
  • 6. The method according to claim 1, wherein the bit-wise information stream is bounded by a series of frames without bit-wise information.
  • 7. The method according to claim 1, wherein the sequence of multi-state frames is displayed on a portion of the digital display.
  • 8. The method of claim 1, wherein the bit-wise information stream comprises a media access control address of the electronic device connected to the digital display.
  • 9. The method of claim 8, further comprising: establishing, by the barcode scanner, a communication link to the electronic device associated with the media access control address.
  • 10. The method of claim 1, wherein detecting the magnitude of a frame comprises detecting the magnitude of the output by averaging a group of pixels of the digital display for each frame of the sequence of multi-state frames.
  • 11. A method of transmitting machine-readable code information with a digital display communicatively coupled to an electronic device, the method comprising the steps of: converting the machine-readable code information into a sequence of multi-state frames, wherein the sequence of multi-state frames comprises a bit wise information stream;detecting a magnitude of an output of the sequence of multi-state frames for a complete transmission of the sequence of multi-state frames from the digital display; anddisplaying, by the digital display, the sequence of multi-state frames.
  • 12. The method according to claim 11, wherein, the sequence of multi-state frames comprises a group of optical states including an “on” state, an “off” state, and an intermediate state, wherein, each optical state represents a different color;on an active window of the digital display, the “on” state is displayed as a white color, the “off” state is displayed as a black color, and the intermediate state is displayed as a gray color;at the “on” state or the “off” state, the sequence of multi-state frames transitions to and from the intermediate state; andat the intermediate state, the sequence of multi-state frames transitions to and from either the “on” state or the “off” state.
  • 13. The method according to claim 12, wherein, the “on” state, the “off” state or the intermediate state have a duration that is at least as long as a frame duration associated with a barcode scanner frame rate.
  • 14. The method according to claim 12, wherein, a transition from the intermediate state to the “on” state or the “off” state and back to a transition state comprises one bit of information; anda transmission period for one bit is equal to two times a frame duration plus two times a sum of a rise time plus a fall time.
  • 15. The method according to claim 12, wherein the intermediate state is displayed as a light steel blue, wherein the light steel blue is represented with an html color code of #B0C4DE.
  • 16. The method according to claim 11, wherein detecting the magnitude of the output of the sequence of multi-state frames comprises detecting the magnitude of the output by averaging pixels of the digital display.
  • 17. The method according to claim 11, wherein the electronic device is a computer.
  • 18. The method according to claim 17, wherein the machine-readable code information comprises a media access control address of the computer.
  • 19. The method according to claim 11, wherein the bit-wise information stream is bounded by a series of frames without bit-wise information.
  • 20. The method according to claim 11, wherein the sequence of multi-state frames comprises a unique identifier of the digital display.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. application Ser. No. 15/491,383, filed Apr. 19, 2017, the entire contents of which are incorporated herein by reference.

US Referenced Citations (463)
Number Name Date Kind
5535147 Jacobs et al. Jul 1996 A
5652602 Fishman et al. Jul 1997 A
6281820 Fields Aug 2001 B1
6832725 Gardiner et al. Dec 2004 B2
7128266 Zhu et al. Oct 2006 B2
7159783 Walczyk et al. Jan 2007 B2
7413127 Ehrhart et al. Aug 2008 B2
7726575 Wang et al. Jun 2010 B2
8249209 Zimman et al. Aug 2012 B2
8294969 Plesko Oct 2012 B2
8317105 Kotlarsky et al. Nov 2012 B2
8322622 Liu Dec 2012 B2
8366005 Kotlarsky et al. Feb 2013 B2
8371507 Haggerty et al. Feb 2013 B2
8376233 Horn et al. Feb 2013 B2
8381979 Franz Feb 2013 B2
8390909 Plesko Mar 2013 B2
8408464 Zhu et al. Apr 2013 B2
8408468 Van et al. Apr 2013 B2
8408469 Good Apr 2013 B2
8424768 Rueblinger et al. Apr 2013 B2
8448863 Xian et al. May 2013 B2
8457013 Essinger et al. Jun 2013 B2
8459557 Havens et al. Jun 2013 B2
8469272 Kearney Jun 2013 B2
8474712 Kearney et al. Jul 2013 B2
8479992 Kotlarsky et al. Jul 2013 B2
8490877 Kearney Jul 2013 B2
8517271 Kotlarsky et al. Aug 2013 B2
8523076 Good Sep 2013 B2
8528818 Ehrhart et al. Sep 2013 B2
8544737 Gomez et al. Oct 2013 B2
8548420 Grunow et al. Oct 2013 B2
8550335 Samek et al. Oct 2013 B2
8550354 Gannon et al. Oct 2013 B2
8550357 Kearney Oct 2013 B2
8556174 Kosecki et al. Oct 2013 B2
8556176 Van et al. Oct 2013 B2
8556177 Hussey et al. Oct 2013 B2
8559767 Barber et al. Oct 2013 B2
8561895 Gomez et al. Oct 2013 B2
8561903 Sauerwein, Jr. Oct 2013 B2
8561905 Edmonds et al. Oct 2013 B2
8565107 Pease et al. Oct 2013 B2
8571307 Li et al. Oct 2013 B2
8579200 Samek et al. Nov 2013 B2
8583924 Caballero et al. Nov 2013 B2
8584945 Wang et al. Nov 2013 B2
8587595 Wang Nov 2013 B2
8587697 Hussey et al. Nov 2013 B2
8588869 Sauerwein et al. Nov 2013 B2
8590789 Nahill et al. Nov 2013 B2
8596539 Havens et al. Dec 2013 B2
8596542 Havens et al. Dec 2013 B2
8596543 Havens et al. Dec 2013 B2
8599271 Havens et al. Dec 2013 B2
8599957 Peake et al. Dec 2013 B2
8600158 Li et al. Dec 2013 B2
8600167 Showering Dec 2013 B2
8602309 Longacre et al. Dec 2013 B2
8608053 Meier et al. Dec 2013 B2
8608071 Liu et al. Dec 2013 B2
8611309 Wang et al. Dec 2013 B2
8615487 Gomez et al. Dec 2013 B2
8621123 Caballero Dec 2013 B2
8622303 Meier et al. Jan 2014 B2
8628013 Ding Jan 2014 B2
8628015 Wang et al. Jan 2014 B2
8628016 Winegar Jan 2014 B2
8629926 Wang Jan 2014 B2
8630491 Longacre et al. Jan 2014 B2
8635309 Berthiaume et al. Jan 2014 B2
8636200 Kearney Jan 2014 B2
8636212 Nahill et al. Jan 2014 B2
8636215 Ding et al. Jan 2014 B2
8636224 Wang Jan 2014 B2
8638806 Wang et al. Jan 2014 B2
8640958 Lu et al. Feb 2014 B2
8640960 Wang et al. Feb 2014 B2
8643717 Li et al. Feb 2014 B2
8646692 Meier et al. Feb 2014 B2
8646694 Wang et al. Feb 2014 B2
8657200 Ren et al. Feb 2014 B2
8659397 Vargo et al. Feb 2014 B2
8668149 Good Mar 2014 B2
8678285 Kearney Mar 2014 B2
8678286 Smith et al. Mar 2014 B2
8682077 Longacre, Jr. Mar 2014 B1
D702237 Oberpriller et al. Apr 2014 S
8687282 Feng et al. Apr 2014 B2
8692927 Pease et al. Apr 2014 B2
8695880 Bremer et al. Apr 2014 B2
8698949 Grunow et al. Apr 2014 B2
8702000 Barber et al. Apr 2014 B2
8717494 Gannon May 2014 B2
8720783 Biss et al. May 2014 B2
8723804 Fletcher et al. May 2014 B2
8723904 Marty et al. May 2014 B2
8727223 Wang May 2014 B2
8740082 Wilz, Sr. Jun 2014 B2
8740085 Furlong et al. Jun 2014 B2
8746563 Hennick et al. Jun 2014 B2
8750445 Peake et al. Jun 2014 B2
8752766 Xian et al. Jun 2014 B2
8756059 Braho et al. Jun 2014 B2
8757495 Qu et al. Jun 2014 B2
8760563 Koziol et al. Jun 2014 B2
8763909 Reed et al. Jul 2014 B2
8777108 Coyle Jul 2014 B2
8777109 Oberpriller et al. Jul 2014 B2
8779898 Havens et al. Jul 2014 B2
8781520 Payne et al. Jul 2014 B2
8783573 Havens et al. Jul 2014 B2
8789757 Barten Jul 2014 B2
8789758 Hawley et al. Jul 2014 B2
8789759 Xian et al. Jul 2014 B2
8794520 Wang et al. Aug 2014 B2
8794522 Ehrhart Aug 2014 B2
8794525 Amundsen et al. Aug 2014 B2
8794526 Wang et al. Aug 2014 B2
8798367 Ellis Aug 2014 B2
8807431 Wang et al. Aug 2014 B2
8807432 Van et al. Aug 2014 B2
8820630 Qu et al. Sep 2014 B2
8822848 Meagher Sep 2014 B2
8824692 Sheerin et al. Sep 2014 B2
8824696 Braho Sep 2014 B2
8842849 Wahl et al. Sep 2014 B2
8844822 Kotlarsky et al. Sep 2014 B2
8844823 Fritz et al. Sep 2014 B2
8849019 Li et al. Sep 2014 B2
D716285 Chaney et al. Oct 2014 S
8851383 Yeakley et al. Oct 2014 B2
8854633 Laffargue et al. Oct 2014 B2
8866963 Grunow et al. Oct 2014 B2
8868421 Braho et al. Oct 2014 B2
8868519 Maloy et al. Oct 2014 B2
8868802 Barten Oct 2014 B2
8868803 Caballero Oct 2014 B2
8870074 Gannon Oct 2014 B1
8879639 Sauerwein, Jr. Nov 2014 B2
8880426 Smith Nov 2014 B2
8881983 Havens et al. Nov 2014 B2
8881987 Wang Nov 2014 B2
8903172 Smith Dec 2014 B2
8908995 Benos et al. Dec 2014 B2
8910870 Li et al. Dec 2014 B2
8910875 Ren et al. Dec 2014 B2
8914290 Hendrickson et al. Dec 2014 B2
8914788 Pettinelli et al. Dec 2014 B2
8915439 Feng et al. Dec 2014 B2
8915444 Havens et al. Dec 2014 B2
8916789 Woodburn Dec 2014 B2
8918250 Hollifield Dec 2014 B2
8918564 Caballero Dec 2014 B2
8925818 Kosecki et al. Jan 2015 B2
8939374 Jovanovski et al. Jan 2015 B2
8942480 Ellis Jan 2015 B2
8944313 Williams et al. Feb 2015 B2
8944327 Meier et al. Feb 2015 B2
8944332 Harding et al. Feb 2015 B2
8950678 Germaine et al. Feb 2015 B2
D723560 Zhou et al. Mar 2015 S
8967468 Gomez et al. Mar 2015 B2
8971346 Sevier Mar 2015 B2
8976030 Cunningham et al. Mar 2015 B2
8976368 El et al. Mar 2015 B2
8978981 Guan Mar 2015 B2
8978983 Bremer et al. Mar 2015 B2
8978984 Hennick et al. Mar 2015 B2
8985456 Zhu et al. Mar 2015 B2
8985457 Soule et al. Mar 2015 B2
8985459 Kearney et al. Mar 2015 B2
8985461 Gelay et al. Mar 2015 B2
8988578 Showering Mar 2015 B2
8988590 Gillet et al. Mar 2015 B2
8991704 Hopper et al. Mar 2015 B2
8996194 Davis et al. Mar 2015 B2
8996384 Funyak et al. Mar 2015 B2
8998091 Edmonds et al. Apr 2015 B2
9002641 Showering Apr 2015 B2
9007368 Laffargue et al. Apr 2015 B2
9010641 Qu et al. Apr 2015 B2
9015513 Murawski et al. Apr 2015 B2
9016576 Brady et al. Apr 2015 B2
D730357 Fitch et al. May 2015 S
9022288 Nahill et al. May 2015 B2
9030964 Essinger et al. May 2015 B2
9033240 Smith et al. May 2015 B2
9033242 Gillet et al. May 2015 B2
9036054 Koziol et al. May 2015 B2
9037344 Chamberlin May 2015 B2
9038911 Xian et al. May 2015 B2
9038915 Smith May 2015 B2
D730901 Oberpriller et al. Jun 2015 S
D730902 Fitch et al. Jun 2015 S
D733112 Chaney et al. Jun 2015 S
9047098 Barten Jun 2015 B2
9047359 Caballero et al. Jun 2015 B2
9047420 Caballero Jun 2015 B2
9047525 Barber et al. Jun 2015 B2
9047531 Showering et al. Jun 2015 B2
9049640 Wang et al. Jun 2015 B2
9053055 Caballero Jun 2015 B2
9053378 Hou et al. Jun 2015 B1
9053380 Xian et al. Jun 2015 B2
9057641 Amundsen et al. Jun 2015 B2
9058526 Powilleit Jun 2015 B2
9064165 Havens et al. Jun 2015 B2
9064167 Xian et al. Jun 2015 B2
9064168 Todeschini et al. Jun 2015 B2
9064254 Todeschini et al. Jun 2015 B2
9066032 Wang Jun 2015 B2
9070032 Corcoran Jun 2015 B2
D734339 Zhou et al. Jul 2015 S
D734751 Oberpriller et al. Jul 2015 S
9082023 Feng et al. Jul 2015 B2
9224022 Ackley et al. Dec 2015 B2
9224027 Van et al. Dec 2015 B2
D747321 London et al. Jan 2016 S
9230140 Ackley Jan 2016 B1
9250712 Todeschini Feb 2016 B1
9258033 Showering Feb 2016 B2
9262633 Todeschini et al. Feb 2016 B1
9310609 Rueblinger et al. Apr 2016 B2
D757009 Oberpriller et al. May 2016 S
9342724 McCloskey et al. May 2016 B2
9375945 Bowles Jun 2016 B1
D760719 Zhou et al. Jul 2016 S
9390596 Todeschini Jul 2016 B1
D762604 Fitch et al. Aug 2016 S
D762647 Fitch et al. Aug 2016 S
9412242 Van et al. Aug 2016 B2
D766244 Zhou et al. Sep 2016 S
9443123 Hejl Sep 2016 B2
9443222 Singel et al. Sep 2016 B2
9478113 Xie et al. Oct 2016 B2
9775101 Beguin et al. Sep 2017 B2
10223626 Ackley Mar 2019 B2
20050254714 Anne Nov 2005 A1
20070009268 Giannopoulos Jan 2007 A1
20070063048 Havens et al. Mar 2007 A1
20080035734 Challa et al. Feb 2008 A1
20090134221 Zhu et al. May 2009 A1
20100177076 Essinger et al. Jul 2010 A1
20100177080 Essinger et al. Jul 2010 A1
20100177707 Essinger et al. Jul 2010 A1
20100177749 Essinger et al. Jul 2010 A1
20110080530 Koo Apr 2011 A1
20110114717 Yu et al. May 2011 A1
20110169999 Grunow et al. Jul 2011 A1
20110202554 Powilleit et al. Aug 2011 A1
20120111946 Golant May 2012 A1
20120168512 Kotlarsky et al. Jul 2012 A1
20120193423 Samek Aug 2012 A1
20120203647 Smith Aug 2012 A1
20120223141 Good et al. Sep 2012 A1
20130043312 Van Horn Feb 2013 A1
20130075168 Amundsen et al. Mar 2013 A1
20130175341 Kearney et al. Jul 2013 A1
20130175343 Good Jul 2013 A1
20130257744 Daghigh et al. Oct 2013 A1
20130257759 Daghigh Oct 2013 A1
20130270346 Xian et al. Oct 2013 A1
20130287258 Kearney Oct 2013 A1
20130292475 Kotlarsky et al. Nov 2013 A1
20130292477 Hennick et al. Nov 2013 A1
20130293539 Hunt et al. Nov 2013 A1
20130293540 Laffargue et al. Nov 2013 A1
20130306728 Thuries et al. Nov 2013 A1
20130306731 Pedrao Nov 2013 A1
20130307964 Bremer et al. Nov 2013 A1
20130308625 Park et al. Nov 2013 A1
20130313324 Koziol et al. Nov 2013 A1
20130313325 Wilz et al. Nov 2013 A1
20130342717 Havens et al. Dec 2013 A1
20140001267 Giordano et al. Jan 2014 A1
20140002828 Laffargue et al. Jan 2014 A1
20140008439 Wang Jan 2014 A1
20140025584 Liu et al. Jan 2014 A1
20140034734 Sauerwein, Jr. Feb 2014 A1
20140036848 Pease et al. Feb 2014 A1
20140039693 Havens et al. Feb 2014 A1
20140042814 Kather et al. Feb 2014 A1
20140049120 Kohtz et al. Feb 2014 A1
20140049635 Laffargue et al. Feb 2014 A1
20140061306 Wu et al. Mar 2014 A1
20140063289 Hussey et al. Mar 2014 A1
20140066136 Sauerwein et al. Mar 2014 A1
20140067692 Ye et al. Mar 2014 A1
20140070005 Nahill et al. Mar 2014 A1
20140071840 Venancio Mar 2014 A1
20140074746 Wang Mar 2014 A1
20140076974 Havens et al. Mar 2014 A1
20140078341 Havens et al. Mar 2014 A1
20140078342 Li et al. Mar 2014 A1
20140078345 Showering Mar 2014 A1
20140098792 Wang et al. Apr 2014 A1
20140100774 Showering Apr 2014 A1
20140100813 Showering Apr 2014 A1
20140103115 Meier et al. Apr 2014 A1
20140104413 McCloskey et al. Apr 2014 A1
20140104414 McCloskey et al. Apr 2014 A1
20140104416 Giordano et al. Apr 2014 A1
20140104451 Todeschini et al. Apr 2014 A1
20140106594 Skvoretz Apr 2014 A1
20140106725 Sauerwein, Jr. Apr 2014 A1
20140108010 Maltseff et al. Apr 2014 A1
20140108402 Gomez et al. Apr 2014 A1
20140108682 Caballero Apr 2014 A1
20140110485 Toa et al. Apr 2014 A1
20140114530 Fitch et al. Apr 2014 A1
20140124577 Wang et al. May 2014 A1
20140124579 Ding May 2014 A1
20140125842 Winegar May 2014 A1
20140125853 Wang May 2014 A1
20140125999 Longacre et al. May 2014 A1
20140129378 Richardson May 2014 A1
20140131438 Kearney May 2014 A1
20140131441 Nahill et al. May 2014 A1
20140131443 Smith May 2014 A1
20140131444 Wang May 2014 A1
20140131445 Ding et al. May 2014 A1
20140131448 Xian et al. May 2014 A1
20140133379 Wang et al. May 2014 A1
20140136208 Maltseff et al. May 2014 A1
20140140585 Wang May 2014 A1
20140151453 Meier et al. Jun 2014 A1
20140152882 Samek et al. Jun 2014 A1
20140158770 Sevier et al. Jun 2014 A1
20140159869 Zumsteg et al. Jun 2014 A1
20140166755 Liu et al. Jun 2014 A1
20140166757 Smith Jun 2014 A1
20140166759 Liu et al. Jun 2014 A1
20140168787 Wang et al. Jun 2014 A1
20140175165 Havens et al. Jun 2014 A1
20140175172 Jovanovski et al. Jun 2014 A1
20140191644 Chaney Jul 2014 A1
20140191913 Ge et al. Jul 2014 A1
20140197238 Liu et al. Jul 2014 A1
20140197239 Havens et al. Jul 2014 A1
20140197304 Feng et al. Jul 2014 A1
20140203087 Smith et al. Jul 2014 A1
20140204268 Grunow et al. Jul 2014 A1
20140214631 Hansen Jul 2014 A1
20140217166 Berthiaume et al. Aug 2014 A1
20140217180 Liu Aug 2014 A1
20140231500 Ehrhart et al. Aug 2014 A1
20140232930 Anderson Aug 2014 A1
20140247315 Marty et al. Sep 2014 A1
20140263493 Amurgis et al. Sep 2014 A1
20140263645 Smith et al. Sep 2014 A1
20140270196 Braho et al. Sep 2014 A1
20140270229 Braho Sep 2014 A1
20140278387 Digregorio Sep 2014 A1
20140282210 Bianconi Sep 2014 A1
20140284384 Lu et al. Sep 2014 A1
20140288933 Braho et al. Sep 2014 A1
20140297058 Barker et al. Oct 2014 A1
20140299665 Barber et al. Oct 2014 A1
20140312121 Lu et al. Oct 2014 A1
20140319220 Coyle Oct 2014 A1
20140319221 Oberpriller et al. Oct 2014 A1
20140326787 Barten Nov 2014 A1
20140332590 Wang et al. Nov 2014 A1
20140344943 Todeschini et al. Nov 2014 A1
20140346233 Liu et al. Nov 2014 A1
20140351317 Smith et al. Nov 2014 A1
20140353373 Van et al. Dec 2014 A1
20140361073 Qu et al. Dec 2014 A1
20140361082 Xian et al. Dec 2014 A1
20140362184 Jovanovski et al. Dec 2014 A1
20140363015 Braho Dec 2014 A1
20140369511 Sheerin et al. Dec 2014 A1
20140374483 Lu Dec 2014 A1
20140374485 Xian et al. Dec 2014 A1
20150001301 Ouyang Jan 2015 A1
20150001304 Todeschini Jan 2015 A1
20150003673 Fletcher Jan 2015 A1
20150009338 Laffargue et al. Jan 2015 A1
20150009610 London et al. Jan 2015 A1
20150014416 Kotlarsky et al. Jan 2015 A1
20150021397 Rueblinger et al. Jan 2015 A1
20150028102 Ren et al. Jan 2015 A1
20150028103 Jiang Jan 2015 A1
20150028104 Ma et al. Jan 2015 A1
20150029002 Yeakley et al. Jan 2015 A1
20150032709 Maloy et al. Jan 2015 A1
20150039309 Braho et al. Feb 2015 A1
20150040378 Saber et al. Feb 2015 A1
20150048168 Fritz et al. Feb 2015 A1
20150049347 Laffargue et al. Feb 2015 A1
20150051992 Smith Feb 2015 A1
20150053766 Havens et al. Feb 2015 A1
20150053768 Wang et al. Feb 2015 A1
20150053769 Thuries et al. Feb 2015 A1
20150062366 Liu et al. Mar 2015 A1
20150063215 Wang Mar 2015 A1
20150063676 Lloyd et al. Mar 2015 A1
20150069130 Gannon Mar 2015 A1
20150071819 Todeschini Mar 2015 A1
20150083800 Li et al. Mar 2015 A1
20150086114 Todeschini Mar 2015 A1
20150088522 Hendrickson et al. Mar 2015 A1
20150096872 Woodburn Apr 2015 A1
20150099557 Pettinelli et al. Apr 2015 A1
20150100196 Hollifield Apr 2015 A1
20150102109 Huck Apr 2015 A1
20150115035 Meier et al. Apr 2015 A1
20150127791 Kosecki et al. May 2015 A1
20150128116 Chen et al. May 2015 A1
20150129659 Feng et al. May 2015 A1
20150133047 Smith et al. May 2015 A1
20150134470 Hejl et al. May 2015 A1
20150136851 Harding et al. May 2015 A1
20150136854 Lu et al. May 2015 A1
20150142492 Kumar May 2015 A1
20150144692 Hejl May 2015 A1
20150144698 Teng et al. May 2015 A1
20150144701 Xian et al. May 2015 A1
20150149946 Benos et al. May 2015 A1
20150161429 Xian Jun 2015 A1
20150169925 Chen et al. Jun 2015 A1
20150169929 Williams et al. Jun 2015 A1
20150186703 Chen et al. Jul 2015 A1
20150193644 Kearney et al. Jul 2015 A1
20150193645 Colavito et al. Jul 2015 A1
20150199957 Funyak et al. Jul 2015 A1
20150204671 Showering Jul 2015 A1
20150210199 Payne Jul 2015 A1
20150220753 Zhu et al. Aug 2015 A1
20150245281 Beguin et al. Aug 2015 A1
20150254485 Feng et al. Sep 2015 A1
20150327012 Bian et al. Nov 2015 A1
20160014251 Hejl Jan 2016 A1
20160021346 Liu et al. Jan 2016 A1
20160040982 Li et al. Feb 2016 A1
20160042241 Todeschini Feb 2016 A1
20160057230 Todeschini et al. Feb 2016 A1
20160109219 Ackley et al. Apr 2016 A1
20160109220 Laffargue et al. Apr 2016 A1
20160109224 Thuries et al. Apr 2016 A1
20160112631 Ackley et al. Apr 2016 A1
20160112643 Laffargue et al. Apr 2016 A1
20160124516 Schoon et al. May 2016 A1
20160125217 Todeschini May 2016 A1
20160125342 Miller et al. May 2016 A1
20160125873 Braho et al. May 2016 A1
20160132708 Tsou May 2016 A1
20160133253 Braho et al. May 2016 A1
20160171720 Todeschini Jun 2016 A1
20160178479 Goldsmith Jun 2016 A1
20160180678 Ackley et al. Jun 2016 A1
20160189087 Morton et al. Jun 2016 A1
20160227912 Oberpriller et al. Aug 2016 A1
20160232891 Pecorari Aug 2016 A1
20160292477 Bidwell Oct 2016 A1
20160294779 Yeakley et al. Oct 2016 A1
20160301924 Tay et al. Oct 2016 A1
20160306769 Kohtz et al. Oct 2016 A1
20160314276 Wilz et al. Oct 2016 A1
20160314294 Kubler et al. Oct 2016 A1
20170177939 Beall et al. Jun 2017 A1
Foreign Referenced Citations (4)
Number Date Country
2013163789 Nov 2013 WO
2013173985 Nov 2013 WO
2014019130 Feb 2014 WO
2014110495 Jul 2014 WO
Non-Patent Literature Citations (28)
Entry
U.S. Appl. No. 29/530,600 for Cyclone filed Jun. 18, 2015 (Vargo et al); 16 pages.
U.S. Appl. No. 29/529,441 for Indicia Reading Device filed Jun. 8, 2015 (Zhou et al.); 14 pages.
U.S. Appl. No. 29/528,890 for Mobile Computer Housing filed Jun. 2, 2015 (Fitch et al.); 61 pages.
U.S. Appl. No. 29/526,918 for Charging Base filed May 14, 2015 (Fitch et al.); 10 pages.
U.S. Appl. No. 29/525,068 for Tablet Computer With Removable Scanning Device filed Apr. 27, 2015 (Schulte et al.); 19 pages.
U.S. Appl. No. 29/523,098 for Handle for a Tablet Computer filed Apr. 7, 2015 (Bidwell et al.); 17 pages.
U.S. Appl. No. 29/516,892 for Table Computer filed Feb. 6, 2015 (Bidwell et al.); 13 pages.
U.S. Appl. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.); 44 pages.
U.S. Appl. No. 14/446,391 for Multifunction Point of Sale Apparatus With Optical Signature Capture filed Jul. 30, 2014 (Good et al.); 37 pages; now abandoned.
U.S. Appl. No. 14/277,337 for Multipurpose Optical Reader, filed May 14, 2014 (Jovanovski et al.); 59 pages; now abandoned.
U.S. Patent Application for Tracking Battery Conditions filed May 4, 2015 (Young et al.); 70 pages, U.S. Appl. No. 14/702,979.
U.S. Patent Application for Terminal Having Illumination and Focus Control filed May 21, 2014 (Liu et al.); 31 pages; now abandoned, U.S. Appl. No. 14/283,282.
U.S. Patent Application for Tactile Switch for a Mobile Electronic Device filed Jun. 16, 2015 (Bamdringa); 38 pages, U.S. Appl. No. 14/740,320.
U.S. Patent Application for System and Method for Regulating Barcode Data Injection Into a Running Application on a Smart Device filed May 1, 2015 (Todeschini et al.); 38 pages, U.S. Appl. No. 14/702,110.
U.S. Patent Application for Optical Pattern Projector filed Jun. 23, 2015 (Thuries et al.); 33 pages, U.S. Appl. No. 14/747,197.
U.S. Patent Application for Method and System to Protect Software-Based Network-Connected Devices From Advanced Persistent Threat filed May 6, 2015 (Hussey et al.); 42 pages, U.S. Appl. No. 14/705,407.
U.S. Patent Application for Intermediate Linear Positioning filed May 5, 2015 (Charpentier et al.); 60 pages, U.S. Appl. No. 14/704,050.
U.S. Patent Application for Indicia-Reading Systems Having an Interface With a User's Nervous System filed Jun. 10, 2015 (Todeschini); 39 pages, U.S. Appl. No. 14/735,717.
U.S. Patent Application for Hands-Free Human Machine Interface Responsive to a Driver of a Vehicle filed May 6, 2015 (Fitch et al.); 44 pages, U.S. Appl. No. 14/705,012.
U.S. Patent Application for Evaluating Image Values filed May 19, 2015 (Ackley); 60 pages, U.S. Appl. No. 14/715,916.
U.S. Patent Application for Dual-Projector Three-Dimensional Scanner filed Jun. 23, 2015 (Jovanovski et al.); 40 pages, U.S. Appl. No. 14/747,490.
U.S. Patent Application for Calibrating a vol. Dimensioner filed Jun. 16, 2015 (Ackley et al.); 63 pages, U.S. Appl. No. 14/740,373.
U.S. Patent Application for Augumented Reality Enabled Hazard Display filed May 19, 2015 (Venkatesha et al.); 35 pages, U.S. Appl. No. 14/715,672.
U.S. Patent Application for Application Independent DEX/UCS Interface filed May 8, 2015 (Pape); 47 pages, U.S. Appl. No. 14/707,123.
U.S. Appl. No. 13/367,978, filed Feb. 7, 2012, (Feng et al.); now abandoned.
Non-Final Rejection dated Jun. 26, 2018 for U.S. Appl. No. 15/491,383.
Notice of Allowance and Fees Due (PTOL-85) dated Oct. 23, 2018 for U.S. Appl. No. 15/491,383.
U.S. Appl. No. 15/491,383, filed Apr. 19, 2017, U.S. Pat. No. 10,223,626, Mar. 5, 2019.
Related Publications (1)
Number Date Country
20190156169 A1 May 2019 US
Continuations (1)
Number Date Country
Parent 15491383 Apr 2017 US
Child 16254057 US