Monitoring worker movement in a warehouse setting

Information

  • Patent Grant
  • 11157869
  • Patent Number
    11,157,869
  • Date Filed
    Friday, August 5, 2016
    7 years ago
  • Date Issued
    Tuesday, October 26, 2021
    2 years ago
Abstract
Systems and methods for monitoring worker movement within a warehouse setting are provided. According to one implementation, a wearable apparatus comprises an audio unit configured to receive audio input signals from a user and convey audio output signals to the user. Also, the wearable apparatus includes a step count unit configured to detect walking movements of the user.
Description
FIELD OF THE INVENTION

The present invention relates to the process of picking, or retrieving, items from a storage area.


BACKGROUND

Generally speaking, warehouses and other similar storage facilities typically include an organized network of racks and shelves on which goods are stored. When a purchasing or shipping order is received, a picker travels throughout the warehouse to retrieve the item or items listed in the order. The picked items can then be packed, if needed, and then delivered to a customer.


An organizational system is normally used in such a storage facility to identify the inventory items and the location of those items within the facility. Often, information associated with the location of inventory items is displayed in the aisles and on the shelves or racks where the items are stored. In order to correctly pick an item, the picker can match information that corresponds to the inventory item with information on the order.


Although the traditional paper-based order picking technique has been somewhat successful over the years, research has found that more efficient ways of picking have been utilized. Also, the occurrence of errors with paper-based picking is higher than more recently developed techniques. As a result of picking errors, incorrect items may be shipped to customers, leading to inventory inaccuracies and other problems. Also, extra time is needed to correct any issues, which is an unproductive use of personnel time. Furthermore, incorrect picking may also lead to customer dissatisfaction. Therefore, companies strive to improve the order picking process to increase efficiency and accuracy of their inventory handling operations.


A significant improvement in order picking includes voice-directed picking. Voice-directed picking is a method that uses verbal instructions for directing the picker to pick the correct items. Voice-directed picking typically involves a headset worn on the head of the picker and a portable computer device clipped to the picker's belt. The headset provides verbal instructions to the picker, which frees the picker's hands and eyes to easily see and pick up the proper items without the need to carry stacks of papers, clipboards, barcode readers, and other cumbersome objects used in other types of picking processes.


In addition, the headset may include a microphone that allows the picker to provide feedback to confirm that the correct item is being picked. For example, the picker may read back some of the information related to the product being picked, such as a number of “check digits.” One issue that may cause picking errors using voice-directed picking, however, is that a picker might memorize the check-digits for some of the more common items and attempt to speak the check digits from memory.


Therefore, a need exists for improving the way in which items are retrieved or picked from a warehouse or storage facility. The present invention provides systems and methods that improve upon the voice-directed item-picking processes already in existence in order to attempt to improve efficiency and further reduce the number of picking errors.


SUMMARY

Accordingly, in one aspect, the present invention embraces a workflow system comprising a plurality of audio-based picker devices. Each audio-based picker device is configured to be worn by a respective user and configured to enable two-way audio communication with the respective user. The workflow system further comprises a host computer configured in wireless communication with the plurality of audio-based picker devices. The host computer is configured to determine a workflow process for each user. Each audio-based picker device includes an inertial unit configured to calculate a step count related to walking movements of the respective user. Furthermore, each audio-based picker device is configured to communicate the workflow process to the respective user based at least on the calculated step count.


In another exemplary embodiment, a wearable apparatus is disclosed. The wearable apparatus includes an audio unit configured to receive audio input signals from a user and convey audio output signals to the user. The wearable apparatus also includes a step count unit configured to detect walking movements of the user.


In yet another exemplary embodiment, a workflow method is provided. The method includes the steps of determining a workflow process for a user, the workflow process including a number of workflow operations. The method also includes transmitting a first workflow operation to a communication unit associated with the user for conveying the first workflow operation to the user. A step count is received from the communication unit associated with the user, the step count being related to walking movements of the user. An audio signal is received from the communication unit associated with the user, the audio signal including verbal responses from the user. The method further includes the step of determining whether or not conditions of the first workflow operation have been met based on the step count and the verbal responses.


In another aspect, the present invention embraces a workflow system that includes at least one audio-based picker device, configured to enable two-way audio communication with the respective user, and a host computer in wireless communication with said at least one audio-based picker device. Typically, the host computer is configured to determine a workflow process for the user, the at least one audio-based picker device includes an inertial unit configured to calculate an estimated distance corresponding to movements of the respective user, and the at least one audio-based picker device is configured to communicate the workflow process to the respective user based at least on the calculated step count.


The foregoing illustrative summary, as well as other exemplary objectives and/or advantages of the invention, and the manner in which the same are accomplished, are further explained within the following detailed description and its accompanying drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a diagram illustrating an audio-based picker device according to various embodiments of the present invention.



FIG. 2 is a block diagram illustrating the circuitry included in the audio-based picker device of FIG. 1 according to various embodiments of the present invention.



FIG. 3 is a diagram illustrating a warehouse management system (WMS) according to various embodiments of the present invention.



FIG. 4 is a diagram illustrating a section of a storage area according to various embodiments of the present invention.



FIG. 5 is a flow diagram illustrating a method of executing an item-picking process according to various embodiments of the present invention.





DETAILED DESCRIPTION

The present invention is directed to systems and methods for assisting a picker in the process of picking inventory items from a storage area. Recently, voice-assisted picking has been developed in the field of inventory management to increase efficiency and to reduce errors in the handling of items stored within a warehouse or other storage facility. In addition to providing verbal instructions to the picker and receiving verbal responses from the picker, the present invention is also configured to monitor movement of the picker.


In order to prevent a picker from simply reciting check digits from memory, the movement monitoring system counts the picker's steps to determine if the picker has actually moved to the correct location to pick the items. When a predetermined distance has been calculated to allow the picker to move to the appropriate location where the next pick is to be made, then the system will allow the picker to read the check digits for that item. Otherwise, if the measurement of the picker's movement does not reach or exceed the predetermined minimum amount, then the system detects an error and instructs the picker to move to the correct pick location.



FIG. 1 is a diagram showing an embodiment of an audio-based picker device 10, which is configured as a portable device to be carried or worn by a picker. According to the embodiment shown in FIG. 1, the audio-based picker device 10 includes a headset 12 and a portable computing device 14. The headset 12 and portable computing device 14 may be connected together by way of a cable 16. In other embodiments, the headset 12 and portable computing device 14 may communicate with each other using a wireless communication protocol, such as Bluetooth™. In the case of wireless communication, the cable 16 can be omitted, thereby eliminating any possibility of an intermediate cable interfering with the picker's arms or hands.


The portable computing device 14 includes a first connector 18, or interface, for coupling with the cable 16. The portable computing device 14 may also include at least a housing 20 and a number of user interface elements. The user interface elements may include, for example, a display unit 22, visual output elements 24 (e.g., LEDs), and input elements 26 (e.g., keypad, buttons, etc.).


The headset 12 includes one or more audio output devices 30 (e.g., headphone speakers), audio input device 32 (e.g., microphone), one or more support straps 34, and support pad 36. The headset 12 may be used in a voice-directed picking process for enabling a user (i.e., the picker) to receive audible picking instructions and to respond with voice commands.


The headset 12 may be worn on the head of the picker in a conventional manner such that the audio output device 30 is next to the picker's ear or ears and the audio input device 32 is near the front of the picker's mouth. The portable computing device 14 may have a clip, clamp, or other suitable fixing unit (not shown) for attaching the portable computing device 14 to the picker's belt, clothing, a pocket of the picker's clothing, a waistband, or other article of clothing or accessory.


The portable computing device 14 may include voice recognition software for converting speech to text and voice synthesizing software for converting text to speech. Also, the portable computing device 14 may include picking assistance software to be used in cooperation with the headset 12, allowing the portable computing device 14 to provide audible picking instructions to the picker and receive audible responses from the picker.


According to the teachings of the present disclosure, the portable computing device 14 may further include step count software for counting the number of steps taken by the picker between picks. For example, if the warehouse management system knows the approximate distance from one location where a first item is located in the warehouse to another location where a next item is located, then an approximation can be made regarding the number of steps that a picker would have to take to travel from the first location to the next one. If the step count is too low for a picker to possibly travel the needed distance, then an error is detected. The system may determine that the picker is reciting check digits from memory without actually moving to the required location to retrieve the items or that the picker is simply in the wrong location.


The portable computing device 14 may be configured to store a list of items to be picked by the picker. For example, using wireless communication in the general vicinity of a warehouse computer, the portable computing device 14 may receive the item list and store it in memory. Using voice software, the portable computing device 14 may provide instructions to the picker regarding information related to the first item to be picked. The instructions may be shown on the display 22, may be audibly expressed through the audio output device 30 using voice commands or tonal commands, or may be indicated in other suitable ways. The instructions may include the name of the item to be picked, a product number, check digits, or other reference characters corresponding to the item, a location of the item, directions of how the picker is to get to the item's location, etc.



FIG. 2 is a block diagram illustrating an embodiment of circuitry of the audio-based picker device 10 of FIG. 1. In some embodiments, the circuitry may be contained completely within the headset 12, in which case the portable computing device 14 may be omitted. In other embodiments, some of the circuitry shown in FIG. 2 may be housed in the headset 12 while some of circuitry may be housed in the portable computing device 14. In still other embodiments, some elements may be housed in both the headset 12 and portable computing device 14.


The audio-based picker device 10, according to the embodiment shown in FIG. 2, includes a processing device 40, a memory device 42, audio input unit 44, audio output unit 46, a wireless communication device 48, and an inertial device 50. The processing device 40 is configured to control the operations of the audio-based picker device 10 and is configured to receive input signals from and provide output signals to the other components of the audio-based picker device 10. It should be noted that the audio-based picker device 10 may include additional components, such as a power source, input devices, output devices, and other components, as would be known by one of ordinary skill in the art.


The processing device 40 may include one or more processors, microprocessors, central processing units, digital signal processors, or other electronic processing components. The processing device 40 may communicate with the other components of the audio-based picker device 10 via one or more bus interfaces.


The memory device 42 may include any combination of volatile and non-volatile memory elements for storing software, data, commands, etc. Picking software may be stored in the memory device 42 for assisting the picker with the picking process, as explained herein. The picking software may be stored on a non-transitory computer-readable medium. According to various embodiments, picking programs may be stored in software, firmware, and/or hardware and executed by the processing device 40.


The inertial device 50 is configured to measure inertia related to the movements of the picker. For example, according to some embodiments, the inertial device 50 may include an acceleration module 52 and a gyroscopic module 54. The acceleration module 52 may include one or more accelerometers and the gyroscopic module 54 may include one or more gyroscopes. According to various embodiments, the acceleration module 52 and gyroscopic module 54 may be incorporated either in the headset 12 for detecting the movements of the picker's head or in the portable computing device 14 for detecting the movements of the picker's body. The accelerometers of the acceleration module 52 may include electromechanical devices for measuring acceleration forces in an x-axis, y-axis, and z-axis with respect to the audio-based picker device 10. The gyroscopes of the gyroscopic module 54 may also include electromechanical devices for measuring angular velocity with respect to the x, y, and z axes.


Among other things, the picking software may include a step count module 56, a voice recognition module 58, and a voice synthesizing module 60. The step count module 56 may be used to analyze the data obtained by the acceleration module 52 and gyroscopic module 54 to determine the steps taken by the picker between two sequential picking locations.


Based on acceleration data, not only can the step count module 56 determine the number of steps taken, based on the rhythm of the walking motion, but the step count module 56 can also determine the walking velocity. For example, based on a certain acceleration in one direction for a certain amount of time, the processing device 40 may utilize the step count module 56 to calculate a velocity. Also, when a deceleration, or acceleration in the opposite direction, is detected, the velocity can be recalculated to determine when the picker comes to a stop. Also, a rest condition can be determined by measuring no additional net acceleration forces over a certain length of time.


The voice recognition module 58, which may be stored in the memory device 42, is configured to receive speech from the picker and convert that speech into text. The text can then be processed by the audio-based picker device 10 and/or transmitted to a warehouse computer that manages the audio-based picking process. Conversely, the voice synthesizing module 60 is configured to receive text signals from the memory device 42 or from the warehouse computer and convert the text into speech, which is conveyed to the audio output unit 46 for providing verbal instructions to the picker. The voice recognition module 58 and voice synthesizing module 60 may be calibrated for a particular user depending on the preferred language, speech patterns, dialect, and other characteristics of the user's speech and preferences.



FIG. 3 is a diagram showing an implementation of a warehouse management system 64. The warehouse management system 64 includes a host computer 66, or warehouse computer, and a plurality of audio-based picker devices 10. The audio-based picker devices 10, which are described above with respect to FIGS. 1 and 2, are carried or worn by a plurality of pickers as they move throughout the warehouse. The host computer 66 provides customized instructions to each of the audio-based picker devices 10 depending on the item-picking routines that each particular picker is needed is follow.


When a picker is within range of the host computer 66, the host computer 66 transmits an item list to the picker's audio-based picker device 10. For example, the wireless communication device 48 of the respective audio-based picker device 10 (see FIG. 2) may be used for communicating with the host computer 66. The item list may include one or more items to be picked. The item list may also include information regarding any product numbers corresponding to the items, the quantity of each item to be picked, the location of the items, etc. When the list is received, the picker can then travel throughout the warehouse to retrieve the items. In some embodiments, audio instructions are provided to the picker to direct him or her to the location of the items.


According to various alternative embodiments, the audio-based picker device 10 may also receive and store distance information from the host computer 66. The distance information may include the approximate distances between each of the subsequent pick locations. These stored distances can then be used to compare with step count information obtained by the audio-based picker device 10 to determine if the picker has traveled the required minimum distance to get to the correct picking location. Therefore, instead of communicating step information back to the host computer 66, as mentioned above with respect to other embodiments, the step count information can be analyzed by the audio-based picker device 10 itself.



FIG. 4 is a diagram showing a small section 70 of a warehouse or other storage facility according to an exemplary embodiment. FIG. 4 shows a rack structure 72 having vertical supports 74 and horizontal shelf units 76. Bins 78 may be placed on the rack structure 72 for storing a number of small inventory items. Other inventory items, such as those having an easily stackable configuration, may be placed directly on the horizontal shelf units 76.


In addition, the rack structure 72 may further include product sheets (not shown) that describe the items in detail and may also include tags 80 or other signage elements for displaying a number of check digits. The check digits may be used to designate a certain area or section 70 of the warehouse or may be used to designate certain items or families of items.


As shown as an example in FIG. 4, the check digits on the tags 80 include different numbers, which may be dependent on any arbitrary warehouse organizational system. In the illustrated example, numbers 255104, 255105, 255106, 255107, 255108, and 255113 are shown on tags 80 and may represent a particular warehouse numbering system to identify different inventory items stored in the various bins 78. It should be noted that the check digits may be different for different warehouses based on any particular organizational technique. The check digits may not necessarily be sequential or in numerical order, as shown, but may depend on the particular organizational system being used in the particular warehouse.


The tags 80 may be placed on any combination of the vertical supports 74, horizontal shelf units 76, and/or bins 78. The tags 80 may be adhered using any suitable adhesion material, such as glue, tape, Velcro, etc., for attaching to the rack structure 72.



FIG. 5 shows an embodiment of a method for controlling a picking process for a single picker during a shift. The method may be performed by the host computer 66 shown in FIG. 3, by the processing device 40 of the audio-based picker device 10 itself, or by another controlling device in the warehouse.


The picking method includes the steps of determining a first item to be picked, as indicated in block 90. The item to be picked can be selected from a group of items to be picked by a single picker during a shift. According to block 92, the method includes determining an approximate distance that the picker must travel to get to the location of the first item. The calculated distance may be obtained from information regarding the location of the inventory items in the warehouse and the starting location of the picker.


The method also includes the step of instructing the picker regarding the location of the first item. The instructions may include warehouse section information, aisle numbers, etc. Also, the instructions may include suggested directions for getting the picker to the proper location in the shortest time possible. When the picker arrives at the picking location, the picker may send a signal indicating that the location has been reached. As indicated in block 96, the host computer 66 receives a notification from the picker device that the picker intends to provide check digits to verify that the picker is indeed at the proper location for picking the first item. After providing instructing according to block 94, there will be a certain amount of time that will elapse before the picker can move to the location to provide the notification according to block 96. This time may also be taken into account.


When the notification according to block 96 is sent from the picker device, the picker device also calculates the step count from a starting point to a current location. The system (e.g., host computer 66) receives the step count information from the picker device, as indicated in block 98. In some embodiments, at least a part of the calculations of step count can be performed by the host computer 66 itself and may include receiving certain information, such accelerometer and gyroscope measurements, from the picker device.


According to decision block 100, the method of FIG. 5 further includes determining if the step count is greater than a minimum threshold that would be needed to cover the approximate distance calculated in step 92. The threshold value may be based on warehouse distance information, time information, approximate stride length of the picker, and other factors that may limit the necessary time and steps that can be achieved by a picker.


If the step count is not sufficient to cover the approximate distance, it can be determined that the picker has not moved to the proper location to pick the first item. In this case, the method proceeds to block 102, which indicates that the system instructs the picker to move to the proper location in order to make the pick. The system may also signal that an error has occurred and transmit the error signal to the picker or to another device, if needed. After this, the method returns back to block 94 to repeat several steps as needed. Otherwise, if it is determined in decision block 100 that the step count is sufficient to cover the distance, the method proceeds to block 104.


The method performs the step indicated in block 104 after it is determined that the picker has moved at least to a location where the first item is located. Therefore, it can be determined that the picker will not simply be reciting check digits from memory without moving from an original starting point. Block 104 indicates that the method further includes receiving check digits from the picker. According to some embodiments, the picker may read the check digits in response to a prompt or after indicating in some other manner that the check digits are to be read.


The method further includes the step of determining, according to decision block 106, whether or not the check digits read by the picker are correct. If not, then the method returns back to block 94 to correct any mistakes that may have been made in the picking process. In some cases, the picker may have misread the check digits and simply needs to read the check digits again.


If it is determined in decision block 106 that the check digits are correct, then it can be determined that the picker is in the correct position for making a pick. At this point, the method proceeds to block 108, which includes instructing the picker to pick a certain quantity of items as needed in the order. After making a successful pick, the system determines, according to decision block 110, whether or not there are more items to be picked. If not, then the picking process ends. However, if more items are to be picked, the method loops back to the beginning and repeats the picking process for the next item to be picked. Also, it should be noted that the step count in the host computer 66 and in the picker device 10 may be reset at this point in order to allow detection of a new step count from the location of the previous pick to the location of the next pick. Also, the distance determined in block 92 may be calculated based on the location of the next pick with respect to the current location regarding the previous pick.


To supplement the present disclosure, this application incorporates entirely by reference the following commonly assigned patents, patent application publications, and patent applications:

  • U.S. Pat. Nos. 6,832,725; 7,128,266;
  • U.S. Pat. Nos. 7,159,783; 7,413,127;
  • U.S. Pat. Nos. 7,726,575; 8,294,969;
  • U.S. Pat. Nos. 8,317,105; 8,322,622;
  • U.S. Pat. Nos. 8,366,005; 8,371,507;
  • U.S. Pat. Nos. 8,376,233; 8,381,979;
  • U.S. Pat. Nos. 8,390,909; 8,408,464;
  • U.S. Pat. Nos. 8,408,468; 8,408,469;
  • U.S. Pat. Nos. 8,424,768; 8,448,863;
  • U.S. Pat. Nos. 8,457,013; 8,459,557;
  • U.S. Pat. Nos. 8,469,272; 8,474,712;
  • U.S. Pat. Nos. 8,479,992; 8,490,877;
  • U.S. Pat. Nos. 8,517,271; 8,523,076;
  • U.S. Pat. Nos. 8,528,818; 8,544,737;
  • U.S. Pat. Nos. 8,548,242; 8,548,420;
  • U.S. Pat. Nos. 8,550,335; 8,550,354;
  • U.S. Pat. Nos. 8,550,357; 8,556,174;
  • U.S. Pat. Nos. 8,556,176; 8,556,177;
  • U.S. Pat. Nos. 8,559,767; 8,599,957;
  • U.S. Pat. Nos. 8,561,895; 8,561,903;
  • U.S. Pat. Nos. 8,561,905; 8,565,107;
  • U.S. Pat. Nos. 8,571,307; 8,579,200;
  • U.S. Pat. Nos. 8,583,924; 8,584,945;
  • U.S. Pat. Nos. 8,587,595; 8,587,697;
  • U.S. Pat. Nos. 8,588,869; 8,590,789;
  • U.S. Pat. Nos. 8,596,539; 8,596,542;
  • U.S. Pat. Nos. 8,596,543; 8,599,271;
  • U.S. Pat. Nos. 8,599,957; 8,600,158;
  • U.S. Pat. Nos. 8,600,167; 8,602,309;
  • U.S. Pat. Nos. 8,608,053; 8,608,071;
  • U.S. Pat. Nos. 8,611,309; 8,615,487;
  • U.S. Pat. Nos. 8,616,454; 8,621,123;
  • U.S. Pat. Nos. 8,622,303; 8,628,013;
  • U.S. Pat. Nos. 8,628,015; 8,628,016;
  • U.S. Pat. Nos. 8,629,926; 8,630,491;
  • U.S. Pat. Nos. 8,635,309; 8,636,200;
  • U.S. Pat. Nos. 8,636,212; 8,636,215;
  • U.S. Pat. Nos. 8,636,224; 8,638,806;
  • U.S. Pat. Nos. 8,640,958; 8,640,960;
  • U.S. Pat. Nos. 8,643,717; 8,646,692;
  • U.S. Pat. Nos. 8,646,694; 8,657,200;
  • U.S. Pat. Nos. 8,659,397; 8,668,149;
  • U.S. Pat. Nos. 8,678,285; 8,678,286;
  • U.S. Pat. Nos. 8,682,077; 8,687,282;
  • U.S. Pat. Nos. 8,692,927; 8,695,880;
  • U.S. Pat. Nos. 8,698,949; 8,717,494;
  • U.S. Pat. Nos. 8,717,494; 8,720,783;
  • U.S. Pat. Nos. 8,723,804; 8,723,904;
  • U.S. Pat. Nos. 8,727,223; D702,237;
  • U.S. Pat. Nos. 8,740,082; 8,740,085;
  • U.S. Pat. Nos. 8,746,563; 8,750,445;
  • U.S. Pat. Nos. 8,752,766; 8,756,059;
  • U.S. Pat. Nos. 8,757,495; 8,760,563;
  • U.S. Pat. Nos. 8,763,909; 8,777,108;
  • U.S. Pat. Nos. 8,777,109; 8,779,898;
  • U.S. Pat. Nos. 8,781,520; 8,783,573;
  • U.S. Pat. Nos. 8,789,757; 8,789,758;
  • U.S. Pat. Nos. 8,789,759; 8,794,520;
  • U.S. Pat. Nos. 8,794,522; 8,794,525;
  • U.S. Pat. Nos. 8,794,526; 8,798,367;
  • U.S. Pat. Nos. 8,807,431; 8,807,432;
  • U.S. Pat. Nos. 8,820,630; 8,822,848;
  • U.S. Pat. Nos. 8,824,692; 8,824,696;
  • U.S. Pat. Nos. 8,842,849; 8,844,822;
  • U.S. Pat. Nos. 8,844,823; 8,849,019;
  • U.S. Pat. Nos. 8,851,383; 8,854,633;
  • U.S. Pat. Nos. 8,866,963; 8,868,421;
  • U.S. Pat. Nos. 8,868,519; 8,868,802;
  • U.S. Pat. Nos. 8,868,803; 8,870,074;
  • U.S. Pat. Nos. 8,879,639; 8,880,426;
  • U.S. Pat. Nos. 8,881,983; 8,881,987;
  • U.S. Pat. Nos. 8,903,172; 8,908,995;
  • U.S. Pat. Nos. 8,910,870; 8,910,875;
  • U.S. Pat. Nos. 8,914,290; 8,914,788;
  • U.S. Pat. Nos. 8,915,439; 8,915,444;
  • U.S. Pat. Nos. 8,916,789; 8,918,250;
  • U.S. Pat. Nos. 8,918,564; 8,925,818;
  • U.S. Pat. Nos. 8,939,374; 8,942,480;
  • U.S. Pat. Nos. 8,944,313; 8,944,327;
  • U.S. Pat. Nos. 8,944,332; 8,950,678;
  • U.S. Pat. Nos. 8,967,468; 8,971,346;
  • U.S. Pat. Nos. 8,976,030; 8,976,368;
  • U.S. Pat. Nos. 8,978,981; 8,978,983;
  • U.S. Pat. Nos. 8,978,984; 8,985,456;
  • U.S. Pat. Nos. 8,985,457; 8,985,459;
  • U.S. Pat. Nos. 8,985,461; 8,988,578;
  • U.S. Pat. Nos. 8,988,590; 8,991,704;
  • U.S. Pat. Nos. 8,996,194; 8,996,384;
  • U.S. Pat. Nos. 9,002,641; 9,007,368;
  • U.S. Pat. Nos. 9,010,641; 9,015,513;
  • U.S. Pat. Nos. 9,016,576; 9,022,288;
  • U.S. Pat. Nos. 9,030,964; 9,033,240;
  • U.S. Pat. Nos. 9,033,242; 9,036,054;
  • U.S. Pat. Nos. 9,037,344; 9,038,911;
  • U.S. Pat. Nos. 9,038,915; 9,047,098;
  • U.S. Pat. Nos. 9,047,359; 9,047,420;
  • U.S. Pat. Nos. 9,047,525; 9,047,531;
  • U.S. Pat. Nos. 9,053,055; 9,053,378;
  • U.S. Pat. Nos. 9,053,380; 9,058,526;
  • U.S. Pat. Nos. 9,064,165; 9,064,167;
  • U.S. Pat. Nos. 9,064,168; 9,064,254;
  • U.S. Pat. Nos. 9,066,032; 9,070,032;
  • U.S. Design Pat. No. D716,285;
  • U.S. Design Pat. No. D723,560;
  • U.S. Design Pat. No. D730,357;
  • U.S. Design Pat. No. D730,901;
  • U.S. Design Pat. No. D730,902;
  • U.S. Design Pat. No. D733,112;
  • U.S. Design Pat. No. D734,339;
  • International Publication No. 2013/163789;
  • International Publication No. 2013/173985;
  • International Publication No. 2014/019130;
  • International Publication No. 2014/110495;
  • U.S. Patent Application Publication No. 2008/0185432;
  • U.S. Patent Application Publication No. 2009/0134221;
  • U.S. Patent Application Publication No. 2010/0177080;
  • U.S. Patent Application Publication No. 2010/0177076;
  • U.S. Patent Application Publication No. 2010/0177707;
  • U.S. Patent Application Publication No. 2010/0177749;
  • U.S. Patent Application Publication No. 2010/0265880;
  • U.S. Patent Application Publication No. 2011/0202554;
  • U.S. Patent Application Publication No. 2012/0111946;
  • U.S. Patent Application Publication No. 2012/0168511;
  • U.S. Patent Application Publication No. 2012/0168512;
  • U.S. Patent Application Publication No. 2012/0193423;
  • U.S. Patent Application Publication No. 2012/0203647;
  • U.S. Patent Application Publication No. 2012/0223141;
  • U.S. Patent Application Publication No. 2012/0228382;
  • U.S. Patent Application Publication No. 2012/0248188;
  • U.S. Patent Application Publication No. 2013/0043312;
  • U.S. Patent Application Publication No. 2013/0082104;
  • U.S. Patent Application Publication No. 2013/0175341;
  • U.S. Patent Application Publication No. 2013/0175343;
  • U.S. Patent Application Publication No. 2013/0257744;
  • U.S. Patent Application Publication No. 2013/0257759;
  • U.S. Patent Application Publication No. 2013/0270346;
  • U.S. Patent Application Publication No. 2013/0287258;
  • U.S. Patent Application Publication No. 2013/0292475;
  • U.S. Patent Application Publication No. 2013/0292477;
  • U.S. Patent Application Publication No. 2013/0293539;
  • U.S. Patent Application Publication No. 2013/0293540;
  • U.S. Patent Application Publication No. 2013/0306728;
  • U.S. Patent Application Publication No. 2013/0306731;
  • U.S. Patent Application Publication No. 2013/0307964;
  • U.S. Patent Application Publication No. 2013/0308625;
  • U.S. Patent Application Publication No. 2013/0313324;
  • U.S. Patent Application Publication No. 2013/0313325;
  • U.S. Patent Application Publication No. 2013/0342717;
  • U.S. Patent Application Publication No. 2014/0001267;
  • U.S. Patent Application Publication No. 2014/0008439;
  • U.S. Patent Application Publication No. 2014/0025584;
  • U.S. Patent Application Publication No. 2014/0034734;
  • U.S. Patent Application Publication No. 2014/0036848;
  • U.S. Patent Application Publication No. 2014/0039693;
  • U.S. Patent Application Publication No. 2014/0042814;
  • U.S. Patent Application Publication No. 2014/0049120;
  • U.S. Patent Application Publication No. 2014/0049635;
  • U.S. Patent Application Publication No. 2014/0061306;
  • U.S. Patent Application Publication No. 2014/0063289;
  • U.S. Patent Application Publication No. 2014/0066136;
  • U.S. Patent Application Publication No. 2014/0067692;
  • U.S. Patent Application Publication No. 2014/0070005;
  • U.S. Patent Application Publication No. 2014/0071840;
  • U.S. Patent Application Publication No. 2014/0074746;
  • U.S. Patent Application Publication No. 2014/0076974;
  • U.S. Patent Application Publication No. 2014/0078341;
  • U.S. Patent Application Publication No. 2014/0078345;
  • U.S. Patent Application Publication No. 2014/0097249;
  • U.S. Patent Application Publication No. 2014/0098792;
  • U.S. Patent Application Publication No. 2014/0100813;
  • U.S. Patent Application Publication No. 2014/0103115;
  • U.S. Patent Application Publication No. 2014/0104413;
  • U.S. Patent Application Publication No. 2014/0104414;
  • U.S. Patent Application Publication No. 2014/0104416;
  • U.S. Patent Application Publication No. 2014/0104451;
  • U.S. Patent Application Publication No. 2014/0106594;
  • U.S. Patent Application Publication No. 2014/0106725;
  • U.S. Patent Application Publication No. 2014/0108010;
  • U.S. Patent Application Publication No. 2014/0108402;
  • U.S. Patent Application Publication No. 2014/0110485;
  • U.S. Patent Application Publication No. 2014/0114530;
  • U.S. Patent Application Publication No. 2014/0124577;
  • U.S. Patent Application Publication No. 2014/0124579;
  • U.S. Patent Application Publication No. 2014/0125842;
  • U.S. Patent Application Publication No. 2014/0125853;
  • U.S. Patent Application Publication No. 2014/0125999;
  • U.S. Patent Application Publication No. 2014/0129378;
  • U.S. Patent Application Publication No. 2014/0131438;
  • U.S. Patent Application Publication No. 2014/0131441;
  • U.S. Patent Application Publication No. 2014/0131443;
  • U.S. Patent Application Publication No. 2014/0131444;
  • U.S. Patent Application Publication No. 2014/0131445;
  • U.S. Patent Application Publication No. 2014/0131448;
  • U.S. Patent Application Publication No. 2014/0133379;
  • U.S. Patent Application Publication No. 2014/0136208;
  • U.S. Patent Application Publication No. 2014/0140585;
  • U.S. Patent Application Publication No. 2014/0151453;
  • U.S. Patent Application Publication No. 2014/0152882;
  • U.S. Patent Application Publication No. 2014/0158770;
  • U.S. Patent Application Publication No. 2014/0159869;
  • U.S. Patent Application Publication No. 2014/0166755;
  • U.S. Patent Application Publication No. 2014/0166759;
  • U.S. Patent Application Publication No. 2014/0168787;
  • U.S. Patent Application Publication No. 2014/0175165;
  • U.S. Patent Application Publication No. 2014/0175172;
  • U.S. Patent Application Publication No. 2014/0191644;
  • U.S. Patent Application Publication No. 2014/0191913;
  • U.S. Patent Application Publication No. 2014/0197238;
  • U.S. Patent Application Publication No. 2014/0197239;
  • U.S. Patent Application Publication No. 2014/0197304;
  • U.S. Patent Application Publication No. 2014/0214631;
  • U.S. Patent Application Publication No. 2014/0217166;
  • U.S. Patent Application Publication No. 2014/0217180;
  • U.S. Patent Application Publication No. 2014/0231500;
  • U.S. Patent Application Publication No. 2014/0232930;
  • U.S. Patent Application Publication No. 2014/0247315;
  • U.S. Patent Application Publication No. 2014/0263493;
  • U.S. Patent Application Publication No. 2014/0263645;
  • U.S. Patent Application Publication No. 2014/0267609;
  • U.S. Patent Application Publication No. 2014/0270196;
  • U.S. Patent Application Publication No. 2014/0270229;
  • U.S. Patent Application Publication No. 2014/0278387;
  • U.S. Patent Application Publication No. 2014/0278391;
  • U.S. Patent Application Publication No. 2014/0282210;
  • U.S. Patent Application Publication No. 2014/0284384;
  • U.S. Patent Application Publication No. 2014/0288933;
  • U.S. Patent Application Publication No. 2014/0297058;
  • U.S. Patent Application Publication No. 2014/0299665;
  • U.S. Patent Application Publication No. 2014/0312121;
  • U.S. Patent Application Publication No. 2014/0319220;
  • U.S. Patent Application Publication No. 2014/0319221;
  • U.S. Patent Application Publication No. 2014/0326787;
  • U.S. Patent Application Publication No. 2014/0332590;
  • U.S. Patent Application Publication No. 2014/0344943;
  • U.S. Patent Application Publication No. 2014/0346233;
  • U.S. Patent Application Publication No. 2014/0351317;
  • U.S. Patent Application Publication No. 2014/0353373;
  • U.S. Patent Application Publication No. 2014/0361073;
  • U.S. Patent Application Publication No. 2014/0361082;
  • U.S. Patent Application Publication No. 2014/0362184;
  • U.S. Patent Application Publication No. 2014/0363015;
  • U.S. Patent Application Publication No. 2014/0369511;
  • U.S. Patent Application Publication No. 2014/0374483;
  • U.S. Patent Application Publication No. 2014/0374485;
  • U.S. Patent Application Publication No. 2015/0001301;
  • U.S. Patent Application Publication No. 2015/0001304;
  • U.S. Patent Application Publication No. 2015/0003673;
  • U.S. Patent Application Publication No. 2015/0009338;
  • U.S. Patent Application Publication No. 2015/0009610;
  • U.S. Patent Application Publication No. 2015/0014416;
  • U.S. Patent Application Publication No. 2015/0021397;
  • U.S. Patent Application Publication No. 2015/0028102;
  • U.S. Patent Application Publication No. 2015/0028103;
  • U.S. Patent Application Publication No. 2015/0028104;
  • U.S. Patent Application Publication No. 2015/0029002;
  • U.S. Patent Application Publication No. 2015/0032709;
  • U.S. Patent Application Publication No. 2015/0039309;
  • U.S. Patent Application Publication No. 2015/0039878;
  • U.S. Patent Application Publication No. 2015/0040378;
  • U.S. Patent Application Publication No. 2015/0048168;
  • U.S. Patent Application Publication No. 2015/0049347;
  • U.S. Patent Application Publication No. 2015/0051992;
  • U.S. Patent Application Publication No. 2015/0053766;
  • U.S. Patent Application Publication No. 2015/0053768;
  • U.S. Patent Application Publication No. 2015/0053769;
  • U.S. Patent Application Publication No. 2015/0060544;
  • U.S. Patent Application Publication No. 2015/0062366;
  • U.S. Patent Application Publication No. 2015/0063215;
  • U.S. Patent Application Publication No. 2015/0063676;
  • U.S. Patent Application Publication No. 2015/0069130;
  • U.S. Patent Application Publication No. 2015/0071819;
  • U.S. Patent Application Publication No. 2015/0083800;
  • U.S. Patent Application Publication No. 2015/0086114;
  • U.S. Patent Application Publication No. 2015/0088522;
  • U.S. Patent Application Publication No. 2015/0096872;
  • U.S. Patent Application Publication No. 2015/0099557;
  • U.S. Patent Application Publication No. 2015/0100196;
  • U.S. Patent Application Publication No. 2015/0102109;
  • U.S. Patent Application Publication No. 2015/0115035;
  • U.S. Patent Application Publication No. 2015/0127791;
  • U.S. Patent Application Publication No. 2015/0128116;
  • U.S. Patent Application Publication No. 2015/0129659;
  • U.S. Patent Application Publication No. 2015/0133047;
  • U.S. Patent Application Publication No. 2015/0134470;
  • U.S. Patent Application Publication No. 2015/0136851;
  • U.S. Patent Application Publication No. 2015/0136854;
  • U.S. Patent Application Publication No. 2015/0142492;
  • U.S. Patent Application Publication No. 2015/0144692;
  • U.S. Patent Application Publication No. 2015/0144698;
  • U.S. Patent Application Publication No. 2015/0144701;
  • U.S. Patent Application Publication No. 2015/0149946;
  • U.S. Patent Application Publication No. 2015/0161429;
  • U.S. Patent Application Publication No. 2015/0169925;
  • U.S. Patent Application Publication No. 2015/0169929;
  • U.S. Patent Application Publication No. 2015/0178523;
  • U.S. Patent Application Publication No. 2015/0178534;
  • U.S. Patent Application Publication No. 2015/0178535;
  • U.S. Patent Application Publication No. 2015/0178536;
  • U.S. Patent Application Publication No. 2015/0178537;
  • U.S. Patent Application Publication No. 2015/0181093;
  • U.S. Patent Application Publication No. 2015/0181109;
  • U.S. patent application Ser. No. 13/367,978 for a Laser Scanning Module Employing an Elastomeric U-Hinge Based Laser Scanning Assembly, filed Feb. 7, 2012 (Feng et al.);
  • U.S. patent application Ser. No. 29/458,405 for an Electronic Device, filed Jun. 19, 2013 (Fitch et al.);
  • U.S. patent application Ser. No. 29/459,620 for an Electronic Device Enclosure, filed Jul. 2, 2013 (London et al.);
  • U.S. patent application Ser. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/150,393 for Indicia-reader Having Unitary Construction Scanner, filed Jan. 8, 2014 (Colavito et al.);
  • U.S. patent application Ser. No. 14/200,405 for Indicia Reader for Size-Limited Applications filed Mar. 7, 2014 (Feng et al.);
  • U.S. patent application Ser. No. 14/231,898 for Hand-Mounted Indicia-Reading Device with Finger Motion Triggering filed Apr. 1, 2014 (Van Horn et al.);
  • U.S. patent application Ser. No. 29/486,759 for an Imaging Terminal, filed Apr. 2, 2014 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/257,364 for Docking System and Method Using Near Field Communication filed Apr. 21, 2014 (Showering);
  • U.S. patent application Ser. No. 14/264,173 for Autofocus Lens System for Indicia Readers filed Apr. 29, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/277,337 for MULTIPURPOSE OPTICAL READER, filed May 14, 2014 (Jovanovski et al.);
  • U.S. patent application Ser. No. 14/283,282 for TERMINAL HAVING ILLUMINATION AND FOCUS CONTROL filed May 21, 2014 (Liu et al.);
  • U.S. patent application Ser. No. 14/327,827 for a MOBILE-PHONE ADAPTER FOR ELECTRONIC TRANSACTIONS, filed Jul. 10, 2014 (Hejl);
  • U.S. patent application Ser. No. 14/334,934 for a SYSTEM AND METHOD FOR INDICIA VERIFICATION, filed Jul. 18, 2014 (Hejl);
  • U.S. patent application Ser. No. 14/339,708 for LASER SCANNING CODE SYMBOL READING SYSTEM, filed Jul. 24, 2014 (Xian et al.);
  • U.S. patent application Ser. No. 14/340,627 for an AXIALLY REINFORCED FLEXIBLE SCAN ELEMENT, filed Jul. 25, 2014 (Rueblinger et al.);
  • U.S. patent application Ser. No. 14/446,391 for MULTIFUNCTION POINT OF SALE APPARATUS WITH OPTICAL SIGNATURE CAPTURE filed Jul. 30, 2014 (Good et al.);
  • U.S. patent application Ser. No. 14/452,697 for INTERACTIVE INDICIA READER, filed Aug. 6, 2014 (Todeschini);
  • U.S. patent application Ser. No. 14/453,019 for DIMENSIONING SYSTEM WITH GUIDED ALIGNMENT, filed Aug. 6, 2014 (Li et al.);
  • U.S. patent application Ser. No. 14/462,801 for MOBILE COMPUTING DEVICE WITH DATA COGNITION SOFTWARE, filed on Aug. 19, 2014 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/483,056 for VARIABLE DEPTH OF FIELD BARCODE SCANNER filed Sep. 10, 2014 (McCloskey et al.);
  • U.S. patent application Ser. No. 14/513,808 for IDENTIFYING INVENTORY ITEMS IN A STORAGE FACILITY filed Oct. 14, 2014 (Singel et al.);
  • U.S. patent application Ser. No. 14/519,195 for HANDHELD DIMENSIONING SYSTEM WITH FEEDBACK filed Oct. 21, 2014 (Laffargue et al.);
  • U.S. patent application Ser. No. 14/519,179 for DIMENSIONING SYSTEM WITH MULTIPATH INTERFERENCE MITIGATION filed Oct. 21, 2014 (Thuries et al.);
  • U.S. patent application Ser. No. 14/519,211 for SYSTEM AND METHOD FOR DIMENSIONING filed Oct. 21, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/519,233 for HANDHELD DIMENSIONER WITH DATA-QUALITY INDICATION filed Oct. 21, 2014 (Laffargue et al.);
  • U.S. patent application Ser. No. 14/519,249 for HANDHELD DIMENSIONING SYSTEM WITH MEASUREMENT-CONFORMANCE FEEDBACK filed Oct. 21, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/527,191 for METHOD AND SYSTEM FOR RECOGNIZING SPEECH USING WILDCARDS IN AN EXPECTED RESPONSE filed Oct. 29, 2014 (Braho et al.);
  • U.S. patent application Ser. No. 14/529,563 for ADAPTABLE INTERFACE FOR A MOBILE COMPUTING DEVICE filed Oct. 31, 2014 (Schoon et al.);
  • U.S. patent application Ser. No. 14/529,857 for BARCODE READER WITH SECURITY FEATURES filed Oct. 31, 2014 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/398,542 for PORTABLE ELECTRONIC DEVICES HAVING A SEPARATE LOCATION TRIGGER UNIT FOR USE IN CONTROLLING AN APPLICATION UNIT filed Nov. 3, 2014 (Bian et al.);
  • U.S. patent application Ser. No. 14/531,154 for DIRECTING AN INSPECTOR THROUGH AN INSPECTION filed Nov. 3, 2014 (Miller et al.);
  • U.S. patent application Ser. No. 14/533,319 for BARCODE SCANNING SYSTEM USING WEARABLE DEVICE WITH EMBEDDED CAMERA filed Nov. 5, 2014 (Todeschini);
  • U.S. patent application Ser. No. 14/535,764 for CONCATENATED EXPECTED RESPONSES FOR SPEECH RECOGNITION filed Nov. 7, 2014 (Braho et al.);
  • U.S. patent application Ser. No. 14/568,305 for AUTO-CONTRAST VIEWFINDER FOR AN INDICIA READER filed Dec. 12, 2014 (Todeschini);
  • U.S. patent application Ser. No. 14/573,022 for DYNAMIC DIAGNOSTIC INDICATOR GENERATION filed Dec. 17, 2014 (Goldsmith);
  • U.S. patent application Ser. No. 14/578,627 for SAFETY SYSTEM AND METHOD filed Dec. 22, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/580,262 for MEDIA GATE FOR THERMAL TRANSFER PRINTERS filed Dec. 23, 2014 (Bowles);
  • U.S. patent application Ser. No. 14/590,024 for SHELVING AND PACKAGE LOCATING SYSTEMS FOR DELIVERY VEHICLES filed Jan. 6, 2015 (Payne);
  • U.S. patent application Ser. No. 14/596,757 for SYSTEM AND METHOD FOR DETECTING BARCODE PRINTING ERRORS filed Jan. 14, 2015 (Ackley);
  • U.S. patent application Ser. No. 14/416,147 for OPTICAL READING APPARATUS HAVING VARIABLE SETTINGS filed Jan. 21, 2015 (Chen et al.);
  • U.S. patent application Ser. No. 14/614,706 for DEVICE FOR SUPPORTING AN ELECTRONIC TOOL ON A USER'S HAND filed Feb. 5, 2015 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/614,796 for CARGO APPORTIONMENT TECHNIQUES filed Feb. 5, 2015 (Morton et al.);
  • U.S. patent application Ser. No. 29/516,892 for TABLE COMPUTER filed Feb. 6, 2015 (Bidwell et al.);
  • U.S. patent application Ser. No. 14/619,093 for METHODS FOR TRAINING A SPEECH RECOGNITION SYSTEM filed Feb. 11, 2015 (Pecorari);
  • U.S. patent application Ser. No. 14/628,708 for DEVICE, SYSTEM, AND METHOD FOR DETERMINING THE STATUS OF CHECKOUT LANES filed Feb. 23, 2015 (Todeschini);
  • U.S. patent application Ser. No. 14/630,841 for TERMINAL INCLUDING IMAGING ASSEMBLY filed Feb. 25, 2015 (Gomez et al.);
  • U.S. patent application Ser. No. 14/635,346 for SYSTEM AND METHOD FOR RELIABLE STORE-AND-FORWARD DATA HANDLING BY ENCODED INFORMATION READING TERMINALS filed Mar. 2, 2015 (Sevier);
  • U.S. patent application Ser. No. 29/519,017 for SCANNER filed Mar. 2, 2015 (Zhou et al.);
  • U.S. patent application Ser. No. 14/405,278 for DESIGN PATTERN FOR SECURE STORE filed Mar. 9, 2015 (Zhu et al.);
  • U.S. patent application Ser. No. 14/660,970 for DECODABLE INDICIA READING TERMINAL WITH COMBINED ILLUMINATION filed Mar. 18, 2015 (Kearney et al.);
  • U.S. patent application Ser. No. 14/661,013 for REPROGRAMMING SYSTEM AND METHOD FOR DEVICES INCLUDING PROGRAMMING SYMBOL filed Mar. 18, 2015 (Soule et al.);
  • U.S. patent application Ser. No. 14/662,922 for MULTIFUNCTION POINT OF SALE SYSTEM filed Mar. 19, 2015 (Van Horn et al.);
  • U.S. patent application Ser. No. 14/663,638 for VEHICLE MOUNT COMPUTER WITH CONFIGURABLE IGNITION SWITCH BEHAVIOR filed Mar. 20, 2015 (Davis et al.);
  • U.S. patent application Ser. No. 14/664,063 for METHOD AND APPLICATION FOR SCANNING A BARCODE WITH A SMART DEVICE WHILE CONTINUOUSLY RUNNING AND DISPLAYING AN APPLICATION ON THE SMART DEVICE DISPLAY filed Mar. 20, 2015 (Todeschini);
  • U.S. patent application Ser. No. 14/669,280 for TRANSFORMING COMPONENTS OF A WEB PAGE TO VOICE PROMPTS filed Mar. 26, 2015 (Funyak et al.);
  • U.S. patent application Ser. No. 14/674,329 for AIMER FOR BARCODE SCANNING filed Mar. 31, 2015 (Bidwell);
  • U.S. patent application Ser. No. 14/676,109 for INDICIA READER filed Apr. 1, 2015 (Huck);
  • U.S. patent application Ser. No. 14/676,327 for DEVICE MANAGEMENT PROXY FOR SECURE DEVICES filed Apr. 1, 2015 (Yeakley et al.);
  • U.S. patent application Ser. No. 14/676,898 for NAVIGATION SYSTEM CONFIGURED TO INTEGRATE MOTION SENSING DEVICE INPUTS filed Apr. 2, 2015 (Showering);
  • U.S. patent application Ser. No. 14/679,275 for DIMENSIONING SYSTEM CALIBRATION SYSTEMS AND METHODS filed Apr. 6, 2015 (Laffargue et al.);
  • U.S. patent application Ser. No. 29/523,098 for HANDLE FOR A TABLET COMPUTER filed Apr. 7, 2015 (Bidwell et al.);
  • U.S. patent application Ser. No. 14/682,615 for SYSTEM AND METHOD FOR POWER MANAGEMENT OF MOBILE DEVICES filed Apr. 9, 2015 (Murawski et al.);
  • U.S. patent application Ser. No. 14/686,822 for MULTIPLE PLATFORM SUPPORT SYSTEM AND METHOD filed Apr. 15, 2015 (Qu et al.);
  • U.S. patent application Ser. No. 14/687,289 for SYSTEM FOR COMMUNICATION VIA A PERIPHERAL HUB filed Apr. 15, 2015 (Kohtz et al.);
  • U.S. patent application Ser. No. 29/524,186 for SCANNER filed Apr. 17, 2015 (Zhou et al.);
  • U.S. patent application Ser. No. 14/695,364 for MEDICATION MANAGEMENT SYSTEM filed Apr. 24, 2015 (Sewell et al.);
  • U.S. patent application Ser. No. 14/695,923 for SECURE UNATTENDED NETWORK AUTHENTICATION filed Apr. 24, 2015 (Kubler et al.);
  • U.S. patent application Ser. No. 29/525,068 for TABLET COMPUTER WITH REMOVABLE SCANNING DEVICE filed Apr. 27, 2015 (Schulte et al.);
  • U.S. patent application Ser. No. 14/699,436 for SYMBOL READING SYSTEM HAVING PREDICTIVE DIAGNOSTICS filed Apr. 29, 2015 (Nahill et al.);
  • U.S. patent application Ser. No. 14/702,110 for SYSTEM AND METHOD FOR REGULATING BARCODE DATA INJECTION INTO A RUNNING APPLICATION ON A SMART DEVICE filed May 1, 2015 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/702,979 for TRACKING BATTERY CONDITIONS filed May 4, 2015 (Young et al.);
  • U.S. patent application Ser. No. 14/704,050 for INTERMEDIATE LINEAR POSITIONING filed May 5, 2015 (Charpentier et al.);
  • U.S. patent application Ser. No. 14/705,012 for HANDS-FREE HUMAN MACHINE INTERFACE RESPONSIVE TO A DRIVER OF A VEHICLE filed May 6, 2015 (Fitch et al.);
  • U.S. patent application Ser. No. 14/705,407 for METHOD AND SYSTEM TO PROTECT SOFTWARE-BASED NETWORK-CONNECTED DEVICES FROM ADVANCED PERSISTENT THREAT filed May 6, 2015 (Hussey et al.);
  • U.S. patent application Ser. No. 14/707,037 for SYSTEM AND METHOD FOR DISPLAY OF INFORMATION USING A VEHICLE-MOUNT COMPUTER filed May 8, 2015 (Chamberlin);
  • U.S. patent application Ser. No. 14/707,123 for APPLICATION INDEPENDENT DEX/UCS INTERFACE filed May 8, 2015 (Pape);
  • U.S. patent application Ser. No. 14/707,492 for METHOD AND APPARATUS FOR READING OPTICAL INDICIA USING A PLURALITY OF DATA SOURCES filed May 8, 2015 (Smith et al.);
  • U.S. patent application Ser. No. 14/710,666 for PRE-PAID USAGE SYSTEM FOR ENCODED INFORMATION READING TERMINALS filed May 13, 2015 (Smith);
  • U.S. patent application Ser. No. 29/526,918 for CHARGING BASE filed May 14, 2015 (Fitch et al.);
  • U.S. patent application Ser. No. 14/715,672 for AUGUMENTED REALITY ENABLED HAZARD DISPLAY filed May 19, 2015 (Venkatesha et al.);
  • U.S. patent application Ser. No. 14/715,916 for EVALUATING IMAGE VALUES filed May 19, 2015 (Ackley);
  • U.S. patent application Ser. No. 14/722,608 for INTERACTIVE USER INTERFACE FOR CAPTURING A DOCUMENT IN AN IMAGE SIGNAL filed May 27, 2015 (Showering et al.);
  • U.S. patent application Ser. No. 29/528,165 for IN-COUNTER BARCODE SCANNER filed May 27, 2015 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/724,134 for ELECTRONIC DEVICE WITH WIRELESS PATH SELECTION CAPABILITY filed May 28, 2015 (Wang et al.);
  • U.S. patent application Ser. No. 14/724,849 for METHOD OF PROGRAMMING THE DEFAULT CABLE INTERFACE SOFTWARE IN AN INDICIA READING DEVICE filed May 29, 2015 (Barten);
  • U.S. patent application Ser. No. 14/724,908 for IMAGING APPARATUS HAVING IMAGING ASSEMBLY filed May 29, 2015 (Barber et al.);
  • U.S. patent application Ser. No. 14/725,352 for APPARATUS AND METHODS FOR MONITORING ONE OR MORE PORTABLE DATA TERMINALS (Caballero et al.);
  • U.S. patent application Ser. No. 29/528,590 for ELECTRONIC DEVICE filed May 29, 2015 (Fitch et al.);
  • U.S. patent application Ser. No. 29/528,890 for MOBILE COMPUTER HOUSING filed Jun. 2, 2015 (Fitch et al.);
  • U.S. patent application Ser. No. 14/728,397 for DEVICE MANAGEMENT USING VIRTUAL INTERFACES CROSS-REFERENCE TO RELATED APPLICATIONS filed Jun. 2, 2015 (Caballero);
  • U.S. patent application Ser. No. 14/732,870 for DATA COLLECTION MODULE AND SYSTEM filed Jun. 8, 2015 (Powilleit);
  • U.S. patent application Ser. No. 29/529,441 for INDICIA READING DEVICE filed Jun. 8, 2015 (Zhou et al.);
  • U.S. patent application Ser. No. 14/735,717 for INDICIA-READING SYSTEMS HAVING AN INTERFACE WITH A USER'S NERVOUS SYSTEM filed Jun. 10, 2015 (Todeschini);
  • U.S. patent application Ser. No. 14/738,038 for METHOD OF AND SYSTEM FOR DETECTING OBJECT WEIGHING INTERFERENCES filed Jun. 12, 2015 (Amundsen et al.);
  • U.S. patent application Ser. No. 14/740,320 for TACTILE SWITCH FOR A MOBILE ELECTRONIC DEVICE filed Jun. 16, 2015 (Bandringa);
  • U.S. patent application Ser. No. 14/740,373 for CALIBRATING A VOLUME DIMENSIONER filed Jun. 16, 2015 (Ackley et al.);
  • U.S. patent application Ser. No. 14/742,818 for INDICIA READING SYSTEM EMPLOYING DIGITAL GAIN CONTROL filed Jun. 18, 2015 (Xian et al.);
  • U.S. patent application Ser. No. 14/743,257 for WIRELESS MESH POINT PORTABLE DATA TERMINAL filed Jun. 18, 2015 (Wang et al.);
  • U.S. patent application Ser. No. 29/530,600 for CYCLONE filed Jun. 18, 2015 (Vargo et al);
  • U.S. patent application Ser. No. 14/744,633 for IMAGING APPARATUS COMPRISING IMAGE SENSOR ARRAY HAVING SHARED GLOBAL SHUTTER CIRCUITRY filed Jun. 19, 2015 (Wang);
  • U.S. patent application Ser. No. 14/744,836 for CLOUD-BASED SYSTEM FOR READING OF DECODABLE INDICIA filed Jun. 19, 2015 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/745,006 for SELECTIVE OUTPUT OF DECODED MESSAGE DATA filed Jun. 19, 2015 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/747,197 for OPTICAL PATTERN PROJECTOR filed Jun. 23, 2015 (Thuries et al.);
  • U.S. patent application Ser. No. 14/747,490 for DUAL-PROJECTOR THREE-DIMENSIONAL SCANNER filed Jun. 23, 2015 (Jovanovski et al.); and
  • U.S. patent application Ser. No. 14/748,446 for CORDLESS INDICIA READER WITH A MULTIFUNCTION COIL FOR WIRELESS CHARGING AND EAS DEACTIVATION, filed Jun. 24, 2015 (Xie et al.).


In the specification and/or figures, typical embodiments of the invention have been disclosed. The present invention is not limited to such exemplary embodiments. The use of the term “and/or” includes any and all combinations of one or more of the associated listed items. The figures are schematic representations and so are not necessarily drawn to scale. Unless otherwise noted, specific terms have been used in a generic and descriptive sense and not for purposes of limitation.

Claims
  • 1. A workflow system comprising: at least one audio-based picker device, configured to enable two-way audio communication with a user and is further configured to communicate a workflow process to the user; anda host computer in wireless communication with the said at least one audio-based picker device, wherein the host computer is configured to determine the workflow process for the user and transmit the workflow process to the at least one audio-based picker device, wherein the workflow process comprises a distance that the user is to travel to a first location of an item from a second location of the at least one audio-based picker device,wherein the at least one audio-based picker device includes an inertial unit configured to: calculate, in response to an indication that the user intends to pick the item, an estimated step count of the user indicating movement of the user from the second location of the at least one audio-based picker device to the first location of the item, and, wherein the at least one audio-based picker device is configured to determine that the user is at the first location of the item when the estimated step count satisfies a minimum threshold step count associated with the distance between the first location of the item and the second location of the at least one audio-based picker device, and wherein the minimum threshold step count is determined based on a function of warehouse distance information, time information, and an approximate stride length of the user,in response to the determination that the estimated step count of the user exceeds the minimum threshold step count, prompt the user to provide a verbal response including check digits associated with the first location, anddetermine that a workflow operation associated to the workflow process is complete when the estimated step count exceeds the minimum threshold step count and when the verbal response includes correct check digits.
  • 2. The workflow system of claim 1, wherein the calculated estimated step count of the user corresponds to an estimated distance travelled by the user between the first location of the item and the second location of the at least one audio-based picker device.
  • 3. The workflow system of claim 1, wherein the at least one audio-based picker device includes an audio unit configured to receive verbal responses from the user and transmit the verbal responses and an estimated distance travelled by the user between the first location of the item and the second location of the at least one audio-based picker device to the host computer.
  • 4. The workflow system of claim 1, wherein the inertial unit of the at least one audio-based picker device includes at least one accelerometer configured to measure acceleration forces along an x-axis, a y-axis, and a z-axis.
  • 5. The workflow system of claim 4, wherein the inertial unit of the at least one audio-based picker device further includes at least one gyroscope configured to measure angular velocity along the x-axis, y-axis, and z-axis.
  • 6. The workflow system of claim 5, wherein the inertial unit of the at least one audio-based picker device further includes a step count module configured to calculate a number of steps taken by the user based on measurements from the at least one accelerometer and the at least one gyroscope.
  • 7. The workflow system of claim 1, wherein the at least one audio-based picker device is configured to convey a first set of audio output signals to the user when it is determined that the estimated step count is below the minimum threshold step count associated with the distance and to convey a second set of audio output signals to the user when it is determined that the estimated step count is above the minimum threshold step count associated with the distance.
  • 8. A workflow method comprising: transmitting a first workflow operation to a communication unit associated with a user for conveying the first workflow operation to the user, wherein the first workflow operation comprises an estimated distance that the user is to travel to a first location of an item from a second location of the communication unit;receiving an estimated step count from the communication unit associated with the user, the estimated step count being related to walking movements of the user and is representative of a user travel distance;receiving a first audio signal from the communication unit associated with the user, the first audio signal including at least one verbal response from the user that indicates that the user intends to pick the item;determining whether conditions of the first workflow operation have been met based on the estimated step count and the verbal responses, wherein determining that the conditions of the first workflow operation have been met comprises determining that the user is at the first location of the item when the estimated step count satisfies a predetermined threshold step count associated with the estimated distance between the first location of the item and the second location of the communication unit, and wherein the predetermined threshold step count is determined based on a function of warehouse distance information, time information, and an approximate stride length of the user; andin response to determining that the estimated step count of the user exceeds the predetermined threshold step count, prompting the user to provide a verbal response including check digits associated with the first location, anddetermining that the conditions of the first workflow operation have been met when the estimated step count exceeds the predetermined threshold step count and when the verbal response includes correct check digits.
  • 9. The workflow method of claim 8, wherein the workflow method further comprises transmitting a second audio signal to the communication unit associated with the user when it is determined that the conditions of the first workflow operation have been met.
  • 10. The workflow method of claim 9, wherein the second audio signal includes at least one verbal instruction to the user that instructs the user to pick a certain quantity of the item.
  • 11. The workflow method of claim 8, wherein the workflow method further comprises transmitting a third audio signal to the communication unit associated with the user when it is determined that the conditions of the first workflow operation have not been met.
  • 12. The workflow method of claim 11, wherein the third audio signal includes at least one verbal instruction to the user that instructs the user to move to a proper location.
  • 13. A non-transitory computer readable medium having instructions for causing a computer to execute a workflow method, the workflow method comprising: transmitting a first workflow operation to a communication unit associated with a user for conveying the first workflow operation to the user, wherein the first workflow operation comprises an estimated distance that the user is to travel to a first location of an item from a second location of the communication unit;receiving an estimated step count from the communication unit associated with the user, the estimated step count being related to walking movements of the user and is representative of a user travel distance;receiving a first audio signal from the communication unit associated with the user, the first audio signal including at least one verbal response from the user that indicates that the user intends to pick the item;determining whether conditions of the first workflow operation have been met based on the estimated step count and the verbal responses, wherein determining that the conditions of the first workflow operation have been met comprises determining that the user is at the first location of the item when the estimated step count satisfies a predetermined threshold step count associated with the estimated distance between the first location of the item and the second location of the communication unit, and wherein the predetermined threshold step count is determined based on a function of warehouse distance information, time information, and an approximate stride length of the user; andin response to determining that the estimated step count of the user exceeds the predetermined threshold step count, prompting the user to provide a verbal response including check digits associated with the first location, anddetermining that the conditions of the first workflow operation have been met when the estimated step count exceeds the predetermined threshold step count and when the verbal response includes correct check digits.
  • 14. The non-transitory computer readable medium of claim 13, wherein the workflow method further comprises transmitting a second audio signal to the communication unit associated with the user when it is determined that the conditions of the first workflow operation have been met.
  • 15. The non-transitory computer readable medium of claim 14, wherein the second audio signal includes at least one verbal instruction to the user that instructs the user to pick a certain quantity of the item.
  • 16. The non-transitory computer readable medium of claim 13, wherein the workflow method further comprises transmitting a third audio signal to the communication unit associated with the user when it is determined that the conditions of the first workflow operation have not been met.
  • 17. The non-transitory computer readable medium of claim 16, wherein the third audio signal includes at least one verbal instruction to the user that instructs the user to move to a proper location.
  • 18. The non-transitory computer readable medium of claim 13, wherein the workflow method further comprises transmitting a second workflow operation to a communication unit associated with the user for conveying the second workflow operation to the user.
US Referenced Citations (489)
Number Name Date Kind
6832725 Gardiner et al. Dec 2004 B2
7128266 Zhu et al. Oct 2006 B2
7159783 Walczyk et al. Jan 2007 B2
7413127 Ehrhart et al. Aug 2008 B2
7726575 Wang et al. Jun 2010 B2
8294969 Plesko Oct 2012 B2
8317105 Kotlarsky et al. Nov 2012 B2
8322622 Liu Dec 2012 B2
8366005 Kotlarsky et al. Feb 2013 B2
8371507 Haggerty et al. Feb 2013 B2
8376233 Van Horn et al. Feb 2013 B2
8381979 Franz Feb 2013 B2
8390909 Plesko Mar 2013 B2
8408464 Zhu et al. Apr 2013 B2
8408468 Horn et al. Apr 2013 B2
8408469 Good Apr 2013 B2
8424768 Rueblinger et al. Apr 2013 B2
8448863 Xian et al. May 2013 B2
8457013 Essinger et al. Jun 2013 B2
8459557 Havens et al. Jun 2013 B2
8469272 Kearney Jun 2013 B2
8474712 Kearney et al. Jul 2013 B2
8479992 Kotlarsky et al. Jul 2013 B2
8490877 Kearney Jul 2013 B2
8517271 Kotlarsky et al. Aug 2013 B2
8523076 Good Sep 2013 B2
8528818 Ehrhart et al. Sep 2013 B2
8544737 Gomez et al. Oct 2013 B2
8548420 Grunow et al. Oct 2013 B2
8550335 Samek et al. Oct 2013 B2
8550354 Gannon et al. Oct 2013 B2
8550357 Kearney Oct 2013 B2
8556174 Kosecki et al. Oct 2013 B2
8556176 Van Horn et al. Oct 2013 B2
8556177 Hussey et al. Oct 2013 B2
8559767 Barber et al. Oct 2013 B2
8561895 Gomez et al. Oct 2013 B2
8561903 Sauerwein Oct 2013 B2
8561905 Edmonds et al. Oct 2013 B2
8565107 Pease et al. Oct 2013 B2
8571307 Li et al. Oct 2013 B2
8579200 Samek et al. Nov 2013 B2
8583924 Caballero et al. Nov 2013 B2
8584945 Wang et al. Nov 2013 B2
8587595 Wang Nov 2013 B2
8587697 Hussey et al. Nov 2013 B2
8588869 Sauerwein et al. Nov 2013 B2
8590789 Nahill et al. Nov 2013 B2
8596539 Havens et al. Dec 2013 B2
8596542 Havens et al. Dec 2013 B2
8596543 Havens et al. Dec 2013 B2
8599271 Havens et al. Dec 2013 B2
8599957 Peake et al. Dec 2013 B2
8600158 Li et al. Dec 2013 B2
8600167 Showering Dec 2013 B2
8602309 Longacre et al. Dec 2013 B2
8608053 Meier et al. Dec 2013 B2
8608071 Liu et al. Dec 2013 B2
8611309 Wang et al. Dec 2013 B2
8615487 Gomez et al. Dec 2013 B2
8621123 Caballero Dec 2013 B2
8622303 Meier et al. Jan 2014 B2
8628013 Ding Jan 2014 B2
8628015 Wang et al. Jan 2014 B2
8628016 Winegar Jan 2014 B2
8629926 Wang Jan 2014 B2
8630491 Longacre et al. Jan 2014 B2
8635309 Berthiaume et al. Jan 2014 B2
8636200 Kearney Jan 2014 B2
8636212 Nahill et al. Jan 2014 B2
8636215 Ding et al. Jan 2014 B2
8636224 Wang Jan 2014 B2
8638806 Wang et al. Jan 2014 B2
8640958 Lu et al. Feb 2014 B2
8640960 Wang et al. Feb 2014 B2
8643717 Li et al. Feb 2014 B2
8646692 Meier et al. Feb 2014 B2
8646694 Wang et al. Feb 2014 B2
8657200 Ren et al. Feb 2014 B2
8659397 Vargo et al. Feb 2014 B2
8668149 Good Mar 2014 B2
8678285 Kearney Mar 2014 B2
8678286 Smith et al. Mar 2014 B2
8682077 Longacre Mar 2014 B1
D702237 Oberpriller et al. Apr 2014 S
8687282 Feng et al. Apr 2014 B2
8692927 Pease et al. Apr 2014 B2
8695880 Bremer et al. Apr 2014 B2
8698949 Grunow et al. Apr 2014 B2
8702000 Barber et al. Apr 2014 B2
8717494 Gannon May 2014 B2
8720783 Biss et al. May 2014 B2
8723804 Fletcher et al. May 2014 B2
8723904 Marty et al. May 2014 B2
8727223 Wang May 2014 B2
8740082 Wilz Jun 2014 B2
8740085 Furlong et al. Jun 2014 B2
8746563 Hennick et al. Jun 2014 B2
8750445 Peake et al. Jun 2014 B2
8752766 Xian et al. Jun 2014 B2
8756059 Braho et al. Jun 2014 B2
8757495 Qu et al. Jun 2014 B2
8760563 Koziol et al. Jun 2014 B2
8763909 Reed et al. Jul 2014 B2
8777108 Coyle Jul 2014 B2
8777109 Oberpriller et al. Jul 2014 B2
8779898 Havens et al. Jul 2014 B2
8781520 Payne et al. Jul 2014 B2
8783573 Havens et al. Jul 2014 B2
8789757 Barten Jul 2014 B2
8789758 Hawley et al. Jul 2014 B2
8789759 Xian et al. Jul 2014 B2
8794520 Wang et al. Aug 2014 B2
8794522 Ehrhart Aug 2014 B2
8794525 Amundsen et al. Aug 2014 B2
8794526 Wang et al. Aug 2014 B2
8798367 Ellis Aug 2014 B2
8807431 Wang et al. Aug 2014 B2
8807432 Van Horn et al. Aug 2014 B2
8820630 Qu et al. Sep 2014 B2
8822848 Meagher Sep 2014 B2
8824692 Sheerin et al. Sep 2014 B2
8824696 Braho Sep 2014 B2
8842849 Wahl et al. Sep 2014 B2
8844822 Kotlarsky et al. Sep 2014 B2
8844823 Fritz et al. Sep 2014 B2
8849019 Li et al. Sep 2014 B2
D716285 Chaney et al. Oct 2014 S
8851383 Yeakley et al. Oct 2014 B2
8854633 Laffargue Oct 2014 B2
8866963 Grunow et al. Oct 2014 B2
8868421 Braho et al. Oct 2014 B2
8868519 Maloy et al. Oct 2014 B2
8868802 Barten Oct 2014 B2
8868803 Caballero Oct 2014 B2
8870074 Gannon Oct 2014 B1
8879639 Sauerwein Nov 2014 B2
8880426 Smith Nov 2014 B2
8881983 Havens et al. Nov 2014 B2
8881987 Wang Nov 2014 B2
8903172 Smith Dec 2014 B2
8908995 Benos et al. Dec 2014 B2
8910870 Li et al. Dec 2014 B2
8910875 Ren et al. Dec 2014 B2
8914290 Hendrickson et al. Dec 2014 B2
8914788 Pettinelli et al. Dec 2014 B2
8915439 Feng et al. Dec 2014 B2
8915444 Havens et al. Dec 2014 B2
8916789 Woodburn Dec 2014 B2
8918250 Hollifield Dec 2014 B2
8918564 Caballero Dec 2014 B2
8925818 Kosecki et al. Jan 2015 B2
8939374 Jovanovski et al. Jan 2015 B2
8942480 Ellis Jan 2015 B2
8944313 Williams et al. Feb 2015 B2
8944327 Meier et al. Feb 2015 B2
8944332 Harding et al. Feb 2015 B2
8950678 Germaine et al. Feb 2015 B2
D723560 Zhou et al. Mar 2015 S
8967468 Gomez et al. Mar 2015 B2
8971346 Sevier Mar 2015 B2
8976030 Cunningham Mar 2015 B2
8976368 Akel et al. Mar 2015 B2
8978981 Guan Mar 2015 B2
8978983 Bremer et al. Mar 2015 B2
8978984 Hennick et al. Mar 2015 B2
8985456 Zhu et al. Mar 2015 B2
8985457 Soule et al. Mar 2015 B2
8985459 Kearney et al. Mar 2015 B2
8985461 Gelay et al. Mar 2015 B2
8988578 Showering Mar 2015 B2
8988590 Gillet et al. Mar 2015 B2
8991704 Hopper et al. Mar 2015 B2
8996194 Davis et al. Mar 2015 B2
8996384 Funyak et al. Mar 2015 B2
8998091 Edmonds et al. Apr 2015 B2
9002641 Showering Apr 2015 B2
9007368 Laffargue et al. Apr 2015 B2
9010641 Qu et al. Apr 2015 B2
9015513 Murawski et al. Apr 2015 B2
9016576 Brady et al. Apr 2015 B2
D730357 Fitch et al. May 2015 S
9022288 Nahill et al. May 2015 B2
9030964 Essinger et al. May 2015 B2
9033240 Smith et al. May 2015 B2
9033242 Gillet et al. May 2015 B2
9036054 Koziol et al. May 2015 B2
9037344 Chamberlin May 2015 B2
9038911 Xian et al. May 2015 B2
9038915 Smith May 2015 B2
D730901 Oberpriller et al. Jun 2015 S
D730902 Fitch et al. Jun 2015 S
D733112 Chaney et al. Jun 2015 S
9047098 Barten Jun 2015 B2
9047359 Caballero et al. Jun 2015 B2
9047420 Caballero Jun 2015 B2
9047525 Barber Jun 2015 B2
9047531 Showering et al. Jun 2015 B2
9049640 Wang et al. Jun 2015 B2
9053055 Caballero Jun 2015 B2
9053378 Hou et al. Jun 2015 B1
9053380 Xian et al. Jun 2015 B2
9057641 Amundsen et al. Jun 2015 B2
9058526 Powilleit Jun 2015 B2
9064165 Havens et al. Jun 2015 B2
9064167 Xian et al. Jun 2015 B2
9064168 Todeschini et al. Jun 2015 B2
9064254 Todeschini et al. Jun 2015 B2
9066032 Wang Jun 2015 B2
9070032 Corcoran Jun 2015 B2
D734339 Zhou et al. Jul 2015 S
D734751 Oberpriller et al. Jul 2015 S
9082023 Feng et al. Jul 2015 B2
9205886 Hickman et al. Dec 2015 B1
9224022 Ackley et al. Dec 2015 B2
9224027 Van Horn et al. Dec 2015 B2
D747321 London et al. Jan 2016 S
9230140 Ackley Jan 2016 B1
9443123 Hejl Jan 2016 B2
9250712 Todeschini Feb 2016 B1
9258033 Showering Feb 2016 B2
9261398 Amundsen et al. Feb 2016 B2
9262633 Todeschini et al. Feb 2016 B1
9262664 Soule et al. Feb 2016 B2
9274806 Barten Mar 2016 B2
9282501 Wang et al. Mar 2016 B2
9292969 Laffargue et al. Mar 2016 B2
9298667 Caballero Mar 2016 B2
9310609 Rueblinger et al. Apr 2016 B2
9319548 Showering et al. Apr 2016 B2
D757009 Oberpriller et al. May 2016 S
9342724 McCloskey May 2016 B2
9342827 Smith May 2016 B2
9355294 Smith et al. May 2016 B2
9367722 Xian et al. Jun 2016 B2
9375945 Bowles Jun 2016 B1
D760719 Zhou et al. Jul 2016 S
9390596 Todeschini Jul 2016 B1
9396375 Qu et al. Jul 2016 B2
9398008 Todeschini et al. Jul 2016 B2
D762604 Fitch et al. Aug 2016 S
D762647 Fitch et al. Aug 2016 S
9407840 Wang Aug 2016 B2
9412242 Van Horn et al. Aug 2016 B2
9418252 Nahill et al. Aug 2016 B2
D766244 Zhou et al. Sep 2016 S
9443222 Singel et al. Sep 2016 B2
9448610 Davis et al. Sep 2016 B2
9478113 Xie et al. Oct 2016 B2
9582696 Barber et al. Feb 2017 B2
9616749 Chamberlin Apr 2017 B2
9618993 Murawski et al. Apr 2017 B2
9715614 Todeschini et al. Jul 2017 B2
9734493 Gomez et al. Aug 2017 B2
10019334 Caballero et al. Jul 2018 B2
10021043 Sevier Jul 2018 B2
10327158 Wang et al. Jun 2019 B2
10410029 Powilleit Sep 2019 B2
20070063048 Havens et al. Mar 2007 A1
20090134221 Zhu et al. May 2009 A1
20100177076 Essinger et al. Jul 2010 A1
20100177080 Essinger et al. Jul 2010 A1
20100177707 Essinger et al. Jul 2010 A1
20100177749 Essinger et al. Jul 2010 A1
20100256950 Morimoto et al. Oct 2010 A1
20100265880 Rautiola et al. Oct 2010 A1
20110169999 Grunow et al. Jul 2011 A1
20110202554 Powilleit et al. Aug 2011 A1
20120111946 Golant May 2012 A1
20120168512 Kotlarsky et al. Jul 2012 A1
20120193423 Samek Aug 2012 A1
20120203647 Smith Aug 2012 A1
20120223141 Good et al. Sep 2012 A1
20130043312 Van Horn Feb 2013 A1
20130075168 Amundsen et al. Mar 2013 A1
20130175341 Kearney et al. Jul 2013 A1
20130175343 Good Jul 2013 A1
20130257744 Daghigh et al. Oct 2013 A1
20130257759 Daghigh Oct 2013 A1
20130270346 Xian et al. Oct 2013 A1
20130287258 Kearney Oct 2013 A1
20130292475 Kotlarsky et al. Nov 2013 A1
20130292477 Hennick et al. Nov 2013 A1
20130293539 Hunt et al. Nov 2013 A1
20130293540 Laffargue et al. Nov 2013 A1
20130306728 Thuries et al. Nov 2013 A1
20130306731 Pedraro Nov 2013 A1
20130307964 Bremer et al. Nov 2013 A1
20130308625 Park et al. Nov 2013 A1
20130313324 Koziol et al. Nov 2013 A1
20130313325 Wilz et al. Nov 2013 A1
20130335273 Pakzad Dec 2013 A1
20130342717 Havens et al. Dec 2013 A1
20140001267 Giordano et al. Jan 2014 A1
20140002828 Laffargue et al. Jan 2014 A1
20140008439 Wang Jan 2014 A1
20140025584 Liu et al. Jan 2014 A1
20140100813 Showering Jan 2014 A1
20140034734 Sauerwein Feb 2014 A1
20140036848 Pease et al. Feb 2014 A1
20140039693 Havens et al. Feb 2014 A1
20140042814 Kather et al. Feb 2014 A1
20140049120 Kohtz et al. Feb 2014 A1
20140049635 Laffargue et al. Feb 2014 A1
20140061306 Wu et al. Mar 2014 A1
20140063289 Hussey et al. Mar 2014 A1
20140066136 Sauerwein et al. Mar 2014 A1
20140067692 Ye et al. Mar 2014 A1
20140070005 Nahill et al. Mar 2014 A1
20140071840 Venancio Mar 2014 A1
20140074746 Wang Mar 2014 A1
20140076974 Havens et al. Mar 2014 A1
20140078341 Havens et al. Mar 2014 A1
20140078342 Li et al. Mar 2014 A1
20140078345 Showering Mar 2014 A1
20140098792 Wang et al. Apr 2014 A1
20140100774 Showering Apr 2014 A1
20140103115 Meier et al. Apr 2014 A1
20140104413 McCloskey et al. Apr 2014 A1
20140104414 McCloskey et al. Apr 2014 A1
20140104416 Giordano et al. Apr 2014 A1
20140104451 Todeschini et al. Apr 2014 A1
20140106594 Skvoretz Apr 2014 A1
20140106725 Sauerwein Apr 2014 A1
20140108010 Maltseff et al. Apr 2014 A1
20140108402 Gomez et al. Apr 2014 A1
20140108682 Caballero Apr 2014 A1
20140110485 Toa et al. Apr 2014 A1
20140114530 Fitch et al. Apr 2014 A1
20140124577 Wang et al. May 2014 A1
20140124579 Ding May 2014 A1
20140125842 Winegar May 2014 A1
20140125853 Wang May 2014 A1
20140125999 Longacre et al. May 2014 A1
20140129378 Richardson May 2014 A1
20140131438 Kearney May 2014 A1
20140131441 Nahill et al. May 2014 A1
20140131443 Smith May 2014 A1
20140131444 Wang May 2014 A1
20140131445 Ding et al. May 2014 A1
20140131448 Xian et al. May 2014 A1
20140133379 Wang et al. May 2014 A1
20140136208 Maltseff et al. May 2014 A1
20140140585 Wang May 2014 A1
20140151453 Meier et al. Jun 2014 A1
20140152882 Samek et al. Jun 2014 A1
20140158770 Sevier et al. Jun 2014 A1
20140159869 Zumsteg et al. Jun 2014 A1
20140166755 Liu et al. Jun 2014 A1
20140166757 Smith Jun 2014 A1
20140166759 Liu et al. Jun 2014 A1
20140168787 Wang et al. Jun 2014 A1
20140175165 Havens et al. Jun 2014 A1
20140175172 Jovanovski et al. Jun 2014 A1
20140191644 Chaney Jul 2014 A1
20140191913 Ge et al. Jul 2014 A1
20140197238 Lui et al. Jul 2014 A1
20140197239 Havens et al. Jul 2014 A1
20140197304 Feng et al. Jul 2014 A1
20140203087 Smith et al. Jul 2014 A1
20140204268 Grunow et al. Jul 2014 A1
20140214631 Hansen Jul 2014 A1
20140217166 Berthiaume et al. Aug 2014 A1
20140217180 Liu Aug 2014 A1
20140231500 Ehrhart et al. Aug 2014 A1
20140232930 Anderson Aug 2014 A1
20140247315 Marty et al. Sep 2014 A1
20140263493 Amurgis et al. Sep 2014 A1
20140263645 Smith et al. Sep 2014 A1
20140267609 Laffargue Sep 2014 A1
20140270196 Braho et al. Sep 2014 A1
20140270229 Braho Sep 2014 A1
20140278387 DiGregorio Sep 2014 A1
20140278391 Braho et al. Sep 2014 A1
20140282210 Bianconi Sep 2014 A1
20140284384 Lu et al. Sep 2014 A1
20140288933 Braho et al. Sep 2014 A1
20140297058 Barker et al. Oct 2014 A1
20140299665 Barber et al. Oct 2014 A1
20140312121 Lu et al. Oct 2014 A1
20140319220 Coyle Oct 2014 A1
20140319221 Oberpriller et al. Oct 2014 A1
20140326787 Barten Nov 2014 A1
20140332590 Wang et al. Nov 2014 A1
20140344943 Todeschini et al. Nov 2014 A1
20140346233 Liu et al. Nov 2014 A1
20140351317 Smith et al. Nov 2014 A1
20140353373 Van Horn et al. Dec 2014 A1
20140361073 Qu et al. Dec 2014 A1
20140361082 Xian et al. Dec 2014 A1
20140362184 Jovanovski et al. Dec 2014 A1
20140363015 Braho Dec 2014 A1
20140369511 Sheerin et al. Dec 2014 A1
20140374483 Lu Dec 2014 A1
20140374485 Xian et al. Dec 2014 A1
20140379346 Aleksic et al. Dec 2014 A1
20150001301 Ouyang Jan 2015 A1
20150001304 Todeschini Jan 2015 A1
20150003673 Fletcher Jan 2015 A1
20150009338 Laffargue et al. Jan 2015 A1
20150009610 London et al. Jan 2015 A1
20150014416 Kotlarsky et al. Jan 2015 A1
20150021397 Rueblinger et al. Jan 2015 A1
20150028102 Ren et al. Jan 2015 A1
20150028103 Jiang Jan 2015 A1
20150028104 Ma et al. Jan 2015 A1
20150029002 Yeakley et al. Jan 2015 A1
20150032709 Maloy et al. Jan 2015 A1
20150039309 Braho et al. Feb 2015 A1
20150040378 Saber et al. Feb 2015 A1
20150048168 Fritz et al. Feb 2015 A1
20150049347 Laffargue et al. Feb 2015 A1
20150051992 Smith Feb 2015 A1
20150053766 Havens et al. Feb 2015 A1
20150053768 Wang et al. Feb 2015 A1
20150053769 Thuries et al. Feb 2015 A1
20150062366 Liu et al. Mar 2015 A1
20150063215 Wang Mar 2015 A1
20150063676 Lloyd et al. Mar 2015 A1
20150069130 Gannon Mar 2015 A1
20150071819 Todeschini Mar 2015 A1
20150083800 Li et al. Mar 2015 A1
20150086114 Todeschini Mar 2015 A1
20150088522 Hendrickson et al. Mar 2015 A1
20150096872 Woodburn Apr 2015 A1
20150099557 Pettinelli et al. Apr 2015 A1
20150100196 Hollifield Apr 2015 A1
20150102109 Huck Apr 2015 A1
20150115035 Meier et al. Apr 2015 A1
20150127791 Kosecki et al. May 2015 A1
20150128116 Chen et al. May 2015 A1
20150129659 Feng et al. May 2015 A1
20150133047 Smith et al. May 2015 A1
20150134470 Hejl et al. May 2015 A1
20150136851 Harding et al. May 2015 A1
20150136854 Lu et al. May 2015 A1
20150142492 Kumar May 2015 A1
20150144692 Hejl May 2015 A1
20150144698 Teng et al. May 2015 A1
20150144701 Xian et al. May 2015 A1
20150149946 Benos et al. May 2015 A1
20150161429 Xian Jun 2015 A1
20150169925 Chang et al. Jun 2015 A1
20150169929 Williams et al. Jun 2015 A1
20150178523 Gelay et al. Jun 2015 A1
20150178534 Jovanovski et al. Jun 2015 A1
20150178535 Bremer et al. Jun 2015 A1
20150178536 Hennick et al. Jun 2015 A1
20150178537 El et al. Jun 2015 A1
20150181093 Zhu et al. Jun 2015 A1
20150181109 Gillet et al. Jun 2015 A1
20150186703 Chen et al. Jul 2015 A1
20150192774 Watanabe Jul 2015 A1
20150193644 Kearney et al. Jul 2015 A1
20150193645 Colavito et al. Jul 2015 A1
20150199957 Funyak et al. Jul 2015 A1
20150204671 Showering Jul 2015 A1
20150210199 Payne Jul 2015 A1
20150220753 Zhu et al. Aug 2015 A1
20150254485 Feng et al. Sep 2015 A1
20150327012 Bian et al. Nov 2015 A1
20160014251 Hejl Jan 2016 A1
20160040982 Li et al. Feb 2016 A1
20160042241 Todeschini Feb 2016 A1
20160050203 Hefetz Feb 2016 A1
20160057230 Todeschini et al. Feb 2016 A1
20160109219 Ackley et al. Apr 2016 A1
20160109220 Laffargue Apr 2016 A1
20160109224 Thuries et al. Apr 2016 A1
20160112631 Ackley et al. Apr 2016 A1
20160112643 Laffargue et al. Apr 2016 A1
20160124516 Schoon et al. May 2016 A1
20160125217 Todeschini May 2016 A1
20160125342 Miller et al. May 2016 A1
20160133253 Braho et al. May 2016 A1
20160171417 Sankaralingham Jun 2016 A1
20160171720 Todeschini Jun 2016 A1
20160176635 Varley Jun 2016 A1
20160178479 Goldsmith Jun 2016 A1
20160180678 Ackley et al. Jun 2016 A1
20160189087 Morton et al. Jun 2016 A1
20160125873 Braho et al. Jul 2016 A1
20160227912 Oberpriller et al. Aug 2016 A1
20160232891 Pecorari Aug 2016 A1
20160292477 Bidwell Oct 2016 A1
20160294779 Yeakley et al. Oct 2016 A1
20160306769 Kohtz et al. Oct 2016 A1
20160314276 Sewell et al. Oct 2016 A1
20160314294 Kubler et al. Oct 2016 A1
Foreign Referenced Citations (4)
Number Date Country
2013163789 Nov 2013 WO
2013173985 Nov 2013 WO
2014019130 Feb 2014 WO
2014110495 Jul 2014 WO
Non-Patent Literature Citations (27)
Entry
U.S. Appl. No. 14/715,916 for Evaluating Image Values filed May 19, 2015 (Ackley); 60 pages.
U.S. Appl. No. 29/525,068 for Tablet Computer With Removable Scanning Device filed Apr. 27, 2015 (Schulte et al.); 19 pages.
U.S. Appl. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.); 44 pages.
U.S. Appl. No. 29/530,600 for Cyclone filed Jun. 18, 2015 (Vargo et al); 16 pages.
U.S. Appl. No. 14/707,123 for Application Independent DEX/UCS Interface filed May 8, 2015 (Pape); 47 pages.
U.S. Appl. No. 14/283,282 for Terminal Having Illumination and Focus Control filed May 21, 2014 (Liu et al.); 31 pages; now abandoned.
U.S. Appl. No. 14/705,407 for Method and System to Protect Software-Based Network-Connected Devices From Advanced Persistent Threat filed May 6, 2015 (Hussey et al.); 42 pages.
U.S. Appl. No. 14/704,050 for Intermediate Linear Positioning filed May 5, 2015 (Charpentier et al.); 60 pages.
U.S. Appl. No. 14/705,012 for Hands-Free Human Machine Interface Responsive to a Driver of a Vehicle filed May 6, 2015 (Fitch et al.); 44 pages.
U.S. Appl. No. 14/715,672 for Augumented Reality Enabled Hazard Display filed May 19, 2015 (Venkatesha et al.); 35 pages.
U.S. Appl. No. 14/735,717 for Indicia-Reading Systems Having an Interface With a User's Nervous System filed Jun. 10, 2015 (Todeschini); 39 pages.
U.S. Appl. No. 14/702,110 for System and Method for Regulating Barcode Data Injection Into a Running Application on a Smart Device filed May 1, 2015 (Todeschini et al.); 38 pages.
U.S. Appl. No. 14/747,197 for Optical Pattern Projector filed Jun. 23, 2015 (Thuries et al.); 33 pages.
U.S. Appl. No. 14/702,979 for Tracking Battery Conditions filed May 4, 2015 (Young et al.); 70 pages.
U.S. Appl. No. 29/529,441 for Indicia Reading Device filed Jun. 8, 2015 (Zhou et al.); 14 pages.
U.S. Appl. No. 14/747,490 for Dual-Projector Three-Dimensional Scanner filed Jun. 23, 2015 (Jovanovski et al.); 40 pages.
U.S. Appl. No. 14/740,320 for Tactile Switch for a Mobile Electronic Device filed Jun. 16, 2015 (Barndringa); 38 pages.
U.S. Appl. No. 14/740,373 for Calibrating a Volume Dimensioner filed Jun. 16, 2015 (Ackley et al.); 63 pages.
U.S. Appl. No. 13/367,978, filed Feb. 7, 2012, (Feng et al.); now abandoned.
U.S. Appl. No. 14/277,337 for Multipurpose Optical Reader, filed May 14, 2014 (Jovanovski et al.); 59 pages; now abandoned.
U.S. Appl. No. 14/446,391 for Multifunction Point of Sale Apparatus With Optical Signature Capture filed Jul. 30, 2014 (Good et al.); 37 pages; now abandoned.
U.S. Appl. No. 29/516,892 for Table Computer filed Feb. 6, 2015 (Bidwell et al.); 13 pages.
U.S. Appl. No. 29/523,098 for Handle for a Tablet Computer filed Apr. 7, 2015 (Bidwell et al.); 17 pages.
U.S. Appl. No. 29/528,890 for Mobile Computer Housing filed Jun. 2, 2015 (Fitch et al.); 61 pages.
U.S. Appl. No. 29/526,918 for Charging Base filed May 14, 2015 (Fitch et al.); 10 pages.
U.S. Patent Application for a Laser Scanning Module Employing an Elastomeric U-Hinge Based Laser Scanning Assembly, filed Feb. 7, 2012 (Feng et al.), U.S. Appl. No. 13/367,978.
U.S. Patent Application for Indicia Reader filed Apr. 1, 2015 (Huck), U.S. Appl. No. 14/676,109.
Related Publications (1)
Number Date Country
20180039935 A1 Feb 2018 US