Dimensioning system with guided alignment

Information

  • Patent Grant
  • 9823059
  • Patent Number
    9,823,059
  • Date Filed
    Wednesday, August 6, 2014
    10 years ago
  • Date Issued
    Tuesday, November 21, 2017
    7 years ago
Abstract
A dimensioning system including a computing device running an alignment software program is disclosed. The alignment software uses range information from a range sensor in order to generate alignment messages. The alignment messages may help a user define a frame of reference and align the dimensioning system's range sensor for improved dimensioning performance.
Description
FIELD OF THE INVENTION

The present invention relates to the field of dimensioning systems, and more specifically, to a system and method for aligning a package dimensioning system.


BACKGROUND

Generally speaking freight carriers calculate shipping costs based on package size and weight (i.e., volumetric weight). This helps prevent lightweight packages that require a large amount of space from becoming unprofitable for the freight carriers.


When printing a shipping label for a package, a freight carrier employee is required to enter the package's size and weight into a software application that uses this information to calculate the cost of shipping. Typically, freight carrier employees derive this information through hand measurements (e.g., with a tape measure) and may weigh the package on a scale. Hand measurements are prone to error, particularly when packages have an irregular shape. These errors may lead to dissatisfaction and/or financial loss. For example, when a shipping company determines, after shipping costs are negotiated, that a package is larger and/or heavier than reported by the customer, additional costs may assessed. In addition, retailers that pass the shipping costs along to customers typically assume the extra shipping costs associated with these errors. As a result, automated dimensioning systems have been developed to bring more accuracy to package volume calculation.


One such automated dimensioning system uses a light projector to project a light pattern (e.g., point cloud) onto objects (e.g., packages) within a field of view. A range camera, physically offset from the light projector, creates a range image from the light pattern reflected from the packages. Software running on a computing device compares the light pattern in the range image to some reference (e.g., a reference image taken during calibration). Through this comparison, the dimensions of a package may be derived.


In order to dimension a specified size range (e.g., a range of package sizes) accurately, the dimensioning system may require a user to position (i.e., align) the range sensor into a particular pose (i.e., height and orientation). This positioning typically takes place during the installation of the dimensioning system. During positioning the pose is computed relative to a reference (i.e., ground) plane that is typically defined prior to positioning. The process of selecting the reference plane and positioning the range sensor is not easily handled by a typical user, but poorly installed range sensors may result in dimensioning errors or the inability to dimension. Therefore, a need exists for a method to assist the user with the selection of a reference plane and the positioning of a range sensor to ensure good performance of the dimensioning system.


SUMMARY

Accordingly, in one aspect, the present invention embraces a package dimensioning system including a range sensor for capturing a series of range images of the range sensor's field of view. The system also includes an adjustable range-sensor support to physically support and position the range sensor in a target pose. A computing device, communicatively coupled to the range sensor, is capable of executing an adjustment software program, which provides adjustment messages to facilitate the adjustment of the range sensor. The adjustment software program configures the computing device to receive the series of range images, process the series of range images to produce the adjustment messages, and transmit the adjustment messages to a display. The display is communicatively coupled to the computing device and displays the series of range images and the adjustment messages.


In another aspect, the present invention embraces a computer implemented method for generating adjustment messages to facilitate the positioning of a range sensor for dimensioning. The method includes the step of recording range images onto a computer-readable storage medium. The method also includes the step of reading the range images from the computer-readable storage medium. In addition, the method includes processing the range images to derive a result. The method further includes the step of generating adjustment messages based on the result.


The foregoing illustrative summary, as well as other exemplary objectives and/or advantages of the invention, and the manner in which the same are accomplished, are further explained within the following detailed description and its accompanying figure (i.e., fig.) set.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 schematically depicts an exemplary range sensor's stereo arrangement of a pattern projector of a range camera for capturing range information as a pixel displacement.



FIG. 2 graphically depicts an exemplary dimensioning system including a platform and an object for dimensioning.



FIG. 3 schematically depicts an exemplary dimensioning system.



FIG. 4 schematically depicts a flowchart of an exemplary adjustment software program.





DETAILED DESCRIPTION

The present invention embraces a dimensioning system to measure items, such as packages for shipment. The dimensioning system typically uses a range sensor, a computing device, and a display for this measurement. The range sensor may be an optical system that acquires information about a field of view or could use another range sensing modality (e.g., ultrasonic). The optical system requires a user to place an item within the system's field of view for measurement and is very easy to use. For the system to give best results, however, the range sensor should be positioned in a range-sensor pose with respect to a frame of reference shared by the package that closely matches a target pose (e.g., below an adjustable threshold value of pose difference). The target pose represents the desired range sensor's position (e.g., height, pitch, roll, and/or yaw) with respect to the frame of reference that ensures good dimensioning performance. The target pose is chosen to allow the range sensor to obtain accurate dimensioning results for a given range of package sizes. For example, a target pose may be established to contain the largest package to be dimensioned, while also ensuring that the smallest package may be resolved sufficiently for accurate measurement. The target pose is also chosen based on the range sensor's resolution, field of view, and/or other limitations (e.g., specular reflections, multipath interference, and/or mixed pixel responses). Similarly, a target pose may be chosen to minimize depth differences between a foreground package and clutter (e.g., background clutter). The target pose may be stored on a computer-readable medium (e.g., non-transitory memory) communicatively coupled to the computing device and is typically set once for a particular application. In certain scenarios, however, this target pose could adjustable. For example, an old target pose could be replaced with a new target pose, or in another embodiment, target poses could be selected by a user to match a particular dimensioning application. For example, a user might want to replace the range sensor with a new range sensor that has a different, sensing modality, resolution, and/or field of view. Here the user could update the target pose to match the new range sensor. Alternatively, the user might want to accommodate a new range of package sizes. Here the user could update the target pose to meet the requirements of the new range of package sizes.


The mathematical representation of a physical pose requires a frame of reference. This frame of reference may be defined with a specified surface (e.g., planar surface) or a line. Alternatively, the frame of reference may be established with a set of 3D points that are arranged in some known way. For example, a pose may be calculated relative to a cylinder placed in front of the camera. Typically, however, a planar surface within the field of view, selected by a user, establishes the frame of reference. For example, the surface that the measured item (e.g., package) rests on during the measurement (e.g., scale or counter-top) may be chosen to serve as the reference surface (i.e., platform). In this way, the orientations of the range-sensor pose and the target pose (each relative to the frame of reference) can be derived mathematically and compared (e.g., compared by rotation matrix or axis-angle representation). For example, rotation vectors, with respect to the frame of reference, could be derived for each pose using Rodrigues' rotation formula. The angle between the two rotation vectors could then be computed and compared to obtain the difference between the range-sensor pose and the target pose (i.e., pose difference).


The process of establishing the platform and aligning the range-sensor pose with a target pose may be made easier through the use of software (e.g., one or more executable files, libraries, and/or scripts) to generate guidance advice for sensor alignment. Here, the adjustment software (i.e., adjustment software program) receives range images from the range sensor and produces feedback (i.e., adjustment messages) to help a user align the range sensor.


Range images are typically single-channel (e.g., gray scale) images that represent the distance between the range camera and the portion of the field of view represented by a pixel. Using these range images, the adjustment software may detect planar surfaces using an algorithm. For example, a random sample consensus (i.e., RANSAC) algorithm may identify planar surfaces within the range sensor's field of view. In the case where more than one planar surface is detected within the range camera's field of view, each planar surface may be indicated in a visual image presented on a display. In one possible embodiment, each planar surface may be indicated by an overlay (e.g., semi-transparent and/or colored overlay) image superimposed on the range image and presented on the display. In other possible embodiments, the reference surface may otherwise highlighted (e.g., an outline). A prompt, generated by the adjustment software, may query a user to indicate which planar surface should be established as the reference surface (i.e., ground plane). Once the ground plane (i.e., platform) is established, the adjustment software may use the range image to compute the range camera's height and orientation (i.e., roll, pitch, and/or yaw) with respect to the ground plane.


During range sensor alignment (e.g., during installation of a package dimensioning system) the adjustment software may use the computed range sensor height and orientation to provide adjustment messages. These adjustment messages include indications of the necessary adjustments in order to align the range sensor's physical pose with the target pose. This feedback may be audible or visual. Visible messages could be text messages or graphical images displayed alone, in addition to, and/or superimposed on other images (e.g., range images, color images, or point-cloud images). By following these adjustment messages, a user may adjust the range-sensor support (e.g., adjustable tripod mount, pole mount, ceiling mount, and/or wall mount) to move the range sensor closer to the target pose. Many adjustment messages may be generated during the alignment process. In one possible embodiment, the process of analyzing range images and providing alignment messages (e.g., “move camera up”) may continue iteratively until the range sensor is aligned with the target pose. Once aligned, the adjustment software may provide an adjustment message indicating that the range sensor is in position, indicating that the user should stop adjusting and secure the support. In another possible embodiment, the software may provide alignment messages that indicate the alignment of the range-sensor pose with the target pose in qualitative terms (e.g., good, better, or best). In still another possible embodiment, the software may simply provide real-time alignment information (e.g., pose difference results in numerical form) and allow the user to decide the ultimate alignment criteria.


While the adjustment software is typically used during the installation of the package dimensioning system, it may also be used periodically after the installation. For example, the adjustment software program may be configured to periodically check the range sensor's pose and compare this with the target pose. If the difference between the two poses (i.e., pose difference) is above some threshold value (e.g., from a misalignment caused by mechanical movement or vibration), the guidance software may provide messages to alert a user that the range sensor is no longer in alignment and that an adjustment is necessary.


The alignment process may happen in real-time with the display rendering real-time range images, while simultaneously displaying adjustment messages. In one embodiment, for example, the adjustment software may guide the user to first adjust range sensor's height and then adjust its orientation (i.e., roll, pitch, and/or yaw) separately and sequentially. In another embodiment, the software may accommodate a user to jointly adjust the range sensor's height and orientation simultaneously.


Three-dimensional (i.e., 3D) sensors (e.g., range sensors) can be utilized effectively in dimensioning applications. The recent advent of relatively low-cost range sensors that can detect and display three-dimensional information has afforded greater opportunity for implementing automated dimensioning on a wider scale. Consequently, the package-dimensioning system disclosed here may include a range sensor to acquire a two-dimensional gray scale image for conveying the range on a pixel by pixel basis (i.e., range image). In an exemplary range image, darker pixels may indicate a point that is a shorter distance away from the range sensor than points represented by lighter pixels.


In the embodiment shown in FIG. 1, the range sensor includes a projector 1 and a range camera 2. The projector 1 may radiate a light pattern onto an item 4 within a field of view 3. The reflected light pattern 6 from the item may be imaged and detected by the range camera 2. If the item's range 7 is changed then the range camera may sense this change as a displacement 5 in detected light pattern. A processor within the range sensor may convert this range information into a range image. In this way the pattern projector and range camera may together help to produce a range image. The projector 1 and the range camera 2 are positioned collinearly and are codirected towards the same field of view (i.e., are positioned in a stereo arrangement). The light from the projector (e.g., the point cloud) may be visible but is typically invisible to the human eye. The range camera is sensitive to this light. In some embodiments, the range sensor also includes a color (i.e., RGB) camera that is sensitive to visible light and which shares the field of view 3 with the projector and range camera. This color camera may be used to display images for a user that are easily interpreted and less confusing than the gray scale range images or point cloud images. In another possible embodiment, the range images may be displayed during the alignment process. In yet another possible embodiment, the raw images including the projected light pattern (point cloud images) may be displayed during the alignment process.


An exemplary package dimensioning system is graphically shown in FIG. 2. Here the range sensor 10 is physically supported and positioned by the range-sensor support 11. The support helps configure the range-sensor pose which is defined by the range sensor's height 12 and orientation. The orientation may include the range sensor's pitch 13, yaw 14, and/or roll 15. A user may configure the range-sensor pose to match a target pose through the use of an adjustable range-sensor support 11. This support is shown in this embodiment as a tripod, though other support mechanisms (e.g., pole-mount, wall-mount, or ceiling-mount) may be used. An adjustment software program executed by a computing device 16 may display adjustment message on a display 13. The range-sensor pose and the target pose are relative to a platform 18 (i.e., reference plane or ground plane) that a package 17 is placed on for dimensioning. The platform may be selected by a user from a plurality of planar surfaces detected within the field of view before the poses are calculated.


The schematic of the package dimensioning system including a computing device 24 for package dimensioning is shown in FIG. 3. Here a range sensor 20 includes a pattern projector 27 for creating a light pattern that can be imaged by the range camera 22 and mathematically transformed into a range image that is transmitted from the range sensor 20 to a processor 23 integrated in the computing device 24 and communicatively coupled to the range sensor. The processor may store the range image in a computer-readable storage medium 25. Adjustment software stored in the storage medium 25 may configure the processor 23 to execute the program steps required for generating the adjustment messages necessary to facilitate the positioning of the range sensor 20 for dimensioning. The processor may transmit the adjustment messages to a display 26. These messages may be displayed along with an image of the range sensor's field of view. This image of the field of view may be the range camera's 22 image or may be a color image created by a color camera 21 configured with the same field of view as the range camera 22. The color camera is optional but may enhance the user's experience, as these images may be easier to understand than the gray scale range images.


As shown in FIG. 4, the adjustment software program 36 operates on range information (e.g., range images 31) to produce some feedback information (e.g., adjustment messages 39). A flowchart illustrating the method for generating adjustment messages to facilitate the positioning of a range sensor for package dimensioning is shown in FIG. 4. The range sensor 30 produces a range image 31. The software analyzes the range image to detect planar surfaces within the field of view. A user may then be prompted to select a reference plane (i.e., platform) from the detected planar surfaces. Alternatively, the software may detect and select a platform automatically. Once the platform 32 is detected, a target pose 33 may be computed based on a stored pose 34. The stored pose information may be information based on the range camera's field of view, the range of expected package sizes, and/or the resolution of the range image. This information may be stored in the computing device's non-transitory, computer-readable storage medium (e.g., hard drive). The platform 32 may also be used to mathematically compute the range-sensor pose 35 from the range image 31 and the platform 32. Mathematically a pose may be defined as a vector relative to the platform 32. The adjustment software program 36 then computes the difference between the target pose 33 and the range-sensor pose 35 to determine a pose difference 37 (e.g., vector difference). If the pose difference is zero (or below some threshold value) then the camera is considered aligned, however if the pose difference is above a threshold value, then a desired action 38 to minimize the pose difference is computed. Based on the desired action 38, an adjustment message 39 is created. This adjustment message is then transmitted with the range camera's image to the display 40 for viewing. The adjustment message could be a text message or a graphical image. In one possible embodiment an arrow graphic indicating the direction to move the range sensor 30 could be overlaid with the range image 31 on the display 40. In another embodiment the adjustment message could provide quantitative measurements (e.g., move camera up 10 cm). In another possible embodiment the adjustment messages may be audio messages transmitted to a speaker for broadcast.


To supplement the present disclosure, this application incorporates entirely by reference the following commonly assigned patents, patent application publications, and patent applications:














U.S. Pat. No. 6,832,725; U.S. Pat. No. 7,128,266;


U.S. Pat. No. 7,159,783; U.S. Pat. No. 7,413,127;


U.S. Pat. No. 7,726,575; U.S. Pat. No. 8,294,969;


U.S. Pat. No. 8,317,105; U.S. Pat. No. 8,322,622;


U.S. Pat. No. 8,366,005; U.S. Pat. No. 8,371,507;


U.S. Pat. No. 8,376,233; U.S. Pat. No. 8,381,979;


U.S. Pat. No. 8,390,909; U.S. Pat. No. 8,408,464;


U.S. Pat. No. 8,408,468; U.S. Pat. No. 8,408,469;


U.S. Pat. No. 8,424,768; U.S. Pat. No. 8,448,863;


U.S. Pat. No. 8,457,013; U.S. Pat. No. 8,459,557;


U.S. Pat. No. 8,469,272; U.S. Pat. No. 8,474,712;


U.S. Pat. No. 8,479,992; U.S. Pat. No. 8,490,877;


U.S. Pat. No. 8,517,271; U.S. Pat. No. 8,523,076;


U.S. Pat. No. 8,528,818; U.S. Pat. No. 8,544,737;


U.S. Pat. No. 8,548,242; U.S. Pat. No. 8,548,420;


U.S. Pat. No. 8,550,335; U.S. Pat. No. 8,550,354;


U.S. Pat. No. 8,550,357; U.S. Pat. No. 8,556,174;


U.S. Pat. No. 8,556,176; U.S. Pat. No. 8,556,177;


U.S. Pat. No. 8,559,767; U.S. Pat. No. 8,599,957;


U.S. Pat. No. 8,561,895; U.S. Pat. No. 8,561,903;


U.S. Pat. No. 8,561,905; U.S. Pat. No. 8,565,107;


U.S. Pat. No. 8,571,307; U.S. Pat. No. 8,579,200;


U.S. Pat. No. 8,583,924; U.S. Pat. No. 8,584,945;


U.S. Pat. No. 8,587,595; U.S. Pat. No. 8,587,697;


U.S. Pat. No. 8,588,869; U.S. Pat. No. 8,590,789;


U.S. Pat. No. 8,596,539; U.S. Pat. No. 8,596,542;


U.S. Pat. No. 8,596,543; U.S. Pat. No. 8,599,271;


U.S. Pat. No. 8,599,957; U.S. Pat. No. 8,600,158;


U.S. Pat. No. 8,600,167; U.S. Pat. No. 8,602,309;


U.S. Pat. No. 8,608,053; U.S. Pat. No. 8,608,071;


U.S. Pat. No. 8,611,309; U.S. Pat. No. 8,615,487;


U.S. Pat. No. 8,616,454; U.S. Pat. No. 8,621,123;


U.S. Pat. No. 8,622,303; U.S. Pat. No. 8,628,013;


U.S. Pat. No. 8,628,015; U.S. Pat. No. 8,628,016;


U.S. Pat. No. 8,629,926; U.S. Pat. No. 8,630,491;


U.S. Pat. No. 8,635,309; U.S. Pat. No. 8,636,200;


U.S. Pat. No. 8,636,212; U.S. Pat. No. 8,636,215;


U.S. Pat. No. 8,636,224; U.S. Pat. No. 8,638,806;


U.S. Pat. No. 8,640,958; U.S. Pat. No. 8,640,960;


U.S. Pat. No. 8,643,717; U.S. Pat. No. 8,646,692;


U.S. Pat. No. 8,646,694; U.S. Pat. No. 8,657,200;


U.S. Pat. No. 8,659,397; U.S. Pat. No. 8,668,149;


U.S. Pat. No. 8,678,285; U.S. Pat. No. 8,678,286;


U.S. Pat. No. 8,682,077; U.S. Pat. No. 8,687,282;


U.S. Pat. No. 8,692,927; U.S. Pat. No. 8,695,880;


U.S. Pat. No. 8,698,949; U.S. Pat. No. 8,717,494;


U.S. Pat. No. 8,717,494; U.S. Pat. No. 8,720,783;


U.S. Pat. No. 8,723,804; U.S. Pat. No. 8,723,904;


U.S. Pat. No. 8,727,223; U.S. Pat. No. D702,237;


International Publication No. 2013/163789;


International Publication No. 2013/173985;


International Publication No. 2014/019130;


U.S. patent application Publication No. 2008/0185432;


U.S. patent application Publication No. 2009/0134221;


U.S. patent application Publication No. 2010/0177080;


U.S. patent application Publication No. 2010/0177076;


U.S. patent application Publication No. 2010/0177707;


U.S. patent application Publication No. 2010/0177749;


U.S. patent application Publication No. 2011/0202554;


U.S. patent application Publication No. 2012/0111946;


U.S. patent application Publication No. 2012/0138685;


U.S. patent application Publication No. 2012/0168511;


U.S. patent application Publication No. 2012/0168512;


U.S. patent application Publication No. 2012/0193407;


U.S. patent application Publication No. 2012/0193423;


U.S. patent application Publication No. 2012/0203647;


U.S. patent application Publication No. 2012/0223141;


U.S. patent application Publication No. 2012/0228382;


U.S. patent application Publication No. 2012/0248188;


U.S. patent application Publication No. 2013/0043312;


U.S. patent application Publication No. 2013/0056285;


U.S. patent application Publication No. 2013/0070322;


U.S. patent application Publication No. 2013/0075168;


U.S. patent application Publication No. 2013/0082104;


U.S. patent application Publication No. 2013/0175341;


U.S. patent application Publication No. 2013/0175343;


U.S. patent application Publication No. 2013/0200158;


U.S. patent application Publication No. 2013/0214048;


U.S. patent application Publication No. 2013/0256418;


U.S. patent application Publication No. 2013/0257744;


U.S. patent application Publication No. 2013/0257759;


U.S. patent application Publication No. 2013/0270346;


U.S. patent application Publication No. 2013/0278425;


U.S. patent application Publication No. 2013/0287258;


U.S. patent application Publication No. 2013/0292474;


U.S. patent application Publication No. 2013/0292475;


U.S. patent application Publication No. 2013/0292477;


U.S. patent application Publication No. 2013/0293539;


U.S. patent application Publication No. 2013/0293540;


U.S. patent application Publication No. 2013/0306728;


U.S. patent application Publication No. 2013/0306730;


U.S. patent application Publication No. 2013/0306731;


U.S. patent application Publication No. 2013/0306734;


U.S. patent application Publication No. 2013/0307964;


U.S. patent application Publication No. 2013/0308625;


U.S. patent application Publication No. 2013/0313324;


U.S. patent application Publication No. 2013/0313325;


U.S. patent application Publication No. 2013/0313326;


U.S. patent application Publication No. 2013/0327834;


U.S. patent application Publication No. 2013/0341399;


U.S. patent application Publication No. 2013/0342717;


U.S. patent application Publication No. 2014/0001267;


U.S. patent application Publication No. 2014/0002828;


U.S. patent application Publication No. 2014/0008430;


U.S. patent application Publication No. 2014/0008439;


U.S. patent application Publication No. 2014/0021256;


U.S. patent application Publication No. 2014/0025584;


U.S. patent application Publication No. 2014/0027518;


U.S. patent application Publication No. 2014/0034723;


U.S. patent application Publication No. 2014/0034734;


U.S. patent application Publication No. 2014/0036848;


U.S. patent application Publication No. 2014/0039693;


U.S. patent application Publication No. 2014/0042814;


U.S. patent application Publication No. 2014/0049120;


U.S. patent application Publication No. 2014/0049635;


U.S. patent application Publication No. 2014/0061305;


U.S. patent application Publication No. 2014/0061306;


U.S. patent application Publication No. 2014/0061307;


U.S. patent application Publication No. 2014/0063289;


U.S. patent application Publication No. 2014/0066136;


U.S. patent application Publication No. 2014/0067692;


U.S. patent application Publication No. 2014/0070005;


U.S. patent application Publication No. 2014/0071840;


U.S. patent application Publication No. 2014/0074746;


U.S. patent application Publication No. 2014/0075846;


U.S. patent application Publication No. 2014/0076974;


U.S. patent application Publication No. 2014/0078341;


U.S. patent application Publication No. 2014/0078342;


U.S. patent application Publication No. 2014/0078345;


U.S. patent application Publication No. 2014/0084068;


U.S. patent application Publication No. 2014/0086348;


U.S. patent application Publication No. 2014/0097249;


U.S. patent application Publication No. 2014/0098284;


U.S. patent application Publication No. 2014/0098792;


U.S. patent application Publication No. 2014/0100774;


U.S. patent application Publication No. 2014/0100813;


U.S. patent application Publication No. 2014/0103115;


U.S. patent application Publication No. 2014/0104413;


U.S. patent application Publication No. 2014/0104414;


U.S. patent application Publication No. 2014/0104416;


U.S. patent application Publication No. 2014/0104451;


U.S. patent application Publication No. 2014/0106594;


U.S. patent application Publication No. 2014/0106725;


U.S. patent application Publication No. 2014/0108010;


U.S. patent application Publication No. 2014/0108402;


U.S. patent application Publication No. 2014/0108682;


U.S. patent application Publication No. 2014/0110485;


U.S. patent application Publication No. 2014/0114530;


U.S. patent application Publication No. 2014/0124577;


U.S. patent application Publication No. 2014/0124579;


U.S. patent application Publication No. 2014/0125842;


U.S. patent application Publication No. 2014/0125853;


U.S. patent application Publication No. 2014/0125999;


U.S. patent application Publication No. 2014/0129378;









  • U.S. patent application Ser. No. 13/367,978 for a Laser Scanning Module Employing An Elastomeric U-Hinge Based Laser Scanning Assembly, filed Feb. 7, 2012 (Feng et al.);

  • U.S. patent application Ser. No. 29/436,337 for an Electronic Device, filed Nov. 5, 2012 (Fitch et al.);

  • U.S. patent application Ser. No. 13/736,139 for an Electronic Device Enclosure, filed Jan. 8, 2013 (Chaney);

  • U.S. patent application Ser. No. 13/771,508 for an Optical Redirection Adapter, filed Feb. 20, 2013 (Anderson);

  • U.S. patent application Ser. No. 13/780,356 for a Mobile Device Having Object-Identification Interface, filed Feb. 28, 2013 (Samek et al.);

  • U.S. patent application Ser. No. 13/852,097 for a System and Method for Capturing and Preserving Vehicle Event Data, filed Mar. 28, 2013 (Barker et al.);

  • U.S. patent application Ser. No. 13/902,110 for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Hollifield);

  • U.S. patent application Ser. No. 13/902,144, for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Chamberlin);

  • U.S. patent application Ser. No. 13/902,242 for a System For Providing A Continuous Communication Link With A Symbol Reading Device, filed May 24, 2013 (Smith et al.);

  • U.S. patent application Ser. No. 13/912,262 for a Method of Error Correction for 3D Imaging Device, filed Jun. 7, 2013 (Jovanovski et al.);

  • U.S. patent application Ser. No. 13/912,702 for a System and Method for Reading Code Symbols at Long Range Using Source Power Control, filed Jun. 7, 2013 (Xian et al.);

  • U.S. patent application Ser. No. 29/458,405 for an Electronic Device, filed Jun. 19, 2013 (Fitch et al.);

  • U.S. patent application Ser. No. 13/922,339 for a System and Method for Reading Code Symbols Using a Variable Field of View, filed Jun. 20, 2013 (Xian et al.);

  • U.S. patent application Ser. No. 13/927,398 for a Code Symbol Reading System Having Adaptive Autofocus, filed Jun. 26, 2013 (Todeschini);

  • U.S. patent application Ser. No. 13/930,913 for a Mobile Device Having an Improved User Interface for Reading Code Symbols, filed Jun. 28, 2013 (Gelay et al.);

  • U.S. patent application Ser. No. 29/459,620 for an Electronic Device Enclosure, filed Jul. 2, 2013 (London et al.);

  • U.S. patent application Ser. No. 29/459,681 for an Electronic Device Enclosure, filed Jul. 2, 2013 (Chaney et al.);

  • U.S. patent application Ser. No. 13/933,415 for an Electronic Device Case, filed Jul. 2, 2013 (London et al.);

  • U.S. patent application Ser. No. 29/459,785 for a Scanner and Charging Base, filed Jul. 3, 2013 (Fitch et al.);

  • U.S. patent application Ser. No. 29/459,823 for a Scanner, filed Jul. 3, 2013 (Zhou et al.);

  • U.S. patent application Ser. No. 13/947,296 for a System and Method for Selectively Reading Code Symbols, filed Jul. 22, 2013 (Rueblinger et al.);

  • U.S. patent application Ser. No. 13/950,544 for a Code Symbol Reading System Having Adjustable Object Detection, filed Jul. 25, 2013 (Jiang);

  • U.S. patent application Ser. No. 13/961,408 for a Method for Manufacturing Laser Scanners, filed Aug. 7, 2013 (Saber et al.);

  • U.S. patent application Ser. No. 14/018,729 for a Method for Operating a Laser Scanner, filed Sep. 5, 2013 (Feng et al.);

  • U.S. patent application Ser. No. 14/019,616 for a Device Having Light Source to Reduce Surface Pathogens, filed Sep. 6, 2013 (Todeschini);

  • U.S. patent application Ser. No. 14/023,762 for a Handheld Indicia Reader Having Locking Endcap, filed Sep. 11, 2013 (Gannon);

  • U.S. patent application Ser. No. 14/035,474 for Augmented-Reality Signature Capture, filed Sep. 24, 2013 (Todeschini);

  • U.S. patent application Ser. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.);

  • U.S. patent application Ser. No. 14/047,896 for Terminal Having Illumination and Exposure Control filed Oct. 7, 2013 (Jovanovski et al.);

  • U.S. patent application Ser. No. 14/053,175 for Imaging Apparatus Having Imaging Assembly, filed Oct. 14, 2013 (Barber);

  • U.S. patent application Ser. No. 14/055,234 for Dimensioning System, filed Oct. 16, 2013 (Fletcher);

  • U.S. patent application Ser. No. 14/053,314 for Indicia Reader, filed Oct. 14, 2013 (Huck);

  • U.S. patent application Ser. No. 14/065,768 for Hybrid System and Method for Reading Indicia, filed Oct. 29, 2013 (Meier et al.);

  • U.S. patent application Ser. No. 14/074,746 for Self-Checkout Shopping System, filed Nov. 8, 2013 (Hejl et al.);

  • U.S. patent application Ser. No. 14/074,787 for Method and System for Configuring Mobile Devices via NFC Technology, filed Nov. 8, 2013 (Smith et al.);

  • U.S. patent application Ser. No. 14/087,190 for Optimal Range Indicators for Bar Code Validation, filed Nov. 22, 2013 (Hejl);

  • U.S. patent application Ser. No. 14/094,087 for Method and System for Communicating Information in an Digital Signal, filed Dec. 2, 2013 (Peake et al.);

  • U.S. patent application Ser. No. 14/101,965 for High Dynamic-Range Indicia Reading System, filed Dec. 10, 2013 (Xian);

  • U.S. patent application Ser. No. 14/118,400 for Indicia Decoding Device with Security Lock, filed Nov. 18, 2013 (Liu);

  • U.S. patent application Ser. No. 14/150,393 for Indicia-reader Having Unitary Construction Scanner, filed Jan. 8, 2014 (Colavito et al.);

  • U.S. patent application Ser. No. 14/154,207 for Laser Barcode Scanner, filed Jan. 14, 2014 (Hou et al.);

  • U.S. patent application Ser. No. 14/154,915 for Laser Scanning Module Employing a Laser Scanning Assembly having Elastomeric Wheel Hinges, filed Jan. 14, 2014 (Havens et al.);

  • U.S. patent application Ser. No. 14/158,126 for Methods and Apparatus to Change a Feature Set on Data Collection Devices, filed Jan. 17, 2014 (Berthiaume et al.);

  • U.S. patent application Ser. No. 14/159,074 for Wireless Mesh Point Portable Data Terminal, filed Jan. 20, 2014 (Wang et al.);

  • U.S. patent application Ser. No. 14/159,509 for MMS Text Messaging for Hand Held Indicia Reader, filed Jan. 21, 2014 (Kearney);

  • U.S. patent application Ser. No. 14/159,603 for Decodable Indicia Reading Terminal with Optical Filter, filed Jan. 21, 2014 (Ding et al.);

  • U.S. patent application Ser. No. 14/160,645 for Decodable Indicia Reading Terminal with Indicia Analysis Functionality, filed Jan. 22, 2014 (Nahill et al.);

  • U.S. patent application Ser. No. 14/161,875 for System and Method to Automatically Discriminate Between Different Data Types, filed Jan. 23, 2014 (Wang);

  • U.S. patent application Ser. No. 14/165,980 for System and Method for Measuring Irregular Objects with a Single Camera filed Jan. 28, 2014 (Li et al.);

  • U.S. patent application Ser. No. 14/166,103 for Indicia Reading Terminal Including Optical Filter filed Jan. 28, 2014 (Lu et al.);

  • U.S. patent application Ser. No. 14/176,417 for Devices and Methods Employing Dual Target Auto Exposure filed Feb. 10, 2014 (Meier et al.);

  • U.S. patent application Ser. No. 14/187,485 for Indicia Reading Terminal with Color Frame Processing filed Feb. 24, 2014 (Ren et al.);

  • U.S. patent application Ser. No. 14/200,405 for Indicia Reader for Size-Limited Applications filed Mar. 7, 2014 (Feng et al.);

  • U.S. patent application Ser. No. 14/222,994 for Method and Apparatus for Reading Optical Indicia Using a Plurality of Data filed Mar. 24, 2014 (Smith et al.);

  • U.S. patent application Ser. No. 14/230,322 for Focus Module and Components with Actuator filed Mar. 31, 2014 (Feng et al.);

  • U.S. patent application Ser. No. 14/231,898 for Hand-Mounted Indicia-Reading Device with Finger Motion Triggering filed Apr. 1, 2014 (Van Horn et al.);

  • U.S. patent application Ser. No. 14/249,497 for Terminal Having Plurality of Operating Modes filed Apr. 10, 2014, Grunow et al.);

  • U.S. patent application Ser. No. 14/250,923 for Reading Apparatus Having Partial Frame Operating Mode filed Apr. 11, 2014, (Deng et al.);

  • U.S. patent application Ser. No. 14/257,174 for Imaging Terminal Having Data Compression filed Apr. 21, 2014, (Barber et al.)

  • U.S. patent application Ser. No. 14/257,364 for Docking System and Method Using Near Field Communication filed Apr. 21, 2014, (Showering);

  • U.S. patent application Ser. No. 14/264,173 for Autofocus Lens System for Indicia Readers filed Apr. 29, 2014, (Ackley et al.);

  • U.S. patent application Ser. No. 14/274,858 for Mobile Printer with Optional Battery Accessory filed May 12, 2014, (Marty et al.);

  • U.S. patent application Ser. No. 14/342,544 for Imaging Based Barcode Scanner Engine with Multiple Elements Supported on a Common Printed Circuit Board filed Mar. 4, 2014 (Liu et al.);

  • U.S. patent application Ser. No. 14/342,551 for Terminal Having Image Data Format Conversion filed Mar. 4, 2014 (Liu et al.);

  • U.S. patent application Ser. No. 14/345,735 for Optical Indicia Reading Terminal with Combined Illumination filed Mar. 19, 2014 (Ouyang);

  • U.S. patent application Ser. No. 29/486,759 for an Imaging Terminal, filed Apr. 2, 2014 (Oberpriller et al.); and

  • U.S. patent application Ser. No. 14/355,613 for Optical Indicia Reading Terminal with Color Image Sensor filed May 1, 2014, (Lu et al.).



In the specification and/or figures, typical embodiments of the invention have been disclosed. The present invention is not limited to such exemplary embodiments. The use of the term “and/or” includes any and all combinations of one or more of the associated listed items. The figures are schematic representations and so are not necessarily drawn to scale. Unless otherwise noted, specific terms have been used in a generic and descriptive sense and not for purposes of limitation.

Claims
  • 1. A dimensioning system comprising: a range sensor for capturing range images of a field of view;an adjustable range-sensor support to physically support and position the range sensor in a range-sensor pose;a computing device communicatively coupled to the range sensor, the computing device capable of executing adjustment software and dimensioning software; anda display communicatively coupled to the computing device;wherein said adjustment software generates adjustment messages to facilitate adjustment of the range-sensor pose, the adjustment software configuring the computing device to: (i) receive the range images from the range sensor, (ii) compute, using at least one of the range images, a pose difference, wherein the pose difference comprises a difference between an initial range-sensor pose and a target range-sensor pose, (iii) generate, based on the pose difference, at least one adjustment message to facilitate adjustment of the range sensor pose to more closely align with the target range-sensor pose, and (iv) transmit the at least one adjustment message to the display-communicatively coupled to the computing device, andwherein the dimensioning software configures the computing device to derive, from at least one of the range images, dimensions of an object in the field of view.
  • 2. The dimensioning system according to claim 1, wherein the adjustment software generates adjustment messages until the pose difference is minimized below an adjustable threshold value.
  • 3. The dimensioning system according to claim 1, wherein the range sensor comprises a stereo arrangement of (i) a pattern projector for projecting a light pattern within a field of view and (ii) a range camera for capturing images of the reflected light pattern.
  • 4. The dimensioning system according to claim 1, wherein the light pattern is invisible.
  • 5. The dimensioning system according to claim 1, wherein the range sensor comprises a color camera for capturing color images of visible light within the field of view.
  • 6. The dimensioning system according to claim 1, wherein the adjustment software further configures the computing device to detect planar surfaces in the range images.
  • 7. The dimensioning system according to claim 6, wherein the adjustment messages comprise indications for a user to select a platform from the detected planar surfaces.
  • 8. The dimensioning system according to claim 1, wherein the target range-sensor pose is stored on a non-transitory computer-readable storage medium communicatively coupled with the computing device.
  • 9. The dimensioning system according to claim 1, wherein the target range-sensor pose is adjustable.
  • 10. A method, implemented by a computer, for operating a dimensioning system, the method comprising: generating adjustment messages to facilitate the positioning of a range sensor of the dimensioning system, the range sensor communicatively coupled to the computer and capable of generating range images, the generating adjustment messages comprising:recording the range images from the range sensor onto a non-transitory computer-readable storage medium;reading the range images from the computer-readable storage medium;processing at least one of the range images to derive a result, comprising computing a difference between an initial range-sensor pose and a target range-sensor pose; andgenerating at least one adjustment message based on the result, the at least one adjustment message being generated to facilitate adjustment of the range sensor to more closely align with the target range-sensor pose; andderiving, from at least one of the range images, dimensions of an object in the field of view.
  • 11. The method according to claim 10, wherein the step of processing comprises identifying at least one planar surface in at least one range image.
  • 12. The method according to claim 11, wherein the step of processing comprises computing the initial range-sensor pose and the target range-sensor pose.
  • 13. The method according to claim 12, wherein the adjustment messages comprise instructions for physically adjusting the range sensor in order to minimize the difference the between the initial range-sensor pose and the target range-sensor pose.
  • 14. The method according to claim 10, wherein the range sensor comprises a projector for projecting a light pattern in a field of view and a range camera for detecting the light pattern in at least part of the field of view, wherein the projector and the range camera are in a stereo arrangement.
  • 15. The method according to claim 10, wherein the adjustment messages comprise range images with superimposed graphics and/or text.
  • 16. The method according to claim 10, wherein the range sensor comprises a color camera.
  • 17. The method according to claim 16, wherein the adjustment messages comprise color images with superimposed graphics and/or text.
  • 18. The method according to claim 10, wherein the adjustment messages comprise audio messages.
  • 19. The method according to claim 10, wherein the adjustment messages comprise visual information displayed on a display communicatively coupled to the computer.
US Referenced Citations (616)
Number Name Date Kind
3971065 Bayer Jul 1976 A
4279328 Ahlbom Jul 1981 A
4398811 Nishioka et al. Aug 1983 A
4495559 Gelatt, Jr. Jan 1985 A
4730190 Win et al. Mar 1988 A
4803639 Steele et al. Feb 1989 A
5184733 Arnarson et al. Feb 1993 A
5220536 Stringer et al. Jun 1993 A
5331118 Jensen Jul 1994 A
5359185 Hanson Oct 1994 A
5384901 Glassner et al. Jan 1995 A
5548707 LoNegro Aug 1996 A
5555090 Schmutz Sep 1996 A
5561526 Huber et al. Oct 1996 A
5590060 Granville et al. Dec 1996 A
5606534 Stringer et al. Feb 1997 A
5619245 Kessler et al. Apr 1997 A
5655095 LoNegro et al. Aug 1997 A
5661561 Wurz et al. Aug 1997 A
5699161 Woodworth Dec 1997 A
5729750 Ishida Mar 1998 A
5730252 Herbinet Mar 1998 A
5732147 Tao Mar 1998 A
5734476 Dlugos Mar 1998 A
5737074 Haga et al. Apr 1998 A
5767962 Suzuki et al. Jun 1998 A
5831737 Stringer et al. Nov 1998 A
5850370 Stringer et al. Dec 1998 A
5850490 Johnson Dec 1998 A
5869827 Rando Feb 1999 A
5870220 Migdal et al. Feb 1999 A
5900611 Hecht May 1999 A
5923428 Woodworth Jul 1999 A
5929856 LoNegro et al. Jul 1999 A
5938710 Lanza et al. Aug 1999 A
5959568 Woolley Sep 1999 A
5960098 Tao Sep 1999 A
5969823 Wurz et al. Oct 1999 A
5978512 Kim et al. Nov 1999 A
5979760 Freyman et al. Nov 1999 A
5988862 Kacyra et al. Nov 1999 A
5991041 Woodworth Nov 1999 A
6009189 Schaack Dec 1999 A
6025847 Marks Feb 2000 A
6049386 Stringer et al. Apr 2000 A
6053409 Brobst et al. Apr 2000 A
6064759 Buckley et al. May 2000 A
6067110 Nonaka et al. May 2000 A
6069696 McQueen et al. May 2000 A
6115114 Berg et al. Sep 2000 A
6137577 Woodworth Oct 2000 A
6177999 Wurz et al. Jan 2001 B1
6189223 Haug Feb 2001 B1
6232597 Kley May 2001 B1
6236403 Chaki May 2001 B1
6246468 Dimsdale Jun 2001 B1
6333749 Reinhardt et al. Dec 2001 B1
6336587 He et al. Jan 2002 B1
6369401 Lee Apr 2002 B1
6373579 Ober et al. Apr 2002 B1
6429803 Kumar Aug 2002 B1
6457642 Good et al. Oct 2002 B1
6507406 Yagi Jan 2003 B1
6517004 Good et al. Feb 2003 B2
6519550 D'Hooge et al. Feb 2003 B1
6674904 McQueen Jan 2004 B1
6705526 Zhu et al. Mar 2004 B1
6781621 Gobush et al. Aug 2004 B1
6824058 Patel et al. Nov 2004 B2
6832725 Gardiner et al. Dec 2004 B2
6858857 Pease et al. Feb 2005 B2
6922632 Foxlin Jul 2005 B2
6971580 Zhu et al. Dec 2005 B2
6995762 Pavlidis et al. Feb 2006 B1
7057632 Yamawaki et al. Jun 2006 B2
7085409 Sawhney et al. Aug 2006 B2
7086162 Tyroler Aug 2006 B2
7104453 Zhu et al. Sep 2006 B1
7128266 Zhu et al. Oct 2006 B2
7137556 Bonner et al. Nov 2006 B1
7159783 Walczyk et al. Jan 2007 B2
7161688 Bonner et al. Jan 2007 B1
7205529 Andersen et al. Apr 2007 B2
7214954 Schopp May 2007 B2
7277187 Smith et al. Oct 2007 B2
7307653 Dutta Dec 2007 B2
7310431 Gokturk et al. Dec 2007 B2
7413127 Ehrhart et al. Aug 2008 B2
7527205 Zhu et al. May 2009 B2
7586049 Wurz Sep 2009 B2
7602404 Reinhardt et al. Oct 2009 B1
7639722 Paxton et al. Dec 2009 B1
7726575 Wang et al. Jun 2010 B2
7780084 Zhang et al. Aug 2010 B2
7788883 Buckley et al. Sep 2010 B2
7974025 Topliss Jul 2011 B2
8027096 Feng et al. Sep 2011 B2
8028501 Buckley et al. Oct 2011 B2
8050461 Shpunt Nov 2011 B2
8055061 Katano Nov 2011 B2
8072581 Breiholz Dec 2011 B1
8102395 Kondo et al. Jan 2012 B2
8132728 Dwinell et al. Mar 2012 B2
8134717 Pangrazio et al. Mar 2012 B2
8149224 Kuo et al. Apr 2012 B1
8194097 Xiao et al. Jun 2012 B2
8212889 Chanas et al. Jul 2012 B2
8228510 Pangrazio et al. Jul 2012 B2
8230367 Bell et al. Jul 2012 B2
8294969 Plesko Oct 2012 B2
8305458 Hara Nov 2012 B2
8310656 Zalewski Nov 2012 B2
8313380 Zalewski et al. Nov 2012 B2
8317105 Kotlarsky et al. Nov 2012 B2
8322622 Liu Dec 2012 B2
8339462 Stec et al. Dec 2012 B2
8350959 Topliss et al. Jan 2013 B2
8351670 Ijiri et al. Jan 2013 B2
8366005 Kotlarsky et al. Feb 2013 B2
8371507 Haggerty et al. Feb 2013 B2
8376233 Van Horn et al. Feb 2013 B2
8381976 Mohideen et al. Feb 2013 B2
8381979 Franz Feb 2013 B2
8390909 Plesko Mar 2013 B2
8408464 Zhu et al. Apr 2013 B2
8408468 Van Horn et al. Apr 2013 B2
8408469 Good Apr 2013 B2
8424768 Rueblinger et al. Apr 2013 B2
8437539 Komatsu et al. May 2013 B2
8441749 Brown et al. May 2013 B2
8448863 Xian et al. May 2013 B2
8457013 Essinger et al. Jun 2013 B2
8459557 Havens et al. Jun 2013 B2
8463079 Ackley et al. Jun 2013 B2
8469272 Kearney Jun 2013 B2
8474712 Kearney et al. Jul 2013 B2
8479992 Kotlarsky et al. Jul 2013 B2
8490877 Kearney Jul 2013 B2
8517271 Kotlarsky et al. Aug 2013 B2
8523076 Good Sep 2013 B2
8528818 Ehrhart et al. Sep 2013 B2
8544737 Gomez et al. Oct 2013 B2
8548420 Grunow et al. Oct 2013 B2
8550335 Samek et al. Oct 2013 B2
8550354 Gannon et al. Oct 2013 B2
8550357 Kearney Oct 2013 B2
8556174 Kosecki et al. Oct 2013 B2
8556176 Van Horn et al. Oct 2013 B2
8556177 Hussey et al. Oct 2013 B2
8559767 Barber et al. Oct 2013 B2
8561895 Gomez et al. Oct 2013 B2
8561903 Sauerwein Oct 2013 B2
8561905 Edmonds et al. Oct 2013 B2
8565107 Pease et al. Oct 2013 B2
8570343 Halstead Oct 2013 B2
8571307 Li et al. Oct 2013 B2
8576390 Nunnink Nov 2013 B1
8579200 Samek et al. Nov 2013 B2
8583924 Caballero et al. Nov 2013 B2
8584945 Wang et al. Nov 2013 B2
8587595 Wang Nov 2013 B2
8587697 Hussey et al. Nov 2013 B2
8588869 Sauerwein et al. Nov 2013 B2
8590789 Nahill et al. Nov 2013 B2
8596539 Havens et al. Dec 2013 B2
8596542 Havens et al. Dec 2013 B2
8596543 Havens et al. Dec 2013 B2
8599271 Havens et al. Dec 2013 B2
8599957 Peake et al. Dec 2013 B2
8600158 Li et al. Dec 2013 B2
8600167 Showering Dec 2013 B2
8602309 Longacre et al. Dec 2013 B2
8608053 Meier et al. Dec 2013 B2
8608071 Liu et al. Dec 2013 B2
8611309 Wang et al. Dec 2013 B2
8615487 Gomez et al. Dec 2013 B2
8621123 Caballero Dec 2013 B2
8622303 Meier et al. Jan 2014 B2
8628013 Ding Jan 2014 B2
8628015 Wang et al. Jan 2014 B2
8628016 Winegar Jan 2014 B2
8629926 Wang Jan 2014 B2
8630491 Longacre et al. Jan 2014 B2
8635309 Berthiaume et al. Jan 2014 B2
8636200 Kearney Jan 2014 B2
8636212 Nahill et al. Jan 2014 B2
8636215 Ding et al. Jan 2014 B2
8636224 Wang Jan 2014 B2
8638806 Wang et al. Jan 2014 B2
8640958 Lu et al. Feb 2014 B2
8640960 Wang et al. Feb 2014 B2
8643717 Li et al. Feb 2014 B2
8646692 Meier et al. Feb 2014 B2
8646694 Wang et al. Feb 2014 B2
8657200 Ren et al. Feb 2014 B2
8659397 Vargo et al. Feb 2014 B2
8668149 Good Mar 2014 B2
8678285 Kearney Mar 2014 B2
8678286 Smith et al. Mar 2014 B2
8682077 Longacre Mar 2014 B1
D702237 Oberpriller et al. Apr 2014 S
8687282 Feng et al. Apr 2014 B2
8692927 Pease et al. Apr 2014 B2
8695880 Bremer et al. Apr 2014 B2
8698949 Grunow et al. Apr 2014 B2
8702000 Barber et al. Apr 2014 B2
8717494 Gannon May 2014 B2
8720783 Biss et al. May 2014 B2
8723804 Fletcher et al. May 2014 B2
8723904 Marty et al. May 2014 B2
8727223 Wang May 2014 B2
8740082 Wilz Jun 2014 B2
8740085 Furlong et al. Jun 2014 B2
8746563 Hennick et al. Jun 2014 B2
8750445 Peake et al. Jun 2014 B2
8752766 Xian et al. Jun 2014 B2
8756059 Braho et al. Jun 2014 B2
8757495 Qu et al. Jun 2014 B2
8760563 Koziol et al. Jun 2014 B2
8763909 Reed et al. Jul 2014 B2
8777108 Coyle Jul 2014 B2
8777109 Oberpriller et al. Jul 2014 B2
8779898 Havens et al. Jul 2014 B2
8781520 Payne et al. Jul 2014 B2
8783573 Havens et al. Jul 2014 B2
8789757 Barten Jul 2014 B2
8789758 Hawley et al. Jul 2014 B2
8789759 Xian et al. Jul 2014 B2
8792688 Unsworth Jul 2014 B2
8794520 Wang et al. Aug 2014 B2
8794522 Ehrhart Aug 2014 B2
8794525 Amundsen et al. Aug 2014 B2
8794526 Wang et al. Aug 2014 B2
8798367 Ellis Aug 2014 B2
8807431 Wang et al. Aug 2014 B2
8807432 Van Horn et al. Aug 2014 B2
8810779 Hilde Aug 2014 B1
8820630 Qu et al. Sep 2014 B2
8897596 Passmore et al. Nov 2014 B1
9014441 Truyen et al. Apr 2015 B2
9082195 Holeva et al. Jul 2015 B2
9142035 Rotman Sep 2015 B1
9233470 Bradski et al. Jan 2016 B1
9299013 Curlander et al. Mar 2016 B1
9424749 Reed et al. Aug 2016 B1
9486921 Straszheim et al. Nov 2016 B1
20010027995 Patel et al. Oct 2001 A1
20010032879 He et al. Oct 2001 A1
20020054289 Thibault et al. May 2002 A1
20020109835 Goetz Aug 2002 A1
20020118874 Chung et al. Aug 2002 A1
20020158873 Williamson Oct 2002 A1
20020167677 Okada et al. Nov 2002 A1
20020179708 Zhu et al. Dec 2002 A1
20020196534 Lizotte et al. Dec 2002 A1
20030038179 Tsikos et al. Feb 2003 A1
20030053513 Vatan et al. Mar 2003 A1
20030063086 Baumberg Apr 2003 A1
20030091227 Chang et al. May 2003 A1
20030156756 Gokturk et al. Aug 2003 A1
20030197138 Pease et al. Oct 2003 A1
20030225712 Cooper et al. Dec 2003 A1
20030235331 Kawaike et al. Dec 2003 A1
20040008259 Gokturk et al. Jan 2004 A1
20040019274 Galloway et al. Jan 2004 A1
20040024754 Mane et al. Feb 2004 A1
20040066329 Zeitfuss et al. Apr 2004 A1
20040073359 Ichijo et al. Apr 2004 A1
20040083025 Yamanouchi et al. Apr 2004 A1
20040089482 Ramsden et al. May 2004 A1
20040098146 Katae et al. May 2004 A1
20040105580 Hager et al. Jun 2004 A1
20040118928 Patel et al. Jun 2004 A1
20040122779 Stickler et al. Jun 2004 A1
20040155975 Hart et al. Aug 2004 A1
20040165090 Ning Aug 2004 A1
20040184041 Schopp Sep 2004 A1
20040211836 Patel et al. Oct 2004 A1
20040214623 Takahashi et al. Oct 2004 A1
20040233461 Armstrong et al. Nov 2004 A1
20040258353 Gluckstad et al. Dec 2004 A1
20050006477 Patel Jan 2005 A1
20050117215 Lange Jun 2005 A1
20050128196 Popescu et al. Jun 2005 A1
20050168488 Montague Aug 2005 A1
20050211782 Martin Sep 2005 A1
20050264867 Cho et al. Dec 2005 A1
20060047704 Gopalakrishnan Mar 2006 A1
20060078226 Zhou Apr 2006 A1
20060108266 Bowers et al. May 2006 A1
20060112023 Horhann May 2006 A1
20060151604 Zhu et al. Jul 2006 A1
20060159307 Anderson et al. Jul 2006 A1
20060159344 Shao et al. Jul 2006 A1
20060232681 Okada Oct 2006 A1
20060255150 Longacre Nov 2006 A1
20060269165 Viswanathan Nov 2006 A1
20060291719 Ikeda et al. Dec 2006 A1
20070003154 Sun et al. Jan 2007 A1
20070025612 Iwasaki et al. Feb 2007 A1
20070031064 Zhao et al. Feb 2007 A1
20070063048 Havens et al. Mar 2007 A1
20070116357 Dewaele May 2007 A1
20070127022 Cohen et al. Jun 2007 A1
20070143082 Degnan Jun 2007 A1
20070153293 Gruhlke et al. Jul 2007 A1
20070171220 Kriveshko Jul 2007 A1
20070177011 Lewin et al. Aug 2007 A1
20070181685 Zhu et al. Aug 2007 A1
20070237356 Dwinell et al. Oct 2007 A1
20070291031 Konev et al. Dec 2007 A1
20070299338 Stevick et al. Dec 2007 A1
20080013793 Hillis et al. Jan 2008 A1
20080035390 Wurz Feb 2008 A1
20080056536 Hildreth et al. Mar 2008 A1
20080062164 Bassi et al. Mar 2008 A1
20080077265 Boyden Mar 2008 A1
20080079955 Storm Apr 2008 A1
20080164074 Wurz Jul 2008 A1
20080185432 Caballero et al. Aug 2008 A1
20080204476 Montague Aug 2008 A1
20080212168 Olmstead et al. Sep 2008 A1
20080247635 Davis et al. Oct 2008 A1
20080273191 Kim et al. Nov 2008 A1
20080273210 Hilde Nov 2008 A1
20080278790 Boesser et al. Nov 2008 A1
20090059004 Bochicchio Mar 2009 A1
20090095047 Patel et al. Apr 2009 A1
20090134221 Zhu et al. May 2009 A1
20090195790 Zhu et al. Aug 2009 A1
20090225333 Bendall et al. Sep 2009 A1
20090237411 Gossweiler et al. Sep 2009 A1
20090268023 Hsieh Oct 2009 A1
20090272724 Gubler Nov 2009 A1
20090273770 Bauhahn Nov 2009 A1
20090313948 Buckley et al. Dec 2009 A1
20090318815 Barnes et al. Dec 2009 A1
20090323084 Dunn et al. Dec 2009 A1
20090323121 Valkenburg Dec 2009 A1
20100035637 Varanasi et al. Feb 2010 A1
20100060604 Zwart et al. Mar 2010 A1
20100091104 Sprigle Apr 2010 A1
20100113153 Yen et al. May 2010 A1
20100118200 Gelman et al. May 2010 A1
20100128109 Banks May 2010 A1
20100161170 Siris Jun 2010 A1
20100171740 Andersen et al. Jul 2010 A1
20100172567 Prokoski Jul 2010 A1
20100177076 Essinger et al. Jul 2010 A1
20100177080 Essinger et al. Jul 2010 A1
20100177707 Essinger et al. Jul 2010 A1
20100177749 Essinger et al. Jul 2010 A1
20100202702 Benos et al. Aug 2010 A1
20100208039 Stettner Aug 2010 A1
20100211355 Horst et al. Aug 2010 A1
20100217678 Goncalves Aug 2010 A1
20100220849 Colbert et al. Sep 2010 A1
20100220894 Ackley et al. Sep 2010 A1
20100223276 Al-Shameri et al. Sep 2010 A1
20100245850 Lee et al. Sep 2010 A1
20100254611 Arnz Oct 2010 A1
20100303336 Abraham Dec 2010 A1
20100315413 Izadi et al. Dec 2010 A1
20100321482 Cleveland Dec 2010 A1
20110019155 Daniel et al. Jan 2011 A1
20110040192 Brenner et al. Feb 2011 A1
20110043609 Choi et al. Feb 2011 A1
20110099474 Grossman et al. Apr 2011 A1
20110169999 Grunow et al. Jul 2011 A1
20110188054 Petronius et al. Aug 2011 A1
20110188741 Sones et al. Aug 2011 A1
20110202554 Powilleit et al. Aug 2011 A1
20110234389 Mellin Sep 2011 A1
20110235854 Berger et al. Sep 2011 A1
20110249864 Venkatesan et al. Oct 2011 A1
20110254840 Halstead Oct 2011 A1
20110279916 Brown et al. Nov 2011 A1
20110286007 Pangrazio et al. Nov 2011 A1
20110286628 Goncalves et al. Nov 2011 A1
20110288818 Thierman Nov 2011 A1
20110301994 Tieman Dec 2011 A1
20110303748 Lemma et al. Dec 2011 A1
20110310227 Konertz et al. Dec 2011 A1
20120024952 Chen Feb 2012 A1
20120056982 Katz et al. Mar 2012 A1
20120057345 Kuchibhotla Mar 2012 A1
20120067955 Rowe Mar 2012 A1
20120074227 Ferren et al. Mar 2012 A1
20120081714 Pangrazio et al. Apr 2012 A1
20120111946 Golant May 2012 A1
20120113223 Hilliges et al. May 2012 A1
20120113250 Farlotti et al. May 2012 A1
20120126000 Kunzig et al. May 2012 A1
20120138685 Qu et al. Jun 2012 A1
20120140300 Freeman Jun 2012 A1
20120168512 Kotlarsky et al. Jul 2012 A1
20120179665 Baarman et al. Jul 2012 A1
20120185094 Rosenstein et al. Jul 2012 A1
20120190386 Anderson Jul 2012 A1
20120193423 Samek Aug 2012 A1
20120197464 Wang et al. Aug 2012 A1
20120203647 Smith Aug 2012 A1
20120218436 Rodriguez et al. Aug 2012 A1
20120223141 Good et al. Sep 2012 A1
20120224026 Bayer et al. Sep 2012 A1
20120228382 Havens et al. Sep 2012 A1
20120236288 Stanley Sep 2012 A1
20120242852 Hayward et al. Sep 2012 A1
20120248188 Kearney Oct 2012 A1
20120256901 Bendall Oct 2012 A1
20120262558 Boger et al. Oct 2012 A1
20120280908 Rhoads et al. Nov 2012 A1
20120282905 Owen Nov 2012 A1
20120282911 Davis et al. Nov 2012 A1
20120284012 Rodriguez et al. Nov 2012 A1
20120284122 Brandis Nov 2012 A1
20120284339 Rodriguez Nov 2012 A1
20120284593 Rodriguez Nov 2012 A1
20120293610 Doepke et al. Nov 2012 A1
20120294549 Doepke Nov 2012 A1
20120299961 Ramkumar et al. Nov 2012 A1
20120300991 Free Nov 2012 A1
20120313848 Galor et al. Dec 2012 A1
20120314030 Datta Dec 2012 A1
20120314058 Bendall et al. Dec 2012 A1
20120316820 Nakazato et al. Dec 2012 A1
20130019278 Sun et al. Jan 2013 A1
20130038881 Pesach et al. Feb 2013 A1
20130038941 Pesach et al. Feb 2013 A1
20130043312 Van Horn Feb 2013 A1
20130050426 Sarmast et al. Feb 2013 A1
20130056285 Meagher Mar 2013 A1
20130070322 Fritz et al. Mar 2013 A1
20130075168 Amundsen et al. Mar 2013 A1
20130094069 Lee et al. Apr 2013 A1
20130101158 Lloyd Apr 2013 A1
20130156267 Muraoka et al. Jun 2013 A1
20130175341 Kearney et al. Jul 2013 A1
20130175343 Good Jul 2013 A1
20130200150 Reynolds et al. Aug 2013 A1
20130200158 Feng et al. Aug 2013 A1
20130201288 Billerbeck et al. Aug 2013 A1
20130208164 Cazier et al. Aug 2013 A1
20130211790 Loveland et al. Aug 2013 A1
20130223673 Davis et al. Aug 2013 A1
20130256418 Havens et al. Oct 2013 A1
20130257744 Daghigh et al. Oct 2013 A1
20130257759 Daghigh Oct 2013 A1
20130270346 Xian et al. Oct 2013 A1
20130278425 Cunningham et al. Oct 2013 A1
20130287258 Kearney Oct 2013 A1
20130291998 Konnerth Nov 2013 A1
20130292475 Kotlarsky et al. Nov 2013 A1
20130292477 Hennick et al. Nov 2013 A1
20130293539 Hunt et al. Nov 2013 A1
20130293540 Laffargue et al. Nov 2013 A1
20130306728 Thuries et al. Nov 2013 A1
20130306730 Brady et al. Nov 2013 A1
20130306731 Pedrao Nov 2013 A1
20130307964 Bremer et al. Nov 2013 A1
20130308013 Li et al. Nov 2013 A1
20130308625 Park et al. Nov 2013 A1
20130313324 Koziol et al. Nov 2013 A1
20130313325 Wilz et al. Nov 2013 A1
20130329012 Bartos Dec 2013 A1
20130329013 Metois et al. Dec 2013 A1
20130341399 Xian et al. Dec 2013 A1
20130342342 Sabre et al. Dec 2013 A1
20130342717 Havens et al. Dec 2013 A1
20140001267 Giordano et al. Jan 2014 A1
20140002828 Laffargue et al. Jan 2014 A1
20140008430 Soule et al. Jan 2014 A1
20140008439 Wang Jan 2014 A1
20140009586 McNamer et al. Jan 2014 A1
20140021259 Moed et al. Jan 2014 A1
20140025584 Liu et al. Jan 2014 A1
20140027518 Edmonds et al. Jan 2014 A1
20140031665 Pinto et al. Jan 2014 A1
20140034731 Gao et al. Feb 2014 A1
20140034734 Sauerwein Feb 2014 A1
20140036848 Pease et al. Feb 2014 A1
20140039693 Havens et al. Feb 2014 A1
20140042814 Kather et al. Feb 2014 A1
20140049120 Kohtz et al. Feb 2014 A1
20140049635 Laffargue et al. Feb 2014 A1
20140058612 Wong et al. Feb 2014 A1
20140061305 Nahill et al. Mar 2014 A1
20140061306 Wu et al. Mar 2014 A1
20140062709 Hyer et al. Mar 2014 A1
20140063289 Hussey et al. Mar 2014 A1
20140066136 Sauerwein et al. Mar 2014 A1
20140067104 Osterhout Mar 2014 A1
20140067692 Ye et al. Mar 2014 A1
20140070005 Nahill et al. Mar 2014 A1
20140071840 Venancio Mar 2014 A1
20140074746 Wang Mar 2014 A1
20140075846 Woodburn Mar 2014 A1
20140076974 Havens et al. Mar 2014 A1
20140078341 Havens et al. Mar 2014 A1
20140078342 Li et al. Mar 2014 A1
20140078345 Showering Mar 2014 A1
20140084068 Gillet et al. Mar 2014 A1
20140091147 Evans et al. Apr 2014 A1
20140097249 Gomez et al. Apr 2014 A1
20140098091 Hori Apr 2014 A1
20140098792 Wang et al. Apr 2014 A1
20140100774 Showering Apr 2014 A1
20140100813 Showering Apr 2014 A1
20140103115 Meier et al. Apr 2014 A1
20140104413 McCloskey et al. Apr 2014 A1
20140104414 McCloskey Apr 2014 A1
20140104416 Giordano Apr 2014 A1
20140104451 Todeschini et al. Apr 2014 A1
20140104664 Lee Apr 2014 A1
20140106594 Skvoretz Apr 2014 A1
20140106725 Sauerwein Apr 2014 A1
20140108010 Maltseff et al. Apr 2014 A1
20140108402 Gomez et al. Apr 2014 A1
20140108682 Caballero Apr 2014 A1
20140110485 Toa et al. Apr 2014 A1
20140114530 Fitch et al. Apr 2014 A1
20140124577 Wang et al. May 2014 A1
20140124579 Ding May 2014 A1
20140125842 Winegar May 2014 A1
20140125853 Wang May 2014 A1
20140125999 Longacre et al. May 2014 A1
20140129378 Richardson May 2014 A1
20140131438 Kearney May 2014 A1
20140131441 Nahill et al. May 2014 A1
20140131443 Smith May 2014 A1
20140131444 Wang May 2014 A1
20140131445 Ding et al. May 2014 A1
20140131448 Xian et al. May 2014 A1
20140133379 Wang et al. May 2014 A1
20140135984 Hirata May 2014 A1
20140136208 Maltseff et al. May 2014 A1
20140139654 Takahashi May 2014 A1
20140140585 Wang May 2014 A1
20140151453 Meier et al. Jun 2014 A1
20140152882 Samek et al. Jun 2014 A1
20140152975 Ko Jun 2014 A1
20140158468 Adami Jun 2014 A1
20140158770 Sevier et al. Jun 2014 A1
20140159869 Zumsteg et al. Jun 2014 A1
20140160329 Ren et al. Jun 2014 A1
20140166755 Liu et al. Jun 2014 A1
20140166757 Smith Jun 2014 A1
20140166759 Liu et al. Jun 2014 A1
20140166760 Meier et al. Jun 2014 A1
20140166761 Todeschini et al. Jun 2014 A1
20140168380 Heidemann et al. Jun 2014 A1
20140168787 Wang et al. Jun 2014 A1
20140175165 Havens et al. Jun 2014 A1
20140175169 Kosecki et al. Jun 2014 A1
20140175172 Jovanovski et al. Jun 2014 A1
20140175174 Barber Jun 2014 A1
20140191644 Chaney Jul 2014 A1
20140191913 Ge et al. Jul 2014 A1
20140192187 Atwell et al. Jul 2014 A1
20140192551 Masaki Jul 2014 A1
20140197238 Liu et al. Jul 2014 A1
20140197239 Havens et al. Jul 2014 A1
20140197304 Feng et al. Jul 2014 A1
20140203087 Smith et al. Jul 2014 A1
20140204268 Grunow et al. Jul 2014 A1
20140205150 Ogawa Jul 2014 A1
20140214631 Hansen Jul 2014 A1
20140217166 Berthiaume et al. Aug 2014 A1
20140217180 Liu Aug 2014 A1
20140225918 Mittal et al. Aug 2014 A1
20140225985 Klusza et al. Aug 2014 A1
20140240464 Lee Aug 2014 A1
20140247279 Nicholas et al. Sep 2014 A1
20140247280 Nicholas et al. Sep 2014 A1
20140267609 Laffargue Sep 2014 A1
20140268093 Tohme et al. Sep 2014 A1
20140270361 Amma et al. Sep 2014 A1
20140306833 Ricci Oct 2014 A1
20140307855 Withagen et al. Oct 2014 A1
20140313527 Askan Oct 2014 A1
20140319219 Liu et al. Oct 2014 A1
20140320408 Zagorsek et al. Oct 2014 A1
20140347553 Ovsiannikov et al. Nov 2014 A1
20140350710 Gopalakrishnan et al. Nov 2014 A1
20140379613 Nishitani et al. Dec 2014 A1
20150009301 Ribnick et al. Jan 2015 A1
20150009338 Laffargue et al. Jan 2015 A1
20150036876 Marrion et al. Feb 2015 A1
20150049347 Laffargue et al. Feb 2015 A1
20150062369 Gehring et al. Mar 2015 A1
20150063676 Lloyd et al. Mar 2015 A1
20150116498 Vartiainen et al. Apr 2015 A1
20150149946 Benos et al. May 2015 A1
20150163474 You Jun 2015 A1
20150204662 Kobayashi et al. Jul 2015 A1
20150213647 Laffargue et al. Jul 2015 A1
20150229838 Hakim et al. Aug 2015 A1
20150269403 Lei et al. Sep 2015 A1
20150276379 Ni et al. Oct 2015 A1
20150301181 Herschbach Oct 2015 A1
20150308816 Laffargue et al. Oct 2015 A1
20150325036 Lee Nov 2015 A1
20150355470 Herschbach Dec 2015 A1
20160048725 Holz et al. Feb 2016 A1
20160063429 Varley et al. Mar 2016 A1
20160090283 Svensson et al. Mar 2016 A1
20160090284 Svensson et al. Mar 2016 A1
20160138247 Conway et al. May 2016 A1
20160138248 Conway et al. May 2016 A1
20160138249 Svensson et al. May 2016 A1
20160169665 Deschenes et al. Jun 2016 A1
20160187186 Coleman et al. Jun 2016 A1
20160187210 Coleman et al. Jun 2016 A1
20160191801 Sivan Jun 2016 A1
20160202478 Masson et al. Jul 2016 A1
20170115490 Hsieh et al. Apr 2017 A1
Foreign Referenced Citations (48)
Number Date Country
2004212587 Apr 2005 AU
201139117 Oct 2008 CN
3335760 Apr 1985 DE
10210813 Oct 2003 DE
102007037282 Mar 2008 DE
1111435 Jun 2001 EP
1443312 Aug 2004 EP
2286932 Feb 2011 EP
2381421 Oct 2011 EP
2533009 Dec 2012 EP
2722656 Apr 2014 EP
2779027 Sep 2014 EP
2833323 Feb 2015 EP
2843590 Mar 2015 EP
2845170 Mar 2015 EP
2966595 Jan 2016 EP
3006893 Mar 2016 EP
3012601 Mar 2016 EP
3007096 Apr 2016 EP
2503978 Jan 2014 GB
2531928 May 2016 GB
H04129902 Apr 1992 JP
2008210276 Sep 2008 JP
2014210646 Nov 2014 JP
20100020115 Feb 2010 KR
20110013200 Feb 2011 KR
20110117020 Oct 2011 KR
20120028109 Mar 2012 KR
9640452 Dec 1996 WO
0114836 Mar 2001 WO
2006095110 Sep 2006 WO
2007015059 Feb 2007 WO
2011017241 Feb 2011 WO
2012175731 Dec 2012 WO
2013021157 Feb 2013 WO
2013033442 Mar 2013 WO
2013163789 Nov 2013 WO
2013166368 Nov 2013 WO
2013173985 Nov 2013 WO
20130184340 Dec 2013 WO
2014019130 Feb 2014 WO
2014102341 Jul 2014 WO
2014110495 Jul 2014 WO
2014149702 Sep 2014 WO
2014151746 Sep 2014 WO
2015006865 Jan 2015 WO
2016020038 Feb 2016 WO
2016061699 Apr 2016 WO
Non-Patent Literature Citations (175)
Entry
U.S. Appl. No. 14/274,858 for Mobile Printer With Optional Battery Accessory, filed May 12, 2014, (Marty et al.); 26 pages.
U.S. Appl. No. 14/264,173 for Autofocus Lens System for Indicia Readers filed Apr. 29, 2014, (Ackley et al.); 39 pages.
U.S. Appl. No. 14/453,019 for Dimensioning System With Guided Alignment, filed Aug. 6, 2014 (Li et al.); 31 pages.
U.S. Appl. No. 14/452,697 for Interactive Indicia Reader , filed Aug. 6, 2014, (Todeschini); 32 pages.
U.S. Appl. No. 14/231,898 for Hand-Mounted Indicia-Reading Device with Finger Motion Triggering filed Apr. 1, 2014 (Van Horn et al.); 36 pages.
U.S. Appl. No. 29/486,759 for an Imaging Terminal, filed Apr. 2, 2014 (Oberpriller et al.); 8 pages.
U.S. Appl. No. 29/436,337 for an Electronic Device, filed Nov. 5, 2012 (Fitch et al.); 19 pages.
U.S. Appl. No. 29/458,405 for an Electronic Device, filed Jun. 19, 2013 (Fitch et al.); 22 pages.
U.S. Appl. No. 29/459,620 for an Electronic Device Enclosure, filed Jul. 2, 2013 (London et al.); 21 pages.
U.S. Appl. No. 29/459,681 for an Electronic Device Enclosure, filed Jul. 2, 2013 (Chaney et al.); 14 pages.
U.S. Appl. No. 29/459,785 for a Scanner and Charging Base, filed Jul. 3, 2013 (Fitch et al.); 21 pages.
U.S. Appl. No. 29/459,823 for a Scanner, filed Jul. 3, 2013 (Zhou et al.); 13 pages.
U.S. Appl. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.); 44 pages.
U.S. Appl. No. 14/340,627 for an Axially Reinforced Flexible Scan Element, filed Jul. 25, 2014 (Reublinger et al.); 41 pages.
U.S. Appl. No. 29/492,903 for an Indicia Scanner, filed Jun. 4, 2014 (Zhou et al.); 23 pages.
U.S. Appl. No. 29/494,725 for an In-Counter Barcode Scanner, filed Jun. 24, 2014 (Oberpriller et al.); 23 pages.
U.S. Appl. No. 14/327,827 for a Mobile-Phone Adapter for Electronic Transactions, filed Jul. 10, 2014 (Hejl); 25 pages.
U.S. Appl. No. 14/334,934 for a System and Method for Indicia Verification, filed Jul. 18, 2014 (Hejl); 38 pages.
U.S. Appl. No. 14/339,708 for Laser Scanning Code Symbol Reading System, filed Jul. 24, 2014 (Xian et al.); 39 pages.
U.S. Appl. No. 14/379,057 for Method of Using Camera Sensor Interface to Transfer Multiple Channels of Scan Data Using an Image Format filed Aug. 15, 2014 (Wang et al.); 28 pages.
U.S. Appl. No. 14/283,282 for Terminal Having Illumination and Focus Control filed May 21, 2014 (Liu et al.); 31 pages.
U.S. Appl. No. 14/490,989, not yet published, filed Sep. 19, 2014, Intermec IP Corporation, Volume Dimensioning System Calibration Systems and Methods.
Wikipedia, YUV description and definition, downloaded from http://www.wikipeida.org/wiki/YUV on Jun. 29, 2012, 10 pages.
YUV Pixel Format, downloaded from http://www.fource.org/yuv.php on Jun. 29, 2012; 13 pages.
YUV to RGB Conversion, downloaded from http://www.fource.org/fccyvrgb.php on Jun. 29, 2012; 5 pages.
Benos et al., “Semi-Automatic Dimensioning with Imager of a Portable Device,” U.S. Appl. No. 61/149,912, filed Feb. 4, 2009 (now expired), 56 pages.
Dimensional Weight—Wikipedia, the Free Encyclopedia, URL=http://en.wikipedia.org/wiki/Dimensional—weight, download date Aug. 1, 2008, 2 pages.
Dimensioning—Wikipedia, the Free Encyclopedia, URL=http://en.wikipedia.org/wiki/Dimensioning, download date Aug. 1, 2008, 1 page.
U.S. Appl. No. 14/519,179, Serge Thuries et al., filed Oct. 21, 2014, not published yet. 40 pages.
U.S. Appl. No. 14/519,249, H. Sprague Ackley et al., filed Oct. 21, 2014, not published yet. 36 pages.
U.S. Appl. No. 14/519,233, Franck Laffargue et al., filed Oct. 21, 2014, not published yet. 34 pages.
U.S. Appl. No. 14/519,211, H. Sprague Ackley et al., filed Oct. 21, 2014, not published yet. 33 pages.
U.S. Appl. No. 14/519,195, Franck Laffargue et al., filed Oct. 21, 2014, not published yet. 35 pages.
U.S. Appl. No. 13/367,978, filed Feb. 7, 2012, (Feng et al.); now abandoned.
U.S. Appl. No. 14/376,472, for an Encoded Information Reading Terminal Including Http Server filed Aug. 4, 2014, (Lu); 30 pages.
U.S. Appl. No. 13/771,508 for an Optical Redirection Adapter, filed Feb. 20, 2013 (Anderson); 26 pages.
U.S. Appl. No. 14/340,716 for an Optical Imager and Method for Correlating a Medication Package With a Patient, filed Jul. 25, 2014 (Ellis); 26 pages.
U.S. Appl. No. 13/852,097 for a System and Method for Capturing and Preserving Vehicle Event Data, filed Mar. 28, 2013 (Barker et al.); 20 pages.
U.S. Appl. No. 13/902,110 for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Hollifield); 29 pages.
U.S. Appl. No. 13/902,144, for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Chamberlin); 23 pages.
U.S. Appl. No. 13/902,242 for a System for Providing a Continuous Communication Link With a Symbol Reading Device, filed May 24, 2013 (Smith et al.); 24 pages.
U.S. Appl. No. 13/912,262 for a Method of Error Correction for 3D Imaging Device, filed Jun. 7, 2013 (Jovanovski et al.); 33 pages.
U.S. Appl. No. 13/912,702 for a System and Method for Reading Code Symbols at Long Range Using Source Power Control, filed Jun. 7, 2013 (Xian et al.); 24 pages.
U.S. Appl. No. 13/922,339 for a System and Method for Reading Code Symbols Using a Variable Field of View, filed Jun. 20, 2013 (Xian et al.); 23 pages.
U.S. Appl. No. 13/927,398 for a Code Symbol Reading System Having Adaptive Autofocus, filed Jun. 26, 2013 (Todeschini); 24 pages.
U.S. Appl. No. 13/930,913 for a Mobile Device Having an Improved User Interface for Reading Code Symbols, filed Jun. 28, 2013 (Gelay et al.); 24 pages.
U.S. Appl. No. 13/933,415 for an Electronic Device Case, filed Jul. 2, 2013 (London et al.); 47 pages.
U.S. Appl. No. 13/947,296 for a System and Method for Selectively Reading Code Symbols, filed Jul. 22, 2013 (Rueblinger et al.); 29 pages.
U.S. Appl. No. 13/950,544 for a Code Symbol Reading System Having Adjustable Object Detection, filed Jul. 25, 2013 (Jiang); 28 pages.
U.S. Appl. No. 13/961,408 for a Method for Manufacturing Laser Scanners, filed Aug. 7, 2013 (Saber et al.); 26 pages.
U.S. Appl. No. 14/018,729 for a Method for Operating a Laser Scanner, filed Sep. 5, 2013 (Feng et al.); 24 pages.
U.S. Appl. No. 14/019,616 for a Device Having Light Source to Reduce Surface Pathogens, filed Sep. 6, 2013 (Todeschini); 23 pages.
U.S. Appl. No. 14/023,762 for a Handheld Indicia Reader Having Locking Endcap, filed Sep. 11, 2013 (Gannon); 31 pages.
U.S. Appl. No. 14/035,474 for Augmented-Reality Signature Capture, filed Sep. 24, 2013 (Todeschini); 33 pages.
U.S. Appl. No. 14/370,267 for Industrial Design for Consumer Device Based Scanning and Mobility, filed Jul. 2, 2014 (Ma et al.); 45 pages.
U.S. Appl. No. 14/336,188 for Method of and System for Detecting Object Weighing Interferences, Filed Jul. 21, 2014 (Amundsen et al.); 34 pages.
U.S. Appl. No. 14/055,234 for Dimensioning System, filed Oct. 16, 2013 (Fletcher); 26 pages.
U.S. Appl. No. 14/053,314 for Indicia Reader, filed Oct. 14, 2013 (Huck); 29 pages.
U.S. Appl. No. 14/065,768 for Hybrid System and Method for Reading Indicia, filed Oct. 29, 2013 (Meier et al.); 22 pages.
U.S. Appl. No. 14/074,746 for Self-Checkout Shopping System, filed Nov. 8, 2013 (Hejl et al.); 26 pages.
U.S. Appl. No. 14/074,787 for Method and System for Configuring Mobile Devices via Nfc Technology, filed Nov. 8, 2013 (Smith et al.); 28 pages.
U.S. Appl. No. 14/087,190 for Optimal Range Indicators for Bar Code Validation, filed Nov. 22, 2013 (Hejl); 27 pages.
U.S. Appl. No. 14/345,735 for Optical Indicia Reading Terminal with Combined Illumination filed Mar. 19, 2014 (Ouyang); 19 pages.
U.S. Appl. No. 14/101,965 for High Dynamic-Range Indicia Reading System, filed Dec. 10, 2013 (Xian); 28 pages.
U.S. Appl. No. 14/462,801 for Mobile Computing Device With Data Cognition Software, filed on Aug. 19, 2014 (Todeschini et al.); 38 pages.
U.S. Appl. No. 14/150,393 for Incicia-reader Having Unitary Construction Scanner, filed Jan. 8, 2014 (Colavito et al.); 28 pages.
U.S. Appl. No. 14/154,207 for Laser Barcode Scanner, filed Jan. 14, 2014 (Hou et al.); 26 pages.
U.S. Appl. No. 14/277,337 for Multipurpose Optical Reader, filed May 14, 2014 (Jovanovski et al.); 59 pages.
U.S. Appl. No. 14/300,276 for Method and System for Considering Information About an Expected Response When Performing Speech Recognition, filed Jun. 10, 2014 (Braho et al.); 31 pages.
U.S. Appl. No. 14/460,829 for Encoded Information Reading Terminal With Wireless Path Selecton Capability, filed Aug. 15, 2014 (Wang et al.); 40 pages.
U.S. Appl. No. 14/342,544 for Imaging Based Barcode Scanner Engine with Multiple Elements Supported on a Common Printed Circuit Board filed Mar. 4, 2014 (Liu et al.); 27 pages.
U.S. Appl. No. 14/257,174 for Reading Apparatus Having Partial Frame Operating Mode filed Apr. 21, 2014, (Barber et al.), 67 pages.
U.S. Appl. No. 14/200,405 for Indicia Reader for Size-Limited Applications filed Mar. 7, 2014 (Feng et al.); 42 pages.
U.S. Appl. No. 14/166,103 for Indicia Reading Terminal Including Optical Filter filed Jan. 28, 2014 (Lu et al.); 29 pages.
U.S. Appl. No. 14/460,387 for Apparatus for Displaying Bar Codes From Light Emitting Display Surfaces filed Aug. 15, 2014 (Van Horn et al.); 40 pages.
U.S. Appl. No. 14/310,226 for Autofocusing Optical Imaging Device filed Jun. 20, 2014 (Koziol et al.); 28 pages.
U.S. Appl. No. 14/305,153 for Indicia Reading System Employing Digital Gain Control filed Jun. 16, 2014 (Xian et al.); 53 pages.
U.S. Appl. No. 14/327,722 for Customer Facing Imaging Systems and Methods for Obtaining Images filed Jul. 10, 2014 (Oberpriller et al,); 39 pages.
U.S. Appl. No. 14/329,303 for Cell Phone Reading Mode Using Image Timer filed Jul. 11, 2014 (Coyle); 22 pages.
U.S. Appl. No. 14/370,237 for Web-Based Scan-Task Enabled System and Method of and Apparatus for Developing and Deploying the Same on a Client-Server Network filed Jul. 2, 2014 (Chen et al.); 65 pages.
U.S. Appl. No. 14/333,588 for Symbol Reading System With Integrated Scale Base filed Jul. 17, 2014 (Barten); 59 pages.
U.S. Appl. No. 14/446,387 for Indicia Reading Terminal Processing Plurality of Frames.
U.S. Appl. No. 14/446,391 for Multifunction Point of Sale Apparatus With Optical Signature Capture filed Jul. 30, 2014 (Good et al.); 37 pages.
Peter Clarke, Actuator Developer Claims Anti-Shake Breakthrough for Smartphone Cams, Electronic Engineering Times, p. 24, May 16, 2011.
U.S. Appl. No. 14/055,234, not yet published, Hand Held Products, Inc. filed Oct. 16, 2013; 26 pages.
U.S. Appl. No. 13/912,262, not yet published, filed Jun. 7, 2013, Hand Held Products Inc., Method of Error Correction for 3D Imaging Device: 33 pages.
European Search Report for application No. EP13186043 (now EP2722656 (Apr. 23, 2014)): Total pp. 7.
International Search Report for PCT/US2013/039438 (WO2013166368), Oct. 1, 2013, 7 pages.
U.S. Appl. No. 14/453,019, not yet published, filed Aug. 6, 2014, Hand Held Products Inc., Dimensioning System With Guided Alignment: 31 pages.
European Office Action for application EP 13186043, dated Jun. 12, 2014(now EP2722656 (Apr. 23, 2014)), Total of 6 pages.
U.S. Appl. No. 14/461,524, not yet published, filed Aug. 18, 2014, Hand Held Products Inc., System and Method for Package Dimensioning: 21 pages.
European Patent Office Action for Application No. 14157971.4-1906, Dated Jul. 16, 2014, 5 pages.
European Patent Search Report for Application No. 14157971.4-1906, Dated Jun. 30, 2014, 6 pages.
Caulier, Yannick et al., “A New Type of Color-Coded Light Structures for an Adapted and Rapid Determination of Point Correspondences for 3D Reconstruction.” Proc. of SPIE, vol. 8082 808232-3; 2011; 8 pages.
Kazantsev, Aleksei et al. “Robust Pseudo-Random Coded Colored STructured Light Techniques for 3D Object Model Recovery”; Rose 2008 IEEE International Workshop on Robotic and Sensors Environments (Oct. 17-18, 2008) , 6 pages.
Mouaddib E. et al. “Recent Progress in Structured Light in order to Solve the Correspondence Problem in Stereo Vision” Proceedings of the 1997 IEEE International Conference on Robotics and Automation, Apr. 1997; 7 pages.
Second Chinese Office Action in related CN Application No. 201520810685.6, Dated Mar. 22, 2016, 5 pages.
European Search Report in related EP Application No. 15190315.0, Dated Apr. 1, 2016, 7 pages.
Second Chinese Office Action in related CN Application No. 2015220810562.2, Dated Mar. 22, 2016, 5 pages. English Translation provided.
European Search Report for related Application EP 151902491, Dated Mar. 22, 2016, 7 pages.
Second Chinese Office Action in related CN Application No. 201520810313.3, Dated Mar. 22, 2016, 5 pages. English Translation provided.
European Search Report for Related EP Application No. 15189214.8, dated Mar. 3, 2016, 9 pages.
European Search Report for related EP Application No. 15188440.0, Dated Mar. 8, 2016, 8 pages.
Search Report and Opinion in related GB Application No. 1517112.7, Dated Feb. 19, 2016, 6 Pages (GB2503978 is a commonly owned now abandoned application).
U.S. Appl. No. 14/800,757 , Eric Todeschini, filed Jul. 16, 2015, not published yet, Dimensioning and Maging Items, 80 pages.
Proesmans, Marc et al. “Active Acquisition of 3D Shape for Moving Objects” 0-7803-3258-X/96 1996 IEEE; 4 pages.
U.S. Appl. No. 14/747,197, Serge Thuries et al., filed Jun. 23, 2015, not published yet, Optical Pattern Projector; 33 pages.
U.S. Appl. No. 14/747,490, Brian L. Jovanovski et al., filed Jun. 23, 2015, not published yet, Dual-Projector Three-Dimensional Scanner; 40 pages.
U.S. Appl. No. 14/715,916, H. Sprague Ackley, filed May 19, 2015, not published yet, Evaluating Image Values; 54 pages.
U.S. Appl. No. 14/793,149, H. Sprague Ackley, filed Jul. 7, 2015, not published yet, Mobile Dimensioner Apparatus for Use in Commerce; 57 pages.
U.S. Appl. No. 14/740,373, H. Sprague Ackley et al., filed Jun. 16, 2015, not published yet, Calibrating a Volume Dimensioner; 63 pages.
U.S. Appl. No. 14/801,023, Tyler Doomenbal et al., filed Jul. 16, 2015, not published yet, Adjusting Dimensioning Results Using Augmented Reality, 39 pages.
Leotta, Matthew, Generic, Deformable Models for 3-D Vehicle Surveillance, May 2010, Doctoral Dissertation, Brown University, Providence RI, 248 pages.
Ward, Benjamin, Interactive 3D Reconstruction from Video, Aug. 2012, Doctoral Thesis, Univesity of Adelaide, Adelaide, South Australia, 157 pages.
Hood, Frederick W.; William A. Hoff, Robert King, Evaluation of an Interactive Technique for Creating Site Models from Range Data, Apr. 27-May 1, 1997 Proceedings of the ANS 7th Topical Meeting on Robotics & Remote Systems, Augusta GA, 9 pages.
Gupta, Alok; Range Image Segmentation for 3-D Objects Recognition, May 1988, Technical Reports (CIS), Paper 736, University of Pennsylvania Department of Computer and Information Science, retrieved from Http://repository.upenn.edu/cis—reports/736, Accessed May 31, 2015, 157 pages.
Reisner-Kollmann,Irene; Anton L. Fuhrmann, Werner Purgathofer, Interactive Reconstruction of Industrial Sites Using Parametric Models, May 2010, Proceedings of the 26th Spring Conference of Computer Graphics SCCG ″10, 8 pages.
Drummond, Tom; Roberto Cipolla, Real-Time Visual Tracking of Complex Structures, Jul. 2002, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, No. 7; 15 pages.
Zhang, Zhaoxiang; Tieniu Tan, Kaiqi Huang, Yunhong Wang; Three-Dimensional Deformable-Model-based Localization and Recognition of Road Vehicles; IEEE Transactions on Image Processing, vol. 21, No. 1, Jan. 2012, 13 pages.
Leotta, Matthew J.; Joseph L. Mundy; Predicting High Resolution Image Edges with a Generic, Adaptive, 3-D Vehicle Model; IEEE Conference on Computer Vision and Pattern Recognition, 2009; 8 pages.
Spiller, Jonathan; Object Localization Using Deformable Templates, Master's Dissertation, University of the Witwatersrand, Johannesburg, South Africa, 2007; 74 pages.
EP Search and Written Opinion Report in related matter EP Application No. 14181437.6, Dated Mar. 26, 2015, 7 pages.
Hetzel, Gunter et al.; “3D Object Recognition from Range Images using Local Feature Histograms,”, Proceedings 2OO1 IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2001. Kauai, Hawaii, Dec. 8-14, 2001; pp. 394-399, XP010584149, ISBN: 978-0-7695-1272-3.
Intention to Grant in counterpart European Application No. 14157971.4 dated Apr. 14, 2015, pp. 1-8.
Decision to Grant in counterpart European Application No. 14157971.4 dated Aug. 6, 2015, pp. 1-2.
Salvi, Joaquim et al. “Pattern Codification Strategies in Structured Light Systems” published in Pattern Recognition; The Journal of the Pattern Recognition Society, Received Mar. 6, 2003; Accepted Oct. 2, 2003; 23 pages.
Office Action in counterpart European Application No. 13186043.9 dated Sep. 30, 2015, pp. 1-7.
Lloyd et al., “System for Monitoring the Condition of Packages Throughout Transit”, U.S. Appl. No. 14/865,575, filed Sep. 25, 2015, 59 pages, not yet published.
James Chamberlin, “System and Method for Picking Validation”, U.S. Appl. No. 14/865,797, filed Sep. 25, 2015, 44 pages, not yet published.
Jovanovski et al., “Image-Stitching for Dimensioning”, U.S. Appl. No. 14/870,488, filed Sep. 30, 2015, 45 pages, not yet published.
Todeschini et al.; “Depth Sensor Based Auto-Focus System for an Indicia Scanner,” U.S. Appl. No. 141872,176, filed Oct. 1, 2015, 44 pages, not yet published.
Wikipedia, “3D projection” Downloaded on Nov. 25, 2015 from www.wikipedia.com, 4 pages.
McCloskey et al., “Methods for Improving the Accuracy of Dimensioning-System Measurements,” U.S. Appl. No. 14/873,613, filed Sep. 2, 2015, 47 pages, not yet published.
Search Report in counterpart European Application No. 15182675.7, Dated Dec. 4, 2015, 10 pages.
McCloskey et al., “Image Transformation for Indicia Reading,” U.S. Appl. No. 14/982,032, filed Oct. 30, 2015, 48 pages, not yet published.
Search Report and Opinion in Related EP Application 15176943.7, Dated Jan. 8, 2016, 7 pages.
Lloyd, Ryan and Scott McCloskey, “Recognition of 3D Package Shapes for Singe Camera Metrology” IEEE Winter conference on Applications of computer Visiona, IEEE, Mar. 24, 2014, pp. 99-106, {retrieved on Jun. 16, 2014}, Authors are employees of common Applicant.
Extended European Search Report in counterpart European Application No. 15182675.7 dated Dec. 4, 2015, pp. 1-10.
Great Britain Search Report for related Application on. GB1517843.7, Dated Feb. 23, 2016; 8 pages.
European Search Report for related EP Application No. 16152477.2, dated May 24, 2016, 8 pages [New Reference cited herein; Reference DE102007037282 A1 and its US Counterparts have been previously cited.].
European Partial Search Report for related EP Application No. 15190306.9, dated May 6, 2016, 8 pages.
Mike Stensvold, “get the Most Out of Variable Aperture Lenses”, published on www.OutdoorPhotogrpaher.com; dated Dec. 7, 2010; 4 pages, [As noted on search report retrieved from URL: http;//www.outdoorphotographer.com/gear/lenses/get-the-most-out-ofvariable-aperture-lenses.html on Feb. 9, 2016].
M. Zahid Gurbuz, Selim Akyokus, Ibrahim Emiroglu, Aysun Guran, An Efficient Algorithm for 3D Rectangular Box Packing, 2009, Applied Automatic Systems: Proceedings of Selected AAS 2009 Papers, pp. 131-134.
Second Chinese Office Action in related CN Application No. 201520810313.3, Dated Mar. 22, 2016, 5 pages. English Translation provided [No references].
European extended Search report in related EP Application 13785171.3, dated Sep. 19, 2016, 8 pages.
El-Hakim et al., “Multicamera vision-based approach to flexible feature measurement for inspection and reverse engineering”, published in Optical Engineering, Society of Photo-Optical Instrumentation Engineers, vol. 32, No. 9, Sep. 1,1993, 15 pages.
El-Hakim et al., “A Knowledge-based Edge/Object Measurement Technique”, Retrieved from the Internet: URL: https://www.researchgate.net/profile/Sabry—E1 -Hakim/publication/44075058—A—Knowledge—Based—EdgeObject—Measurement—Technique/links/00b4953b5faa7d3304000000.pdf [retrieved on Jul. 15, 2016] dated Jan. 1, 1993, 9 pages.
United Kingdom Search Report in related application GB1517842.9, dated Apr. 8, 2016, 8 pages.
European Extended search report in related EP Application No. 15190306.9, Dated Sep. 9, 2016, 15 pages (only new references are cited; remaining references were cited with partial search report in same application dated May 6, 2016).
Collings et al., “The Applications and Technology of Phase-Only Liquid Crystal on Silicon Devices”, Journal of Display Technology, IEEE Service Center, New, York, NY, US, vol. 7, No. 3, Mar. 1, 2011 (Mar. 1, 2011), pp. 112-119.
United Kingdom combined Search and Examination Report in related GB Application No. 1607394.2, Dated Oct. 19, 2016, 7 pages.
European Extended Search Report in Related EP Application No. 16172995.9, Dated Aug. 22, 2016, 11 pages (Only new references have been cited; U.S. Pat. No. 8,463,079 (formerly U.S. Publication 2010/0220894) and U.S. Publication 2001/0027955 have been previously cited.).
European Search Report from related EP Application No. 16168216.6, Dated Oct. 20, 2016, 8 pages [New reference cited above; U.S. Publication 2014/0104413 has been previously cited].
European Extended Search Report in related EP Application No. 16173429.8, dated Dec. 1, 2016, 8 pages [Only new references cited: US 2013/0038881 was previously cited].
Extended European Search Report in related EP Application No. 16175410.0, dated Dec. 13, 2016, 5 pages.
Padzensky, Ron; “Augmera; Gesture Control”, Dated Apr. 18, 2015, 15 pages.
Grabowski, Ralph; “New Commands in AutoCADS 2010: Part 11 Smoothing 3D Mesh Objects” Dated 2011 (per examiner who cited reference), 6 pages.
Theodoropoulos, Gabriel; “Using Gesture Recognizers to Handle Pinch, Rotate, Pan, Swipe, and Tap Gestures” dated Aug. 25, 2014, 34 pages.
Wikipedia, “Microlens”, Downloaded from https://en.wikipedia.org/wiki/Microlens, pp. 3.
Fukaya et al., “Characteristics of Speckle Random Pattern and Its Applications”, pp. 317-327, Nouv. Rev. Optique, t.6, n.6. (1975).
European extended search report in related EP Application 16190833.0, dated Mar. 9, 2017, 8 pages [only new art has been cited; US Publication 2014/0034731 was previously cited].
United Kingdom Combined Search and Examination Report in related Application No. GB1620676.5, dated Mar. 8, 2017, 6 pages [References have been previosly cited; WO2014/151746, WO2012/1175731, US2014/0313527, GB2503978].
European Exam Report in related EP Application No. 16168216.6, dated Feb. 27, 2017, 5 pages, [References have been previously cited;WO2011/017241 and US 2014/0104413].
Chinese Notive of Re-examination in related CN Application No. 201520810313.3, dated Mar. 14, 2017, 7 pages, [Translation provided; no new references].
Thorslab, Examiner Cited NPL in Advisory Action dated Apr. 12, 2017 in related commonly owned application, downloaded from https://www.thorlabs.com/newgrouppage9.cfm?objectgroup—id=6430, 4 pages.
EKSMA Optics, Examiner Cited NPL in Advisory Action dated Apr. 13, 2017 in related commonly owned application, downloaded from http://eksmaoptics.com/optical-systems/f-theta-lenses/f-theta-lens-for-1064-nm/, 2 pages.
Sill Optics, Examiner Cited NPL in Advisory Action dated Apr. 12, 2017 in related commonly owned application, http://www.silloptics.de/1/products/sill-encyclopedia/laser-optics/f-theta-lenses/, 4 pages.
Extended European search report in related EP Application 16199707.7, dated Apr. 10, 2017, 15 pages.
Ulusoy et al., One-Shot Scanning using De Bruijn Spaced Grids, 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, 7 pages [Cited In EP Extended search report dated Apr. 10, 2017].
European Written Exam in related EP Application 1516943.7, dated Apr. 12, 2017, 6 pages [Art has been previosly cited in this matter].
European Written Exam in related EP Application 15188440.0, dated Apr. 21, 2017, 6 pages [No new Art].
Ralph Grabowski, “Smothing 3D Mesh Objects,” New Commands in AutoCAD 2010: Part 11, Examiner Cited art in related matter Non Final Office Action dated May 19, 2017; 6 pages.
EP Search Report in Related EP Application No. 17171844 dated Sep. 18, 2017. 4 pages.
European Search Report in related EP Application No. 17175357.7, dated Aug. 17, 2017, pp. 1-7.
Boavida et al., “Dam monitoring using combined terrestrial imaging systems”, 2009 Civil Engineering Survey De/Jan. 2009, pp. 33-38.
Related Publications (1)
Number Date Country
20160040982 A1 Feb 2016 US