System and method for package dimensioning

Information

  • Patent Grant
  • 9464885
  • Patent Number
    9,464,885
  • Date Filed
    Monday, August 18, 2014
    10 years ago
  • Date Issued
    Tuesday, October 11, 2016
    8 years ago
Abstract
A system and method for package dimensioning is provided. The package-dimensioning system includes an image capturing subsystem for acquiring information about an object within the image-capturing subsystem's field of view. A features-computation module analyzes object information and compiles a feature set describing the object's surface features. A classification module analyzes the feature set and categorizes the object's shape. A shape-estimation module estimates the dimensions of the object.
Description
FIELD OF THE INVENTION

The present invention relates to systems and methods for determining the dimensions of packages. More particularly, the present invention relates to a dimensioning system and method for determining package dimensions having reduced cost, better accuracy, and increased speed.


BACKGROUND

In the retail shipping environment, dimensional weight (e.g., volumetric weight) is typically employed to determine the shipping rate for a particular package. Traditionally, measurements of a package's dimensions for purposes of determining dimensional weight are obtained by hand measuring the package with a measuring device such as a tape measure. This approach consumes the time of the shipping company's personnel and is also subject to measurement errors (e.g., incorrectly reading or recording the measurement).


An attractive automated system for package dimensioning has proved elusive. Obtaining an accurate three-dimensional representation of the package has typically required the use of at least two imaging devices to obtain images of the package at different angles. This approach introduces complexities into the process due to the need to assimilate the two images for analysis and, because it requires two imaging devices, it tends to be expensive.


Therefore, a need exists for an inexpensive, but accurate, automated system for determining the dimensions of objects such as packages for shipping.


SUMMARY

Accordingly, in one aspect, the present invention embraces a package-dimensioning system. The package-dimensioning system includes an image-capturing subsystem for acquiring information about an object within the image-capturing subsystem's field of view. The package-dimensioning system also includes a features-computation module for analyzing object information acquired by the image-capturing subsystem and compiling a feature set describing the object's surface features. The package-dimensioning system also includes a classification module for analyzing the feature set describing the object's surface features and for categorizing the object's shape. The package-dimensioning system also includes a shape-estimation module for estimating dimensions of the object. The shape-estimation module includes a plurality of shape-specific submodules adapted for estimating the dimensions of particular shapes.


In one embodiment, the image-capturing subsystem includes a three-dimensional range camera.


In another embodiment, the image-capturing subsystem is configured to generate a point cloud with respect to the object.


In yet another embodiment, the features-computation module is configured to analyze the point cloud.


In yet another embodiment, the features-computation module is configured to compile a relative histogram describing the object's surface features.


In yet another alternative embodiment, the features-computation module is configured to compile a relative histogram describing the object's surface features.


In yet another alternative embodiment, the classification module is configured to select the shape-specific submodule for estimating the dimensions of the object.


In another aspect, the invention embraces a package-dimensioning system that includes an image-capturing subsystem for acquiring information about a plurality of objects within the image-capturing subsystem's field of view. The package-dimensioning system also includes a features-computation module for analyzing object information acquired by the image-capturing subsystem and for compiling a feature set describing each object's surface features. The package-dimensioning system also includes a classification module for analyzing the feature set describing each object's surface features and for categorizing each object's shape. The package-dimensioning system also includes a shape-estimation module for estimating dimensions of each object. The shape-estimation module includes a plurality of shape-specific submodules adapted for estimating the dimensions of particular shapes.


In another embodiment, the features-computation module is configured to differentiate the information about the plurality of objects into discrete object-specific information.


In yet another aspect, the invention embraces a method for estimating the dimensions of a package. An image of an object is acquired. A feature set describing the object's surface features is compiled. The object's shape is categorized based on the feature set. The dimensions of the object are estimated based on the categorization of the object's shape.


The foregoing illustrative summary, as well as other exemplary objectives and/or advantages of the invention, and the manner in which the same are accomplished, are further explained within the following detailed description and its accompanying drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram illustrating an exemplary package-dimensioning system according to the present invention.



FIG. 2 is a block diagram illustrating an exemplary alternative embodiment of an exemplary package-dimensioning system according to the present invention.



FIG. 3 is a block diagram illustrating an exemplary package-dimensioning system according to the present invention.





DETAILED DESCRIPTION

The present invention embraces a package-dimensioning system. The package-dimensioning system is typically used in measuring an object's dimensions for the purpose of determining the appropriate shipping rate to charge in shipping transactions. In retail shipping operations, in particular, the dimensions of an object (e.g., package, box, container, shipping tube, etc.) are often used to determine the shipping fare where the shipping fare is based on the dimensional weight of an object.


Traditionally, the dimensional weight is calculated by hand-measuring the dimensions of the object to be shipped (e.g., with a tape measure). Hand-measuring tends to be slow, personnel-intensive, and subject to human error. Automation of the dimensioning process would alleviate these shortcomings of the hand-measuring approach. Because the estimated dimensions are used for a commercial transaction, most countries require certification of the system's accuracy under typical operating conditions (e.g., the intended-use case). Consequently, the automation must provide reliable, accurate dimensioning of objects.


Three-dimensional cameras (e.g., 3D cameras), such as range cameras, can be utilized effectively in dimensioning applications. The high cost of these range cameras, however, has precluded their widespread use. The recent advent of relatively low-cost range cameras has afforded greater opportunity for implementing automated dimensioning on a wider scale. Consequently, the package-dimensioning system according to the present invention uses an image capturing subsystem (e.g., a range camera) to acquire a two-dimensional image showing the distance to points within the field of view. The acquired image (e.g., range image) typically has pixel values which correspond to the distance from the range camera. For example, brighter pixels indicate the point is a shorter distance away than points represented by darker pixels. In this way, the range image is a two-dimensional image that provides depth information regarding a three-dimensional scene.


The image acquired by the image capturing subsystem is analyzed to identify the object within the image and to identify certain surface features of the object. The package-dimensioning system analyzes the surface features to categorize the shape of the object (e.g., cuboid, cylinder, or prism) and to estimate its dimensions. Typically, the processing of the image that is acquired by the image capturing subsystem is performed by a computer, which typically has a central processing unit (CPU) and a memory.


Referring now to FIGS. 1 and 2, the package-dimensioning system 100 according to the present invention includes an image-capturing subsystem 110. The image-capturing subsystem 110 acquires information about an object within the image-capturing subsystem's 110 field of view. Typically, the image-capturing subsystem includes an imaging device (e.g., camera, stereo camera, range camera 110A, lidar). The imaging device may be any sensor that provides information from which a computer can build an organized 3D point cloud. An organized point cloud has 3D points arranged in a 2D matrix, which corresponds to the rasterization of the range image. Typically, the image-capturing system is configured to generate a point cloud with respect to the object within its field of view. The point cloud usually contains information regarding the positioning of points in a three-dimensional space (e.g., X, Y, Z coordinates) within the field of view of the image-capturing subsystem 110 when the range image is acquired, including points on the surface of the object 112. The information regarding the positioning of points on the surface of the object is referred to as object information.


A features-computation module 120 analyzes object information acquired by the image-capturing subsystem 110. Based upon the analysis of the object information, the features-computation module 120 compiles a feature set describing the object's 112 features. Compiling a feature set typically involves consideration of two features of points on the object 112: curvature c and orientation θ relative to the ground plane. The orientation of each point relative to the ground plane is typically measured by the angle, θ, between the local surface normal and the ground's normal. In other words, the feature set for each object 112 includes computations of two features, curvature c and orientation θ, for each point in the point cloud representing the surface of the object 112.


Object information relating to an image of a regular box, for instance, will be dominated by points with zero curvature (c=0) and with orientations parallel or orthogonal to the ground (θ=π/2 radians or θ=0 radians). By way of further example, object information relating to an image of a cylinder lying flat will have non-zero curvature (c>0) and a continuous range of orientation with respect to the ground (e.g., θ=[0 radians, π radians]). The feature set as a whole, therefore, describes an object's surface features (e.g., planar surfaces, curved surfaces). Typically, the feature set takes the form of a relative histogram (e.g., 2D relative histogram). By way of example, a typical histogram might use (i) ten uniform bins for curvature c, ranging from 0 to 0.08 and (ii) 20 uniform bins for orientation θ, varying from 0 to π radians.


The package-dimensioning system 100 according to the present invention also includes a classification module 130. The classification module 130 analyzes the feature set (e.g., relative histogram) describing the object's surface features. Based on the analysis of the feature set, the classification module 130 categorizes the object's shape. Typically, the system 100 takes advantage of the fact that the domain of shipped objects 112 is generally limited, with the vast majority of shipped objects 112 being cuboids, cylinders, or prisms. Given this, the classification module 130 is typically limited to a relatively small (e.g., between about 4 and 8) number of shape categories. By way of example, and without intending to limit the disclosure, the classification module 130 may be configured to categorize an object's shape as a rectangular box, a right circular cylinder lying flat, a right circular cylinder standing vertically, a right regular prism with triangular bases lying flat, or a right regular prism with triangular bases standing vertically. Providing a limited number of potential shape categories increases the likelihood that the classification module 130 will be able to successfully categorize the object's 112 shape because the differences between the feature sets associated with each shape category are sufficiently distinct to avoid shape confusion (e.g., tendency to associate more than one shape with a given feature set). This approach of limiting the number of shapes categories therefore increases system accuracy while increasing system speed through decreased computational complexity (e.g., by limiting the number of shape categories under consideration when attempting to match a feature set with a shape category).


The package-dimensioning system 100 also includes a shape-estimation module 140. The shape-estimation module 140 estimates the dimensions of the object 112. Typically, the shape-estimation module includes a plurality of shape-specific submodules 145 adapted for estimating the dimensions of particular shapes. For example, and without intending to limit the disclosure to any particular embodiment, upon the classification module's 130 categorization of the object's 112 shape as a rectangular box (e.g., by analysis of the feature set derived from the point cloud associated with the object's surface), the shape-estimation module 140 would utilize the shape-specific submodule 145 adapted for estimating the dimensions of rectangular boxes. In this example, the rectangular box version of the shape-specific submodule 145 would analyze the object information and output an estimate of the dimensions (e.g., length, width, and height) of the rectangular box object. Typically, the shape-estimation module 140 utilizes object information relating to the distance of various points on the object from the range camera's 110A to estimate the object's dimensions.


Referring now to FIG. 3, the package-dimensioning system 100 embraced by the present invention may be adapted to estimate the dimensions of a plurality of objects 112 within the field of view of the image-capturing subsystem 110. To accomplish this, the system 100 typically must segment the objects 112 that are above the ground plane (e.g., supporting surface) within the field of view of the image-capturing subsystem 110. Typically, Euclidian clustering is employed to segment the objects 112 in an organized point cloud, whereupon the system 100 processes each segmented portion of the point cloud associated with a distinct object in the manner described above.


To supplement the present disclosure, this application incorporates entirely by reference the following patents, patent application publications, and patent applications: U.S. Pat. No. 6,832,725; U.S. Pat. No. 7,159,783; U.S. Pat. No. 7,128,266; U.S. Pat. No. 7,413,127; U.S. Pat. No. 7,726,575; U.S. Pat. No. 8,390,909; U.S. Pat. No. 8,294,969; U.S. Pat. No. 8,408,469; U.S. Pat. No. 8,408,468; U.S. Pat. No. 8,381,979; U.S. Pat. No. 8,408,464; U.S. Pat. No. 8,317,105; U.S. Pat. No. 8,366,005; U.S. Pat. No. 8,424,768; U.S. Pat. No. 8,322,622; U.S. Pat. No. 8,371,507; U.S. Pat. No. 8,376,233; U.S. Pat. No. 8,457,013; U.S. Pat. No. 8,448,863; U.S. Pat. No. 8,459,557; U.S. Pat. No. 8,469,272; U.S. Pat. No. 8,474,712; U.S. Pat. No. 8,479,992; U.S. Pat. No. 8,490,877; U.S. Pat. No. 8,517,271; U.S. Pat. No. 8,523,076; U.S. Pat. No. 8,528,819; U.S. Patent Application Publication No. 2012/0111946; U.S. Patent Application Publication No. 2012/0223141; U.S. Patent Application Publication No. 2012/0193423; U.S. Patent Application Publication No. 2012/0203647; U.S. Patent Application Publication No. 2012/0248188; U.S. Patent Application Publication No. 2012/0228382; U.S. Patent Application Publication No. 2012/0193407; U.S. Patent Application Publication No. 2012/0168511; U.S. Patent Application Publication No. 2012/0168512; U.S. Patent Application Publication No. 2010/0177749; U.S. Patent Application Publication No. 2010/0177080; U.S. Patent Application Publication No. 2010/0177707; U.S. Patent Application Publication No. 2010/0177076; U.S. Patent Application Publication No. 2009/0134221; U.S. Patent Application Publication No. 2012/0318869; U.S. Patent Application Publication No. 2013/0043312; U.S. Patent Application Publication No. 2013/0068840; U.S. Patent Application Publication No. 2013/0070322; U.S. Patent Application Publication No. 2013/0075168; U.S. Patent Application Publication No. 2013/0056285; U.S. Patent Application Publication No. 2013/0075464; U.S. Patent Application Publication No. 2013/0082104; U.S. Patent Application Publication No. 2010/0225757; U.S. Patent Application Publication No. 2013/0175343; U.S. patent application Ser. No. 13/347,193 for a HYBRID-TYPE BIOPTICAL LASER SCANNING AND DIGITAL IMAGING SYSTEM EMPLOYING DIGITAL IMAGER WITH FIELD OF VIEW OVERLAPPING FIELD OF FIELD OF LASER SCANNING SUBSYSTEM, filed Jan. 10, 2012 (Kearney et al.); U.S. patent application Ser. No. 13/367,047 for LASER SCANNING MODULES EMBODYING SILICONE SCAN ELEMENT WITH TORSIONAL HINGES, filed Feb. 6, 2012 (Feng et al.); U.S. patent application Ser. No. 13/400,748 for a LASER SCANNING BAR CODE SYMBOL READING SYSTEM HAVING INTELLIGENT SCAN SWEEP ANGLE ADJUSTMENT CAPABILITIES OVER THE WORKING RANGE OF THE SYSTEM FOR OPTIMIZED BAR CODE SYMBOL READING PERFORMANCE, filed Feb. 21, 2012 (Wilz); U.S. patent application Ser. No. 13/432,197 for a LASER SCANNING SYSTEM USING LASER BEAM SOURCES FOR PRODUCING LONG AND SHORT WAVELENGTHS IN COMBINATION WITH BEAM-WAIST EXTENDING OPTICS TO EXTEND THE DEPTH OF FIELD THEREOF WHILE RESOLVING HIGH RESOLUTION BAR CODE SYMBOLS HAVING MINIMUM CODE ELEMENT WIDTHS, filed Mar. 28, 2012 (Havens et al.); U.S. patent application Ser. No. 13/492,883 for a LASER SCANNING MODULE WITH ROTATABLY ADJUSTABLE LASER SCANNING ASSEMBLY, filed Jun. 10, 2012 (Hennick et al.); U.S. patent application Ser. No. 13/367,978 for a LASER SCANNING MODULE EMPLOYING AN ELASTOMERIC U-HINGE BASED LASER SCANNING ASSEMBLY, filed Feb. 7, 2012 (Feng et al.); U.S. patent application Ser. No. 13/852,097 for a System and Method for Capturing and Preserving Vehicle Event Data, filed Mar. 28, 2013 (Barker et al.); U.S. patent application Ser. No. 13/780,356 for a Mobile Device Having Object-Identification Interface, filed Feb. 28, 2013 (Samek et al.); U.S. patent application Ser. No. 13/780,158 for a Distraction Avoidance System, filed Feb. 28, 2013 (Sauerwein); U.S. patent application Ser. No. 13/784,933 for an Integrated Dimensioning and Weighing System, filed Mar. 5, 2013 (McCloskey et al.); U.S. patent application Ser. No. 13/785,177 for a Dimensioning System, filed Mar. 5, 2013 (McCloskey et al.); U.S. patent application Ser. No. 13/780,196 for Android Bound Service Camera Initialization, filed Feb. 28, 2013 (Todeschini et al.); U.S. patent application Ser. No. 13/792,322 for a Replaceable Connector, filed Mar. 11, 2013 (Skvoretz); U.S. patent application Ser. No. 13/780,271 for a Vehicle Computer System with Transparent Display, filed Feb. 28, 2013 (Fitch et al.); U.S. patent application Ser. No. 13/736,139 for an Electronic Device Enclosure, filed Jan. 8, 2013 (Chaney); U.S. patent application Ser. No. 13/771,508 for an Optical Redirection Adapter, filed Feb. 20, 2013 (Anderson); U.S. patent application Ser. No. 13/750,304 for Measuring Object Dimensions Using Mobile Computer, filed Jan. 25, 2013; U.S. patent application Ser. No. 13/471,973 for Terminals and Methods for Dimensioning Objects, filed May 15, 2012; U.S. patent application Ser. No. 13/895,846 for a Method of Programming a Symbol Reading System, filed Apr. 10, 2013 (Corcoran); U.S. patent application Ser. No. 13/867,386 for a Point of Sale (POS) Based Checkout System Supporting a Customer-Transparent Two-Factor Authentication Process During Product Checkout Operations, filed Apr. 22, 2013 (Cunningham et al.); U.S. patent application Ser. No. 13/888,884 for an Indicia Reading System Employing Digital Gain Control, filed May 7, 2013 (Xian et al.); U.S. patent application Ser. No. 13/895,616 for a Laser Scanning Code Symbol Reading System Employing Multi-Channel Scan Data Signal Processing with Synchronized Digital Gain Control (SDGC) for Full Range Scanning, filed May 16, 2013 (Xian et al.); U.S. patent application Ser. No. 13/897,512 for a Laser Scanning Code Symbol Reading System Providing Improved Control over the Length and Intensity Characteristics of a Laser Scan Line Projected Therefrom Using Laser Source Blanking Control, filed May 20, 2013 (Brady et al.); U.S. patent application Ser. No. 13/897,634 for a Laser Scanning Code Symbol Reading System Employing Programmable Decode Time-Window Filtering, filed May 20, 2013 (Wilz, Sr. et al.); U.S. patent application Ser. No. 13/902,242 for a System For Providing A Continuous Communication Link With A Symbol Reading Device, filed May 24, 2013 (Smith et al.); U.S. patent application Ser. No. 13/902,144, for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Chamberlin); U.S. patent application Ser. No. 13/902,110 for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Hollifield); U.S. patent application Ser. No. 13/912,262 for a Method of Error Correction for 3D Imaging Device, filed Jun. 7, 2013 (Jovanovski et al.); U.S. patent application Ser. No. 13/912,702 for a System and Method for Reading Code Symbols at Long Range Using Source Power Control, filed Jun. 7, 2013 (Xian et al.); U.S. patent application Ser. No. 13/922,339 for a System and Method for Reading Code Symbols Using a Variable Field of View, filed Jun. 20, 2013 (Xian et al.); U.S. patent application Ser. No. 13/927,398 for a Code Symbol Reading System Having Adaptive Autofocus, filed Jun. 26, 2013 (Todeschini); U.S. patent application Ser. No. 13/930,913 for a Mobile Device Having an Improved User Interface for Reading Code Symbols, filed Jun. 28, 2013 (Gelay et al.); U.S. patent application Ser. No. 13/933,415 for an Electronic Device Case, filed Jul. 2, 2013 (London et al.); U.S. patent application Ser. No. 13/947,296 for a System and Method for Selectively Reading Code Symbols, filed Jul. 22, 2013 (Rueblinger et al.); U.S. patent application Ser. No. 13/950,544 for a Code Symbol Reading System Having Adjustable Object Detection, filed Jul. 25, 2013 (Jiang); U.S. patent application Ser. No. 13/961,408 for a Method for Manufacturing Laser Scanners, filed Aug. 7, 2013 (Saber et al.); U.S. patent application Ser. No. 13/973,315 for a Symbol Reading System Having Predictive Diagnostics, filed Aug. 22, 2013 (Nahill et al.); U.S. patent application Ser. No. 13/973,354 for a Pairing Method for Wireless Scanner via RFID, filed Aug. 22, 2013 (Wu et al.); and U.S. patent application Ser. No. 13/974,374 for Authenticating Parcel Consignees with Indicia Decoding Devices, filed Aug. 23, 2013 (Ye et al.).


* * *

In the specification and/or figures, typical embodiments of the invention have been disclosed. The present invention is not limited to such exemplary embodiments. The use of the term “and/or” includes any and all combinations of one or more of the associated listed items. The figures are schematic representations and so are not necessarily drawn to scale. Unless otherwise noted, specific terms have been used in a generic and descriptive sense and not for purposes of limitation.

Claims
  • 1. A package-dimensioning system, comprising: an imager for acquiring information about an object within the imager field of view;a features-computation module for analyzing object information acquired by the imager and compiling a feature set describing the object's surface features;a classification module for analyzing the feature set describing the object's surface features and for categorizing the object's shape; anda shape-estimation module for estimating dimensions of the object, the shape-estimation module comprising a plurality of shape-specific submodules adapted for estimating the dimensions of particular shapes.
  • 2. The package-dimensioning system according to claim 1, wherein the imager comprises a range camera.
  • 3. The package-dimensioning system according to claim 1, wherein the imager is configured to generate a range image with respect to the object.
  • 4. The package-dimensioning system according to claim 3, wherein the features-computation module is configured to analyze the range image.
  • 5. The package-dimensioning system according to claim 1, wherein the features-computation module is configured to compile a relative histogram describing the object's surface features.
  • 6. The package-dimensioning system according to claim 1, wherein the classification module is configured to select the shape-specific submodule for estimating the dimensions of the object.
  • 7. A package-dimensioning system, comprising: an imager for acquiring information about a plurality of objects within the imager field of view;a features-computation module for analyzing object information acquired by the imager and for compiling a feature set describing each object's surface features;a classification module for analyzing the feature set describing each object's surface features and for categorizing each object's shape; anda shape-estimation module for estimating dimensions of each object, the shape-estimation module comprising a plurality of shape-specific submodules adapted for estimating the dimensions of particular shapes.
  • 8. The package-dimensioning system according to claim 7, wherein the imager comprises a range camera.
  • 9. The package-dimensioning system according to claim 7, wherein the imager is configured to generate a range image with respect to each object.
  • 10. The package-dimensioning system according to claim 9, wherein the features-computation module is configured to analyze the range image associated with each object.
  • 11. The package-dimensioning system according to claim 7, wherein the features-computation module is configured to compile a relative histogram describing each object's surface features.
  • 12. The package-dimensioning system according to claim 7, wherein the classification module is configured to select the shape-specific submodule for estimating the dimensions of each object.
  • 13. The package-dimensioning system according to claim 7, wherein the features-computation module is configured to differentiate the information about the plurality of objects into discrete object-specific information.
  • 14. A method for estimating the dimensions of a package, comprising: acquiring an image of an object with an imager;compiling a feature set describing the object's surface features with a features-computation module;categorizing the object's shape based on the feature set with a classification module; andestimating the dimensions of the object based on the categorization of the object's shape using a shape-estimation module comprising a plurality of shape-specific submodules adapted for estimating the dimensions of particular shapes.
  • 15. The method according to claim 14, wherein the image of an object is acquired using a range camera.
  • 16. The method according to claim 14, wherein the feature set comprises a relative histogram describing the object's surface features.
CROSS-REFERENCE TO RELATED APPLICATION

The present application claims the benefit of U.S. Patent Application No. 61/872,299 for a System and Method for Package Dimensioning filed Aug. 30, 2013. The foregoing patent application is hereby incorporated by reference in its entirety.

US Referenced Citations (426)
Number Name Date Kind
3971065 Bayer Jul 1976 A
4398811 Nishioka et al. Aug 1983 A
4495559 Gelatt, Jr. Jan 1985 A
4730190 Win et al. Mar 1988 A
5220536 Stringer et al. Jun 1993 A
5331118 Jensen Jul 1994 A
5359185 Hanson Oct 1994 A
5548707 LoNegro Aug 1996 A
5555090 Schmutz Sep 1996 A
5590060 Granville et al. Dec 1996 A
5606534 Stringer et al. Feb 1997 A
5655095 LoNegro et al. Aug 1997 A
5661561 Wurz et al. Aug 1997 A
5699161 Woodworth Dec 1997 A
5729750 Ishida Mar 1998 A
5730252 Herbinet Mar 1998 A
5734476 Dlugos Mar 1998 A
5737074 Haga et al. Apr 1998 A
5831737 Stringer et al. Nov 1998 A
5850370 Stringer et al. Dec 1998 A
5869827 Rando Feb 1999 A
5870220 Migdal et al. Feb 1999 A
5900611 Hecht May 1999 A
5923428 Woodworth Jul 1999 A
5929856 LoNegro et al. Jul 1999 A
5969823 Wurz et al. Oct 1999 A
5978512 Kim et al. Nov 1999 A
5979760 Freyman et al. Nov 1999 A
5991041 Woodworth Nov 1999 A
6025847 Marks Feb 2000 A
6049386 Stringer et al. Apr 2000 A
6053409 Brobst et al. Apr 2000 A
6064759 Buckley et al. May 2000 A
6067110 Nonaka et al. May 2000 A
6069696 McQueen et al. May 2000 A
6137577 Woodworth Oct 2000 A
6177999 Wurz et al. Jan 2001 B1
6236403 Chaki May 2001 B1
6246468 Dimsdale Jun 2001 B1
6336587 He et al. Jan 2002 B1
6369401 Lee Apr 2002 B1
6373579 Ober et al. Apr 2002 B1
6429803 Kumar Aug 2002 B1
6457642 Good et al. Oct 2002 B1
6507406 Yagi et al. Jan 2003 B1
6517004 Good et al. Feb 2003 B2
6674904 McQueen Jan 2004 B1
6705526 Zhu et al. Mar 2004 B1
6781621 Gobush et al. Aug 2004 B1
6824058 Patel et al. Nov 2004 B2
6832725 Gardiner et al. Dec 2004 B2
6858857 Pease et al. Feb 2005 B2
6971580 Zhu et al. Dec 2005 B2
6995762 Pavlidis et al. Feb 2006 B1
7085409 Sawhney et al. Aug 2006 B2
7086162 Tyroler Aug 2006 B2
7104453 Zhu et al. Sep 2006 B1
7128266 Zhu et al. Oct 2006 B2
7137556 Bonner et al. Nov 2006 B1
7159783 Walczyk et al. Jan 2007 B2
7161688 Bonner et al. Jan 2007 B1
7214954 Schopp May 2007 B2
7277187 Smith et al. Oct 2007 B2
7307653 Dutta Dec 2007 B2
7310431 Gokturk et al. Dec 2007 B2
7413127 Ehrhart et al. Aug 2008 B2
7527205 Zhu et al. May 2009 B2
7586049 Wurz Sep 2009 B2
7602404 Reinhardt et al. Oct 2009 B1
7726575 Wang et al. Jun 2010 B2
7780084 Zhang et al. Aug 2010 B2
7788883 Buckley et al. Sep 2010 B2
7974025 Topliss Jul 2011 B2
8027096 Feng et al. Sep 2011 B2
8028501 Buckley et al. Oct 2011 B2
8050461 Shpunt et al. Nov 2011 B2
8055061 Katano Nov 2011 B2
8102395 Kondo et al. Jan 2012 B2
8132728 Dwinell et al. Mar 2012 B2
8134717 Pangrazio et al. Mar 2012 B2
8149224 Kuo et al. Apr 2012 B1
8194097 Xiao et al. Jun 2012 B2
8212889 Chanas et al. Jul 2012 B2
8228510 Pangrazio et al. Jul 2012 B2
8294969 Plesko Oct 2012 B2
8305458 Hara Nov 2012 B2
8310656 Zalewski Nov 2012 B2
8313380 Zalewski et al. Nov 2012 B2
8317105 Kotlarsky et al. Nov 2012 B2
8322622 Liu Dec 2012 B2
8339462 Stec et al. Dec 2012 B2
8350959 Topliss et al. Jan 2013 B2
8366005 Kotlarsky et al. Feb 2013 B2
8371507 Haggerty et al. Feb 2013 B2
8376233 Horn et al. Feb 2013 B2
8381976 Mohideen et al. Feb 2013 B2
8381979 Franz Feb 2013 B2
8390909 Plesko Mar 2013 B2
8408464 Zhu et al. Apr 2013 B2
8408468 Van Horn et al. Apr 2013 B2
8408469 Good Apr 2013 B2
8424768 Rueblinger et al. Apr 2013 B2
8437539 Komatsu et al. May 2013 B2
8441749 Brown et al. May 2013 B2
8448863 Xian et al. May 2013 B2
8457013 Essinger et al. Jun 2013 B2
8459557 Havens et al. Jun 2013 B2
8463079 Ackley et al. Jun 2013 B2
8469272 Kearney Jun 2013 B2
8474712 Kearney et al. Jul 2013 B2
8479992 Kotlarsky et al. Jul 2013 B2
8490877 Kearney Jul 2013 B2
8517271 Kotlarsky et al. Aug 2013 B2
8523076 Good Sep 2013 B2
8528818 Ehrhart et al. Sep 2013 B2
8544737 Gomez et al. Oct 2013 B2
8548420 Grunow et al. Oct 2013 B2
8550335 Samek et al. Oct 2013 B2
8550354 Gannon et al. Oct 2013 B2
8550357 Kearney Oct 2013 B2
8556174 Kosecki et al. Oct 2013 B2
8556176 Van Horn et al. Oct 2013 B2
8556177 Hussey et al. Oct 2013 B2
8559767 Barber et al. Oct 2013 B2
8561895 Gomez et al. Oct 2013 B2
8561903 Sauerwein Oct 2013 B2
8561905 Edmonds et al. Oct 2013 B2
8565107 Pease et al. Oct 2013 B2
8570343 Halstead Oct 2013 B2
8571307 Li et al. Oct 2013 B2
8576390 Nunnink Nov 2013 B1
8579200 Samek et al. Nov 2013 B2
8583924 Caballero et al. Nov 2013 B2
8584945 Wang et al. Nov 2013 B2
8587595 Wang Nov 2013 B2
8587697 Hussey et al. Nov 2013 B2
8588869 Sauerwein et al. Nov 2013 B2
8590789 Nahill et al. Nov 2013 B2
8596539 Havens et al. Dec 2013 B2
8596542 Havens et al. Dec 2013 B2
8596543 Havens et al. Dec 2013 B2
8599271 Havens et al. Dec 2013 B2
8599957 Peake et al. Dec 2013 B2
8600158 Li et al. Dec 2013 B2
8600167 Showering Dec 2013 B2
8602309 Longacre et al. Dec 2013 B2
8608053 Meier et al. Dec 2013 B2
8608071 Liu et al. Dec 2013 B2
8611309 Wang et al. Dec 2013 B2
8615487 Gomez et al. Dec 2013 B2
8621123 Caballero Dec 2013 B2
8622303 Meier et al. Jan 2014 B2
8628013 Ding Jan 2014 B2
8628015 Wang et al. Jan 2014 B2
8628016 Winegar Jan 2014 B2
8629926 Wang Jan 2014 B2
8630491 Longacre et al. Jan 2014 B2
8635309 Berthiaume et al. Jan 2014 B2
8636200 Kearney Jan 2014 B2
8636212 Nahill et al. Jan 2014 B2
8636215 Ding et al. Jan 2014 B2
8636224 Wang Jan 2014 B2
8638806 Wang et al. Jan 2014 B2
8640958 Lu et al. Feb 2014 B2
8640960 Wang et al. Feb 2014 B2
8643717 Li et al. Feb 2014 B2
8646692 Meier et al. Feb 2014 B2
8646694 Wang et al. Feb 2014 B2
8657200 Ren et al. Feb 2014 B2
8659397 Vargo et al. Feb 2014 B2
8668149 Good Mar 2014 B2
8678285 Kearney Mar 2014 B2
8678286 Smith et al. Mar 2014 B2
8682077 Longacre Mar 2014 B1
D702237 Oberpriller et al. Apr 2014 S
8687282 Feng et al. Apr 2014 B2
8692927 Pease et al. Apr 2014 B2
8695880 Bremer et al. Apr 2014 B2
8698949 Grunow et al. Apr 2014 B2
8702000 Barber et al. Apr 2014 B2
8717494 Gannon May 2014 B2
8720783 Biss et al. May 2014 B2
8723804 Fletcher et al. May 2014 B2
8723904 Marty et al. May 2014 B2
8727223 Wang May 2014 B2
8810779 Hilde Aug 2014 B1
20010027995 Patel et al. Oct 2001 A1
20010032879 He et al. Oct 2001 A1
20020118874 Chung et al. Aug 2002 A1
20020158873 Williamson Oct 2002 A1
20020167677 Okada et al. Nov 2002 A1
20020179708 Zhu et al. Dec 2002 A1
20030038179 Tsikos et al. Feb 2003 A1
20030053513 Vatan et al. Mar 2003 A1
20030063086 Baumberg Apr 2003 A1
20030091227 Chang et al. May 2003 A1
20030156756 Gokturk et al. Aug 2003 A1
20030197138 Pease et al. Oct 2003 A1
20030235331 Kawaike et al. Dec 2003 A1
20040008259 Gokturk et al. Jan 2004 A1
20040024754 Mane et al. Feb 2004 A1
20040066329 Zeitfuss et al. Apr 2004 A1
20040118928 Patel et al. Jun 2004 A1
20040155975 Hart et al. Aug 2004 A1
20040165090 Ning Aug 2004 A1
20040184041 Schopp Sep 2004 A1
20040211836 Patel et al. Oct 2004 A1
20040233461 Armstrong et al. Nov 2004 A1
20050006477 Patel Jan 2005 A1
20050117215 Lange Jun 2005 A1
20050128196 Popescu et al. Jun 2005 A1
20060112023 Horhann May 2006 A1
20060151604 Zhu et al. Jul 2006 A1
20060159307 Anderson et al. Jul 2006 A1
20060159344 Shao et al. Jul 2006 A1
20060232681 Okada Oct 2006 A1
20060255150 Longacre Nov 2006 A1
20060269165 Viswanathan Nov 2006 A1
20070031064 Zhao et al. Feb 2007 A1
20070063048 Havens et al. Mar 2007 A1
20070127022 Cohen et al. Jun 2007 A1
20070171220 Kriveshko Jul 2007 A1
20070181685 Zhu et al. Aug 2007 A1
20070237356 Dwinell et al. Oct 2007 A1
20070299338 Stevick et al. Dec 2007 A1
20080013793 Hillis et al. Jan 2008 A1
20080035390 Wurz Feb 2008 A1
20080056536 Hildreth et al. Mar 2008 A1
20080077265 Boyden Mar 2008 A1
20080164074 Wurz Jul 2008 A1
20080185432 Caballero et al. Aug 2008 A1
20080212168 Olmstead et al. Sep 2008 A1
20080247635 Davis et al. Oct 2008 A1
20080273191 Kim et al. Nov 2008 A1
20080278790 Boesser et al. Nov 2008 A1
20090059004 Bochicchio Mar 2009 A1
20090134221 Zhu et al. May 2009 A1
20090225333 Bendall et al. Sep 2009 A1
20090237411 Gossweiler et al. Sep 2009 A1
20090268023 Hsieh Oct 2009 A1
20090272724 Gubler Nov 2009 A1
20090273770 Bauhahn et al. Nov 2009 A1
20090313948 Buckley et al. Dec 2009 A1
20090323084 Dunn et al. Dec 2009 A1
20100035637 Varanasi et al. Feb 2010 A1
20100128109 Banks May 2010 A1
20100161170 Siris Jun 2010 A1
20100172567 Prokoski Jul 2010 A1
20100177076 Essinger et al. Jul 2010 A1
20100177080 Essinger et al. Jul 2010 A1
20100177707 Essinger et al. Jul 2010 A1
20100177749 Essinger et al. Jul 2010 A1
20100202702 Benos et al. Aug 2010 A1
20100208039 Stettner Aug 2010 A1
20100217678 Goncalves Aug 2010 A1
20100220894 Ackley et al. Sep 2010 A1
20100254611 Arnz Oct 2010 A1
20100303336 Abraham Dec 2010 A1
20110040192 Brenner et al. Feb 2011 A1
20110043609 Choi et al. Feb 2011 A1
20110099474 Grossman et al. Apr 2011 A1
20110169999 Grunow et al. Jul 2011 A1
20110202554 Powilleit et al. Aug 2011 A1
20110249864 Venkatesan et al. Oct 2011 A1
20110254840 Halstead Oct 2011 A1
20110279916 Brown et al. Nov 2011 A1
20110286007 Pangrazio et al. Nov 2011 A1
20110288818 Thierman Nov 2011 A1
20110301994 Tieman Dec 2011 A1
20120024952 Chen Feb 2012 A1
20120056982 Katz et al. Mar 2012 A1
20120074227 Ferren et al. Mar 2012 A1
20120081714 Pangrazio et al. Apr 2012 A1
20120111946 Golant May 2012 A1
20120138685 Qu et al. Jun 2012 A1
20120140300 Freeman Jun 2012 A1
20120168511 Kotlarsky et al. Jul 2012 A1
20120168512 Kotlarsky et al. Jul 2012 A1
20120179665 Baarman et al. Jul 2012 A1
20120185094 Rosenstein et al. Jul 2012 A1
20120193407 Barten Aug 2012 A1
20120193423 Samek Aug 2012 A1
20120197464 Wang et al. Aug 2012 A1
20120203647 Smith Aug 2012 A1
20120218436 Rodriguez et al. Aug 2012 A1
20120223141 Good et al. Sep 2012 A1
20120224026 Bayer et al. Sep 2012 A1
20120228382 Havens et al. Sep 2012 A1
20120242852 Hayward et al. Sep 2012 A1
20120248188 Kearney Oct 2012 A1
20120262558 Boger et al. Oct 2012 A1
20120280908 Rhoads et al. Nov 2012 A1
20120282905 Owen Nov 2012 A1
20120282911 Davis et al. Nov 2012 A1
20120284012 Rodriguez et al. Nov 2012 A1
20120284122 Brandis Nov 2012 A1
20120284339 Rodriguez Nov 2012 A1
20120284593 Rodriguez Nov 2012 A1
20120293610 Doepke et al. Nov 2012 A1
20120294549 Doepke Nov 2012 A1
20120299961 Ramkumar et al. Nov 2012 A1
20120300991 Free Nov 2012 A1
20120313848 Galor et al. Dec 2012 A1
20120314030 Datta Dec 2012 A1
20120314058 Bendall et al. Dec 2012 A1
20120316820 Nakazato et al. Dec 2012 A1
20130043312 Van Horn Feb 2013 A1
20130050426 Sarmast et al. Feb 2013 A1
20130056285 Meagher Mar 2013 A1
20130070322 Fritz et al. Mar 2013 A1
20130075168 Amundsen et al. Mar 2013 A1
20130082104 Kearney et al. Apr 2013 A1
20130094069 Lee et al. Apr 2013 A1
20130101158 Lloyd et al. Apr 2013 A1
20130175341 Kearney et al. Jul 2013 A1
20130175343 Good Jul 2013 A1
20130200158 Feng et al. Aug 2013 A1
20130214048 Wilz Aug 2013 A1
20130256418 Havens et al. Oct 2013 A1
20130257744 Daghigh et al. Oct 2013 A1
20130257759 Daghigh Oct 2013 A1
20130270346 Xian et al. Oct 2013 A1
20130278425 Cunningham et al. Oct 2013 A1
20130287258 Kearney Oct 2013 A1
20130291998 Konnerth Nov 2013 A1
20130292474 Xian et al. Nov 2013 A1
20130292475 Kotlarsky et al. Nov 2013 A1
20130292477 Hennick et al. Nov 2013 A1
20130293539 Hunt et al. Nov 2013 A1
20130293540 Laffargue et al. Nov 2013 A1
20130306728 Thuries et al. Nov 2013 A1
20130306730 Brady et al. Nov 2013 A1
20130306731 Pedraro Nov 2013 A1
20130306734 Xian et al. Nov 2013 A1
20130307964 Bremer et al. Nov 2013 A1
20130308013 Li et al. Nov 2013 A1
20130308625 Corcoran Nov 2013 A1
20130313324 Koziol et al. Nov 2013 A1
20130313325 Wilz et al. Nov 2013 A1
20130313326 Ehrhart Nov 2013 A1
20130327834 Hennick et al. Dec 2013 A1
20130329012 Bartos Dec 2013 A1
20130341399 Xian et al. Dec 2013 A1
20130342717 Havens et al. Dec 2013 A1
20140001267 Giordano et al. Jan 2014 A1
20140002828 Laffargue et al. Jan 2014 A1
20140008430 Soule et al. Jan 2014 A1
20140008439 Wang Jan 2014 A1
20140021256 Qu et al. Jan 2014 A1
20140021259 Moed et al. Jan 2014 A1
20140025584 Liu et al. Jan 2014 A1
20140027518 Edmonds et al. Jan 2014 A1
20140034723 Van Horn et al. Feb 2014 A1
20140034731 Gao et al. Feb 2014 A1
20140034734 Sauerwein Feb 2014 A1
20140036848 Pease et al. Feb 2014 A1
20140039693 Havens et al. Feb 2014 A1
20140042814 Kather et al. Feb 2014 A1
20140049120 Kohtz et al. Feb 2014 A1
20140049635 Laffargue et al. Feb 2014 A1
20140061305 Nahill et al. Mar 2014 A1
20140061306 Wu et al. Mar 2014 A1
20140061307 Wang et al. Mar 2014 A1
20140063289 Hussey et al. Mar 2014 A1
20140066136 Sauerwein et al. Mar 2014 A1
20140067104 Osterhout Mar 2014 A1
20140067692 Ye et al. Mar 2014 A1
20140070005 Nahill et al. Mar 2014 A1
20140071840 Venancio Mar 2014 A1
20140074746 Wang Mar 2014 A1
20140075846 Woodburn Mar 2014 A1
20140076974 Havens et al. Mar 2014 A1
20140078341 Havens et al. Mar 2014 A1
20140078342 Li et al. Mar 2014 A1
20140078345 Showering Mar 2014 A1
20140084068 Gillet et al. Mar 2014 A1
20140086348 Peake et al. Mar 2014 A1
20140097249 Gomez et al. Apr 2014 A1
20140098284 Oberpriller et al. Apr 2014 A1
20140098792 Wang et al. Apr 2014 A1
20140100774 Showering Apr 2014 A1
20140100813 Showering Apr 2014 A1
20140103115 Meier et al. Apr 2014 A1
20140104413 McCloskey et al. Apr 2014 A1
20140104414 McCloskey et al. Apr 2014 A1
20140104416 Giordano et al. Apr 2014 A1
20140104451 Todeschini et al. Apr 2014 A1
20140104664 Lee Apr 2014 A1
20140106594 Skvoretz Apr 2014 A1
20140106725 Sauerwein Apr 2014 A1
20140108010 Maltseff et al. Apr 2014 A1
20140108402 Gomez et al. Apr 2014 A1
20140108682 Caballero Apr 2014 A1
20140110485 Toa et al. Apr 2014 A1
20140114530 Fitch et al. Apr 2014 A1
20140124577 Wang et al. May 2014 A1
20140124579 Ding May 2014 A1
20140125842 Winegar May 2014 A1
20140125853 Wang May 2014 A1
20140125999 Longacre et al. May 2014 A1
20140129378 Richardson May 2014 A1
20140131438 Kearney May 2014 A1
20140131441 Nahill et al. May 2014 A1
20140131445 Ding et al. May 2014 A1
20140133379 Wang et al. May 2014 A1
20140140585 Wang May 2014 A1
20140151453 Meier et al. Jun 2014 A1
20140152975 Ko Jun 2014 A1
20140160329 Ren et al. Jun 2014 A1
20140192187 Atwell et al. Jul 2014 A1
20140192551 Masaki Jul 2014 A1
20140225985 Klusza et al. Aug 2014 A1
20140267609 Laffargue Sep 2014 A1
20140268093 Tohme et al. Sep 2014 A1
20140270361 Amma Sep 2014 A1
20140313527 Askan Oct 2014 A1
20140320408 Zagorsek et al. Oct 2014 A1
20140350710 Gopalkrishnan et al. Nov 2014 A1
20150009301 Ribnick et al. Jan 2015 A1
20150009338 Laffargue et al. Jan 2015 A1
20150049347 Laffargue et al. Feb 2015 A1
20150063676 Lloyd et al. Mar 2015 A1
20150149946 Benos et al. May 2015 A1
20150163474 You Jun 2015 A1
20150213647 Laffargue et al. Jul 2015 A1
20150308816 Laffargue et al. Oct 2015 A1
Foreign Referenced Citations (35)
Number Date Country
2004212587 Apr 2005 AU
3335760 Apr 1985 DE
10210813 Oct 2003 DE
102007037282 Mar 2008 DE
1443312 Aug 2004 EP
2381421 Oct 2011 EP
2533009 Dec 2012 EP
2722656 Apr 2014 EP
2779027 Sep 2014 EP
2843590 Mar 2015 EP
2845170 Mar 2015 EP
2966595 Jan 2016 EP
3006893 Mar 2016 EP
3012601 Mar 2016 EP
3007096 Apr 2016 EP
2503978 Jan 2014 GB
2531928 May 2016 GB
H04129902 Apr 1992 JP
2008210276 Sep 2008 JP
2014210646 Nov 2014 JP
20110013200 Feb 2011 KR
20110117020 Oct 2011 KR
20120028109 Mar 2012 KR
2006095110 Sep 2006 WO
2007015059 Feb 2007 WO
2012175731 Dec 2012 WO
2013021157 Feb 2013 WO
2013033442 Mar 2013 WO
2013163789 Nov 2013 WO
2013166368 Nov 2013 WO
2013173985 Nov 2013 WO
2014019130 Feb 2014 WO
2014149702 Sep 2014 WO
2014151746 Sep 2014 WO
2016020038 Feb 2016 WO
Non-Patent Literature Citations (123)
Entry
Peter Clarke, Actuator Developer Claims Anti-Shake Breakthrough for Smartphone Cams, Electronic Engineering Times, p. 24, May 16, 2011.
U.S. Appl. No. 14/055,234, not yet published, Hand Held Products, Inc. Filed Oct. 16, 2013; 26 pages.
U.S. Appl. No. 13/912,262, not yet published, Filed Jun. 7, 2013, Hand Held Products Inc., Method of Error Correction for 3D Imaging Device: 33 pages.
European Search Report for application No. EP13186043 (now EP2722656 (Apr. 23, 2014)): Total pp. 7.
International Search Report for PCT/US2013/039438 (WO2013166368), Oct. 1, 2013, 7 pages.
U.S. Appl. No. 14/453,019, not yet published, filed Aug. 6, 2014, Hand Held Products Inc., Dimensioning System With Guided Alignment: 31 pages.
European Office Action for application EP 13186043, dated Jun. 12, 2014(now EP2722656 (Apr. 23, 2014)), Total of 6 pages.
U.S. Appl. No. 14/461,524, not yet published, filed Aug. 18, 2014, Hand Held Products Inc., System and Method for Package Dimensioning: 21 pages.
European Patent Office Action for Application No. 14157971.4-1906, Dated Jul. 16, 2014, 5 pages.
European Patent Search Report for Application No. 14157971.4-1906, Dated Jun. 30, 2014, 6 pages.
Caulier, Yannick et al., “A New Type of Color-Coded Light Structures for an Adapted and Rapid Determination of Point Correspondences for 3D Reconstruction.” Proc. of SPIE, vol. 8082 808232-3; 2011; 8 pages.
Kazantsev, Aleksei et al. “Robust Pseudo-Random Coded Colored STructured Light Techniques for 3D Object Model Recovery”; ROSE 2008 IEEE International Workshop on Robotic and Sensors Environments (Oct. 17-18, 2008) , 6 pages.
Mouaddib E. et al. “Recent Progress in Structured Light in order to Solve the Correspondence Problem in Stereo Vision” Proceedings of the 1997 IEEE International Conference on Robotics and Automation, Apr. 1997; 7 pages.
U.S. Appl. No. 14/519,179, Serge Thuries et al., filed Oct. 21, 2014, not published yet. 40 pages.
U.S. Appl. No. 14/519,249, H. Sprague Ackley et al., filed Oct. 21, 2014, not published yet. 36 pages.
U.S. Appl. No. 14/519,233, Franck Laffargue et al., filed Oct. 21, 2014, not published yet. 34 pages.
U.S. Appl. No. 14/519,211, H. Sprague Ackley et al., filed Oct. 21, 2014, not published yet. 33 pages.
U.S. Appl. No. 14/519,195, Franck Laffargue et al., filed Oct. 21, 2014, not published yet. 35 pages.
U.S. Appl. No. 14/274,858 for Mobile Printer With Optional Battery Accessory, filed May 12, 2014, (Marty et al.), 26 pages.
U.S. Appl. No. 14/264,173 for Autofocus Lens System for Indicia Readers filed Apr. 29, 2014, (Ackley et al.), 39 pages.
U.S. Appl. No. 14/230,322 for Focus Module and Components with Actuator filed Mar. 31, 2014 (Feng et al.); 92 pages.
U.S. Appl. No. 14/222,994 for Method and Apparatus for Reading Optical Indicia Using a Plurality of Data filed Mar. 24, 2014 (Smith et al.); 30 pages.
U.S. Appl. No. 14/231,898 for Hand-Mounted Indicia-Reading Device with Finger Motion Triggering filed Apr. 1, 2014 (Van Horn et al.); 36 pages.
U.S. Appl. No. 29/486,759 for an Imaging Terminal, filed Apr. 2, 2014 (Oberpriller et al.); 8 pages.
U.S. Appl. No. 29/436,337 for an Electronic Device, filed Nov. 5, 2012 (Fitch et al.); 19 pages.
U.S. Appl. No. 29/458,405 for an Electronic Device, filed Jun. 19, 2013 (Fitch et al.); 22 pages.
U.S. Appl. No. 29/459,620 for an Electronic Device Enclosure, filed Jul. 2, 2013 (London et al.); 21 pages.
U.S. Appl. No. 29/459,681 for an Electronic Device Enclosure, filed Jul. 2, 2013 (Chaney et al.); 14 pages.
U.S. Appl. No. 29/459,785 for a Scanner and Charging Base, filed Jul. 3, 2013 (Fitch et al.); 21 pages.
U.S. Appl. No. 29/459,823 for a Scanner, filed Jul. 3, 2013 (Zhou et al.); 13 pages.
U.S. Appl. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.); 44 pages.
U.S. Appl. No. 13/367,978, filed Feb. 7, 2012, (Feng et al.); now abandoned.
U.S. Appl. No. 13/736,139 for an Electronic Device Enclosure, filed Jan. 8, 2013 (Chaney); 40 pages.
U.S. Appl. No. 13/771,508 for an Optical Redirection Adapter, filed Feb. 20, 2013 (Anderson); 26 pages.
U.S. Appl. No. 13/780,356 for a Mobile Device Having Object Identification Interface, filed Feb. 28, 2013 (Samek et al.); 21 pages.
U.S. Appl. No. 13/852,097 for a System and Method for Capturing and Preserving Vehicle Event Data, filed Mar. 28, 2013 (Barker et al.); 20 pages.
U.S. Appl. No. 13/902,110 for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Hollifield); 29 pages.
U.S. Appl. No. 13/902,144, for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Chamberlin); 23 pages.
U.S. Appl. No. 13/902,242 for a System for Providing a Continuous Communication Link With a Symbol Reading Device, filed May 24, 2013 (Smith et al.); 24 pages.
U.S. Appl. No. 13/912,262 for a Method of Error Correction for 3D Imaging Device, filed Jun. 7, 2013 (Jovanovski et al.); 33 pages.
U.S. Appl. No. 13/912,702 for a System and Method for Reading Code Symbols at Long Range Using Source Power Control, filed Jun. 7, 2013 (Xian et al.); 24 pages.
U.S. Appl. No. 13/922,339 for a System and Method for Reading Code Symbols Using a Variable Field of View, filed Jun. 20, 2013 (Xian et al.); 23 pages.
U.S. Appl. No. 13/927,398 for a Code Symbol Reading System Having Adaptive Autofocus, filed Jun. 26, 2013 (Todeschini); 24 pages.
U.S. Appl. No. 13/930,913 for a Mobile Device Having an Improved User Interface for Reading Code Symbols, filed Jun. 28, 2013 (Gelay et al.); 24 pages.
U.S. Appl. No. 13/933,415 for an Electronic Device Case, filed Jul. 2, 2013 (London et al.); 47 pages.
U.S. Appl. No. 13/947,296 for a System and Method for Selectively Reading Code Symbols, filed Jul. 22, 2013 (Rueblinger et al.); 29 pages.
U.S. Appl. No. 13/950,544 for a Code Symbol Reading System Having Adjustable Object Detection, filed Jul. 25, 2013 (Jiang); 28 pages.
U.S. Appl. No. 13/961,408 for a Method for Manufacturing Laser Scanners, filed Aug. 7, 2013 (Saber et al.); 26 pages.
U.S. Appl. No. 14/018,729 for a Method for Operating a Laser Scanner, filed Sep. 5, 2013 (Feng et al.); 24 pages.
U.S. Appl. No. 14/019,616 for a Device Having Light Source to Reduce Surface Pathogens, filed Sep. 6, 2013 (Todeschini); 23 pages.
U.S. Appl. No. 14/023,762 for a Handheld Indicia Reader Having Locking Endcap, filed Sep. 11, 2013 (Gannon); 31 pages.
U.S. Appl. No. 14/035,474 for Augmented-Reality Signature Capture, filed Sep. 24, 2013 (Todeschini); 33 pages.
U.S. Appl. No. 14/047,896 for Terminal Having Illumination and Exposure Control filed Oct. 7, 2013 (Jovanovski et al.); 32 pages.
U.S. Appl. No. 14/053,175 for Imaging Apparatus Having Imaging Assembly, filed Oct. 14, 2013 (Barber); 39 pages.
U.S. Appl. No. 14/055,234 for Dimensioning System, filed Oct. 16, 2013 (Fletcher); 26 pages.
U.S. Appl. No. 14/053,314 for Indicia Reader, filed Oct. 14, 2013 (Huck); 29 pages.
U.S. Appl. No. 14/065,768 for Hybrid System and Method for Reading Indicia, filed Oct. 29, 2013 (Meier et al.); 22 pages.
U.S. Appl. No. 14/074,746 for Self-Checkout Shopping System, filed Nov. 8, 2013 (Hejl et al.); 26 pages.
U.S. Appl. No. 14/074,787 for Method and System for Configuring Mobile Devices via NFC Technology, filed Nov. 8, 2013 (Smith et al.); 28 pages.
U.S. Appl. No. 14/087,190 for Optimal Range Indicators for Bar Code Validation, filed Nov. 22, 2013 (Hejl); 27 pages.
U.S. Appl. No. 14/345,735 for Optical Indicia Reading Terminal with Combined Illumination filed Mar. 19, 2014 (Ouyang); 19 pages.
U.S. Appl. No. 14/101,965 for High Dynamic-Range Indicia Reading System, filed Dec. 10, 2013 (Xian); 28 pages.
U.S. Appl. No. 14/118,400 for Indicia Decoding Device with Security Lock, filed Nov. 18, 2013 (Liu); 28 pages.
U.S. Appl. No. 14/150,393 for Incicia-reader Having Unitary Construction Scanner, filed Jan. 8, 2014 (Colavito et al.); 28 pages.
U.S. Appl. No. 14/154,207 for Laser Barcode Scanner, filed Jan. 14, 2014 (Hou et al.); 26 pages.
U.S. Appl. No. 14/154,915 for Laser Scanning Module Employing a Laser Scanning Assembly having Elastomeric Wheel Hinges, filed Jan. 14, 2014 (Havens et al.); 24 pages.
U.S. Appl. No. 14/158,126 for Methods and Apparatus to Change a Feature Set on Data Collection Devices, filed Jan. 17, 2014 (Berthiaume et al.); 53 pages.
U.S. Appl. No. 14/342,551 for Terminal Having Image Data Format Conversion filed Mar. 4, 2014 (Lui et al.); 25 pages.
U.S. Appl. No. 14/342,544 for Imaging Based Barcode Scanner Engine with Multiple Elements Supported on a Common Printed Circuit Board filed Mar. 4, 2014 (Liu et al.); 27 pages.
U.S. Appl. No. 14/257,174 for Reading Apparatus Having Partial Frame Operating Mode filed Apr. 21, 2014, (Barber et al.), 67 pages.
U.S. Appl. No. 14/200,405 for Indicia Reader for Size-Limited Applications filed Mar. 7, 2014 (Feng et al.); 42 pages.
U.S. Appl. No. 14/166,103 for Indicia Reading Terminal Including Optical Filter filed Jan. 28, 2014 (Lu et al.); 29 pages.
U.S. Appl. No. 14/490,989, not yet published, Filed Sep. 19, 2014, Intermec IP Corporation, Volume Dimensioning System Calibration Systems and Methods: Now published as US 2015/0009338 A1, on Jan. 8, 2015. Laffargue et al.
Wikipedia, YUV description and definition, downloaded from http://www.wikipeida.org/wiki/YUV on Jun. 29, 2012, 10 pages.
YUV Pixel Format, downloaded from http://www.fource.org/yuv.php on Jun. 29, 2012; 13 pages.
YUV to RGB Conversion, downloaded from http://www.fource.org/fccyvrgb.php on Jun. 29, 2012; 5 pages.
Benos et al., “Semi-Automatic Dimensioning with Imager of a Portable Device,” U.S. Appl. No. 61/149,912, filed Feb. 4, 2009 (now expired), 56 pages.
Dimensional Weight—Wikipedia, the Free Encyclopedia, URL=http://en.wikipedia.org/wiki/Dimensional—weight, download date Aug. 1, 2008, 2 pages.
Dimensioning—Wikipedia, the Free Encyclopedia, URL=http://en.wikipedia.org/wiki/Dimensioning, download date Aug. 1, 2008, 1 page.
European Search Report for related EP Application No. 15188440.0, Dated Mar. 8, 2016, 8 pages.
Lloyd, Ryan and Scott McCloskey, “Recognition of 3D Package Shapes for Singe Camera Metrology” IEEE Winter Conference on Applications of computer Visiona, IEEE, Mar. 24, 2014, pp. 99-106, {retrieved on Jun. 16, 2014}, Authors are employees of common Applicant.
European Search Report for Related EP Application No. 15189214.8, dated Mar. 3, 2016, 9 pages.
Second Chinese Office Action in related CN Application No. 2015220810562.2, Dated Mar. 22, 2016, 5 pages. English Translation provided [No references].
United Kingdom Search Report in related application GB1517842.9, dated Apr. 8, 2016, 8 pages.
Search Report and Opinion in related GB Application No. 1517112.7, Dated Feb. 19, 2016, 6 Pages (GB2503978 is a commonly owned now abandoned application and not cited above).
Second Chinese Office Action in related CN Application No. 201520810313.3, Dated Mar. 22, 2016, 5 pages. English Translation provided [No references].
European Search Report in related EP Application No. 15190315.0, Dated Apr. 1, 2016, 7 pages. [Commonly owned Reference 2014/0104416 has been previously cited].
U.S. Appl. No. 14/800,757, Eric Todeschini, filed Jul. 16, 2015, not published yet, Dimensioning and Imaging Items, 80 pages.
Proesmans, Marc et al. “Active Acquisition of 3D Shape for Moving Objects” 0-7803-3258-X/96 1996 IEEE; 4 pages.
U.S. Appl. No. 14/747,197, Serge Thuries et al., filed Jun. 23, 2015, not published yet, Optical Pattern Projector; 33 pages.
U.S. Appl. No. 14/747,490, Brian L. Jovanovski et al., filed Jun. 23, 2015, not published yet, Dual-Projector Three-Dimensional Scanner; 40 pages.
U.S. Appl. No. 14/715,916, H. Sprague Ackley, filed May 19, 2015, not published yet, Evaluating Image Values; 54 pages.
U.S. Appl. No. 14/793,149, H. Sprague Ackley, filed Jul. 7, 2015, not published yet, Mobile Dimensioner Apparatus for Use in Commerce; 57 pages.
U.S. Appl. No. 14/740,373, H. Sprague Ackley et al., filed Jun. 16, 2015, not published yet, Calibrating A Volume Dimensioner; 63 pages.
U.S. Appl. No. 14/801,023, Tyler Doomenbal et al., filed Jul. 16, 2015, not published yet, Adjusting Dimensioning Results Using Augmented Reality, 39 pages.
Leotta, Matthew, Generic, Deformable Models for 3-D Vehicle Surveillance, May 2010, Doctoral Dissertation, Brown University, Providence RI, 248 pages.
Ward, Benjamin, Interactive 3D Reconstruction from Video, Aug. 2012, Doctoral Thesis, Univesity Of Adelaide, Adelaide, South Australia, 157 pages.
Hood, Frederick W.; William A. Hoff, Robert King, Evaluation of An Interactive Technique for Creating Site Models from Range Data, Apr. 27-May 1, 1997 Proceedings of the ANS 7th Topical Meeting on Robotics & Remote Systems, Augusta GA, 9 pages.
Gupta, Alok; Range Image Segmentation for 3-D Objects Recognition, May 1988, Technical Reports (CIS), Paper 736, University of Pennsylvania Department of Computer and Information Science, retrieved from Http://repository.upenn.edu/cis—reports/736, Accessed May 31, 2015, 157 pages.
Reisner-Kollmann,lrene; Anton L. Fuhrmann, Werner Purgathofer, Interactive Reconstruction of Industrial Sites Using Parametric Models, May 2010, Proceedings of the 26th Spring Conference of Computer Graphics SCCG 10, 8 pages.
Drummond, Tom; Roberto Cipolla, Real-Time Visual Tracking of Complex Structures, Jul. 2002, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, No. 7; 15 pages.
Zhang, Zhaoxiang; Tieniu Tan, Kaiqi Huang, Yunhong Wang; Three-Dimensional Deformable-Model-based Localization and Recognition of Road Vehicles; IEEE Transactions on Image Processing, vol. 21, No. 1, Jan. 2012, 13 pages.
Leotta, Matthew J.; Joseph L. Mundy; Predicting High Resolution Image Edges with a Generic, Adaptive, 3-D Vehicle Model; IEEE Conference on Computer Vision and Pattern Recognition, 2009; 8 pages.
Spiller, Jonathan; Object Localization Using Deformable Templates, Master's Dissertation, University of the Witwatersrand, Johannesburg, South Africa, 2007; 74 pages.
EP Search and Written Opinion Report in related matter EP Application No. 14181437.6, Dated Mar. 26, 2015, 7 pages.
Hetzel, Gunter et al.; “3D Object Recognition from Range Images using Local Feature Histograms,”, Proceedings 2001 IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2001. Kauai, Hawaii, Dec. 8-14, 2001; pp. 394-399, XP010584149, ISBN: 978-0-7695-1272-3.
Intention to Grant in counterpart European Application No. 14157971.4 dated Apr. 14, 2015, pp. 1-8.
Decisions to Grant in counterpart European Application No. 14157971.4 dated Aug. 6, 2015, pp. 1-2.
Salvi, Joaquim et al. “Pattern Codification Strategies in Structured Light Systems” published in Pattern Recognition; The Journal of the Pattern Recognition Society, Received Mar. 6, 2003; Accepted Oct. 2, 2003; 23 pages.
Office Action in counterpart European Application No. 13186043.9 dated Sep. 30, 2015, pp. 1-7.
Lloyd et al., “System for Monitoring the Condition of Packages Throughout Transit”, U.S. Appl. No. 14/865,575, filed Sep. 25, 2015, 59 pages, not yet published.
James Chamberlin, “System and Method for Picking Validation”, U.S. Appl. No. 14/865,797, filed Sep. 25, 2015, 44 pages, not yet published.
Jovanovski et al., “Image-Stitching for Dimensioning”, U.S. Appl. No. 14/870,488, filed Sep. 30, 2015, 45 pages, not yet published.
Todeschini et al.; “Depth Sensor Based Auto-Focus System for an Indicia Scanner,” U.S. Appl. No. 14/872,176, filed Oct. 1, 2015, 44 pages, not yet published.
Wikipedia, “3D projection” Downloaded on Nov. 25, 2015 from www.wikipedia.com, 4 pages.
McCloskey et al., “Methods for Improving the Accuracy of Dimensioning-System Measurements,” U.S. Appl. No. 14/873,613, filed Sep. 2, 2015, 47 pages, not yet published.
Search Report in counterpart European Application No. 15182675.7, Dated Dec. 4, 2015, 10 pages.
McCloskey et al., “Image Transformation for Indicia Reading,” U.S. Appl. No. 14/982,032, filed Oct. 30, 2015, 48 pages, not yet published.
Search Report and Opinion in Related EP Application 15176943.7, Dated Jan. 8, 2016, 8 pages, (US Application 2014/0049635 has been previously cited).
Second Chinese Office Action in related CN Application No. 201520810685.6, Dated Mar. 22, 2016, 5 pages, no references.
Great Britain Search Report for related Application On. GB1517843.7, Dated Feb. 23, 2016; 8 pages.
European Search Report for related Application EP 15190249.1, Dated Mar. 22, 2016, 7 pages.
M. Zahid Gurbuz, Selim Akyokus, Ibrahim Emiroglu, Aysun Guran, An Efficient Algorithm for 3D Rectangular Box Packing, 2009, Applied Automatic Systems: Proceedings of Selected AAS 2009 Papers, pp. 131-134.
Related Publications (1)
Number Date Country
20150063676 A1 Mar 2015 US
Provisional Applications (1)
Number Date Country
61872299 Aug 2013 US