Measuring object dimensions using mobile computer

Information

  • Patent Grant
  • 9939259
  • Patent Number
    9,939,259
  • Date Filed
    Friday, January 25, 2013
    13 years ago
  • Date Issued
    Tuesday, April 10, 2018
    7 years ago
Abstract
Devices, methods, and software are disclosed for determining dimensions of a physical object using a mobile computer equipped with a motion sensing device. In an illustrative embodiment, the mobile computer can comprise a microprocessor, a memory, a user interface, a motion sensing device, and a dimensioning program executable by the microprocessor. The processor can be in communicative connection with executable instructions for enabling the processor for various steps. One step includes initiating a trajectory tracking mode responsive to receiving a first user interface action. Another step includes tracking the mobile computer's trajectory along a surface of a physical object by storing in the memory a plurality of motion sensing data items outputted by the motion sensing device. Another step includes exiting the trajectory tracking mode responsive to receiving a second user interface action. Another step includes calculating three dimensions of a minimum bounding box corresponding to the physical object.
Description
FIELD OF THE INVENTION

The present disclosure relates to mobile computers in general and in particular to a mobile computer having a motion sensing device.


BACKGROUND

In several logistics use cases, dimensions of a package or a parcel need to be determined, for example, in order to ascertain that maximum size limits imposed by a carrier have not been exceeded, or to calculate shipping costs based on estimated volumetric weight.


The discussion above is merely provided for general background information and is not intended to be used as an aid in determining the scope of the claimed subject matter.


SUMMARY

Systems are disclosed that in various embodiments include devices, methods, and/or software for determining dimensions of a physical object using a mobile computer. In an illustrative embodiment, the mobile computer can comprise a microprocessor, a memory, a user interface, a motion sensing device, and a dimensioning program executable by the microprocessor.


The processor can be in communicative connection with executable instructions for enabling the processor for various steps. One step includes initiating a trajectory tracking mode responsive to receiving a first user interface action. Another step includes tracking the mobile computer's trajectory along a surface of a physical object by storing in the memory a plurality of motion sensing data items outputted by the motion sensing device. Another step includes exiting the trajectory tracking mode responsive to receiving a second user interface action. Another step includes calculating at least one dimension of the physical object based on the plurality of motion sensing data items.


In some embodiments, the mobile computer running the dimensioning program can be further configured to calculate three dimensions of a minimum perimeter bounding box corresponding to the physical object, a minimum surface bounding box corresponding to the physical object, or a minimum volume bounding box corresponding to the physical object.


In some embodiments, the mobile computer running the dimensioning program can be further configured to construct the minimum bounding box by cycling through all edges Ei connecting two points of a plurality of points defined by the plurality of motion sensing data items; constructing, for each edge Ei, a bounding box having an edge E collinear with the selected edge Ei and at least one corner coinciding with one ends of the selected edge Ei; and selecting, among the constructed bounding boxes, a minimum perimeter bounding box corresponding to the physical object, a minimum surface bounding box corresponding to the physical object, or a minimum volume bounding box corresponding to the physical object.


In some embodiments, the motion sensing device can be provided by at least one accelerometer. Alternatively, the motion sensing device can be provided by at least three accelerometers configured to measure proper acceleration values along at least three mutually-perpendicular directions. Alternatively, the motion sensing device can be provided by a 9-DOF (degree of freedom) motion sensing unit containing a 3-axis accelerometer, a 3-axis magnetometer, and a 3-axis gyroscope.


In some embodiments, motion sensing device can comprise a gyroscope configured to determine orientation of the mobile computer.


In some embodiments, each motion sensing data item of the plurality of motion sensing data items can be stored in memory along with a corresponding timestamp.


This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter. The claimed subject matter is not limited to implementations that solve any or all disadvantages noted in the background.





BRIEF DESCRIPTION OF THE DRAWINGS

The features described herein can be better understood with reference to the drawings described below. The drawings are not necessarily to scale, emphasis instead generally being placed upon illustrating the principles of the invention. In the drawings, like numerals are used to indicate like parts throughout the various views.



FIG. 1 schematically illustrates one embodiment of the dimensioning method described herein.



FIG. 2 depicts a simplified, mixed perspective and diagram view of a system including a mobile computer, in accordance with an illustrative embodiment.



FIG. 3 depicts a schematic block diagram of one embodiment of a mobile computer described herein.



FIG. 4 depicts a diagram of one embodiment of a state machine implemented by the mobile computer running the object dimensioning program described herein.





The drawings are not necessarily to scale, emphasis instead generally being placed upon illustrating the principles of various embodiments. In the drawings, like numerals are used to indicate like parts throughout the various views.


DETAILED DESCRIPTION OF ILLUSTRATIVE EMBODIMENTS

Existing physical object dimensioning methods typically employ specialized equipment, and hence the dimensioning is usually performed at shipping facilities, i.e., after the logistics workflow for a particular parcel has already been initiated. Upon determining the parcel dimensions, the carrier may determine that a parcel is too large and should be returned to the consignor. In another example, the shipping cost determined based on the volumetric weight can exceed a previously estimated shipping cost. These and other issues could have been avoided if the consignor had means to determine the parcel dimensions before initiating the shipment workflow for the parcel (e.g., before delivering the parcel to a shipping facility).


Modern mobile computers are often equipped with at least one motion sensing device, such as an accelerometer and/or a gyroscope, which can be used for determining the spatial position and orientation of the device relative to a known point of origin. Disclosed herein is a simple, fast, and reasonably accurate method of determining dimensions of physical items (including irregularly shaped items) employing a mobile computer equipped with at least one motion sensing device.


A “mobile computer” herein shall refer to a portable programmable device for data processing, including a central processing unit (CPU), a memory, and at least one communication interface. A mobile computer can be provided, e.g., by a personal digital assistant (PDA), a portable data terminal (PDT), or a smart phone.


In one illustrative embodiment of the dimensioning method described herein and schematically shown in FIG. 1, mobile computer 100 can be configured to execute a program for physical object dimensioning for implementing various methods described herein. A user of the mobile computer 100 can position the computer at one corner 102a of the item 110 dimensions of which are to be determined, and initiate the dimensioning operation via a user interface action, for example, by pressing a button or issuing a voice command. The user can then move the mobile computer 100 along the edges 104a-104c of the item, and the mobile computer can track its own motion by recording the inputs from the motion sensing device to produce a series of spatial points representing the path of the computer, including points along the edges 104a-104c and points corresponding to the corners 102a-102d. Having positioned the mobile computer successively at each extremity (e.g., a corner) of the item, the user can then indicate via a user interface action that the dimensioning operation is complete. The order in which the extremities of the object are visited does not matter, as long as all the extremities have been visited.


Upon the user's completion of the measurement process, the mobile computer can examine the trajectory data recorded during the measurement process, and calculate the dimensions of a bounding box using known computational geometry methods.


“Bounding box of a set of N points” herein shall refer to a rectangular parallelepiped within which all the points lie.


In some embodiments, the mobile computer can be configured to construct the minimum perimeter bounding box, i.e., a bounding box having the minimum perimeter among all possible bounding boxes for the given set of points. Alternatively, the mobile computer can be configured to construct the minimum area bounding box, i.e., a bounding box having the minimum surface area among all possible bounding boxes for the given set of points. Alternatively, the mobile computer can be configured to construct the minimum volume bounding box, i.e., a bounding box having the minimum surface area among all possible bounding boxes for the given set of points.


The dimensions of the constructed minimum bounding box can then be compared to the carrier's requirements, and/or can be used to calculate the volumetric weight of the item.


In another use case, a user of mobile computer 100 may need to define a selected one dimension of a physical object, for example, the length of the edge 1104a. The user can position mobile computer 100 at the corner 102a of the item 110 and initiate the dimensioning operation via a user interface action, for example, by pressing a button or issuing a voice command. The user can then move the mobile computer 100 along the edge 104a, and the mobile computer can track its own motion by recording the inputs from the motion sensing device to produce a series of spatial points representing the path of the computer. Having positioned the mobile computer successively at both ends of the edge 104a, the user can then indicate via a user interface action that the dimensioning operation is complete. Upon the user's completion of the measurement process, the mobile computer can examine the trajectory data recorded during the measurement process, and calculate the distance between the first and the last point of the recorded trajectory.


In some embodiments, mobile computer 100 can transmit the stored trajectory to one or more external servers 2000, 3000 over a network. Responsive to receiving the motion sensing data from mobile computer 100, one more external servers 2000, 3000 can construct a minimum bounding box based on the data received and transmit the box dimensions in a response transmitted over the network to mobile computer 100.



FIG. 2 depicts a system 1000 for measuring dimensions of a physical object in accordance with an illustrative embodiment that includes a mobile computer 100, depicted here in perspective view. Mobile computer 100 is depicted as a hand held style mobile computer in the illustrative embodiment of FIG. 2, and may also take the form of a smartphone, a mobile phone, a tablet or netbook computer, a laptop computer, an e-book reader, an indicia scanning terminal, or any of a wide range of other types of digital devices equipped with communication interfaces and motion sensing devices, in various embodiments. In the illustrative embodiment of FIG. 2, mobile computer 100 includes user interface elements including display 1222, pointer mechanism 1224, and keyboard 1226 disposed on a hand held housing 1014. Two of the keys on keyboard 1226 are designated as a dimensioning program start key 1227 and an enter key 1228, though which keys are used for such functions and where they are disposed on mobile computer 100 may be arbitrarily selected and may differ from the illustrative depiction in FIG. 2.


Display 1222 in various embodiments can incorporate a touch panel for navigation and virtual actuator selection in which case a user interface of mobile computer 100 can be provided by display 1222. In another embodiment, a mobile device may be devoid of a display and can be in a gun style form factor. In various embodiments, mobile computer 100 may itself constitute a system for determining dimensions of physical objects, and in various embodiments, mobile computer 100 in combination with one or more external servers 2000, 3000 (depicted in block diagram), which may be connected over a network 2500, may together serve as a system for determining dimensions of physical objects. In the description herein, system 1000 may be described as being enabled or configured for various features, characteristics, or functions; and in various embodiments this may refer to mobile computer 100 alone, or in communication or cooperation with other elements of system 1000, being enabled or configured for those features, characteristics, or functions. Various elements of FIG. 2 are further described below.



FIG. 3 depicts a schematic block diagram of an illustrative embodiment of mobile computer 100. Mobile computer 100 can include at least one processor 1060. One or more processors 1060 may illustratively be or include a central processing unit (CPU), a complex programmable logic device (CPLD), an application-specific integrated circuit (ASIC), a field programmable gate array (FPGA), or any type of circuit capable of processing logic operations, in accordance with various embodiments. Processor 1060 may be in communicative connection with executable instructions for enabling the processor 1060 for various steps, e.g., the steps illustratively depicted in FIG. 4, in accordance with an illustrative embodiment of an object dimensioning method described herein.


Mobile computer 100 can further include one or more memory components 1085 communicatively coupled to processor 1060 via system bus 1500. Memory components 1085 can include at least a first memory component, illustratively such as RAM 1080, that is operatively capable of at least temporarily or transiently storing data, while other memory components may be used in various embodiments. Memory components 1085 can further include a nonvolatile memory such as EPROM 1082, a memory storage device 1084, and any of a variety of other types of memory components, in various embodiments. Memory storage device 1084 may illustratively be or include a flash memory, a hard disc drive, any type of RAM, EPROM, EEPROM, DVD-ROM, CD-ROM, or other type of ROM, optical disc, magnetic disc, magnetic cassette, magnetic tape, or any other type of volatile or non-volatile or removable or non-removable memory or data storage components, in illustrative embodiments.


System bus 1500 can provide for bus arbitration, and can be provided by any of a variety of bus structures such as a memory bus or memory controller, a peripheral bus, or a local bus, using any of a variety of architectures, in various embodiments. For example, this may include a Peripheral Component Interconnect (PCI) or Mezzanine bus, an Industry Standard Architecture (ISA) bus, an Enhanced Industry Standard Architecture (EISA) bus, a Micro Channel Architecture (MCA) bus, a Video Electronics Standards Association (VESA) bus, or other bus architectures, in various embodiments. Mobile computer 100 may include a direct memory access unit (DMA) 1070 for routing image information read out from image sensor 1032 that has been subject to conversion to RAM 1080, in various embodiments. Other embodiments of the system bus architecture and/or direct memory access components providing for efficient data transfer between the image sensor 1032 and RAM 1080 may be encompassed in various embodiments.


Mobile computer 100 may further include at least one motion sensing device 1208. In some embodiments, the motion sensing device can be provided by at least three accelerometers configured to measure proper acceleration values along at least three mutually-perpendicular axes. Alternatively, the motion sensing device can be provided by a 9-DOF (degree of freedom) motion sensing unit containing a 3-axis accelerometer, a 3-axis magnetometer, and 3-axis gyroscope sensors.


Mobile computer 100 may further include an imaging assembly 900 comprising lens assembly 250 and image sensor integrated circuit 1040. Image sensor 1032 may include multiple pixel image sensor array 1033 having pixels arranged in rows and columns of pixels, associated column circuitry 1034 and row circuitry 1035. Associated with the image sensor 1032 may be amplifier circuitry 1036 (amplifier), and an analog to digital converter 1037 which converts image information in the form of analog signals read out of image sensor array 1033 into image information in the form of digital signals. Image sensor 1032 can also have an associated timing and control circuit 1038 for use in controlling e.g., the exposure period of image sensor 1032, gain applied to the amplifier 1036. The noted circuit components 1032, 1036, 1037, and 1038 may be packaged into a common image sensor integrated circuit 1040, in this illustrative embodiment. Image sensor integrated circuit 1040 may incorporate fewer than the noted number of components, in various embodiments.


Mobile computer 100 may further include an illumination subsystem 800 for illumination of target area and projection of an illumination pattern 1260, in various embodiments.


Mobile computer 100 may include various interface circuits for coupling various of the peripheral devices to system address/data bus (system bus) 1500, for communication with processor 1060 also coupled to system bus 1500. Mobile computer 100 may include interface circuit 1028 for coupling image sensor timing and control circuit 1038 to system bus 1500, interface circuit 1102 for coupling electrical power input unit 1202 to system bus 1500, interface circuit 1106 for coupling illumination light source bank control circuit 1206 to system bus 1500, and interface circuit 1108 for coupling motion sensing device 1208 to system bus 1500. Mobile computer 100 may also include a display 1222 coupled to system bus 1500 and in communication with processor 1060, via interface 1122, as well as pointer mechanism 1224 in communication with processor 1060 via interface 1124 connected to system bus 1500. Mobile computer 100 may also include keyboard 1226 coupled to system bus 1500. Keyboard 1226 may be in communication with processor 1060 via interface 1126 connected to system bus 1500.


The object dimensioning steps described herein can be distributed among mobile computer 100, servers 2000 and/or 3000 and one embodiment may be executed entirely by mobile computer 100. In such an embodiment, system 1000 may be regarded as being provided by mobile computer 100.



FIG. 4 illustrates a state diagram of one illustrative embodiment a method 200 of operating a mobile computer 100 for determining dimensions of a physical object.


Upon initialization, the method can transition into the Idle state. Responsive to receiving a user interface action indicating the start of a dimensioning operation, the method can transition into the Trajectory Tracking state.


Responsive to receiving, while being in the Trajectory Tracking state, a message containing motion sensing data from a motion sensing device, the method can transition to the Store MSD state. Upon storing in memory the motion sensing data and the current time, the method can return to the Trajectory Tracking state.


Responsive to receiving, while being in the Trajectory Tracking state, a user interface action indicating the completion of the dimensioning operation, the method can transition into the Calculate Dimensions state. In the Calculate Dimensions state, the method can calculate at least one dimension of the physical object. In some embodiments, the method can calculate at least three dimensions of the physical object along three mutually perpendicular directions based on the plurality of motion sensing data items stored in the memory. In some embodiments, the method can construct a minimum bounding box (e.g., a minimum perimeter bounding box, a minimum surface bounding box, or a minimum volume bounding box). The method can output the calculation result, for example, onto the display 1222.


In a further aspect, the minimum bounding box can be constructed by employing O'Rourke's algorithm cycling through all edges connecting two points of the given set of points, and for each edge Ei constructing a bounding box having an edge E collinear with the selected edge Ei and at least one corner of the box coinciding with one of the ends of the selected edge Ei, and finally, selecting the minimum bounding box (e.g., the minimum perimeter bounding box) among the bounding boxes constructed.


A skilled artisan would appreciate the fact that other methods of constructing a minimum bounding box are within the scope of this disclosure.


A small sample of illustrative devices, systems, apparatuses, or methods that are described herein is as follows:


A1. A mobile computer comprising: a microprocessor; a memory; a user interface; a motion sensing device; a dimensioning program executable by said microprocessor; wherein said mobile computer running said dimensioning program is configured to initiate a trajectory tracking mode responsive to receiving a first user interface action; wherein said mobile computer operating in said trajectory tracking mode is configured to track own trajectory along a surface of a physical object by storing in said memory a plurality of motion sensing data items outputted by said motion sensing device; wherein said mobile computer running said dimensioning program is further configured to exit said trajectory tracking mode responsive to receiving a second user interface action; and wherein said mobile computer running said dimensioning program is further configured to calculate at least one dimension of said physical object based on said plurality of motion sensing data items. A2. The mobile computer of (A1), wherein said mobile computer running said dimensioning program is further configured to calculate three dimensions of at least one of: a minimum perimeter bounding box corresponding to said physical object, a minimum surface bounding box corresponding to said physical object, a minimum volume bounding box corresponding to said physical object. A3. The mobile computer of (A2), wherein said mobile computer running said dimensioning program is further configured to construct said minimum bounding box by cycling through all edges Ei connecting two points of a plurality of points defined by said plurality of motion sensing data items; wherein said mobile computer running said dimensioning program is further configured, for each edge Ei, to construct a bounding box having an edge E collinear with said selected edge Ei, said bounding box further having and at least one corner coinciding with one ends of said selected edge Ei; and wherein said mobile computer running said dimensioning program is further configured to select, among said constructed bounding boxes, one of: a minimum perimeter bounding box corresponding to said physical object, a minimum surface bounding box corresponding to said physical object, a minimum volume bounding box corresponding to said physical object. A4. The mobile computer of (A1), wherein said motion sensing device is provided by at least one accelerometer. A5. The mobile computer of (A1), wherein said motion sensing device is provided by at least three accelerometers configured to measure proper acceleration values along at least three mutually-perpendicular directions. A6. The mobile computer of (A1), wherein said motion sensing device is provided by a 9-DOF (degree of freedom) motion sensing unit containing a 3-axis accelerometer, a 3-axis magnetometer, and a 3-axis gyroscope. A7. The mobile computer of (A1), wherein said motion sensing device comprises a gyroscope configured to determine orientation of said mobile computer. A8. The mobile computer of (A1), wherein each motion sensing data item of said plurality of motion sensing data items is stored in memory along with a corresponding timestamp.


B1. A method of estimating dimensions of a physical object by a mobile computer including a microprocessor, a memory, and a motion sensing device, said method comprising the steps of: said mobile computer initiating a trajectory tracking mode responsive to receiving a first user interface action; said mobile computer operating in said trajectory tracking mode tracking own trajectory along a surface of a physical object by storing in said memory a plurality of motion sensing data items outputted by said motion sensing device; said mobile computer exiting said trajectory tracking mode responsive to receiving a second user interface action; and said mobile computer calculating at least one dimension of said physical object based on said plurality of motion sensing data items. B2. The method of (B1), wherein said step of calculating at least one dimension of said physical object comprises calculating three dimensions of at least one of: a minimum perimeter bounding box corresponding to said physical object, a minimum surface bounding box corresponding to said physical object, a minimum volume bounding box corresponding to said physical object. B3. The method of claim B2, wherein said step of constructing said minimum bounding box comprises: cycling through all edges Ei connecting two points of a plurality of points defined by said plurality of motion sensing data items; for each edge Ei, constructing a bounding box having an edge E collinear with said selected edge Ei, said bounding box further having and at least one corner coinciding with one ends of said selected edge Ei; and selecting, among said constructed bounding boxes, one of: a minimum perimeter bounding box corresponding to said physical object, a minimum surface bounding box corresponding to said physical object, a minimum volume bounding box corresponding to said physical object. B4. The method of (B1), wherein said motion sensing device is provided by at least one accelerometer. B5. The method of (B1), wherein said motion sensing device is provided by at least three accelerometers configured to measure proper acceleration values along at least three mutually-perpendicular directions. B6. The method of (B1), wherein said motion sensing device is provided by a 9-DOF (degree of freedom) motion sensing unit containing a 3-axis accelerometer, a 3-axis magnetometer, and a 3-axis gyroscope. B7. The method of (B1), wherein said motion sensing device comprises a gyroscope configured to determine orientation of said mobile computer. B8. The method of (B1), wherein said step of storing in said memory a plurality of motion sensing data items comprises storing a timestamp corresponding to each motion sensing data item of said a plurality of motion sensing data items.


While the present invention has been described with reference to a number of specific embodiments, it will be understood that the true spirit and scope of the invention should be determined only with respect to claims that can be supported by the present specification. Further, while in numerous cases herein wherein systems and apparatuses and methods are described as having a certain number of elements it will be understood that such systems, apparatuses and methods can be practiced with fewer than or greater than the mentioned certain number of elements. Also, while a number of particular embodiments have been described, it will be understood that features and aspects that have been described with reference to each particular embodiment can be used with each remaining particularly described embodiment.

Claims
  • 1. A mobile computer comprising: a microprocessor;a memory;a user interface;a motion sensing device configured to determine spatial position and orientation, the motion sensing device comprising at least one of an accelerometer and/or a gyroscope;a dimensioning program executable by said microprocessor;wherein said mobile computer running said dimensioning program is configured to initiate a trajectory tracking mode responsive to receiving a first user interface action;wherein said mobile computer operating in said trajectory tracking mode is configured to record a trajectory of said mobile computer at least partially around a physical object comprising storing in said memory a plurality of motion sensing data items outputted by said motion sensing device, and said plurality of motion sensing data items defining a series of spatial points that represent the recorded trajectory of said mobile computer at least partially around the physical object;wherein said mobile computer running said dimensioning program is further configured to exit said trajectory tracking mode responsive to receiving a second user interface action; andwherein said mobile computer running said dimensioning program is further configured to calculate dimensions of said physical object based on at least some spatial points of said series of spatial points that represent the recorded trajectory of said mobile computer at least partially around the physical object.
  • 2. The mobile computer of claim 1, wherein said mobile computer running said dimensioning program is further configured to calculate three dimensions of at least one of: a minimum perimeter bounding box corresponding to said physical object, a minimum surface bounding box corresponding to said physical object, a minimum volume bounding box corresponding to said physical object.
  • 3. The mobile computer of claim 2, wherein said mobile computer running said dimensioning program is further configured to construct said minimum bounding box by cycling through all edges Ei connecting two spatial points of the series of spatial points defined by said plurality of motion sensing data items; wherein said mobile computer running said dimensioning program is further configured, for each edge Ei, to construct a bounding box having an edge E collinear with said selected edge Ei, said bounding box further having and at least one corner coinciding with one end of said selected edge E Ei; andwherein said mobile computer running said dimensioning program is further configured to select, among said constructed bounding boxes, one of: a minimum perimeter bounding box corresponding to said physical object, a minimum surface bounding box corresponding to said physical object, a minimum volume bounding box corresponding to said physical object.
  • 4. The mobile computer of claim 1, wherein said motion sensing device comprises a 3-axis accelerometer.
  • 5. The mobile computer of claim 1, wherein said motion sensing device comprises is provided by at least three accelerometers configured to measure proper acceleration values along at least three mutually-perpendicular directions.
  • 6. The mobile computer of claim 1, wherein said motion sensing device comprises by a 9-DOF (degree of freedom) motion sensing unit containing a 3-axis accelerometer, a 3-axis magnetometer, and a 3-axis gyroscope.
  • 7. The mobile computer of claim 1, wherein said motion sensing device comprises a 3-axis gyroscope configured to determine orientation of said mobile computer.
  • 8. The mobile computer of claim 1, wherein each motion sensing data item of said plurality of motion sensing data items is stored in memory along with a corresponding timestamp.
  • 9. The mobile computer of claim 1, wherein the physical object comprises a package or a parcel.
  • 10. The mobile computer of claim 1, wherein the series of spatial points comprise points along edges of the physical object and/or points corresponding to corners of the physical object.
  • 11. A method of estimating dimensions of a physical object by a mobile computer including a microprocessor, a memory, and a motion sensing device, said method comprising the steps of: said mobile computer initiating a trajectory tracking mode responsive to receiving a first user interface action;said mobile computer operating in said trajectory tracking mode recording a trajectory of said mobile computer at least partially around a physical object comprising storing in said memory a plurality of motion sensing data items outputted by said motion sensing device, said motion sensing device comprising at least one of an accelerometer and/or a gyroscope and configured to determine spatial position and orientation, and said plurality of motion sensing data items defining a series of spatial points that represent the recorded trajectory of said mobile computer at least partially around the physical object;said mobile computer exiting said trajectory tracking mode responsive to receiving a second user interface action; andsaid mobile computer calculating dimensions of said physical object based on at least some spatial points of said series of spatial points that represent the recorded trajectory of said mobile computer at least partially around the physical object.
  • 12. The method of claim 11, wherein said step of calculating dimensions of said physical object comprises calculating three dimensions of at least one of: a minimum perimeter bounding box corresponding to said physical object, a minimum surface bounding box corresponding to said physical object, a minimum volume bounding box corresponding to said physical object.
  • 13. The method of claim 11, wherein said step of constructing said minimum bounding box comprises: cycling through all edges Ei connecting two spatial points of the series of spatial points defined by said plurality of motion sensing data items;for each edge Ei, constructing a bounding box having an edge E collinear with said selected edge Ei, said bounding box further having and at least one corner coinciding with one end of said selected edge Ei, andselecting, among said constructed bounding boxes, one of: a minimum perimeter bounding box corresponding to said physical object, a minimum surface bounding box corresponding to said physical object, a minimum volume bounding box corresponding to said physical object.
  • 14. The method of claim 11, wherein said motion sensing device comprises a 3-axis accelerometer.
  • 15. The method of claim 11, wherein said motion sensing device comprises at least three accelerometers configured to measure proper acceleration values along at least three mutually-perpendicular directions.
  • 16. The method of claim 11, wherein said motion sensing device comprises a 9-DOF (degree of freedom) motion sensing unit containing a 3-axis accelerometer, a 3-axis magnetometer, and a 3-axis gyroscope.
  • 17. The method of claim 11, wherein said motion sensing device comprises a 3-axis gyroscope configured to determine orientation of said mobile computer.
  • 18. The method of claim 11, wherein said step of storing in said memory said plurality of motion sensing data items comprises storing a timestamp corresponding to each motion sensing data item of said plurality of motion sensing data items.
  • 19. The method of claim 11, wherein the physical object comprises a package or a parcel.
  • 20. The method of claim 11, wherein the series of spatial points comprise points along edges of the physical object and/or points corresponding to corners of the physical object.
CROSS REFERENCE TO RELATED APPLICATIONS

This application claims priority to U.S. Provisional Application No. 61/709,606 filed Oct. 4, 2012 entitled, “Measuring Object Dimensions Using Mobile Computer.” The above application is incorporated herein by reference in its entirety.

US Referenced Citations (668)
Number Name Date Kind
3955073 Carew May 1976 A
3971065 Bayer Jul 1976 A
4026031 Siddall et al. May 1977 A
4279328 Ahlbom Jul 1981 A
4398811 Nishioka et al. Aug 1983 A
4495559 Gelatt, Jr. Jan 1985 A
4613866 Blood Sep 1986 A
4730190 Win et al. Mar 1988 A
4803639 Steele et al. Feb 1989 A
5161313 Rijlaarsadam Nov 1992 A
5184733 Amarson et al. Feb 1993 A
5198648 Hibbard Mar 1993 A
5220536 Stringer et al. Jun 1993 A
5331118 Jensen Jul 1994 A
5359185 Hanson Oct 1994 A
5384901 Glassner et al. Jan 1995 A
5477622 Skalnik Dec 1995 A
5548707 LoNegro Aug 1996 A
5555090 Schmutz Sep 1996 A
5561526 Huber et al. Oct 1996 A
5590060 Granville et al. Dec 1996 A
5606534 Stringer et al. Feb 1997 A
5619245 Kessler et al. Apr 1997 A
5655095 LoNegro et al. Aug 1997 A
5661561 Wurz et al. Aug 1997 A
5699161 Woodworth Dec 1997 A
5729750 Ishida Mar 1998 A
5730252 Herbinet Mar 1998 A
5732147 Tao Mar 1998 A
5734476 Dlugos Mar 1998 A
5737074 Haga et al. Apr 1998 A
5748199 Palm May 1998 A
5767962 Suzuki et al. Jun 1998 A
5831737 Stringer et al. Nov 1998 A
5850370 Stringer et al. Dec 1998 A
5850490 Johnson Dec 1998 A
5869827 Rando Feb 1999 A
5870220 Migdal et al. Feb 1999 A
5900611 Hecht May 1999 A
5923428 Woodworth Jul 1999 A
5929856 LoNegro et al. Jul 1999 A
5938710 Lanza et al. Aug 1999 A
5959568 Woolley Sep 1999 A
5960098 Tao Sep 1999 A
5969823 Wurz et al. Oct 1999 A
5978512 Kim et al. Nov 1999 A
5979760 Freyman et al. Nov 1999 A
5988862 Kacyra et al. Nov 1999 A
5991041 Woodworth Nov 1999 A
6009189 Schaack Dec 1999 A
6025847 Marks Feb 2000 A
6035067 Ponticos Mar 2000 A
6049386 Stringer et al. Apr 2000 A
6053409 Brobst et al. Apr 2000 A
6064759 Buckley et al. May 2000 A
6067110 Nonaka et al. May 2000 A
6069696 McQueen et al. May 2000 A
6115114 Berg et al. Sep 2000 A
6137577 Woodworth Oct 2000 A
6177999 Wurz et al. Jan 2001 B1
6189223 Haug Feb 2001 B1
6232597 Kley May 2001 B1
6236403 Chaki May 2001 B1
6246468 Dimsdale Jun 2001 B1
6333749 Reinhardt et al. Dec 2001 B1
6336587 He et al. Jan 2002 B1
6369401 Lee Apr 2002 B1
6373579 Ober et al. Apr 2002 B1
6429803 Kumar Aug 2002 B1
6457642 Good et al. Oct 2002 B1
6507406 Yagi et al. Jan 2003 B1
6517004 Good et al. Feb 2003 B2
6519550 D'Hooge Feb 2003 B1
6674904 McQueen Jan 2004 B1
6705526 Zhu et al. Mar 2004 B1
6781621 Gobush et al. Aug 2004 B1
6824058 Patel et al. Nov 2004 B2
6832725 Gardiner et al. Dec 2004 B2
6858857 Pease et al. Feb 2005 B2
6922632 Foxlin Jul 2005 B2
6971580 Zhu et al. Dec 2005 B2
6995762 Pavlidis et al. Feb 2006 B1
7057632 Yamawaki et al. Jun 2006 B2
7085409 Sawhney et al. Aug 2006 B2
7086162 Tyroler Aug 2006 B2
7104453 Zhu et al. Sep 2006 B1
7128266 Marlton et al. Oct 2006 B2
7137556 Bonner Nov 2006 B1
7159783 Walczyk et al. Jan 2007 B2
7161688 Bonner et al. Jan 2007 B1
7205529 Andersen et al. Apr 2007 B2
7214954 Schopp May 2007 B2
7277187 Smith et al. Oct 2007 B2
7307653 Dutta Dec 2007 B2
7310431 Gokturk et al. Dec 2007 B2
7413127 Ehrhart et al. Aug 2008 B2
7509529 Colucci et al. Mar 2009 B2
7527205 Zhu et al. May 2009 B2
7586049 Wurz Sep 2009 B2
7602404 Reinhardt et al. Oct 2009 B1
7639722 Paxton et al. Dec 2009 B1
7726575 Wang et al. Jun 2010 B2
7780084 Zhang et al. Aug 2010 B2
7788883 Buckley et al. Sep 2010 B2
7974025 Topliss Jul 2011 B2
8027096 Feng et al. Sep 2011 B2
8028501 Buckley et al. Oct 2011 B2
8050461 Shpunt et al. Nov 2011 B2
8055061 Katano Nov 2011 B2
8072581 Breiholz Dec 2011 B1
8102395 Kondo et al. Jan 2012 B2
8132728 Dwinell et al. Mar 2012 B2
8134717 Pangrazio et al. Mar 2012 B2
8149224 Kuo et al. Apr 2012 B1
8194097 Xiao et al. Jun 2012 B2
8201737 Palacios Durazo et al. Jun 2012 B1
8212889 Chanas et al. Jul 2012 B2
8228510 Pangrazio et al. Jul 2012 B2
8230367 Bell et al. Jul 2012 B2
8294969 Plesko Oct 2012 B2
8305458 Hara Nov 2012 B2
8310656 Zalewski Nov 2012 B2
8313380 Zalewski et al. Nov 2012 B2
8317105 Kotlarsky et al. Nov 2012 B2
8322622 Suzhou et al. Dec 2012 B2
8339462 Stec et al. Dec 2012 B2
8350959 Topliss et al. Jan 2013 B2
8351670 Ijiri et al. Jan 2013 B2
8366005 Kotlarsky et al. Feb 2013 B2
8371507 Haggerty et al. Feb 2013 B2
8376233 Van Horn et al. Feb 2013 B2
8381976 Mohideen et al. Feb 2013 B2
8381979 Franz Feb 2013 B2
8390909 Plesko Mar 2013 B2
8408464 Zhu et al. Apr 2013 B2
8408468 Horn et al. Apr 2013 B2
8408469 Good Apr 2013 B2
8424768 Rueblinger et al. Apr 2013 B2
8437539 Komatsu et al. May 2013 B2
8441749 Brown et al. May 2013 B2
8448863 Xian et al. May 2013 B2
8457013 Essinger et al. Jun 2013 B2
8459557 Havens et al. Jun 2013 B2
8463079 Ackley et al. Jun 2013 B2
8469272 Kearney Jun 2013 B2
8474712 Kearney et al. Jul 2013 B2
8479992 Kotlarsky et al. Jul 2013 B2
8490877 Kearney Jul 2013 B2
8517271 Kotlarsky et al. Aug 2013 B2
8523076 Good Sep 2013 B2
8528818 Ehrhart et al. Sep 2013 B2
8544737 Gomez et al. Oct 2013 B2
8548420 Grunow et al. Oct 2013 B2
8550335 Samek et al. Oct 2013 B2
8550354 Gannon et al. Oct 2013 B2
8550357 Kearney Oct 2013 B2
8556174 Kosecki et al. Oct 2013 B2
8556176 Van Horn et al. Oct 2013 B2
8556177 Hussey et al. Oct 2013 B2
8559767 Barber et al. Oct 2013 B2
8561895 Gomez et al. Oct 2013 B2
8561903 Sauerwein Oct 2013 B2
8561905 Edmonds et al. Oct 2013 B2
8565107 Pease et al. Oct 2013 B2
8570343 Halstead Oct 2013 B2
8571307 Li et al. Oct 2013 B2
8576390 Nunnink Nov 2013 B1
8579200 Samek et al. Nov 2013 B2
8583924 Caballero et al. Nov 2013 B2
8584945 Wang et al. Nov 2013 B2
8587595 Wang Nov 2013 B2
8587697 Hussey et al. Nov 2013 B2
8588869 Sauerwein et al. Nov 2013 B2
8590789 Nahill et al. Nov 2013 B2
8594425 Gurman et al. Nov 2013 B2
8596539 Havens et al. Dec 2013 B2
8596542 Havens et al. Dec 2013 B2
8596543 Havens et al. Dec 2013 B2
8599271 Havens et al. Dec 2013 B2
8599957 Peake et al. Dec 2013 B2
8600158 Li et al. Dec 2013 B2
8600167 Showering Dec 2013 B2
8602309 Longacre et al. Dec 2013 B2
8608053 Meier et al. Dec 2013 B2
8608071 Liu et al. Dec 2013 B2
8611309 Wang et al. Dec 2013 B2
8615487 Gomez et al. Dec 2013 B2
8621123 Caballero Dec 2013 B2
8622303 Meier et al. Jan 2014 B2
8628013 Ding Jan 2014 B2
8628015 Wang et al. Jan 2014 B2
8628016 Winegar Jan 2014 B2
8629926 Wang Jan 2014 B2
8630491 Longacre et al. Jan 2014 B2
8635309 Berthiaume et al. Jan 2014 B2
8636200 Kearney Jan 2014 B2
8636212 Nahill et al. Jan 2014 B2
8636215 Ding et al. Jan 2014 B2
8636224 Wang Jan 2014 B2
8638806 Wang et al. Jan 2014 B2
8640958 Lu et al. Feb 2014 B2
8640960 Wang et al. Feb 2014 B2
8643717 Li et al. Feb 2014 B2
8646692 Meier et al. Feb 2014 B2
8646694 Wang et al. Feb 2014 B2
8657200 Ren et al. Feb 2014 B2
8659397 Vargo et al. Feb 2014 B2
8668149 Good Mar 2014 B2
8678285 Kearney Mar 2014 B2
8678286 Smith et al. Mar 2014 B2
8682077 Longacre Mar 2014 B1
D702237 Oberpriller et al. Apr 2014 S
8687282 Feng et al. Apr 2014 B2
8692927 Pease et al. Apr 2014 B2
8695880 Bremer et al. Apr 2014 B2
8698949 Grunow et al. Apr 2014 B2
8702000 Barber et al. Apr 2014 B2
8717494 Gannon May 2014 B2
8720783 Biss et al. May 2014 B2
8723804 Fletcher et al. May 2014 B2
8723904 Marty et al. May 2014 B2
8727223 Wang May 2014 B2
8736909 Sato et al. May 2014 B2
8740082 Wilz Jun 2014 B2
8740085 Furlong et al. Jun 2014 B2
8746563 Hennick et al. Jun 2014 B2
8750445 Peake et al. Jun 2014 B2
8752766 Xian et al. Jun 2014 B2
8756059 Braho et al. Jun 2014 B2
8757495 Qu et al. Jun 2014 B2
8760563 Koziol et al. Jun 2014 B2
8777108 Coyle Jul 2014 B2
8777109 Oberpriller et al. Jul 2014 B2
8779898 Havens et al. Jul 2014 B2
8781520 Payne et al. Jul 2014 B2
8783573 Havens et al. Jul 2014 B2
8789757 Barten Jul 2014 B2
8789758 Hawley et al. Jul 2014 B2
8789759 Xian et al. Jul 2014 B2
8792688 Unsworth Jul 2014 B2
8794520 Wang et al. Aug 2014 B2
8794522 Ehrhart Aug 2014 B2
8794525 Amundsen et al. Aug 2014 B2
8794526 Wang et al. Aug 2014 B2
8798367 Ellis Aug 2014 B2
8807431 Wang et al. Aug 2014 B2
8807432 Van Horn et al. Aug 2014 B2
8810779 Hilde Aug 2014 B1
8820630 Qu et al. Sep 2014 B2
8897596 Passmore et al. Nov 2014 B1
8928896 Kennington et al. Jan 2015 B2
9014441 Truyen et al. Apr 2015 B2
9082195 Holeva et al. Jul 2015 B2
9142035 Rotman Sep 2015 B1
9171278 Kong et al. Oct 2015 B1
9233470 Bradski et al. Jan 2016 B1
9235899 Kirmani et al. Jan 2016 B1
9299013 Curlander et al. Mar 2016 B1
9424749 Reed et al. Aug 2016 B1
9486921 Straszheim et al. Nov 2016 B1
20010027995 Patel et al. Oct 2001 A1
20010032879 He et al. Oct 2001 A1
20020054289 Thibault et al. May 2002 A1
20020067855 Chiu et al. Jun 2002 A1
20020109835 Goetz Aug 2002 A1
20020118874 Chung et al. Aug 2002 A1
20020158873 Williamson Oct 2002 A1
20020167677 Okada et al. Nov 2002 A1
20020179708 Zhu et al. Dec 2002 A1
20020196534 Lizotte et al. Dec 2002 A1
20030038179 Tsikos et al. Feb 2003 A1
20030053513 Vatan et al. Mar 2003 A1
20030063086 Baumberg Apr 2003 A1
20030078755 Leutz et al. Apr 2003 A1
20030091227 Chang et al. May 2003 A1
20030156756 Gokturk et al. Aug 2003 A1
20030197138 Pease et al. Oct 2003 A1
20030225712 Cooper et al. Dec 2003 A1
20030235331 Kawaike et al. Dec 2003 A1
20040008259 Gokturk et al. Jan 2004 A1
20040019274 Galloway et al. Jan 2004 A1
20040024754 Mane et al. Feb 2004 A1
20040066329 Zeitfuss et al. Apr 2004 A1
20040073359 Ichijo et al. Apr 2004 A1
20040083025 Yamanouchi et al. Apr 2004 A1
20040089482 Ramsden et al. May 2004 A1
20040098146 Katae et al. May 2004 A1
20040105580 Hager et al. Jun 2004 A1
20040118928 Patel et al. Jun 2004 A1
20040122779 Stickler et al. Jun 2004 A1
20040155975 Hart et al. Aug 2004 A1
20040165090 Ning Aug 2004 A1
20040184041 Schopp Sep 2004 A1
20040211836 Patel et al. Oct 2004 A1
20040214623 Takahashi et al. Oct 2004 A1
20040233461 Armstrong et al. Nov 2004 A1
20040258353 Gluckstad et al. Dec 2004 A1
20050006477 Patel Jan 2005 A1
20050117215 Lange Jun 2005 A1
20050128193 Popescu et al. Jun 2005 A1
20050128196 Popescu et al. Jun 2005 A1
20050168488 Montague Aug 2005 A1
20050211782 Martin Sep 2005 A1
20050257748 Kriesel et al. Nov 2005 A1
20050264867 Cho et al. Dec 2005 A1
20060047704 Gopalakrishnan Mar 2006 A1
20060078226 Zhou Apr 2006 A1
20060108266 Bowers et al. May 2006 A1
20060112023 Horhann May 2006 A1
20060151604 Zhu Jul 2006 A1
20060159307 Anderson et al. Jul 2006 A1
20060159344 Shao et al. Jul 2006 A1
20060213999 Wang et al. Sep 2006 A1
20060232681 Okada Oct 2006 A1
20060255150 Longacre Nov 2006 A1
20060269165 Viswanathan Nov 2006 A1
20060291719 Ikeda et al. Dec 2006 A1
20070003154 Sun et al. Jan 2007 A1
20070025612 Iwasaki et al. Feb 2007 A1
20070031064 Zhao et al. Feb 2007 A1
20070063048 Havens et al. Mar 2007 A1
20070116357 Dewaele May 2007 A1
20070127022 Cohen et al. Jun 2007 A1
20070143082 Degnan Jun 2007 A1
20070153293 Gruhlke et al. Jul 2007 A1
20070171220 Kriveshko Jul 2007 A1
20070177011 Lewin et al. Aug 2007 A1
20070181685 Zhu et al. Aug 2007 A1
20070237356 Dwinell et al. Oct 2007 A1
20070291031 Konev et al. Dec 2007 A1
20070299338 Stevick et al. Dec 2007 A1
20080013793 Hillis et al. Jan 2008 A1
20080035390 Wurz Feb 2008 A1
20080056536 Hildreth et al. Mar 2008 A1
20080062164 Bassi et al. Mar 2008 A1
20080077265 Boyden Mar 2008 A1
20080079955 Storm Apr 2008 A1
20080164074 Wurz Jun 2008 A1
20080185432 Caballero et al. Aug 2008 A1
20080204476 Montague Aug 2008 A1
20080212168 Olmstead et al. Sep 2008 A1
20080247635 Davis et al. Oct 2008 A1
20080273191 Kim et al. Nov 2008 A1
20080273210 Hilde Nov 2008 A1
20080278790 Boesser et al. Nov 2008 A1
20090059004 Bochicchio Mar 2009 A1
20090081008 Somin et al. Mar 2009 A1
20090095047 Patel et al. Apr 2009 A1
20090134221 Zhu et al. May 2009 A1
20090195790 Zhu et al. Aug 2009 A1
20090225333 Bendall et al. Sep 2009 A1
20090237411 Gossweiler et al. Sep 2009 A1
20090268023 Hsieh Oct 2009 A1
20090272724 Gubler Nov 2009 A1
20090273770 Bauhahn et al. Nov 2009 A1
20090313948 Buckley et al. Dec 2009 A1
20090318815 Barnes et al. Dec 2009 A1
20090323084 Dunn et al. Dec 2009 A1
20090323121 Valkenburg Dec 2009 A1
20100035637 Varanasi et al. Feb 2010 A1
20100060604 Zwart et al. Mar 2010 A1
20100091104 Sprigle Apr 2010 A1
20100113153 Yen May 2010 A1
20100118200 Gelman et al. May 2010 A1
20100128109 Banks May 2010 A1
20100161170 Siris Jun 2010 A1
20100171740 Andersen et al. Jul 2010 A1
20100172567 Prokoski Jul 2010 A1
20100177076 Essinger et al. Jul 2010 A1
20100177080 Essinger et al. Jul 2010 A1
20100177707 Essinger et al. Jul 2010 A1
20100177749 Essinger et al. Jul 2010 A1
20100202702 Benos et al. Aug 2010 A1
20100208039 Stettner Aug 2010 A1
20100211355 Horst et al. Aug 2010 A1
20100217678 Goncalves Aug 2010 A1
20100220849 Colbert et al. Sep 2010 A1
20100220894 Ackley et al. Sep 2010 A1
20100223276 Al-Shameri et al. Sep 2010 A1
20100245850 Lee et al. Sep 2010 A1
20100254611 Arnz Oct 2010 A1
20100303336 Abraham Dec 2010 A1
20100315413 Izadi et al. Dec 2010 A1
20100321482 Cleveland Dec 2010 A1
20110019155 Daniel et al. Jan 2011 A1
20110040192 Brenner et al. Feb 2011 A1
20110043609 Choi et al. Feb 2011 A1
20110099474 Grossman et al. Apr 2011 A1
20110169999 Grunow et al. Jul 2011 A1
20110188054 Petronius et al. Aug 2011 A1
20110188741 Sones et al. Aug 2011 A1
20110202554 Powilleit et al. Aug 2011 A1
20110234389 Mellin Sep 2011 A1
20110235854 Berger et al. Sep 2011 A1
20110249864 Venkatesan et al. Oct 2011 A1
20110254840 Halstead Oct 2011 A1
20110279916 Brown et al. Nov 2011 A1
20110286007 Pangrazio et al. Nov 2011 A1
20110286628 Goncalves et al. Nov 2011 A1
20110288818 Thierman Nov 2011 A1
20110301994 Tieman Dec 2011 A1
20110303748 Lemma et al. Dec 2011 A1
20110310227 Konertz et al. Dec 2011 A1
20120024952 Chen Feb 2012 A1
20120056982 Katz et al. Mar 2012 A1
20120057345 Kuchibhotla Mar 2012 A1
20120067955 Rowe Mar 2012 A1
20120074227 Ferren et al. Mar 2012 A1
20120081714 Pangrazio et al. Apr 2012 A1
20120111946 Golant May 2012 A1
20120113223 Hilliges et al. May 2012 A1
20120126000 Kunzig et al. May 2012 A1
20120138685 Qu et al. Jun 2012 A1
20120140300 Freeman Jun 2012 A1
20120168512 Kotlarsky et al. Jul 2012 A1
20120179665 Baarman et al. Jul 2012 A1
20120185094 Rosenstein et al. Jul 2012 A1
20120190386 Anderson Jul 2012 A1
20120193423 Samek Aug 2012 A1
20120197464 Wang et al. Aug 2012 A1
20120203647 Smith Aug 2012 A1
20120218436 Rhoads et al. Aug 2012 A1
20120223141 Good et al. Sep 2012 A1
20120224026 Bayer et al. Sep 2012 A1
20120224743 Rodriguez et al. Sep 2012 A1
20120228382 Havens et al. Sep 2012 A1
20120236288 Stanley Sep 2012 A1
20120242852 Hayward et al. Sep 2012 A1
20120113250 Farlotti et al. Oct 2012 A1
20120248188 Kearney Oct 2012 A1
20120256901 Bendall Oct 2012 A1
20120261474 Kawashime et al. Oct 2012 A1
20120262558 Boger et al. Oct 2012 A1
20120280908 Rhoads et al. Nov 2012 A1
20120282905 Owen Nov 2012 A1
20120282911 Davis et al. Nov 2012 A1
20120284012 Rodriguez et al. Nov 2012 A1
20120284122 Brandis Nov 2012 A1
20120284339 Rodriguez Nov 2012 A1
20120284593 Rodriguez Nov 2012 A1
20120293610 Doepke et al. Nov 2012 A1
20120293625 Schneider et al. Nov 2012 A1
20120294549 Doepke Nov 2012 A1
20120299961 Ramkumar et al. Nov 2012 A1
20120300991 Mikio Nov 2012 A1
20120313848 Galor et al. Dec 2012 A1
20120314030 Datta Dec 2012 A1
20120314058 Bendall et al. Dec 2012 A1
20120316820 Nakazato et al. Dec 2012 A1
20130019278 Sun et al. Jan 2013 A1
20130038881 Pesach et al. Feb 2013 A1
20130038941 Pesach et al. Feb 2013 A1
20130043312 Van Horn Feb 2013 A1
20130050426 Sarmast et al. Feb 2013 A1
20130056285 Meagher Mar 2013 A1
20130070322 Fritz et al. Mar 2013 A1
20130075168 Amundsen et al. Mar 2013 A1
20130094069 Lee et al. Apr 2013 A1
20130101158 Lloyd et al. Apr 2013 A1
20130156267 Muraoka et al. Jun 2013 A1
20130175341 Kearney et al. Jul 2013 A1
20130175343 Good Jul 2013 A1
20130200150 Reynolds et al. Aug 2013 A1
20130200158 Feng et al. Aug 2013 A1
20130201288 Billerbeck et al. Aug 2013 A1
20130208164 Cazier et al. Aug 2013 A1
20130211790 Loveland et al. Aug 2013 A1
20130223673 Davis et al. Aug 2013 A1
20130256418 Havens et al. Oct 2013 A1
20130257744 Daghigh et al. Oct 2013 A1
20130257759 Daghigh Oct 2013 A1
20130270346 Xian et al. Oct 2013 A1
20130277430 Zumsteg Oct 2013 A1
20130278425 Cunningham et al. Oct 2013 A1
20130287258 Kearney Oct 2013 A1
20130291998 Konnerth Nov 2013 A1
20130292475 Kotlarsky et al. Nov 2013 A1
20130292477 Hennick et al. Nov 2013 A1
20130293539 Hunt et al. Nov 2013 A1
20130293540 Laffargue et al. Nov 2013 A1
20130306728 Thuries et al. Nov 2013 A1
20130306730 Brady et al. Nov 2013 A1
20130306731 Pedraro Nov 2013 A1
20130307964 Bremer et al. Nov 2013 A1
20130308013 Li et al. Nov 2013 A1
20130308625 Corcoran Nov 2013 A1
20130313324 Koziol et al. Nov 2013 A1
20130313325 Wilz et al. Nov 2013 A1
20130329012 Bartos Dec 2013 A1
20130329013 Metois Dec 2013 A1
20130341399 Xian et al. Dec 2013 A1
20130342343 Harring et al. Dec 2013 A1
20130342717 Havens et al. Dec 2013 A1
20140001267 Giordano et al. Jan 2014 A1
20140002828 Laffargue et al. Jan 2014 A1
20140008430 Soule et al. Jan 2014 A1
20140008439 Wang Jan 2014 A1
20140009586 McNamer et al. Jan 2014 A1
20140021259 Moed et al. Jan 2014 A1
20140025584 Liu et al. Jan 2014 A1
20140027518 Edmonds et al. Jan 2014 A1
20140031665 Pinto et al. Jan 2014 A1
20140034731 Gao et al. Feb 2014 A1
20140034734 Sauerwein Feb 2014 A1
20140036848 Pease et al. Feb 2014 A1
20140039674 Motoyama et al. Feb 2014 A1
20140039693 Havens et al. Feb 2014 A1
20140042814 Kather et al. Feb 2014 A1
20140049120 Kohtz et al. Feb 2014 A1
20140049635 Laffargue et al. Feb 2014 A1
20140058612 Wong et al. Feb 2014 A1
20140061305 Nahill et al. Mar 2014 A1
20140061306 Wu et al. Mar 2014 A1
20140062709 Hyer et al. Mar 2014 A1
20140063289 Hussey et al. Mar 2014 A1
20140064624 Kim et al. Mar 2014 A1
20140066136 Sauerwein et al. Mar 2014 A1
20140067104 Osterhout Mar 2014 A1
20140067692 Ye et al. Mar 2014 A1
20140070005 Nahill et al. Mar 2014 A1
20140071430 Hansen et al. Mar 2014 A1
20140071840 Venancio Mar 2014 A1
20140074746 Wang Mar 2014 A1
20140075846 Woodburn Mar 2014 A1
20140076974 Havens et al. Mar 2014 A1
20140078341 Havens et al. Mar 2014 A1
20140078342 Li et al. Mar 2014 A1
20140078345 Showering Mar 2014 A1
20140079297 Tadayon et al. Mar 2014 A1
20140084068 Gillet et al. Mar 2014 A1
20140091147 Evans et al. Apr 2014 A1
20140097238 Ghazizadeh Apr 2014 A1
20140097249 Gomez et al. Apr 2014 A1
20140098091 Hori Apr 2014 A1
20140098792 Wang et al. Apr 2014 A1
20140100774 Showering Apr 2014 A1
20140100813 Showering Apr 2014 A1
20140103115 Meier et al. Apr 2014 A1
20140104413 McCloskey et al. Apr 2014 A1
20140104414 McCloskey et al. Apr 2014 A1
20140104416 Li et al. Apr 2014 A1
20140104451 Todeschini et al. Apr 2014 A1
20140104664 Lee Apr 2014 A1
20140106594 Skvoretz Apr 2014 A1
20140106725 Sauerwein Apr 2014 A1
20140108010 Maltseff et al. Apr 2014 A1
20140108402 Gomez et al. Apr 2014 A1
20140108682 Caballero Apr 2014 A1
20140110485 Toa et al. Apr 2014 A1
20140114530 Fitch et al. Apr 2014 A1
20140121438 Kearney May 2014 A1
20140121445 Ding et al. May 2014 A1
20140124577 Wang et al. May 2014 A1
20140124579 Ding May 2014 A1
20140125842 Winegar May 2014 A1
20140125853 Wang May 2014 A1
20140125999 Longacre et al. May 2014 A1
20140129378 Richardson May 2014 A1
20140131441 Nahill et al. May 2014 A1
20140131443 Smith May 2014 A1
20140131444 Wang May 2014 A1
20140131448 Xian et al. May 2014 A1
20140133379 Wang et al. May 2014 A1
20140135984 Hirata May 2014 A1
20140136208 Maltseff et al. May 2014 A1
20140139654 Takahashi May 2014 A1
20140140585 Wang May 2014 A1
20140142398 Patil et al. May 2014 A1
20140151453 Meier et al. Jun 2014 A1
20140152882 Samek et al. Jun 2014 A1
20140152975 Ko Jun 2014 A1
20140158468 Adami Jun 2014 A1
20140158770 Sevier et al. Jun 2014 A1
20140159869 Zumsteg et al. Jun 2014 A1
20140160329 Ren et al. Jun 2014 A1
20140166755 Liu et al. Jun 2014 A1
20140166757 Smith Jun 2014 A1
20140166759 Liu et al. Jun 2014 A1
20140166760 Meier et al. Jun 2014 A1
20140166761 Todeschini et al. Jun 2014 A1
20140168380 Heidemann et al. Jun 2014 A1
20140168787 Wang et al. Jun 2014 A1
20140175165 Havens et al. Jun 2014 A1
20140175172 Jovanovski et al. Jun 2014 A1
20140175174 Barber Jun 2014 A1
20140177931 Kocherscheidt et al. Jun 2014 A1
20140191644 Chaney Jul 2014 A1
20140191913 Ge et al. Jul 2014 A1
20140192187 Atwell et al. Jul 2014 A1
20140192551 Masaki Jul 2014 A1
20140197238 Lui et al. Jul 2014 A1
20140197239 Havens et al. Jul 2014 A1
20140197304 Feng et al. Jul 2014 A1
20140201126 Zadeh et al. Jul 2014 A1
20140203087 Smith et al. Jul 2014 A1
20140204268 Grunow et al. Jul 2014 A1
20140205150 Ogawa Jul 2014 A1
20140214631 Hansen Jul 2014 A1
20140217166 Berthiaume et al. Aug 2014 A1
20140217180 Liu Aug 2014 A1
20140225918 Mittal et al. Aug 2014 A1
20140225985 Klusza et al. Aug 2014 A1
20140240454 Lee Aug 2014 A1
20140247279 Nicholas et al. Sep 2014 A1
20140247280 Nicholas et al. Sep 2014 A1
20140267609 Laffargue Sep 2014 A1
20140268093 Tohme et al. Sep 2014 A1
20140270361 Amma et al. Sep 2014 A1
20140306833 Ricci Oct 2014 A1
20140307855 Withagen et al. Oct 2014 A1
20140313527 Askan Oct 2014 A1
20140319219 Liu et al. Oct 2014 A1
20140320408 Zagorsek et al. Oct 2014 A1
20140333775 Naikal et al. Nov 2014 A1
20140347533 Ovsiannikov et al. Nov 2014 A1
20140350710 Gopalkrishnan et al. Nov 2014 A1
20140379613 Nishitani et al. Dec 2014 A1
20150009100 Haneda et al. Jan 2015 A1
20150009301 Ribnick et al. Jan 2015 A1
20150009338 Laffargue et al. Jan 2015 A1
20150036876 Marrion et al. Feb 2015 A1
20150042791 Metois et al. Feb 2015 A1
20150049347 Laffargue et al. Feb 2015 A1
20150062369 Gehring et al. Mar 2015 A1
20150063676 Lloyd et al. Mar 2015 A1
20150116498 Vartiainen et al. Apr 2015 A1
20150149946 Benos et al. May 2015 A1
20150163474 You Jun 2015 A1
20150178900 Kim et al. Jun 2015 A1
20150201181 Moore et al. Jul 2015 A1
20150204662 Kobayashi et al. Jul 2015 A1
20150213647 Laffargue et al. Jul 2015 A1
20150229838 Hakim et al. Aug 2015 A1
20150269403 Lei et al. Sep 2015 A1
20150276379 Ni et al. Oct 2015 A1
20150308816 Laffargue et al. Oct 2015 A1
20150316368 Moench et al. Nov 2015 A1
20150325036 Lee Nov 2015 A1
20150355470 Herschbach Dec 2015 A1
20160048725 Holz et al. Feb 2016 A1
20160063429 Varley et al. Mar 2016 A1
20160070982 Jachalsky et al. Mar 2016 A1
20160088287 Sadi et al. Mar 2016 A1
20160090283 Svensson et al. Mar 2016 A1
20160090284 Svensson et al. Mar 2016 A1
20160101936 Chamberlin Apr 2016 A1
20160102975 McCloskey et al. Apr 2016 A1
20160104019 Todeschini et al. Apr 2016 A1
20160104274 Jovanovski et al. Apr 2016 A1
20160109219 Ackley et al. Apr 2016 A1
20160109220 Laffargue et al. Apr 2016 A1
20160109224 Thuries et al. Apr 2016 A1
20160112631 Ackley et al. Apr 2016 A1
20160112643 Laffargue et al. Apr 2016 A1
20160138247 Conway et al. May 2016 A1
20160138248 Conway et al. May 2016 A1
20160138249 Svensson et al. May 2016 A1
20160169665 Deschenes et al. Jun 2016 A1
20160187186 Coleman et al. Jun 2016 A1
20160187210 Coleman et al. Jun 2016 A1
20160191801 Sivan Jun 2016 A1
20160202478 Masson et al. Jul 2016 A1
20160343176 Ackley Nov 2016 A1
20160370220 Ackley et al. Dec 2016 A1
20170010141 Ackley Jan 2017 A1
20170017301 Doornenbal et al. Jan 2017 A1
20170115490 Hsieh et al. Apr 2017 A1
20170182942 Hardy et al. Jun 2017 A1
Foreign Referenced Citations (62)
Number Date Country
2004212587 Apr 2005 AU
201139117 Oct 2008 CN
3335760 Apr 1985 DE
10210813 Oct 2003 DE
102007037282 Mar 2008 DE
1111435 Jun 2001 EP
1443312 Aug 2004 EP
2013117 Jan 2009 EP
2286932 Feb 2011 EP
2372648 Oct 2011 EP
2381421 Oct 2011 EP
2533009 Dec 2012 EP
2533009 Dec 2012 EP
2562715 Feb 2013 EP
2722656 Apr 2014 EP
2779027 Sep 2014 EP
2833323 Feb 2015 EP
2843590 Mar 2015 EP
2845170 Mar 2015 EP
2966595 Jan 2016 EP
3006893 Mar 2016 EP
3012601 Mar 2016 EP
3007096 Apr 2016 EP
2503978 Jan 2014 GB
2503978 Jan 2014 GB
252053 Oct 2015 GB
2531928 May 2016 GB
H04129902 Apr 1992 JP
200696457 Apr 2006 JP
2007084162 Apr 2007 JP
2008210276 Sep 2008 JP
2014210646 Nov 2014 JP
2014210646 Nov 2014 JP
2015174705 Oct 2015 JP
20100020115 Feb 2010 KR
20110013200 Feb 2011 KR
20110117020 Oct 2011 KR
20120028109 Mar 2012 KR
9640452 Dec 1996 WO
0077726 Dec 2000 WO
0114836 Mar 2001 WO
2006095110 Sep 2006 WO
2007015059 Feb 2007 WO
200712554 Nov 2007 WO
2011017241 Feb 2011 WO
2012175731 Dec 2012 WO
2013021157 Feb 2013 WO
2013033442 Mar 2013 WO
2013163789 Nov 2013 WO
2013166368 Nov 2013 WO
2013173985 Nov 2013 WO
20130184340 Dec 2013 WO
2014019130 Feb 2014 WO
2014023697 Feb 2014 WO
2014102341 Jul 2014 WO
2014110495 Jul 2014 WO
2014149702 Sep 2014 WO
2014151746 Sep 2014 WO
2014151746 Sep 2014 WO
2015006865 Jan 2015 WO
2016020038 Feb 2016 WO
2016061699 Apr 2016 WO
Non-Patent Literature Citations (112)
Entry
Joseph O'Rourke, “Finding Minimal Enclosing Boxes”, 1985, International Journal of Computer and Information Sciences, vol. 14, No. 3 pp. 183-199.
Oberpriller et al., “Electronic Device Case” submitted Dec. 26, 2014 pp. 44.
Boavida et al., Dam monitoring using combined terrestrial imaging systems, 2009 Civil Engineering Survey Dec./Jan. 2009 pp. 33-38.
European Patent Office Action for Application No. 14157971.4-1906, dated Jul. 16, 2014, 5 pages.
European Patent Search Report for Application No. 14157971.4-1906, dated Jun. 30, 2014, 6 pages.
Caulier, Yannick et al., “A New Type of Color-Coded Light Structures for an Adapted and Rapid Determination of Point Correspondences for 3D Reconstruction.” Proc. of SPIE, vol. 8082 808232-3; 2011; 8 pages.
Kazantsev, Aleksei et al. “Robust Pseudo-Random Coded Colored STructured Light Techniques for 3D Object Model Recovery”; ROSE 2008 IEEE International Workshop on Robotic and Sensors Environments (Oct. 17-18, 2008) , 6 pages.
Mouaddib E. et al. “Recent Progress in Structured Light in order to Solve the Correspondence Problem in Stereo Vision” Proceedings of the 1997 IEEE International Conference on Robotics and Automation, Apr. 1997; 7 pages.
Peter Clarke, Actuator Developer Claims Anti-Shake Breakthrough For Smartphone Cams, Electronic Engineering Times, p. 24, May 16, 2011.
European Search Report for application No. EP13186043 (now EP2722656 (Apr. 23, 2014)): Total pp. 7.
International Search Report for PCT/US2013/039438 (WO2013166368), dated Oct. 1, 2013, 7 pages.
European Office Action for application EP 13186043, dated Jun. 12, 2014(now EP2722656 (Apr. 23, 2014)), Total of 6 pages.
Wikipedia, YUV description and definition, downloaded from http://www.wikipeida.org/wiki/YUV on Jun. 29, 2012, 10 pages.
YUV Pixel Format, downloaded from http://www.fource.org/yuv.php on Jun. 29, 2012; 13 pages.
YUV to RGB Conversion, downloaded from http://www.fource.org/fccyvrgb.php on Jun. 29, 2012; 5 pages.
Benos et al., “Semi-Automatic Dimensioning with Imager of a Portable Device,” U.S. Appl. No. 61/149,912, filed Feb. 4, 2009 (now expired), 56 pages.
Dimensional Weight—Wikipedia, the Free Encyclopedia, URL=http://en.wikipedia.org/wiki/Dimensional_weight, download date Aug. 1, 2008, 2 pages.
Dimensioning—Wikipedia, the Free Encyclopedia, URL=http://en.wikipedia.org/wiki/Dimensioning, download date Aug. 1, 2008, 1 page.
Search Report and Opinion in related GB Application No. 1517112.7, dated Feb. 19, 2016 6 Pages (GB2503978 is a commonly owned now abandoned application and not cited above).
Extended European Search Report in counterpart European Application No. 15182675.7 dated Dec. 4, 2015, pp. 1-10 (References previously cited).
Proesmans, Marc et al. “Active Acquisition of 3D Shape for Moving Objects” 0-7803-3258-X/96 1996 IEEE; 4 pages.
Leotta, Matthew, Generic, Deformable Models for 3-D Vehicle Surveillance, May 2010, Doctoral Dissertation, Brown University, Providence RI, 248 pages.
Ward, Benjamin, Interactive 3D Reconstruction from Video, Aug. 2012, Doctoral Thesis, Univesity Of Adelaide, Adelaide, South Australia, 157 pages.
Hood, Frederick W.; William A. Hoff, Robert King, Evaluation of An Interactive Technique for Creating Site Models from Range Data, Apr. 27-May 1, 1997 Proceedings of the ANS 7th Topical Meeting on Robotics & Remote Systems, Augusta GA, 9 pages.
Gupta, Alok; Range Image Segmentation for 3-D Objects Recognition, May 1988, Technical Reports (CIS), Paper 736, University of Pennsylvania Department of Computer and Information Science, retrieved from Http://repository.upenn.edu/cis_reports/736, Accessed May 31, 2015, 157 pages.
Reisner-Kollmann,Irene; Anton L Fuhrmann, Werner Purgathofer, Interactive Reconstruction of Industrial Sites Using Parametric Models, May 2010, Proceedings of the 26th Spring Conference of Computer Graphics SCCG ″10, 8 pages.
Drummond, Tom; Roberto Cipolla, Real-Time Visual Tracking of Complex Structures, Jul. 2002, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, No. 7; 15 pages.
Zhang, Zhaoxiang; Tieniu Tan, Kaiqi Huang, Yunhong Wang; Three-Dimensional Deformable-Model-based Localization and Recognition of Road Vehicles; IEEE Transactions on Image Processing, vol. 21, No. 1, Jan. 2012, 13 pages.
Leotta, Matthew J.; Joseph L. Mundy; Predicting High Resolution Image Edges with a Generic, Adaptive, 3-D Vehicle Model; IEEE Conference on Computer Vision and Pattern Recognition, 2009; 8 pages.
Spiller, Jonathan; Object Localization Using Deformable Templates, Master's Dissertation, University of the Witwatersrand, Johannesburg, South Africa, 2007; 74 pages.
EP Search and Written Opinion Report in related matter EP Application No. 14181437.6, dated Mar. 26, 2015, 7 pages.
Hetzel, Gunter et al.; “3D Object Recognition from Range Images using Local Feature Histograms,”, Proceedings 2OO1 IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2001. Kauai, Hawaii, Dec. 8-14, 2001; pp. 394-399, XP010584149, ISBN: 978-0-7695-1272-3.
Intention to Grant in counterpart European Application No. 14157971.4 dated Apr. 14, 2015, pp. 1-8.
Decision to Grant in counterpart European Application No. 14157971.4 dated Aug. 6, 2015, pp. 1-2.
Salvi, Joaquim et al. “Pattern Codification Strategies in Structured Light Systems” published in Pattern Recognition; The Journal of the Pattern Recognition Society, Received Mar. 6, 2003; Accepted Oct. 2, 2003; 23 pages.
James Chamberlin, “System and Method for Picking Validation”, U.S. Appl. No. 14/865,797, filed Sep. 25, 2015, 44 pages, not yet published.
Jovanovski et al., “Image-Stitching for Dimensioning”, U.S. Appl. No. 14/870,488, filed Sep. 30, 2015, 45 pages, not yet published.
Todeschini et al.; “Depth Sensor Based Auto-Focus System for an Indicia Scanner,” U.S. Appl. No. 14/872,176, filed Oct. 1, 2015, 44 pages, not yet published.
McCloskey et al., “Methods for Improving the Accuracy of Dimensioning-System Measurements,” U.S. Appl. No. 14/873,613, filed Sep. 2, 2015, 47 pages, not yet published.
McCloskey et al., “Image Transformation for Indicia Reading,” U.S. Appl. No. 14/982,032, filed Oct. 30, 2015, 48 pages, not yet published.
M.Zahid Gurbuz, Selim Akyokus, Ibrahim Emiroglu, Aysun Guran, An Efficient Algorithm for 3D Rectangular Box Packing, 2009, Applied Automatic Systems: Proceedings of Selected AAS 2009 Papers, pp. 131-134 [Examiner cited art in related US matter with Notice of Allowance dated Aug. 11, 2016].
U.S. Appl. No. 15/182,636; H. Sprague Ackley et al., Automatic Mode Switching In a Volumer Dimensioner, not yet published, filed Jun. 15, 2016, 53 pages.
European Search Report for related EP Application No. 16152477.2, dated May 24, 2016, 8 pages. [New Reference cited herein; Reference DE102007037282 A1 and its US Counterparts have been previously cited.].
Search Report and Opinion in related GB Application No. 1517112.7, dated Feb. 19, 2016, 6 Pages.
“A one-step intrinsic and extrinsic calibration method for taster line scanner operation in coordinate measuring machines”, dated Apr. 1, 2009, Measurement Science and Technology, IOP, Bristol, GB, vol. 20, No. 4; 12 pages.
Search Report and Opinion in Related EP Application 15176943.7, dated Jan. 8, 2016, 8 pages, (US Application 2014/0049635 has been previously cited).
Second Chinese Office Action in related CN Application No. 201520810685.6, dated Mar. 22, 2016, 5 pages, no references.
European Search Report in related EP Application No. 15190315.0, dated Apr. 1, 2016, 7 pages [Commonly owned Reference 2014/0104416 has been previously cited].
Second Chinese Office Action in related CN Application No. 2015220810562.2, dated Mar. 22, 2016, 5 pages. English Translation provided [No references].
European Search Report for related Application EP 151902491, dated Mar. 22, 2016, 7 pages.
Second Chinese Office Action in related CN Application No. 201520810313.3, dated Mar. 22, 2016, 5 pages. English Translation provided [No references].
European Extended Search Report in Related EP Application No. 16172995.9, dated Aug. 22, 2016, 11 pages (Only new references have been cited; U.S. Pat. No. 8,463,079 (formerly U.S. Publication 2010/0220894) and U.S. Publication 2001/0027955 have been previously cited.).
European Search Report from related EP Application No. 16168216.6, dated Oct. 20, 2016, 8 pages [New reference cited above; U.S. Publication 2014/0104413 has been previously cited].
United Kingdom combined Search and Examination Report in related GB Application No. 1607394.2, dated Oct. 19, 2016, 7 pages.
Lloyd, Ryan and Scott McCloskey, “Recognition of 3D Package Shapes for Singe Camera Metrology” IEEE Winter conference on Applications of computer Visiona, IEEE, Mar. 24, 2014, pp. 99-106, {retrieved on Jun. 16, 2014}, Authors are employees of common Applicant.
European Search Report for Related EP Application No. 15189214.8, dated Mar. 3, 2016, 9 pages.
Santolaria et al. “A one-step intrinsic and extrinsic calibration method for laster line scanner operation in coordinate measuring machines”, dated Apr. 1, 2009, Measurement Science and Technology, IOP, Bristol, GB, vol. 20, No. 4; 12 pages.
Search Report and Opinion in Related EP Application 15176943.7, dated Jan. 8, 2016, 8 pages.
European Search Report for related EP Application No. 15188440.0, dated Mar. 8, 2016, 8 pages.
United Kingdom Search Report in related application GB1517842.9, dated Apr. 8, 2016, 8 pages.
Great Britain Search Report for related Application On. GB1517843.7, dated Feb. 23, 2016; 8 pages.
Office Action in counterpart European Application No. 13186043.9 dated Sep. 30, 2015, pp. 1-7.
Lloyd et al., “System for Monitoring the Condition of Packages Throughout Transit”, U.S. Appl. No. 14/865,575, filed Sep. 25, 2015, 59 pages, not yet published.
McCloskey et al., “Image Transformation for Indicia Reading,” U.S. Appl. No. 14/928,032, filed Oct. 30, 2015, 48 pages, not yet published.
Great Britain Combined Search and Examination Report in related Application GB1517842.9, dated Apr. 8, 2016, 8 pages.
Search Report in counterpart European Application No. 15182675.7, dated Dec. 4, 2015, 10 pages.
Wikipedia, “3D projection” Downloaded on Nov. 25, 2015 from www.wikipedia.com, 4 pages.
M.Zahid Gurbuz, Selim Akyokus, Ibrahim Emiroglu, Aysun Guran, An Efficient Algorithm for 3D Rectangular Box Packing, 2009, Applied Automatic Systems: Proceedings of Selected AAS 2009 Papers, pp. 131-134.
European Extended Search Report in Related EP Application No. 16172995.9, dated Aug. 22, 2016, 11 pages.
European Extended search report in related EP Application No. 15190306.9, dated Sep. 9, 2016, 15 pages.
Collings et al., “The Applications and Technology of Phase-Only Liquid Crystal on Silicon Devices”, Journal of Display Technology, IEEE Service Center, New, York, NY, US, vol. 7, No. 3, Mar. 1, 2011 (Mar. 1, 2011), pp. 112-119.
European extended Search report in related EP Application 13785171.3, dated Sep. 19, 2016, 8 pages.
El-Hakim et al., “Multicamera vision-based approach to flexible feature measurement for inspection and reverse engineering”, published in Optical Engineering, Society of Photo-Optical Instrumentation Engineers, vol. 32, No. 9, Sep. 1, 1993, 15 pages.
El-Hakim et al., “A Knowledge-based Edge/Object Measurement Technique”, Retrieved from the Internet: URL: https://www.researchgate.net/profile/Sabry_E1 -Hakim/publication/44075058_A_Knowledge_Based_EdgeObject_Measurement_Technique/links/00b4953b5faa7d3304000000.pdf [retrieved on Jul. 15, 2016] dated Jan. 1, 1993, 9 pages.
H. Sprague Ackley, “Automatic Mode Switching in a Volume Dimensioner”, U.S. Appl. No. 15/182,636, filed Jun. 15, 2016, 53 pages, Not yet published.
Bosch Tool Corporation, “Operating/Safety Instruction for DLR 130”, dated Feb. 2, 2009, 36 pages.
European Search Report for related EP Application No. 16152477.2, dated May 24, 2016, 8 pages.
Mike Stensvold, “Get the Most Out of Variable Aperture Lenses”, published on www.OutdoorPhotogrpaher.com; dated Dec. 7, 2010; 4 pages, [As noted on search report retrieved from URL: http;//www.outdoorphotographer.com/gear/lenses/get-the-most-out-ofvariable-aperture-lenses.html on Feb. 9, 2016].
Padzensky, Ron; “Augmera; Gesture Control”, Dated Apr. 18, 2015, 15 pages [Examiner Cited Art in Office Action dated Jan. 20. 2017 in related Application.].
Grabowski, Ralph; “New Commands in AutoCADS 2010: Part 11 Smoothing 3D Mesh Objects” Dated 2011 (per examiner who cited reference), 6 pages, [Examiner Cited Art in Office Action dated Jan. 20, 2017 in related Application.].
Theodoropoulos, Gabriel; “Using Gesture Recognizers to Handle Pinch, Rotate, Pan, Swipe, and Tap Gestures” dated Aug. 25, 2014, 34 pages, [Examiner Cited Art in Office Action dated Jan. 20, 2017 in related Application].
European Extended Search Report in related EP Application No. 16190017.0, dated Jan. 4, 2017, 6 pages.
European Extended Search Report in related EP Application No. 16173429.8, dated Dec. 1, 2016, 8 pages [Only new references cited: US 2013/0038881 was previously cited].
Extended European Search Report in related EP Application No. 16175410.0, dated Dec. 13, 2016, 5 pages.
Wikipedia, “Microlens”, Downloaded from https://en.wikipedia.org/wiki/Microlens, pp. 3. {Cited by Examiner in Feb. 9, 2017 Final Office Action in related matter}.
Fukaya et al., “Characteristics of Speckle Random Pattern and Its Applications”, pp. 317-327, Nouv. Rey. Optique, t.6, n.6. (1975) {Cited by Examiner in Feb. 9, 2017 Final Office Action in related matter: downloaded Mar. 2, 2017 from http://iopscience.iop.org}.
European Examination report in related EP Application No. 14181437.6, dated Feb. 8, 2017, 5 pages [References have been previously cited].
European extended search report in related EP Application 16190833.0, dated Mar. 9, 2017, 8 pages [only new art has been cited; US Publication 2014/0034731 was previously cited].
United Kingdom Combined Search and Examination Report in related Application No. GB1620676.5, dated Mar. 8, 2017, 6 pages [References have been previously cited; WO2014/151746, WO2012/175731, US 2014/0313527, GB2503978].
European Exam Report in related , EP Application No. 16168216.6, dated Feb. 27, 2017, 5 pages, [References have been previously cited; WO2011/017241 and US 2014/0104413].
Thorlabs, Examiner Cited NPL in Advisory Action dated Apr. 12, 2017 in related commonly owned application, downloaded from https://www.thorlabs.com/newgrouppage9.cfm?objectgroup_id=6430, 4 pages.
EKSMA Optics, Examiner Cited NPL in Advisory Action dated Apr. 12, 2017 in related commonly owned application, downloaded from http://eksmaoptics.com/optical-systems/f-theta-lenses/f-theta-lens-for-1064-nm/, 2 pages.
Sill Optics, Examiner Cited NPL in Advisory Action dated Apr. 12, 2017 in related commonly owned application, http://www.silloptics.de/1/products/sill/encyclopedia/laser-optics/f-theta-lenses/, 4 pages.
Chinese Notice of Reexamination in related Chinese Application 201520810313.3, dated Mar. 14, 2017, English Computer Translation provided, 7 pages [No new art cited].
Extended European search report in related EP Application 16199707.7, dated Apr. 10, 2017, 15 pages.
Ulusoy et al., One-Shot Scanning using De Bruijn Spaced Grids, 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, 7 pages [Cited in EP Extended search report dated Apr. 10, 2017].
European Exam Report in related EP Application No. 15176943.7, dated Apr. 12, 2017, 6 pages [Art previously cited in this matter].
European Exam Report in related EP Application No. 15188440.0, dated Apr. 21, 2017, 4 pages [No new art to cite].
Ralph Grabowski, “Smothing 3D Mesh Objects,” New Commands in AutoCAD 2010: Part 11, Examiner Cited art in related matter Non Final Office Action dated May 19, 2017; 6 pages.
European Exam Report in related EP Application No. 16152477.2, dated Jun. 20, 2017, 4 pages.
European Exam Report in related EP Applciation 16172995.9, dated Jul. 6, 2017, 9 pages.
United Kingdom Search Report in related Application No. GB1700338.5, dated Jun. 30, 2017, 5 pages.
European Search Report in related EP Application No. 17175357.7, dated Aug. 17, 2017, pp. 1-7.
EP Search Report in related EP Application No. 17171844 dated Sep. 18, 2017. 4 pages.
EP Extended Search Report in related EP Applicaton No. 17174843.7 dated Oct. 17, 2017, 5 pages.
UK Further Exam Report in related UK Application No. GB1517842.9, dated Sep. 1, 2017, 5 pages.
Ulusoy, Ali Osman et al.; “One-Shot Scanning using De Bruijn Spaced Grids”, Brown University; 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, pp. 1786-1792 [Cited in EPO Search Report dated Dec. 5, 2017}.
Extended European Search report in related EP Application No. 17189496.7 dated Dec. 5, 2017; 9 pages.
Boavida et al., “Dam monitoring using combined terrestrial imaging systems”, 2009 Civil Engineering Survey Dec./Jan. 2009, pp. 33-38 {Cited in Notice of Allowance dated Sep. 15, 2017 in related matter}.
Extended European Search report in related Ep Application No. 17190323.0 dated Jan. 19, 2018; 6 pages.
Examination Report in related GB Application No. GB1517843.7, dated Jan. 19, 2018, 4 pages.
Examination Report in related EP Application No. 15190315, dated Jan. 26, 2018, 6 pages.
Related Publications (1)
Number Date Country
20140100813 A1 Apr 2014 US
Provisional Applications (1)
Number Date Country
61709606 Oct 2012 US