Distributed volumetric cargo sensor system

Information

  • Patent Grant
  • 11299219
  • Patent Number
    11,299,219
  • Date Filed
    Thursday, August 15, 2019
    5 years ago
  • Date Issued
    Tuesday, April 12, 2022
    2 years ago
Abstract
A cargo detection apparatus installed within a cargo container includes multiple distance sensors and one or more processors in communication with the distance sensors. The distance sensors, which are configured for mounting on or adjacent to a ceiling of the cargo container, generate distance signals. Each distance signal indicates a distance between a corresponding one of the distance sensors and a surface within the interior space of the cargo container. The one or more processors are operable to execute instructions to compare the distance indicated by each of the distance signals to a distance threshold, and generate a cargo-present indication if the distance is less than the distance threshold. The distance sensors may be laser-ranging time-of-flight sensors or ultrasonic sensors.
Description
FIELD

This invention relates to the field of cargo transportation. More particularly, this invention relates to a system for providing a volumetric occupancy reading to indicate the load status of a cargo container, such as a cargo trailer.


BACKGROUND

Knowledge of the volumetric occupancy of cargo containers, such as cargo trailers, is important to cargo carriers. If a dispatcher knows that there is room to accept more cargo in a particular trailer that is en route to a destination, the dispatcher can divert the trailer to pick up a load at a nearby customer's facility. In this way, owners of trailers can make more efficient use of their assets, increasing profitability and reducing waste.


Previous solutions were capable only of determining if cargo was present or absent in a container. Although optical imaging techniques have been used to determine the load state of cargo containers, such techniques have been unreliable and difficult to implement. Optical imaging techniques are prone to inaccuracy due to potentially large distances separating the imaging device from a distant portion of the trailer. These techniques require significant computing resources to provide an accurate cargo occupancy indication.


What is needed, therefore, is a cargo sensor system that can be used on a loaded trailer or other cargo container to determine the amount of loaded cargo without reliance on optical imaging techniques.


SUMMARY

The above and other needs are met by a cargo sensor system that uses distance sensors distributed throughout a cargo trailer or other type of cargo container. In a preferred embodiment, each distance sensor is a laser time-of-flight (ToF) sensor. However, the distance sensor may be any sensor that can determine the distance between the sensor and an impeding object, such as an ultrasonic sensor.


Embodiments described herein also provide a method for detecting cargo within a cargo container that has a ceiling, a floor, and an interior space disposed between the ceiling and the floor. In a preferred embodiment, the method includes:

    • (a) mounting multiple distance sensors on or adjacent to the ceiling of the cargo container;
    • (b) the distance sensors generating distance signals that each indicate a distance between a corresponding one of the distance sensors and a surface within the interior space of the cargo container;
    • (c) for each of the distance signals, comparing the distance to a distance threshold; and
    • (d) generating a cargo-present indication if the distance is less than the distance threshold.


In some embodiments, each of the distance sensors has a field of view, and

    • step (a) includes distributing the distance sensors across the ceiling of the cargo container such that their fields of view define multiple cargo detection areas distributed across the floor of the cargo container, wherein each cargo detection area corresponds to a portion of a total area of the floor of the cargo container, and
    • step (b) includes generating the distance signals such that each indicate a distance between a corresponding one of the distance sensors and a surface located within a corresponding one of the cargo detection areas.


In some embodiments, the method includes:

    • (e) determining how many cargo-present indications are generated in step (d); and
    • (f) calculating how much of the total area of the floor of the cargo container is occupied by cargo based on how many cargo-present indications are determined in step (e).


In some embodiments, the method includes:

    • (e) for each of the cargo-present indications generated in step (d), determining a cargo height value based on the distance between the corresponding distance sensor and the surface located within the corresponding cargo detection area; and
    • (f) calculating a total volume of the interior space of the cargo container occupied by cargo based on the cargo height values and area values for the portions of the total area of the floor of the cargo container in which cargo is present.


In some embodiments, step (a) includes mounting one or more sensor modules on or adjacent to the ceiling of the cargo container, wherein each of the one or more sensor modules contain one or more of the distance sensors.


In some embodiments, step (a) includes distributing the sensor modules across the length of a central portion of the ceiling of the cargo container.


In some embodiments, step (a) includes distributing the sensor modules across the length of the cargo container adjacent to locations at which the ceiling of the cargo container meets a wall of the cargo container.


In some embodiments, step (a) includes mounting one or more sensor modules on a door of the cargo container, so that the distance sensors contained in the sensor modules mounted on the door each have a field of view directed toward a back wall of the cargo container when the door is in a closed position and directed toward the floor of the cargo container when the door is in an open position.


In some embodiments, each of the distance signals generated in step (b) indicate a distance between a corresponding one of the distance sensors and either a surface of a piece of cargo or the floor of the cargo container.


In some embodiments, the ceiling of the cargo container is disposed at a height above the floor of the cargo container, and step (c) includes comparing the distance to a distance threshold that is less than the height.


In some embodiments, the method includes providing the cargo-present indications generated in step (d) to a tracker unit mounted on the cargo container, and based on the cargo-present indications, transmitting cargo loading information from the tracker unit to a backend server via a data communication network.


In another aspect, embodiments described herein provide an apparatus for detecting cargo within a cargo container that has a ceiling, a floor, and an interior space disposed between the ceiling and the floor. A preferred embodiment of the apparatus includes multiple distance sensors and one or more processors in communication with the distance sensors. The distance sensors, which are configured for mounting on or adjacent to the ceiling of the cargo container, generate distance signals. Each distance signal indicates a distance between a corresponding one of the distance sensors and a surface within the interior space of the cargo container. The one or more processors are operable to execute instructions to compare the distance indicated by each of the distance signals to a distance threshold, and generate a cargo-present indication if the distance is less than the distance threshold.


In some embodiments, each of the distance sensors has a field of view, and they are mounted to the ceiling of the cargo container such that their fields of view define multiple cargo detection areas distributed across the floor of the cargo container. Each cargo detection area corresponds to a portion of the total area of the floor of the cargo container. Each of the distance signals indicate a distance between a corresponding one of the distance sensors and a surface located within a corresponding one of the cargo detection areas.


In some embodiments, the one or more processors are operable to execute instructions to calculate how much of the total area of the floor of the cargo container is occupied by cargo based how many cargo-present indications have been generated.


In some embodiments, the one or more processors are operable to execute instructions to:

    • for each cargo-present indication, determine a cargo height value based on the distance between the corresponding distance sensor and the surface located within the corresponding cargo detection area; and
    • calculate the total volume of the interior space of the cargo container occupied by cargo based on cargo height values and area values for the portions of the total area of the floor of the cargo container in which cargo is present.


In some embodiments, the sensor modules are configured for mounting on or adjacent to the ceiling of the cargo container, and they each contain one or more distance sensors.


In some embodiments, the one or more processors are components of the one or more sensor modules.


In some embodiments, each distance sensor has a field of view that is aligned in a different direction from fields of view of other distance sensors contained in one and the same sensor module.


In some embodiments, the sensor modules are distributed across the length of the central portion of the ceiling of the cargo container.


In some embodiments, the sensor modules are distributed across the length of the cargo container adjacent to locations at which the ceiling of the cargo container meets a wall of the cargo container.


In some embodiments, one or more of the sensor modules are mounted on a door of the cargo container. The distance sensors contained in the sensor modules mounted on the door each have a field of view directed toward a back wall of the cargo container when the door is in a closed position, and directed toward the floor of the cargo container when the door is in an open position.


In some embodiments, each of the distance signals indicate a distance between a corresponding one of the distance sensors and either a surface of a piece of cargo or the floor of the cargo container.


In some embodiments, the ceiling of the cargo container is disposed at a height above the floor of the cargo container, and the one or more processors are operable to execute instructions to compare the distance to a distance threshold that is less than the height.


Some embodiments include a tracker unit configured for mounting on the cargo container. The tracker unit is operable to generate cargo loading information based on the cargo-present indication, and to transmit the cargo loading information from the tracker unit to a backend server via a data communication network.


In some embodiments, the one or more processors are components of the tracker unit.


In some embodiments, the one or more processors are components of the backend server.


In some embodiments, the distance sensors comprise laser-ranging time-of-flight sensors or ultrasonic sensors.





BRIEF DESCRIPTION OF THE DRAWINGS

Other embodiments of the invention will become apparent by reference to the detailed description in conjunction with the figures, wherein elements are not to scale, so as to more clearly show the details, wherein like reference numbers indicate like elements throughout the several views, and wherein:



FIG. 1 depicts a cargo sensing system according to an embodiment of the invention;



FIG. 2 depicts a typical cargo trailer attached to a tractor;



FIG. 3A depicts a side elevation view of the interior of a cargo trailer with cargo sensing modules installed therein according to a preferred embodiment;



FIG. 3B depicts a top plan view of a cargo trailer with cargo sensing modules installed therein according to a preferred embodiment;



FIG. 3C depicts an end elevation view of a cargo trailer with cargo sensing modules installed therein according to a preferred embodiment;



FIGS. 4A and 4B depict a sensor module according to a preferred embodiment;



FIGS. 5A and 5B depict cargo detection areas according to a preferred embodiment;



FIG. 6 depicts a top plan view of a cargo trailer with cargo detection areas according to a preferred embodiment;



FIG. 7A depicts a side elevation view of a cargo trailer with the trailer door closed and with cargo sensing modules installed therein according to a preferred embodiment;



FIG. 7B depicts a side elevation view of a cargo trailer with the trailer door open and with cargo sensing modules installed therein according to a preferred embodiment;



FIG. 8 depicts a sensor module according to an alternative embodiment;



FIG. 9 depicts a top plan view of a cargo trailer with cargo detection areas according to an alternative embodiment;



FIG. 10A depicts a side elevation view of the interior of a cargo trailer with cargo sensing modules installed therein according to an alternative embodiment;



FIG. 10B depicts a top plan view of a cargo trailer with cargo sensing modules installed therein according to an alternative embodiment; and



FIG. 10C depicts an end elevation view of a cargo trailer with cargo sensing modules installed therein according to an alternative embodiment.





DETAILED DESCRIPTION


FIG. 1 depicts an embodiment of a cargo sensing system 10. Generally, the system 10 includes one or more sensor modules 12 distributed throughout a cargo container, such as the trailer 20 depicted in FIG. 2. Each sensor module 12 includes one or more distance sensors 14, which in a preferred embodiment are laser-ranging time-of-flight (ToF) sensors, such as the ST Micro VL53L1X manufactured by ST Microelectronics. In other embodiments, the distance sensor may implement other technology to determine the distance between the sensor and an impeding object. In some embodiments, the sensors 14 are ultrasonic sensors. In a preferred embodiment, each sensor module 12 includes a processor 38 for processing information from the distance sensor(s) 14, and local area network communication electronics 36 for communicating with other system components as described hereinafter. The local network communications electronics 36 may comprise one or more interfaces to implement wired communications protocols, such as Inter-integrated Circuit (I2C), Serial Peripheral Interface (SPI), Universal Asynchronous Receiver/Transmitter (UART), or Ethernet, or to implement wireless communications protocols, such as Wi-Fi, Bluetooth, or Bluetooth Low Energy.


Data generated by the sensor modules 12 are provided to a trailer tracker unit 18 and/or a backend server 24 for processing as described hereinafter. The tracker unit 18 monitors the location and health of the trailer 20, and sends the current cargo loading state of the trailer to the backend server 24 based on cargo detected using the sensor modules 12. The tracker unit 18, which is typically located in or on the trailer 20, preferably includes a processor 28, GPS receiver electronics 30 for determining location coordinates of the trailer 20, local area network communication electronics 32 for communicating with the sensor modules 12, and wide area network communication electronics 34 (such as a wireless cellular data modem) for communicating with the backend server 24. The local network communications electronics 30 may comprise one or more interfaces to implement wired communications protocols, such as I2C, SPI, UART, or Ethernet, or to implement wireless communications protocols, such as Wi-Fi, Bluetooth, or Bluetooth Low Energy. The tracker unit 18 may also interface with multiple external sensors, such as door sensors and temperature sensors.


The sensor modules 12 preferably communicate with the tracker unit 18 and/or the backend server 24 through a network 16, which may comprise a local area network in the trailer 20, a wireless data communication network, the Internet, or a combination of such networks. The sensor modules 12 may also communicate with the tracker unit 18 directly through serial or parallel interfaces. Each sensor module 12 preferably includes a local processor that communicates with processors of other sensor modules 12 through the network 16 or through a series of serial ports.


The preferred ToF sensor 14 has an angular field of view of about 25 degrees. A single one of these sensors mounted on the ceiling of the trailer 20 and pointing directly downward has a field of view that covers about half of the width of the floor area of the trailer. In a preferred embodiment depicted in FIGS. 4A and 4B, each sensor module 12 incorporates a faceted structure having four or more sensors 14 with their lines-of-sight angularly offset from each other by about 12 degrees in four or more different directions. As depicted in FIG. 5A, this multi-sensor module 12 provides a sensor pattern comprising four or more detection areas 22 that cover the trailer floor from wall-to-wall. This pattern allows each sensor module 12 to provide an accurate reading of the cumulative height of cargo containers in the module's detection areas. A fifth sensor 14 may be added pointing directly down to cover the open spot in the middle (see FIG. 5B). In one embodiment, a sensor module positioned closest to the trailer rear door may have an additional sensor pointing toward the door, thereby allowing that particular module to detect whether the door is open or closed, either by sensing ambient light, sensing the distance to the door, or a combination of both.


In an embodiment depicted in FIGS. 3A-3C, 4A, 4B, and 6, six sensor modules 12 are distributed along the length of the ceiling of the trailer 20, each including four sensors 14, thereby covering substantially the entire floor of the trailer 20 with twenty-four sensors. In some embodiments that do not require this level of accuracy, the floor of the trailer may be more sparsely sampled and still provide acceptable results.


Although a preferred embodiment disperses sensors evenly across the length of the trailer 20, other embodiments include more sensors in areas of the trailer 20 that are more likely to contain cargo (such as in the front and the back) and fewer sensors in the broad middle area of the trailer 20. For example, as depicted in FIG. 9, one embodiment includes sensor modules near the front and rear sections of the trailer that provide four detection areas, and sensor modules in the center that provide three detection areas.


In many trailers, the rear door is a roll-up door. In such a situation, the raised door may block one or more sensor modules mounted on the ceiling nearest the rear end of the trailer, such as the sensor modules 12a as shown in FIGS. 7A and 7B. In an embodiment designed to address this situation, a sensor module 12c may be mounted near the bottom of the door 26 and a sensor module 12b may be mounted near the top of the door 26. The sensor module 12c preferably includes a sensor that points down toward the floor when the door is closed. This sensor module 12c may be used to determine whether the door is closed or open, and to determine how far the door is open based on the distance from the sensor module 12c to the floor. When the door 26 is in a fully open position (FIG. 7B), the sensor modules 12b and 12c mounted near the top and bottom of the door face vertically down toward the floor. In this door-open position, the sensor modules 12b and 12c may be used for cargo detection, just as the sensor modules 12a that are blocked by the open door 26 were used.


In some embodiments, each sensor module 12 includes one or more accelerometers that can be used to determine the orientation of the sensors 14—whether they are facing horizontally or vertically—by sensing the direction of the acceleration of gravity relative to the position of the sensor module 12. Using this orientation information, the trailer tracker unit 18 or back-end server system 24 can determine how to treat the information provided by each sensor module 12.


In an embodiment depicted in FIG. 8, a sensor module 12 includes multiple sensors 14 distributed along the length of an elongate housing. This embodiment of the sensor module 12 may be approximately four feet long, with sensors 14 on each extreme end and one or more in the middle. Interface electronics within the module 12 provide for communications between the individual sensors 14, and communications with the tracker unit 18.


As shown in FIG. 3A, each sensor module 12 detects a distance between the sensor module 12 and the closest surfaces that are within the module's detection zone. The detected surfaces may be the floor of the trailer or surfaces of pieces of cargo. For example, one sensor within the sensor module designated as Module 2 detects a distance d1 to the floor of the trailer, and another sensor within Module 2 detects a distance d2 to the top surface of a cargo container. One sensor within the sensor module designated as Module 3 detects a distance d3 to the top surface of a cargo container, and another sensor within Module 3 detects a distance d4 to the floor.


In a preferred embodiment, distance measurements from each detection zone, and/or the areas 22 within each detection zone, are reported either to the tracker unit 18 or via the tracker unit 18 to the back-end server 24. Using the distance measurements, a process executed by the back-end server 24 or by a processor in the tracker unit 18 determines one or more of:

    • (1) the total floor area of the trailer 20 that is occupied by cargo;
    • (2) the total volume of the trailer 20 that is occupied;
    • (3) which detection areas of the trailer floor are occupied; and
    • (4) the height to which each detection area of the trailer is loaded.


The occupied floor area (1) may be determined by counting the number of detection areas that report a distance (d1, d2, d3, d4, etc.) that is less than some threshold value related to the full height of the interior of the trailer, and multiplying that number by the floor area within each detection area. For example, if the threshold value is 110 inches, a reported distance of less than 110 inches indicates that the corresponding detection area 22 is occupied by cargo. In the exemplary trailer 20 depicted in FIG. 6, there are twenty-four detection areas 22, each occupying about 13 square feet of floor area. If sensors in twenty of the detection areas 22 report distances that are less than the threshold value, then at least about 260 ft2 of the trailer floor space is occupied by cargo. (20×13 ft2=260 ft2)


The occupied trailer volume (2) may be determined by multiplying the size of each occupied detection area by the height above the floor of the detected cargo surface, and adding them all up. For example, if the twenty occupied areas from the previous example all have detected cargo surfaces that are at least 90 inches above the floor, then at least about 1950 ft3 of the trailer volume is occupied. (20×13 ft2×90/12 ft=1950 ft3)


Other information may be inferred as well, such as weight based on knowledge of the type of cargo and its total volume, and other relevant information. This information can be used in concert with an imaging cargo sensor to provide images at relevant times based on the position and state of the cargo load. The back end server 24 may also use this data to build a three-dimensional visualization of the cargo in the trailer, such as by displaying blocks based on x-position and y-position along with height in the z-direction.


In preferred embodiments, the sensor modules 12 are located near the center of the ceiling of the trailer. In an alternative embodiment depicted in FIGS. 10A-10C, sensor modules 12 may also be mounted in the corners of the trailer ceiling looking toward the opposite corners. This could provide a view with a larger detection areas 22 using fewer sensors.


The foregoing description of preferred embodiments for this invention have been presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise form disclosed. Obvious modifications or variations are possible in light of the above teachings. The embodiments are chosen and described in an effort to provide the best illustrations of the principles of the invention and its practical application, and to thereby enable one of ordinary skill in the art to utilize the invention in various embodiments and with various modifications as are suited to the particular use contemplated. All such modifications and variations are within the scope of the invention as determined by the appended claims when interpreted in accordance with the breadth to which they are fairly, legally, and equitably entitled.

Claims
  • 1. A method for detecting cargo within a cargo container that has a ceiling, a floor, and an interior space disposed between the ceiling and the floor, the method comprising: (a) receiving a plurality of distance signals generated by a plurality of distance sensors mounted on or adjacent to the ceiling of the cargo container, each of the distance signals indicating a distance between a corresponding one of the distance sensors and a surface within the interior space of the cargo container;(b) determining a load state of the cargo container based on one or more of the plurality of the distance signals;(c) for each of the distance signals, determining a cargo height value based on the distance between the corresponding distance sensor and the surface located within the corresponding cargo detection area; and(d) calculating a total volume of the interior space of the cargo container occupied by cargo based on the cargo height values and area values for the portions of the total area of the floor of the cargo container in which cargo is present.
  • 2. The method of claim 1 wherein: each of the distance sensors has a field of view, and the plurality of distance sensors are distributed across the ceiling of the cargo container such that their fields of view define a plurality of cargo detection areas distributed across the floor of the cargo container, wherein each cargo detection area corresponds to a portion of a total area of the floor of the cargo container; andeach of the plurality of distance signals indicates a distance between a corresponding one of the distance sensors and a surface located within a corresponding one of the cargo detection areas.
  • 3. The method of claim 2 further comprising calculating how much of the total area of the floor of the cargo container is occupied by cargo based on the distances indicated by the plurality of distance signals.
  • 4. The method of claim 1 wherein each of the plurality of distance signals indicate a distance between a corresponding one of the distance sensors and either a surface of a piece of cargo or the floor of the cargo container.
  • 5. The method of claim 1 wherein the ceiling of the cargo container is disposed at a height above the floor of the cargo container, and wherein the load state determination step comprises comparing the distance to a distance threshold that is less than the height.
  • 6. The method of claim 1 further comprising providing one or more load state indications to a tracker unit mounted on the cargo container, and based on the one or more load state indications, transmitting load state information from the tracker unit to a backend server via a data communication network.
  • 7. An apparatus for detecting cargo within a cargo container that has a ceiling, a floor, and an interior space disposed between the ceiling and the floor, the apparatus comprising: a plurality of distance sensors configured for mounting on or adjacent to the ceiling of the cargo container, the plurality of distance sensors generating a plurality of distance signals, each of the distance signals indicating a distance between a corresponding one of the distance sensors and a surface within the interior space of the cargo container; andone or more processors in communication with the plurality of distance sensors, the one or more processors operable to execute instructions to: generate a load state indication based on one or more of the plurality of distance signals;for each load state indication, determine a cargo height value based on the distance between the corresponding distance sensor and the surface located within the corresponding cargo detection area; andcalculate a total volume of the interior space of the cargo container occupied by cargo based on cargo height values and area values for the portions of the total area of the floor of the cargo container in which cargo is present.
  • 8. The apparatus of claim 7 wherein each of the plurality of distance sensors has a field of view,the plurality of distance sensors are mounted to the ceiling of the cargo container such that their fields of view define a plurality of cargo detection areas distributed across the floor of the cargo container,each cargo detection area corresponds to a portion of a total area of the floor of the cargo container, andeach of the distance signals indicate a distance between a corresponding one of the distance sensors and a surface located within a corresponding one of the cargo detection areas.
  • 9. The apparatus of claim 7 wherein the one or more processors are operable to execute instructions to calculate how much of the total area of the floor of the cargo container is occupied by cargo based at least in part on the load state indication.
  • 10. The apparatus of claim 7 further comprising one or more sensor modules configured for mounting on or adjacent to the ceiling of the cargo container, each of the one or more sensor modules containing two or more of the plurality of distance sensors, wherein each of the two or more distance sensors in each sensor module has a field of view that is aligned in a different direction from fields of view of other distance sensors in one and the same sensor module.
  • 11. The apparatus of claim 10 wherein the one or more processors are components of the one or more sensor modules.
  • 12. The apparatus of claim 10 wherein multiple sensor modules are distributed across a length of a central portion of the ceiling of the cargo container.
  • 13. The apparatus of claim 10 wherein multiple sensor modules are distributed across a length of the cargo container adjacent to locations at which the ceiling of the cargo container meets a wall of the cargo container.
  • 14. The apparatus of claim 10 wherein: one or more of the sensor modules are mounted on a door of the cargo container;the one or more distance sensors contained in the one or more sensor modules mounted on the door each have a field of view directed toward a back wall of the cargo container when the door is in a closed position; andthe one or more distance sensors contained in the one or more sensor modules mounted on the door each have a field of view directed toward the floor of the cargo container when the door is in an open position.
  • 15. The apparatus of claim 10 wherein at least one of the one or more sensor modules contains an accelerometer used to determine an orientation of the at least one sensor module.
  • 16. The apparatus of claim 7 wherein each of the distance signals indicate a distance between a corresponding one of the distance sensors and either a surface of a piece of cargo or the floor of the cargo container.
  • 17. The apparatus of claim 7 wherein the ceiling of the cargo container is disposed at a height above the floor of the cargo container, and wherein the one or more processors are operable to execute instructions to compare the distance to a distance threshold that is less than the height.
  • 18. The apparatus of claim 7 further comprising a tracker unit configured for mounting on the cargo container, the tracker unit operable to generate cargo loading information based on the load state indication, and transmit the cargo loading information from the tracker unit to a backend server via a data communication network.
  • 19. The apparatus of claim 18 wherein the one or more processors are components of the tracker unit.
  • 20. The apparatus of claim 18 wherein the one or more processors are components of the backend server.
  • 21. The apparatus of claim 7 wherein the plurality of distance sensors comprise laser-ranging time-of-flight sensors or ultrasonic sensors.
RELATED APPLICATIONS

This nonprovisional application claims priority to provisional patent application Ser. No. 62/765,256 filed Aug. 20, 2018, titled Distributed Volumetric Cargo Sensor, the entire contents of which are incorporated herein by reference.

US Referenced Citations (191)
Number Name Date Kind
906021 Herrick Dec 1908 A
4633407 Freienstein et al. Dec 1986 A
4837700 Ando et al. Jun 1989 A
5119301 Shimizu et al. Jun 1992 A
5289369 Hirshberg Feb 1994 A
5299132 Wortham Mar 1994 A
5307277 Hirano Apr 1994 A
5870029 Otto et al. Feb 1999 A
5877956 Frank et al. Mar 1999 A
6025774 Forbes Feb 2000 A
6240365 Bunn May 2001 B1
6249217 Forbes Jun 2001 B1
6510381 Grounds et al. Jan 2003 B2
6512465 Flick Jan 2003 B2
6701234 Vogelsang Mar 2004 B1
6771970 Dan Aug 2004 B1
6816090 Teckchandani et al. Nov 2004 B2
6930638 Lloyd et al. Aug 2005 B2
6931309 Phelan et al. Aug 2005 B2
6985087 Soliman Jan 2006 B2
7034683 Ghazarian Apr 2006 B2
7091835 Boulay et al. Aug 2006 B2
7102510 Boling et al. Sep 2006 B2
7170390 Quiñones et al. Jan 2007 B2
7174243 Lightner et al. Feb 2007 B1
7177738 Diaz Feb 2007 B2
7215282 Boling et al. May 2007 B2
7266378 Norta et al. Sep 2007 B2
7346439 Bodin Mar 2008 B2
7366551 Hartley Apr 2008 B1
7405658 Richards Jul 2008 B2
7546151 Hartley Jun 2009 B2
7574195 Krasner et al. Aug 2009 B2
7593999 Nathanson Sep 2009 B2
7675423 Boling et al. Mar 2010 B2
7701363 Zlojutro Apr 2010 B1
7725216 Kim May 2010 B2
7817033 Motoyama Oct 2010 B2
7818098 Koepf et al. Oct 2010 B2
7830305 Boling et al. Nov 2010 B2
7893818 Smoyer et al. Feb 2011 B2
7970496 Koepf et al. Jun 2011 B2
8018332 Boling et al. Sep 2011 B2
8126601 Kapp et al. Feb 2012 B2
8237591 Holcomb et al. Aug 2012 B2
8330626 Adelson Dec 2012 B1
8330817 Foster Dec 2012 B1
8368561 Welch et al. Feb 2013 B2
8452673 Boling et al. May 2013 B2
8462021 Welch et al. Jun 2013 B2
8510200 Pearlman et al. Aug 2013 B2
8527135 Lowrey et al. Sep 2013 B2
8565963 Burke Oct 2013 B2
8612137 Harris et al. Dec 2013 B2
8626152 Farrell et al. Jan 2014 B2
8655544 Fletcher et al. Feb 2014 B2
8671063 Ehrman et al. Mar 2014 B2
8725326 Kapp et al. May 2014 B2
8760274 Boling et al. Jun 2014 B2
8799461 Herz et al. Aug 2014 B2
8933802 Baade Jan 2015 B2
8970701 Lao Mar 2015 B2
9008894 Bishop et al. Apr 2015 B2
9049564 Muetzel et al. Jun 2015 B2
9060213 Jones Jun 2015 B2
9070271 Baade et al. Jun 2015 B2
9316737 Baade Apr 2016 B2
9332404 Boling et al. May 2016 B2
9516394 Carlo et al. Dec 2016 B2
9551788 Epler Jan 2017 B2
9779379 Hall et al. Oct 2017 B2
9779449 Meyer et al. Oct 2017 B2
10089598 Reeder et al. Oct 2018 B2
10169822 Jarvis et al. Jan 2019 B2
10223744 Brady et al. Mar 2019 B2
10232823 Bobay et al. Mar 2019 B1
10255824 Pearlman et al. Apr 2019 B2
10311315 Drazan et al. Jun 2019 B2
20010018639 Bunn Aug 2001 A1
20010034577 Grounds et al. Oct 2001 A1
20020000916 Richards Jan 2002 A1
20020014978 Flick Feb 2002 A1
20020059126 Ricciardi May 2002 A1
20020082025 Baese et al. Jun 2002 A1
20020184062 Diaz Dec 2002 A1
20020186144 Meunier Dec 2002 A1
20030083060 Menendez May 2003 A1
20030151501 Teckchandani et al. Aug 2003 A1
20030151507 Andre et al. Aug 2003 A1
20030174067 Soliman Sep 2003 A1
20040093291 Bodin May 2004 A1
20040130440 Boulay et al. Jul 2004 A1
20040143378 Vogelsang Jul 2004 A1
20040162063 Quinones et al. Aug 2004 A1
20040225557 Phelan et al. Nov 2004 A1
20040246177 Lloyd et al. Dec 2004 A1
20050021199 Zimmerman et al. Jan 2005 A1
20050026627 Boling et al. Feb 2005 A1
20050134504 Harwood et al. Jun 2005 A1
20050215194 Boling et al. Sep 2005 A1
20050237166 Chen Oct 2005 A1
20060007038 Boling et al. Jan 2006 A1
20060055561 Kamali et al. Mar 2006 A1
20060087411 Chang Apr 2006 A1
20060129290 Zimmerman et al. Jun 2006 A1
20070013779 Gin et al. Jan 2007 A1
20070152844 Hartley et al. Jul 2007 A1
20070167147 Krasner et al. Jul 2007 A1
20070290923 Norta et al. Dec 2007 A1
20080015748 Nagy Jan 2008 A1
20080147245 Koepf et al. Jun 2008 A1
20080162045 Lee Jul 2008 A1
20080176537 Smoyer et al. Jul 2008 A1
20080186135 Boling et al. Aug 2008 A1
20080198018 Hartley Aug 2008 A1
20080278314 Miller et al. Nov 2008 A1
20080287151 Fjelstad et al. Nov 2008 A1
20080294302 Basir Nov 2008 A1
20090043445 Bishop et al. Feb 2009 A1
20090079591 Motoyama Mar 2009 A1
20090112394 Lepejian et al. Apr 2009 A1
20090140887 Breed et al. Jun 2009 A1
20090224966 Boling et al. Sep 2009 A1
20100094482 Schofield et al. Apr 2010 A1
20100103042 Bishop et al. Apr 2010 A1
20100117868 Wiemeersch et al. May 2010 A1
20100191412 Kim Jul 2010 A1
20100265104 Zlojutro Oct 2010 A1
20100299020 Koepf et al. Nov 2010 A1
20110016514 Carlo et al. Jan 2011 A1
20110090075 Armitage et al. Apr 2011 A1
20110093159 Boling et al. Apr 2011 A1
20110143669 Farrell et al. Jun 2011 A1
20110241903 Welch et al. Oct 2011 A1
20120041618 Sun et al. Feb 2012 A1
20120077475 Holcomb et al. Mar 2012 A1
20120078497 Burke Mar 2012 A1
20120197484 Nath et al. Aug 2012 A1
20120299721 Jones Nov 2012 A1
20120299755 Jones Nov 2012 A1
20130059607 Herz et al. Mar 2013 A1
20130066757 Lovelace et al. Mar 2013 A1
20130088371 Welch et al. Apr 2013 A1
20130100286 Lao Apr 2013 A1
20130113637 Sin et al. May 2013 A1
20130127617 Baade et al. May 2013 A1
20130141249 Peariman et al. Jun 2013 A1
20130144667 Ehrman et al. Jun 2013 A1
20130144770 Boling et al. Jun 2013 A1
20130144771 Boling et al. Jun 2013 A1
20130144805 Boling et al. Jun 2013 A1
20130147617 Boling et al. Jun 2013 A1
20130159214 Boling et al. Jun 2013 A1
20130185193 Boling et al. Jul 2013 A1
20130249713 Adelson Sep 2013 A1
20130297199 Kapp et al. Nov 2013 A1
20130302757 Pearlman et al. Nov 2013 A1
20140012634 Pearlman et al. Jan 2014 A1
20140036072 Lyall et al. Feb 2014 A1
20140052605 Beerle et al. Feb 2014 A1
20140074692 Beede et al. Mar 2014 A1
20140095061 Hyde Apr 2014 A1
20140125500 Baade May 2014 A1
20140125501 Baade May 2014 A1
20140220966 Muetzel et al. Aug 2014 A1
20140267688 Aich et al. Sep 2014 A1
20140280658 Boling et al. Sep 2014 A1
20150006207 Jarvis et al. Jan 2015 A1
20150019270 Jarvis et al. Jan 2015 A1
20150024727 Hale-Pletka et al. Jan 2015 A1
20150032291 Lowrey et al. Jan 2015 A1
20150066362 Meyer et al. Mar 2015 A1
20150067312 Lewandowski et al. Mar 2015 A1
20150095255 Hall et al. Apr 2015 A1
20150168173 Lewis-Evans et al. Jun 2015 A1
20150172518 Lucas et al. Jun 2015 A1
20150186991 Meyer et al. Jul 2015 A1
20150260529 Petersen Sep 2015 A1
20150332525 Harris et al. Nov 2015 A1
20150356497 Reeder et al. Dec 2015 A1
20150373487 Miller et al. Dec 2015 A1
20160225072 Brady et al. Aug 2016 A1
20160282466 Epler Sep 2016 A1
20170262717 Drazan et al. Sep 2017 A1
20180222504 Birch Aug 2018 A1
20180300967 Winograd Oct 2018 A1
20180352198 Raasch et al. Dec 2018 A1
20190005442 Reeder et al. Jan 2019 A1
20190061692 Bobay et al. Feb 2019 A1
20190114577 Kilburn Apr 2019 A1
20190279494 Raasch et al. Sep 2019 A1
Foreign Referenced Citations (61)
Number Date Country
2609106 Oct 2008 CA
2683208 Nov 2008 CA
2837320 Nov 2012 CA
2856796 May 2013 CA
2867447 Sep 2013 CA
2826902 Mar 2014 CA
2828835 Apr 2014 CA
2832185 May 2014 CA
2846134 Sep 2014 CA
2921908 Jul 2007 CN
101240734 Aug 2008 CN
101734228 Jun 2010 CN
101192322 Jul 2012 CN
102779407 Nov 2012 CN
103813477 May 2014 CN
104931066 Sep 2015 CN
4423328 Jan 1996 DE
0096252 May 1987 EP
0451482 Oct 1991 EP
0519630 Dec 1992 EP
0393935 Mar 1995 EP
0744727 Feb 1997 EP
1191500 Mar 2002 EP
1384635 Jan 2004 EP
2418461 Feb 2012 EP
2006123891 May 2006 JP
2014170000 Sep 2014 JP
2009011420 Mar 2010 MX
2010001545 Aug 2010 MX
1984001823 May 1984 WO
1999063357 Dec 1999 WO
2000070530 Nov 2000 WO
2001024393 Apr 2001 WO
2001059601 Aug 2001 WO
2002089077 Nov 2002 WO
2003034089 Apr 2003 WO
2003036462 May 2003 WO
2003079717 Sep 2003 WO
2003012473 Mar 2004 WO
2004075090 Sep 2004 WO
2004102536 Jun 2005 WO
2005086933 Sep 2005 WO
2006028995 Mar 2006 WO
2006028995 Feb 2007 WO
2007146449 Dec 2007 WO
2008034097 Mar 2008 WO
2007146449 Oct 2008 WO
2008121612 Oct 2008 WO
2008141456 Nov 2008 WO
2008144411 Nov 2008 WO
2005086933 Dec 2008 WO
2009021117 Feb 2009 WO
2009058972 Jul 2009 WO
2009097595 Aug 2009 WO
2010047887 Apr 2010 WO
2012162358 Nov 2012 WO
2012162450 Nov 2012 WO
2013078291 May 2013 WO
2013138798 Sep 2013 WO
2014008752 Jan 2014 WO
2016061355 Apr 2016 WO
Related Publications (1)
Number Date Country
20200055553 A1 Feb 2020 US
Provisional Applications (1)
Number Date Country
62765256 Aug 2018 US