This invention relates generally to the location measurement field, and more specifically to a new and useful method and system for location measurement analysis in the location measurement field.
Typical methods and systems for location measurement analysis require location data with reliably high accuracy and/or precision, and/or require additional metadata, to enable determination of useful and/or reliable information. Accordingly, such methods and systems may not be able to rely on location data of low or inconsistent accuracy and/or precision, such as some data received from geopositioning systems. Further, some methods and systems for location measurement analysis may rely (in part or in whole) on information received from users, such as information regarding geofence positions, shapes, and/or sizes, but may not be able to adequately reconcile such information with some or all received location data (e.g., location data of low or inconsistent accuracy and/or precision). Thus, there is a need in the location measurement field to create a new and useful method and system for location measurement analysis.
The following description of the preferred embodiments of the invention is not intended to limit the invention to these preferred embodiments, but rather to enable any person skilled in the art to make and use this invention.
1. Overview.
A method 10 for location measurement analysis preferably includes associating a location measurement with a geofence S100. The method 10 can optionally include training a classifier S200 and/or any other suitable elements.
A container management system 20 preferably includes a computing system (e.g., remote server), and can additionally or alternatively include one or more containers, one or more location sensors associated with each container, and/or any other suitable elements (e.g., as shown in
The method is preferably performed using the system 20, but can additionally or alternatively be performed by any other suitable system.
2. System.
The containers can include dumpsters (e.g., front load containers, roll off containers, etc.), shipping containers (e.g., intermodal freight containers, unit load devices, etc.), sections of a vehicle (e.g., land, sea, air, and/or space vehicle) such as vehicle cargo holds, rooms of a structure (e.g., a fixed structure such as a building), and/or any other suitable containers.
Each container is preferably associated with one or more location sensors (e.g., GNSS and/or other geopositioning modules, such as receivers for one or more of GPS, GLONASS, BeiDou, Galileo, etc.; local positioning modules, such as modules enabling techniques such as triangulation, trilateration, multilateration, etc.), preferably location sensors configured to determine the location (e.g., geospatial position) of the container. The location sensor is preferably affixed to the container, but can additionally or alternatively have any other suitable arrangement with respect to the container.
Each container can optionally include one or more content sensors (e.g., integrated with and/or separate from the location sensor). The content sensor is preferably configured to sense (e.g., image) the interior of the container that it is associated with (e.g., image and/or otherwise sense the contents of the container), more preferably configured to sense substantially all of the interior but alternatively configured to image any suitable portion thereof. The content sensor preferably has a fixed position and/or orientation relative to the container (e.g., is mechanically coupled to the container, preferably by a fixed coupling) but can alternatively have any other suitable spatial relationship with respect to the container (e.g., as shown in
The content sensor preferably includes one or more imaging devices. The imaging device is preferably an optical sensor (e.g., camera), but can additionally or alternatively include an ultrasound imaging device and/or any other suitable imaging devices. Examples of optical sensors include a monocular camera, stereocamera, multi-lens or multi-view camera, color camera (e.g., an RGB camera) such as a charge coupled device (CCD) or a camera including a CMOS sensor, grayscale camera, multispectral camera (narrow band or wide band), hyperspectral camera, ultraspectral camera, spectral camera, spectrometer, time of flight camera, high-, standard-, or low-dynamic range cameras, range imaging system (e.g., LIDAR system), active light system (e.g., wherein a light, such as an IR LED, is pulsed and directed at the subject and the reflectance difference measured by a sensor, such as an IR sensor), thermal sensor, infra-red imaging sensor, projected light system, full spectrum sensor, high dynamic range sensor, or any other suitable imaging system. The optical sensor is preferably configured to capture a 2-dimensional or 3-dimensional image, but can alternatively capture any measurement having any other suitable dimension. The image is preferably single, multi-pixel, time-averaged or sum total measurement of the intensity of a signal emitted or reflected by objects within a field of view, but can alternatively be a video (e.g., a set of images or frames), or any other suitable measurement. The image preferably has a resolution (e.g., cycles per millimeter, line pairs per millimeter, lines of resolution, contrast vs. cycles/mm, modulus of the OTF, or any other suitable measure) capable of resolving a 1 cm3 object at a sensor distance of at least 10 feet from the object, but can alternatively have a higher or lower resolution.
The content sensor can optionally include one or more emitters that are configured to emit electromagnetic signals, audio signals, compounds, or any other suitable interrogator that the content sensor is configured to measure. However, the content sensor can additionally or alternatively measure signals from the ambient environment. Examples of sensor-emitter pairs include LIDAR systems, time-of-flight systems, ultrasound systems, radar systems, X-ray systems, and/or any other suitable systems. In embodiments in which the content sensor includes an emitter, the content sensor can optionally include a reference sensor that measures the ambient environment signals (e.g., wherein the content sensor measurement can be corrected by the reference sensor measurement).
The content sensor can optionally include a lens that functions to adjust the optical properties of the incident signal on the sensor. For example, the optical sensor can include a fish-eye lens to broaden the area monitored by the optical sensor, wherein the resultant distortion is known and can be adjusted for during image processing. However, the lens can be a wavelength filter, polarizing filter, or any other suitable lens. The content sensor can additionally or alternatively include a physical or digital filter, such as a noise filter that corrects for interferences in the measurement.
The location sensors and/or content sensors can optionally include one or more communication modules. The communication module preferably functions to communicate data from the content sensor to a second system (e.g., the computing system). The data can be measurements from the location sensor (e.g., location measurements), measurements from the content sensor (and/or any other suitable components), processed measurements, instructions, pickup requests, and/or any other suitable data. The second system can be a device, server system, or any other suitable computing system. The second system can be remote or wired to the communication system. Examples of the second system include a mobile device (e.g., smartphone, tablet, computer), server system, or any other suitable computing system. The communication system can be a wireless or wired communication system. The communication system can be a cellular, WiFi, Zigbee, Z-Wave, near-field communication system (e.g., Bluetooth, RF, NFC, etc.), Ethernet, powerline communication, or any other suitable communication system. The communication system is preferably operable in a standby or off mode, wherein the communication system consumes power at a rate less than a threshold rate, and an on or communication mode, wherein the communication system consumes power at a rate required to communicate data. However, the communication system can be operable in any other suitable mode.
The location sensors and/or content sensors can optionally include one or more auxiliary sensors, such as IMU sensors (e.g., accelerometer, gyroscope, magnetometer, etc.), weight sensors, audio sensors, and/or any other suitable auxiliary sensors.
However, the system 20 can additionally or alternatively include any other suitable elements in any suitable arrangement.
3. Method.
3.1 Associating a Location Measurement with a Geofence.
Associating a location measurement with a geofence S100 preferably functions to assign a location measurement to a nearby geofence. S100 preferably includes: receiving a location measurement S110; selecting candidate geofences S120; determining features S130; and/or classifying the location S140 (e.g., as shown in
S100 is preferably performed in response to receiving the location measurement (e.g., automatically performed in response, performed immediately in response, etc.). However, S100 can additionally or alternatively be performed periodically, sporadically, in response to any other suitable trigger, and/or with any other suitable timing. For example, each time S100 is performed, it can be performed for each location measurement received (e.g., by the computing system performing S100) since the previous performance of S100. However, S100 can additionally or alternatively be performed with any other suitable timing for any other suitable location measurement or measurements.
3.1.1 Receiving a Location Measurement.
Receiving a location measurement S110 preferably functions to provide a location measurement to be assigned to a geofence. The location measurement is preferably received from a location sensor, such as a location sensor associated with a container of the system. The location measurement is preferably associated with a location (e.g., geospatial position such as latitude, longitude, and/or elevation). For example, S110 can include receiving a GPS post (e.g., from a GPS receiver associated with a container of the system). However, S110 can additionally or alternatively include receiving one or more location measurements in any other suitable manner.
3.1.2 Selecting Candidate Geofences.
Selecting candidate geofences S120 preferably functions to select one or more geofences that may be associated with the location. The geofences are preferably selected from a set of geofences defined in a region (e.g., a municipal region, a container service area, etc.). The set of geofences (and/or the subset selected in S120) can include user-defined geofences, automatically determined geofences (e.g., geofences determined based on a set of location measurements), and/or any other suitable geofences.
S120 preferably includes selecting geofences close to the location (e.g., the geofences closest to the location). Geofence proximity is preferably determined based on the shortest distance between the location and an edge of each geofence (“edge distance”). The edge distance associated with a location inside a geofence is preferably considered to be negative, such that a first location inside a geofence is considered to have a shorter edge distance than a second location outside the geofence, even if the absolute value of the first location's edge distance is greater than the second location's. Alternatively, the proximity can be determined based on a distance between the location and the center of each geofence (“center distance”), and/or based on any other suitable proximity metric. In a first variation, S120 includes selecting all geofences within a threshold distance (e.g., 0.1, 0.2, 0.5, 1, 2, 3, 5, 10, 0.1-0.5, 0.5-2, or 2-10 km, etc.) of the location. In a second variation, S120 includes selecting a predetermined number of geofences (e.g., 1, 2, 3, 4, 5, 6-10, etc.), preferably by selecting the closest geofences (e.g., the 2 closest geofences, such as shown by way of example in
3.1.3 Determining Features.
Determining features S130 preferably functions to determine one or more features to be used by a classifier. Such features can include, for example, one or more geofence features associated with each candidate geofence. The geofence features can include size (e.g., length scale such as radius, width, perimeter, convex hull perimeter, etc.; area; convex hull area; etc.), center and/or edge distance to the location, one or more distances to other geofences (e.g., other candidate geofences, geofences not selected as candidate geofences, etc.), sizes (e.g., length scale such as radius, width, perimeter, convex hull perimeter, etc.; area; convex hull area; etc.) of other geofences (e.g., other candidate geofences, geofences not selected as candidate geofences, etc.), and/or any other suitable features associated with one or more geofences (e.g., as shown in
In a first example, the distance to other candidate geofences is defined as the shortest distance between geofence edges (“edge-edge distance”). In a second example, the distance is defined as the distance between geofence centers (“center-center distance”). The other geofences to which the distances are defined can include: the closest geofence or non-candidate geofence to the geofence under consideration, the non-candidate geofence with the shortest average distance to each of the candidate geofences, the closest non-candidate geofence to the location, and/or any other suitable geofences. However, the features can additionally or alternatively include any other suitable distances.
The geofence features can additionally or alternatively include derived features (e.g., features determined based on one or more other geofence features, such as the features described above, and/or determined based on any other suitable features). In a first example, the derived features include one or more logarithms, such as logarithms of one or more lengths (e.g., radius, distance, etc.). In a second example, the derived features include one or more ratios (e.g., edge distance divided by radius, ratios between radii of other geofences, etc.). However, the features can additionally or alternatively include any other suitable derived features.
S130 can additionally or alternatively include determining location measurement features, such as measurement uncertainty (e.g., GPS uncertainty, such as uncertainty reported by a GPS receiver), historical location measurements (e.g., previous measurements received from the same location sensor and/or received from another location sensor associated with the first location sensor, such as another location sensor associated with the same container), and/or any other suitable measurement features. However, S130 can additionally or alternatively include determining any other suitable features in any suitable manner.
3.1.4 Classifying the Location.
Classifying the location S140 preferably functions to determine a geofence to associate with the location measurement. S140 is preferably performed using a computerized classifier (e.g., trained such as described below regarding S200). The computerized classifier is preferably a random forest classifier (or variant thereof, such as a kernel random forest classifier). However, the classifier can additionally or alternatively include one or more support vector machines and/or any other suitable classifiers.
The classifier preferably outputs one of the candidate geofences, no candidate geofence, and/or optionally an indication of classification ambiguity (e.g., when the classifier determines that two or more of the candidate geofences may be associated with the location). However, the classifier can additionally or alternatively provide any other suitable outputs.
In one example, the classifier results in associations such as shown in
S100 can additionally or alternatively include any other suitable elements performed in any suitable manner.
3.2 Training the Classifier.
Training the classifier S200 preferably functions to generate a classifier for associating location measurements with geofences. S200 preferably includes receiving a training dataset S210 and/or training the classifier S240 (e.g., as shown in
3.2.1 Receiving a Training Dataset.
Receiving a training dataset S210 preferably functions to receive data for training a classifier. The training dataset preferably includes a set of data elements. Each data element preferably includes a location measurement and a set of possible geofences, and can optionally include additional information. The location measurement is preferably in the same form as the location measurements provided in S100, but can additionally or alternatively include any other suitable location measurements and/or purported location measurements (e.g., location data generated for the training dataset rather than measured). The set of possible geofences preferably includes one or more geofences that the location measurement may be associated with (e.g., geofences near the location measurement, all geofences within an area such as a municipality or service area, etc.). Each geofence of the set preferably defines a shape and position (e.g., a radius and a location of the geofence center). The set of geofences (and/or elements thereof) can be shared between multiple location measurements (e.g., all location measurements of the training dataset, a subset thereof, etc.), can differ between location measurements, and/or can be shared or different in any other suitable manner. The possible geofences can be the same as the geofences used in S120, can include geofences from S120 and additional geofences, can differ entirely from the geofences of S120, and/or can include any other suitable geofences.
The additional information can include location measurement features (e.g., as described in S130), associations with users, customers, waste types, and/or any other suitable associations, and/or can include any other suitable information associated with the location measurement, geofences, and/or other aspects of the data elements. However, the data element can additionally or alternatively include any other suitable information.
The dataset (or a subset of the elements thereof) can optionally be labeled. For example, each data element can include one or more of: a geofence association, an indication of no associated geofence, and/or an indication of uncertainty, such as an ambiguous classification.
However S210 can additionally or alternatively include receiving any other suitable information in any other suitable manner.
3.2.2 Modifying the Training Dataset.
Training the classifier S200 can optionally include modifying the training dataset S220. S220 can include generating new data elements and/or modifying existing elements. S220 is preferably performed by adding perturbations (e.g., random perturbations) to the location measurements, geofence positions, geofence sizes, and/or any other suitable aspects of the data elements.
The new and/or modified elements generated in S220 are preferably labeled in S230 (e.g., wherein S230 is performed after S220, such as in response to S220). However, the modified elements can alternatively retain their prior labels (e.g., labels associated with the dataset received in S210), and/or can be associated with labels in any other suitable manner.
3.2.3 Labeling the Training Dataset.
S200 can optionally include labeling the training dataset S230. S230 is preferably performed if the training dataset received in S210 is not labeled (or only partially labeled, such as including one or more data elements that are not labeled), and/or performed if S220 was performed (e.g., wherein S230 functions to label new points generated in S220 and/or points modified by performance of S220).
In a first embodiment, S230 includes automatically labeling the training dataset (or a subset thereof) using an alternative classifier (e.g., classifier different than the classifier trained in S200). In a first example of this embodiment, the alternative classifier relies on additional information (e.g., information not necessarily available for performing S100, such as information not received in S110). For example, this information can include associations with users, customers, waste types, and/or any other suitable information (e.g., such as described above regarding S210). In a second example of this embodiment, the alternative classifier has higher computational requirements (e.g., requires more processor time and/or other computational resources) than the classifier to be trained in S200. However, the alternative classifier can additionally or alternatively include any other suitable computerized classifier.
In a second embodiment, S230 includes receiving labels from human classifiers. For example S230 can include presenting a location measurement and nearby geofences to one or more human classifiers (e.g., presented as a graphical representation of the location measurement and the geofences). In this embodiment, S230 can further include receiving a geofence selection, an indication of no assignment, and/or an indication of ambiguity from the human classifiers, and preferably includes labeling the location measurement based on the information received from the human classifiers.
However S230 can additionally or alternatively include labeling the training dataset in any other suitable manner.
3.2.4 Generating the Classifier.
Generating the classifier S240 preferably functions to generate the trained computerized classifier (e.g., for use in S140). The classifier is preferably trained based on the labeled training data (e.g., training dataset received in S210, dataset modified in S220, and/or dataset labeled in S230), but can additionally or alternatively be trained based on any other suitable information. Training the classifier preferably includes determining features associated with the labeled training data (e.g., features such as described above regarding S130 and/or any other suitable features). S240 preferably includes training the classifier such as according to established techniques known in the art for training classifiers of the classifier type (e.g., using a random forest training algorithm to train a random forest classifier). However, S240 can additionally or alternatively include training a classifier in any other suitable manner.
However S200 and/or the method 10 can additionally or alternatively include any other suitable elements performed in any suitable manner.
An alternative embodiment preferably implements the some or all of above methods in a computer-readable medium storing computer-readable instructions. The instructions are preferably executed by computer-executable components preferably integrated with a communication routing system. The communication routing system may include a communication system, routing system and a pricing system. The computer-readable medium may be stored on any suitable computer readable media such as RAMs, ROMs, flash memory, EEPROMs, optical devices (CD or DVD), hard drives, floppy drives, or any suitable device. The computer-executable component is preferably a processor but the instructions may alternatively or additionally be executed by any suitable dedicated hardware device.
Although omitted for conciseness, embodiments of the system and/or method can include every combination and permutation of the various system components and the various method processes, wherein one or more instances of the method and/or processes described herein can be performed asynchronously (e.g., sequentially), concurrently (e.g., in parallel), or in any other suitable order by and/or using one or more instances of the systems, elements, and/or entities described herein.
The FIGURES illustrate the architecture, functionality and operation of possible implementations of systems, methods and computer program products according to preferred embodiments, example configurations, and variations thereof. In this regard, each block in the flowchart or block diagrams may represent a module, segment, step, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block can occur out of the order noted in the FIGURES. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
As a person skilled in the art will recognize from the previous detailed description and from the figures and claims, modifications and changes can be made to the preferred embodiments of the invention without departing from the scope of this invention defined in the following claims.
This application claims the benefit of U.S. Provisional Application Ser. No. 62/841,447, filed on 1 May 2019, which is incorporated in its entirety by this reference.
Number | Name | Date | Kind |
---|---|---|---|
4485831 | Ungerleider | Dec 1984 | A |
4646793 | Sherratt | Mar 1987 | A |
5257577 | Clark | Nov 1993 | A |
5299493 | Durbin et al. | Apr 1994 | A |
5609193 | Steckler | Mar 1997 | A |
5708424 | Orlando et al. | Jan 1998 | A |
5812060 | Despain et al. | Sep 1998 | A |
5927142 | Mercer | Jul 1999 | A |
5964258 | Schoenbauer | Oct 1999 | A |
6123017 | Little et al. | Sep 2000 | A |
6125213 | Morimoto | Sep 2000 | A |
6369715 | Bennett et al. | Apr 2002 | B2 |
6561085 | Durbin et al. | May 2003 | B1 |
6822565 | Thomas et al. | Nov 2004 | B2 |
6951615 | Tripodi et al. | Oct 2005 | B2 |
7032820 | Kreiner et al. | Apr 2006 | B2 |
7313464 | Perreault et al. | Dec 2007 | B1 |
7421112 | Calver et al. | Sep 2008 | B2 |
7423541 | Miller | Sep 2008 | B2 |
7441569 | Lease | Oct 2008 | B2 |
7728730 | Langlois et al. | Jun 2010 | B2 |
7853142 | Meyers et al. | Dec 2010 | B2 |
7999688 | Healey et al. | Aug 2011 | B2 |
8185277 | Flood et al. | May 2012 | B2 |
8339269 | Sherron | Dec 2012 | B2 |
8565536 | Liu | Oct 2013 | B2 |
8620393 | Bornstein et al. | Dec 2013 | B2 |
9301094 | Jajoo et al. | Mar 2016 | B2 |
9352887 | Poss et al. | May 2016 | B2 |
9952600 | Gurin | Apr 2018 | B2 |
10064007 | Deluca et al. | Aug 2018 | B1 |
10405133 | Merjanian et al. | Sep 2019 | B1 |
10416311 | Huang et al. | Sep 2019 | B2 |
10559208 | Mcalpine et al. | Feb 2020 | B1 |
10621873 | Spiel et al. | Apr 2020 | B1 |
10798522 | Benjamin | Oct 2020 | B1 |
20020108507 | May et al. | Aug 2002 | A1 |
20040129781 | Kreiner et al. | Jul 2004 | A1 |
20040199401 | Wagner et al. | Oct 2004 | A1 |
20050080520 | Kline et al. | Apr 2005 | A1 |
20050083197 | Glenn et al. | Apr 2005 | A1 |
20050126958 | Bohlig et al. | Jun 2005 | A1 |
20060196937 | Kreiner et al. | Sep 2006 | A1 |
20060261964 | Maetzke | Nov 2006 | A1 |
20070133980 | Meyers et al. | Jun 2007 | A1 |
20070260466 | Casella et al. | Nov 2007 | A1 |
20080061125 | Langlois et al. | Mar 2008 | A1 |
20080061977 | Maruca et al. | Mar 2008 | A1 |
20090014363 | Gonen et al. | Jan 2009 | A1 |
20090161907 | Healey et al. | Jun 2009 | A1 |
20090321511 | Browne | Dec 2009 | A1 |
20100001867 | Rodrigue et al. | Jan 2010 | A1 |
20100017127 | Pepitone et al. | Jan 2010 | A1 |
20100042940 | Monday et al. | Feb 2010 | A1 |
20100069035 | Johnson | Mar 2010 | A1 |
20100092089 | Wilson et al. | Apr 2010 | A1 |
20100217715 | Lipcon | Aug 2010 | A1 |
20100287073 | Kocis et al. | Nov 2010 | A1 |
20100312601 | Lin | Dec 2010 | A1 |
20110000295 | Kritlow | Jan 2011 | A1 |
20110054979 | Cova | Mar 2011 | A1 |
20110137812 | Sherga | Jun 2011 | A1 |
20110148634 | Putz | Jun 2011 | A1 |
20110175739 | Mcfeeters | Jul 2011 | A1 |
20120010746 | Sundholm | Jan 2012 | A1 |
20120192646 | Arnold et al. | Aug 2012 | A1 |
20120209783 | Smith et al. | Aug 2012 | A1 |
20120314059 | Hoffmann et al. | Dec 2012 | A1 |
20130006522 | Vellaikal et al. | Jan 2013 | A1 |
20130096731 | Tamari et al. | Apr 2013 | A1 |
20130180892 | Marrapodi et al. | Jul 2013 | A1 |
20130278067 | Poss et al. | Oct 2013 | A1 |
20130295970 | Sheshadri et al. | Nov 2013 | A1 |
20130345969 | Udeshi | Dec 2013 | A1 |
20140005039 | Ohishi et al. | Jan 2014 | A1 |
20140050397 | Badholm et al. | Feb 2014 | A1 |
20140074298 | Jambeck et al. | Mar 2014 | A1 |
20140214697 | Mcsweeney | Jul 2014 | A1 |
20140229501 | Josefiak | Aug 2014 | A1 |
20140266698 | Hall et al. | Sep 2014 | A1 |
20140351328 | Woods et al. | Nov 2014 | A1 |
20150081212 | Mitchell et al. | Mar 2015 | A1 |
20150088792 | O'Neill et al. | Mar 2015 | A1 |
20150148077 | Jelle | May 2015 | A1 |
20150186497 | Patton et al. | Jul 2015 | A1 |
20150271126 | Menayas et al. | Sep 2015 | A1 |
20150271639 | Ziskind et al. | Sep 2015 | A1 |
20150310606 | Shreve et al. | Oct 2015 | A1 |
20150339864 | Herron | Nov 2015 | A1 |
20150348252 | Mask | Dec 2015 | A1 |
20160003627 | Bonhomme | Jan 2016 | A1 |
20160195602 | Meadow | Jul 2016 | A1 |
20160212590 | Decorte et al. | Jul 2016 | A1 |
20160266258 | Huang et al. | Sep 2016 | A1 |
20160292862 | Mask | Oct 2016 | A1 |
20170109027 | Shubs et al. | Apr 2017 | A1 |
20170124643 | Haimi et al. | May 2017 | A1 |
20170223499 | Jernigan | Aug 2017 | A1 |
20170289754 | Anderson et al. | Oct 2017 | A1 |
20170353829 | Kumar et al. | Dec 2017 | A1 |
20170366616 | Rodrigues Nascimento et al. | Dec 2017 | A1 |
20180027371 | Austraat et al. | Jan 2018 | A1 |
20180075108 | Park et al. | Mar 2018 | A1 |
20180082279 | Vasgaard et al. | Mar 2018 | A1 |
20180108023 | Stewart et al. | Apr 2018 | A1 |
20180165970 | Namgoong et al. | Jun 2018 | A1 |
20180174067 | Spiro | Jun 2018 | A1 |
20180189835 | Deluca et al. | Jul 2018 | A1 |
20180192237 | Privitera et al. | Jul 2018 | A1 |
20180232767 | Garg et al. | Aug 2018 | A1 |
20180262903 | Ryan et al. | Sep 2018 | A1 |
20180268377 | Mofardin et al. | Sep 2018 | A1 |
20180312095 | Eletrabi | Nov 2018 | A1 |
20180338001 | Pereira Cabral et al. | Nov 2018 | A1 |
20180338031 | Subramanian | Nov 2018 | A1 |
20180374046 | Powers et al. | Dec 2018 | A1 |
20190007484 | Chen et al. | Jan 2019 | A1 |
20190026915 | Seaman et al. | Jan 2019 | A1 |
20190102726 | Ushiki et al. | Apr 2019 | A1 |
20190164081 | Deluca et al. | May 2019 | A1 |
20190313204 | Ayoub et al. | Oct 2019 | A1 |
20200100052 | Ellis | Mar 2020 | A1 |
20200145783 | Hanada | May 2020 | A1 |
20200162842 | Jones | May 2020 | A1 |
20200252742 | Yeon et al. | Aug 2020 | A1 |
20200344567 | Wirola et al. | Oct 2020 | A1 |
20210006936 | Wirola | Jan 2021 | A1 |
Number | Date | Country |
---|---|---|
101482742 | Aug 2011 | CN |
2284104 | Feb 2011 | EP |
2386686 | Sep 2003 | GB |
2012015664 | Feb 2012 | WO |
2012164098 | Dec 2012 | WO |
2014079586 | May 2014 | WO |
Entry |
---|
Vicentini et al. Sensorized waste collection container for content estimation and collection optimization. Waste Management 29 (2009) [retrieved on Oct. 29, 2014], Retrieved from the Internet: pp. 1467-1472. |
Gates, Jason , “Fullness Monitoring for Waste- Image-based vs. Ultrasonic Sensors”, https://medium.com/@compology/fullness-monitoring-for-waste-image-based-vs-ultrasonic-sensors-29f360bf01e8, dated Jul. 11, 2017. |
Karpathy, Andrej , “CS231n Convolutional Neural Networks for Visual Recognition”, http://cs231n.github.io/neural-networks-2/#losses. |
Lamba, Marshall , “One Shot Learning with Siamese Networks using Keras”, https://towardsdatascience.com/one-shot-leaming-with-siamese-networks-using-keras-17f34e75bb3d, dated Jan. 20, 2019. |
Number | Date | Country | |
---|---|---|---|
62841447 | May 2019 | US |