Recognizing an object (e.g., a person or a car) by means of a single recognition system employing a facial recognition camera or license plate recognition system can lead to a high error rate due to varying environmental conditions or partial object occlusion.
This invention proposes a data fusion system that combines the results of many image capture devices or detectors as well as travel time of the object between them in order to improve the recognition rate.
For this purpose, the recognition task is considered a classification problem, where a class corresponds to a particular object from a black list. The classification problem is solved by way of a Bayesian fusion algorithm, which operates on learned statistical models of recognition scores, travel time, and prior information. All these statistical models are generated machine learning algorithms from training data and can be combined in arbitrary fusion using modular data structure.
The subject matter regarded as the invention is particularly pointed out and distinctly claimed in the concluding portion of the specification. The invention, its method of operation, features, and advantages may best be understood with reference to the following detailed description in view of the accompanying drawings in which:
It will be appreciated that for clarity figure elements may not be been drawn to scale and reference numerals may be repeated among the figures indicating corresponding or analogous elements.
In the following description, numerous details are set forth in order to provide a thorough understanding of the invention. However, it will be understood that the present invention may be practiced without these specific details. Furthermore, well-known methods, procedures, and components have not been described in detail so as not to obscure the present invention.
The present invention is a fusion-based object recognition system achieving object recognition by fusing data from a plurality of component object recognition systems together with temporal data.
Referring now to the figures,
System 1 also includes one or more clocks 16 linked to detectors 5, one or more processors 10, one or more long term storage disks 15, memory 20 loaded with, inter alia, a Bayesian fusion engine 25, fusion task orchestration module 30, a database of priors plus transition and likelihood models 35, a fusion task container 40, and image data 45.
It should be appreciated that any of the data may be stored in part or in entirely in disks 15.
Furthermore, system 1 includes user-input device 55, and an output device 60. Input device 55 may be implemented as touch-screens, keyboard, microphone, pointer device, or other devices providing such functionality and output device 60 may be implemented as any one or a combination of devices like a display screen, speakers or headphones, or other devices providing such functionality.
Memory 20 may be implemented as Random Access Memory (RAM) or other memory types providing such functionality.
As shown, memory 20 is loaded with Fusion Task Orchestration module 35. Orchestration module 35 is configured to assign incoming recognition data to an active fusion task or creates a new fusion task for it.
As shown, memory 20 is also loaded with Bayesian Fusion Engine 25. Fusion Engine 25 is configured to calculate a probability value for each class, where a class can correspond to a particular object (e.g., probability of object A is x % and of object B is y %) or a class can be binary (e.g., probability of object A is x %, and of not being object A is 100−x %).
In a certain embodiment Bayesian Fusion Engine 25 includes a time update component 25D configured to calculate the probability of an object moving from detectors 5i to 5j in time t, for example. Based on this transition probability, the class probability can be adjusted. The transition probability is provided by a Transition Model learned a priori from training data in an unsupervised fashion based on time measurements between various linked detectors without annotations; i.e. in a totally automated fashion. In certain embodiments, the transition model 25D distinguishes three phases of transitions of an object:
Bayesian Fusion Engine 25 also includes a second component configured to update measurements. The score values of the detections of a fusion task are combined using Bayes' rule. The likelihood of a score is calculated by means of a likelihood model, according to an embodiment. Construction of arbitrary complex likelihood models, a modular structure for combining multiple predefined likelihood models is employed using basic operations used for combining learned likelihood models, e.g., sum, product, scale, constant. A prior learned by means of a supervised machine learning algorithm. Especially for score values, Gaussian mixture models are learned by means of an EM algorithm with automatic model selection. To improve the model selection (i.e., the optimal number of mixture components), a novel initialization procedure is employed that avoids the standard random clustering (K means) initialization and uses mixture splitting instead.
If the fusion task merely contains a single detection corresponding to a newly created fusion task and no time update is performed (no travel information available). Instead, a so-called prior is used to initialize the fusion. The prior is also learned a prior by means of supervised machine learning or statistical methods according to certain embodiments.
The present fusion-based object-recognition system 1 has application for the detection of various types of objects as noted above. Without diminishing in scope, system operation will be described in terms of detecting and identifying a person using facial recognition cameras.
By way of example and in reference to
System 1 or the detectors 5 compare the captured image with each of the five template images on the black list stored in image data 45 and calculates a similarity value called score for each of the five people on the black list. The score is typically a value between 0 and 1, where 0 means “not matching at all” and 1 means “perfect match”.
After calculating the scores, system 1 creates a detection D1 parameter which includes three elements, according to an embodiment:
Detection parameter is passed to the Fusion Task Orchestration module 35. This module is responsible for assigning the detection to an appropriate Fusion Task A Fusion Task reflects a data structure that stores the results of the fusion and is associated to particular persons being part of the score list. All open fusion tasks are stored in the Task Container 40. In certain embodiments, a Fusion Task can be closed after a pre-defined time out.
In this example, task container is empty 40. Thus, Fusion Task Orchestration module 35 creates a new Fusion Task T1 that is associated with person of ID numbers 2 and 5, as these are the two elements of the detection. The detection is assigned to the Task and then the Task is sent to the Bayesian Fusion Engine. In a certain embodiment, two new Fusion Tasks are created; a first task associated with the person of ID number 2 and second task associated with the person of ID number 5. It should be appreciate that the orchestration module 35 in certain embodiments will deice if there is a fusion task open and if not, will open one.
Bayesian Fusion Engine 25 then calculates the probability of which person from the black list has actually been observed by a detector 5 by utilizing the Bayesian rule as shown in
Referring now to the processing steps depicted in section “A” of
where p (i) is the prior probability of person i 25A, p(D1|i) is the likelihood of detection D1 25B, and p (D1) is a normalization constant. Often the prior probability is uniform and thus, p(i)=1/6. It is important to note that i=0 is not the ID of an actual person from the black list. It can be interpreted as “person not part of the black list” and thus, p(0(D1) is the probability that the observed person is none of the people on the black list.
The likelihood model 25B reflects the probability if detection D1 belongs to person i. This model has to be learned a priori by means of machine learning algorithm. For score values, a good choice for a likelihood model is a Gaussian mixture.
For this example the evaluation of the likelihood and performing the above Bayesian rule results in the following values for p (i|D1):
If in addition the credibility value should be used, there are several options. For example, the credibility value is also evaluated by means of likelihood model 25B and thus, learning likelihood model 25B is not only based on the scores but also on the credibility values. Another way is to use a modified version of the above formula according to
The probability values for this version are listed in the right-hand column of the above table.
The Fusion Result 40B, i.e., the probability values, are stored as a Fusion Task 40A and the Task is stored in Task Container 40 as depicted in
At time instant t=10 seconds, another detector 5 observes again the same person. (Detector 5 is unaware that this person was observed previously.) The above described process of matching with the black list is repeated and leads to the following detection D2:
This detection D2 goes to Orchestration module 35, as shown in
If for instance detection D2 would have a different score list than the one above, e.g.,
Task T1 with detection D2 goes to the Fusion Engine 25, executes processing of section “B” in
First, a time update 75 is performed according to
p−(i|D1)=A(Δt)·p(i|D1)
Where A(Δt) is a so-called transition matrix depending on the time difference Δt and p−(i|D1) is the predicted probability. The transition matrix is based on a statistical model that reflects the distribution of travel times between two detectors 5i and 5j.
For simplicity let the transition matrix be A(Δt) the identity matrix, which can be interpreted as “plausible transition” of a person between both recognition systems. Thus, it holds that p−(i|D1)=p(i|D1). Next, the measurement update 70, i.e., Bayesian rule, is performed in order to calculate the probability values p(i|D1, D2) according to
with values as in the following table:
Hence, the probability that the observed person is actually person 2 became significantly higher.
Fusion task result 40B is output to an output device 60 and may be displayed as text or various forms of graphical representations, according to certain embodiment.
The fusion-based object-recognition system advantageously enables unsupervised learning of transition models from travel times between recognition systems, incorporation of transition model into fusion to provide higher detection accuracy and a lower false alarm rate.
Bayesian fusion of multiple scores and travel time enables fusion of detections that can be performed sequentially, thereby reducing computation time relative to batch approaches. The modular structure of likelihood and transition models based on “basic operators” allows simplified creation of complex models and the incorporation of additional information (e.g., expert knowledge, sensor information), and incorporation of credibility values into fusion.
It should be appreciated that combinations of various features set forth in different examples are also included within the scope of present invention.
While certain features of the invention have been illustrated and described herein, many modifications, substitutions, changes, and equivalents will now occur to those of ordinary skill in the art. It is, therefore, to be understood that the appended claims are intended to cover all such modifications and changes as fall within the true spirit of the invention.
Number | Date | Country | Kind |
---|---|---|---|
10201403293T | Jun 2014 | SG | national |
Number | Name | Date | Kind |
---|---|---|---|
6944342 | Stahl et al. | Sep 2005 | B1 |
7298877 | Collins et al. | Nov 2007 | B1 |
8468111 | Tgavalekos et al. | Jun 2013 | B1 |
20080162389 | Aboutalib | Jul 2008 | A1 |
20140316736 | Strohbach et al. | Oct 2014 | A1 |
Number | Date | Country |
---|---|---|
1983666 | Oct 2010 | EP |
2615561 | Jul 2013 | EP |
2704065 | Mar 2014 | EP |
WO 2008057451 | May 2008 | WO |
WO 2013072282 | May 2013 | WO |
WO 2013083823 | Jun 2013 | WO |
WO 2014023607 | Feb 2014 | WO |
Entry |
---|
Anastasi G et al.: “Energy conservation in wireless sensor networks: A survey”, Ad Hoc Network, Elsevier, vol. 7, No. 3, May 1, 2009, pp. 537-568, XP 026185309, ISSN: 1570-7805, DOI: 10.1016/J.ADHOC.2008.06.003 [retrieved on Jul. 29, 2008]. |
Boulis A et al.: “Aggregation in sensor networks: an energy accuracy trade-off”, New Frontiers in Telecommunications : 2003 IEEE International Conference on Communications ; ICC 2003 ; May 11-15, 2003, Anchorage, Alaska, USA; [IEEE International Conference on Communications], IEEE Operations Center, Piscataway, NJ, Jan. 1, 2003 (Jan. 1, 2003), pp. 128-138, XP010642624, ISBN: 978-0-7803-7802-5. |
Deshpande C. et al.: Model-Driven Data Acquisition in Sensor Networks, In Proceedings of the 30th Very Large Data base Conference (VLDB 2004) Toronto, Canada, Aug. 2004. |
Dieter Fox et al.: “Bayesian Filtering for Location Estimation”, IEEE Pervasive Computing, IEEE Service Center, Los Alamitos, CA, US , vol. 2, No. 3 Sep. 30, 2003 (Sep. 30, 2003), pp. 24-33, XP002598055, ISSN: 1536-1268. |
Emiliano Miluzzo et al: “CaliBree: A Self-calibration System for Mobile Sensor Networks”, Distributed Computing in Sensor Systems : 4th IEEE International Conference, Dcoss 2008, Santorini Island, Greece, Jun. 11-14, 2008; Proceedings; [Lecture Notes in Computer Science], Springer, Berlin, Heidelberg, vol. 5067, Jun. 11, 2008 (Jun. 11, 2008), pp. 314-331, XP002668228, ISBN: 978-3-540-69169-3. |
European Search Report from EP11189417 dated Feb. 28, 2012. |
European Search Report EP11192890 dated May 7, 2012. |
European Search Report from EP2615561 dated Jul. 23, 2012. |
European Search Report EP12179412 dated Dec. 21, 2012. |
Feng J et al.: “Model-based calibration for sensor networks”, Proceedings of Ieee Sensors 2003 (Ieee Cat. No. 03ch37498) IEEE Piscataway, Nj, Usa; [IEEE International Conference on Sensors], IEEE, vol. CONF. 2, Oct. 22, 2003 (Oct. 22, 2003) , pp. 737-742vol. 2, XP010691005, DOI: 10.1109/ICSENS.2003.1279039 ISBN: 978-0-7803-8133-9. |
Guestrin C et al.: “Distributed Regression: An Efficient framework for Modeling Sensor Network Data”, Information Processing in Sensor networks, 2004, IPSN 2004, Third International Symposium, Apr. 26, 2004, pp. 1-10, XP009070537, ISBN: 978-1-58113-846-7. |
Halkidi M et al.: “Clustering validity checking methods”, ACM Sigmod Record, vol. 31, No. 3, Sep. 1, 2002, XP55032760, ISSN: 0163-5808, DOI: 10.1145/601858.601862. |
Harrington B et al.: “Energy-Efficient Map Interpolation for Sensor Fields Using Kriging”, IEEE Transactions on Mobile Computing, IEE Service center, Los Alamitos, CA, US, vol. 8, No. 5, May 1, 2009, pager 622-635, XP011335350, ISSN: 1536-1233, DOI: 10.1109/TMC.2008.167. |
Hou Chaojun et al.: “Cluster Based Routing Scheme for Distributed Regression in Wireless Sensor Network: Gaussian Eliminations”. High Performance Computing and Communications, 2008. HPCC '08. 10th IEEE International Conference on, IEEE Piscataway, NJ, USA, pp. 813-818, XP031344297, ISBN: 978-0-7695-3352-0. |
Hou Chaojun et al.: “Energy-efficient in routing scheme for distributed regression in wireless sensor network”, Control Conference 2008, CCC 2008, 27th Chinese, IEEE, Piscataway, NJ, USA, pp. 354-358, XP031311542, ISBN: 978-7-900719-70-6. |
International Search Report together with Written Opinion from PCT/EP2012/072409 dated Jan. 14, 2013. |
International Search Report together with Written Opinion from PCT/EP2012/074875 dated Feb. 6, 2013. |
International Search Report together with Written Opinion from PCT/EP2013/065960 dated Sep. 9, 2013. |
International Search Report together with Written Opinion from PCT/EP2015/063436 dated Aug. 12, 2015. |
Sanford Jessica Feng et al.: “On-line sensor calibration and error modeling using single actuator stimulus”, Aerospace Conference, 2009 IEEE, IEEE, Piscataway, NJ, USA, Mar. 7, 2009 (Mar. 7, 2009), pp. 1-11, XP031450069, ISBN: 978-1-4244-2621-8. |
Jing Zhou et al.: “FloodNet: Coupling Adaptive Sampling with Energy Aware Routing in a Flood warning System”, Journal of Computer Science and Technology, Kluwer Academic Publishers Bo, vol. 22, No. 1. Feb. 14, 2007. pp. 121-130, XP019485538, ISSN: 1860-4749, DOI: 10.1007/S11390-007-9017-7. |
Ko Jonathan et al.: “GP-BayesFilters: Bayesian filtering using Gaussian process prediction and observation models”, Intelligent Robots and Systems, 2008. IROS 2008. IEEE/RSJ International Conference on, IEEE, Piscataway, NJ, USA, Sep. 22, 2008 (Sep. 22, 2008), pp. 3471-3476, XP031348720, ISBN: 978-1-4244-2057-5. |
Mahawaththa M D W S et al.: “Location aware queries for sensor network”, Advance in ICT for Emerging regions (ICTER) 2011 International Conference on , IEEE. Sep. 1, 2011, pager 2-8, XP031990376. DOI: 10.1109/ICTER.2011.6075018. ISBN: 978-1-4577-1113-8. |
Mourikis A and Roumeliotis S.I.: “Optimal Sensor Scheduling for Resource-Constrained Localization of Mobile Robot Formations”, in IEEE Transactions on Robotics, 22(5), Oct. 2006, pp. 917-931. |
Mukhopadhyay S et al.: “Model based error correction for wireless sensor networks”, Sensor and Ad Hoc Communications and Networks, 2004. IEEE Secon 2004. 2004 First Annual Ieee Communications Society Conference on Santa Clara, CA, USA Oct. 4-7, 2004, Piscataway, NJ, USA,IEEE, Oct. 4, 2004 (Oct. 4, 2004), pp. 575-584, XP010759635, DOI: 10.1109/SAHCN.2004.1381960, ISBN: 978-0-7803-8796-6. |
Osborne M A et al.: “Towards Real-Time Information Processing of Sensor Network Data Using Computationally Efficient Multi-output Gaussian Processes”, Information Processing in Sensor Networks, 2008. IPSN '08. International Conference on, IEEE, Piscataway, NJ, USA, Apr. 22, 2008 (Apr. 22, 2008), pp. 109-120, XP031246704, ISBN: 978-0-7695-3157-1. |
Rasmussen Carl E: Gaussian Processes for Machine Learning, MIT Press 2006. |
Santini S.: “Adaptive Sensor Selection Algorithms for Wireless Sensor Networks”, PhD thesis, ETH Zurich, Switzerland, 2009. |
Number | Date | Country | |
---|---|---|---|
20150363643 A1 | Dec 2015 | US |