This application is based on and claims priority under 35 USC 119 from Japanese Patent Application No. 2016-062499 filed Mar. 25, 2016.
The present invention relates to a hand-raising detection device, a non-transitory computer readable medium, and a hand-raising detection method.
According to an aspect of the invention, there is provided a hand-raising detection device including a converter and a detection unit. The converter performs conversion of a predetermined space including a person into an overhead view image by using a result of three dimensional measurement performed on the predetermined space. The detection unit performs detection of a hand-raising action by using a silhouette image of the person in the overhead view image resulting from the conversion performed by the converter.
An exemplary embodiment of the present invention will be described in detail based on the following figures, wherein:
Hereinafter, an example of an exemplary embodiment will be described in detail with reference to the drawings.
A hand-raising detection device 10 according to the exemplary embodiment detects an action of raising a hand performed by a person in a shop, in front of digital signage, in a meeting room, or the like. For example, the hand-raising detection device 10 is provided to detect whether a person raises a hand to access a product in a store or a gesture such as pointing at a large screen.
As illustrated in
Further, not only a monitor 22 but also a memory device such as a hard disk (HDD) 20 is connected to the bus 21. The monitor 22 displays various pieces of information and images, and the HDD 20 stores various pieces of data such as various databases.
The ROM 14 stores a hand-raising detection program (described later in detail) for detecting an action of raising a hand performed by a person and other programs. The CPU 12 runs these programs to thereby execute various processes.
A three-dimensional (3D) sensor 24 is connected to the I/O 18. The 3D sensor 24 three-dimensionally scans a predetermined space including the 3D sensor 24 and thereby three-dimensionally measures the space. For example, laser light or the like is radiated in a scanning manner, and the 3D sensor 24 receives reflected light. The 3D sensor 24 detects a distance to an object from which the laser light is reflected and thereby measures 3D points. Results of the 3D scanning performed by the 3D sensor 24 are stored in the HDD 20 as a 3D point cloud. The locations, inclinations, and the like that are known regarding points in the space are used, for the 3D sensor 24, and the 3D coordinates of the scanned 3D points may he converted into room coordinates.
Note that the computer included in the hand-raising detection device 10 also includes not only the aforementioned components but also input devices such as a keyboard and other peripheral devices. In addition, the I/O 18 may be connected to a network or the like in addition to the aforementioned components, and the computer thus may perform communication with another device connected to the network.
The hand-raising detection program has functions of a 3D point cloud 30, an image converter 32, a silhouette extraction unit 34, an ellipse fitting unit 36, a hand-raising detection unit 38, and a fingertip detection unit 40.
The 3D point cloud 30 is a set of pieces of 3D point data detected as a result of the 3D scanning performed by the 3D sensor 24 on the predetermined space including the 3D sensor 24. The detected set of pieces of 3D point data is stored as a 3D point cloud in the HDD 50. The result of detection performed by the 3D sensor 24 is stored occasionally, and a hand-raised state changing from a hand-lowered state, for example, as illustrated in
The image converter 32 converts the 3D point data of the 3D point cloud 30 into a depth image of the person viewed from above in the room (hereinafter, referred to as an overhead depth image). For example, an image of a person with a hand not raised in
The silhouette extraction unit 34 silhouettes the image resulting from the conversion performed by the image converter 32 and extracts s region around a person as a silhouette image. Note that silhouetting herein corresponds to a process of causing a region inside the contour of an image to have a single color. For example, in the exemplary embodiment, binarization is performed in the following manner based on the image data regarding the overhead depth image resulting from the conversion performed by the image converter 32. Specifically, an image data value lower than a threshold is regarded as an image data value corresponding to the height of a floor, and a black pixel is used for the image data value. A white pixel is used for an image data value higher than the floor height. The image resulting from the conversion is thereby silhouetted to represent a specific region in the overhead depth image in a single color. The region corresponding to the person (portion higher than the floor) is extracted as a silhouette image. Specifically, a person with a hand lowered as in
The ellipse fitting unit 36 fits, to the person region extracted by the silhouette extraction unit 34, an ellipse that is circumscribed about the person region. For example, an ellipse is fitted to the person with a hand lowered as in
The hand-raising detection unit 38 detects a hand-raising action by using the silhouette image extracted by the silhouette extraction unit 34 and the ellipse fitted to the silhouette image by the ellipse fitting unit 36. For example, the hand-raising detection unit 38 detects the hand-raising action by using an area ratio between the area of the silhouette image and the area of the fitted ellipse. Specifically, as illustrated in
As illustrated in
Note that if the person raises a hand to a location higher than the head, the fingertip detection unit 40 may confuse the fingertip location with the head location in the fingertip location detection. Accordingly, the head location may be tracked, and a suddenly changed location may be judged to be the fingertip location. Alternatively, since the head is not located at the end of the ellipse unlike a fingertip, this may be utilized to distinguish the fingertip location from the head location.
In addition, when detecting the fingertip, the fingertip detection unit 40 may detect the orientation of the face to enhance the reliability of a fingertip detection result. For example, as illustrated in
A specific process performed by the hand-raising detection device 10 according to the exemplary embodiment in the configuration as described above will be described.
In step 100, the image converter 32 converts 3D point data in the 3D point cloud 30 into a depth image viewed from above in the room, and the process proceeds to step 102.
In step 102, it is judged whether the hand-raising detection unit 38 detects a hand-raising action. Based on the detection result, the silhouette extraction unit 34 extracts a silhouette image, and the ellipse fitting unit 36 fits an ellipse to a person region. A silhouette area ratio is obtained by dividing the area of the silhouette image by the area of the fitted ellipse, and it is judged whether the obtained silhouette area ratio is smaller than the predetermined threshold. If the judgment has a negative result, the process returns to step 100 to repeat the aforementioned steps. If the judgment has an affirmative result, the process proceeds to step 104.
In step 104, the fingertip detection unit 40 detects the fingertip location, and the process proceeds to step 106. In other words, the fingertip detection unit 40 detects, as the head location, the location that is highest in the hand-lowered state in the overhead depth image resulting from the conversion performed by the image converter 32 and also detects, as the fingertip location, the location that is farthest from the head location in the ellipse fitted by the ellipse fitting unit 36.
In step 106, the fingertip detection unit 40 displays, on the monitor 22, a video of an operator, a pointer for the fingertip, and the like, and the process proceeds to step 108.
In step 108, the hand-raising detection unit 38 judges whether the hand-raising action is stopped within a predetermined period of time. Whether the hand-raising action ends within the predetermined period of time is judged in order to judge whether the hand-raising action is wrongly detected. If the judgment has an affirmative result, the process proceeds to step 110. If the judgment has a negative result, the process proceeds to step 112.
In step 110, the hand-raising detection unit 38 determines that the hand-raising action is wrongly detected. The hand-raising detection unit 38 thus performs feedback and stops user operation. The process returns to step 100 to repeat the aforementioned steps. That is, if the hand-raising action is stopped within the predetermined period of time, it is judged that the hand-raising action has been wrongly judged. Since the detection sensitivity for the hand-raising action is too high, feedback such as adjusting the threshold or the like for the silhouette area ratio provided for judging the hand-raising action is performed, and user operations such as clicking displayed content are stopped.
In step 112, the fingertip detection unit 40 continues an interaction operation based on the fingertip. The process then returns to step 100 to repeat the aforementioned steps. That is, the fingertip detection unit 40 detects, as the head location, the location that is highest in the hand-lowered state in the overhead depth image resulting from the conversion performed by the image converter 32 and also detects, as the fingertip location, the location that is farthest from the head location in the ellipse fitted by the ellipse fitting unit 36. The fingertip location detected by the fingertip detection unit 40 is used as a pointing operation and other operations. If the fingertip location is found, it is understood which part of a shelf in the store a person touches or what a person is pointing at in the meeting room. As described above, when detecting a fingertip, the fingertip detection unit 48 may further detect the orientation of the face to check whether the fingertip location is correct.
In addition, the height within which the hand-raising action is detected by the hand-raising detection unit 38 may be limited in the exemplary embodiment described above. For example, as illustrated in
In addition, the process executed by the hand-raising detection device 10 according to the exemplary embodiment may be stored as a program in a storage medium for distribution.
The exemplary embodiment is not limited to the exemplary embodiment described above. It goes without saying that the exemplary embodiment may be performed after any of various modifications is made without departing from the spirit of the exemplary embodiment.
The foregoing description of the exemplary embodiment of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Obviously, many modifications and variations will be apparent to practitioners skilled in the art. The embodiment was chosen and described in order to best explain the principles of the invention and its practical applications, thereby enabling others skilled in the art to understand the invention for various embodiments and with the various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
2016-062499 | Mar 2016 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
7324080 | Hu | Jan 2008 | B1 |
7427140 | Ma | Sep 2008 | B1 |
7853288 | Ma | Dec 2010 | B2 |
8125163 | Dunn | Feb 2012 | B2 |
8508155 | Schuch | Aug 2013 | B2 |
8829815 | Dunn | Sep 2014 | B2 |
8830189 | Rimon | Sep 2014 | B2 |
8896531 | Minnen | Nov 2014 | B2 |
8933882 | Burachas | Jan 2015 | B2 |
9030129 | Dunn | May 2015 | B2 |
9167655 | Dunn | Oct 2015 | B2 |
9529446 | Hung | Dec 2016 | B2 |
9812047 | Schuch | Nov 2017 | B2 |
9867253 | Dunn | Jan 2018 | B2 |
9924583 | Schuch | Mar 2018 | B2 |
10133454 | Wilson | Nov 2018 | B2 |
20070222910 | Hu | Sep 2007 | A1 |
20090021461 | Hu | Jan 2009 | A1 |
20090061945 | Ma | Mar 2009 | A1 |
20090289580 | Dunn | Nov 2009 | A1 |
20100231563 | Dunn | Sep 2010 | A1 |
20100235786 | Maizels | Sep 2010 | A1 |
20110279397 | Rimon | Nov 2011 | A1 |
20110283199 | Schuch | Nov 2011 | A1 |
20120051658 | Tong | Mar 2012 | A1 |
20120062736 | Xiong | Mar 2012 | A1 |
20120086344 | Schuch | Apr 2012 | A1 |
20120268364 | Minnen | Oct 2012 | A1 |
20120299891 | Fujiwara | Nov 2012 | A1 |
20120326963 | Minnen | Dec 2012 | A1 |
20130016080 | Dunn | Jan 2013 | A1 |
20140184494 | Burachas | Jul 2014 | A1 |
20140285477 | Cho | Sep 2014 | A1 |
20140304647 | Maizels | Oct 2014 | A1 |
20140340375 | Dunn | Nov 2014 | A1 |
20150169947 | Kawaguchi | Jun 2015 | A1 |
20150245443 | Dunn | Aug 2015 | A1 |
20150297986 | Dal Mutto | Oct 2015 | A1 |
20160018957 | Wilson | Jan 2016 | A1 |
20160037606 | Dunn | Feb 2016 | A1 |
20160187991 | Hung | Jun 2016 | A1 |
20160334666 | Liu | Nov 2016 | A1 |
20160335705 | Williams | Nov 2016 | A1 |
20160338181 | Schuch | Nov 2016 | A1 |
20160338182 | Schuch | Nov 2016 | A1 |
20160351133 | Kim | Dec 2016 | A1 |
20160358538 | Schuch | Dec 2016 | A1 |
20180012566 | Lin | Jan 2018 | A1 |
20180061297 | Schuch | Mar 2018 | A1 |
Number | Date | Country |
---|---|---|
2005-242759 | Sep 2005 | JP |
2007-048232 | Feb 2007 | JP |
Number | Date | Country | |
---|---|---|---|
20170277943 A1 | Sep 2017 | US |