This application is based upon and claims the benefit of priority of the prior Japanese Patent Application No. 2015-209261, filed on Oct. 23, 2015, the entire contents of which are incorporated herein by reference.
The embodiments discussed herein are related to a technique for detecting a line of sight of a subject.
There is a technique for detecting a line of sight of a person, who is a subject, by a corneal reflection method using a near infrared light source and a near infrared camera for example, Takehiko Ohno et al., An Eye Tracking System Based on Eye Ball Model-Toward Realization of Gaze Controlled Input Device—, Information Processing Society of Japan Technical Reports 2001-HI-93, 2001, pp 47-54). In the corneal reflection method, a reflection is generated on a cornea using a near infrared light source, and a center position of the corneal reflection and a center position of a pupil are obtained from image processing. In the corneal reflection method, the line of sight of a subject is detected from a relationship between the center position of the corneal reflection and the center position of a pupil.
Also, if a subject whose line of sight is to be detected is a person who is wearing glasses, for example, near infrared light irradiated from a near infrared light source is reflected on the lens surface of the glasses. Following, a reflection on the lens surface of glasses is referred to as a glasses reflection. Various proposals have been made on the line of sight detection techniques in consideration of a glasses reflection (for example, Japanese Laid-open Patent Publication Nos. 2006-167256 and 2003-339642).
According to an aspect of the invention, a line of sight detection system includes a light source that emits light toward a subject, a light detector configured to detect reflected light from the subject, and a processor configured to change an emission pattern of the light emitted from the light source toward the subject, identify a line of sight position of the subject based on a change of the reflected light from the subject when the emission pattern of the light is changed, and output information of the line of sight position.
The object and advantages of the invention will be realized and attained by means of the elements and combinations particularly pointed out in the claims.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are not restrictive of the invention, as claimed.
In an image captured in a state in which a glasses reflection occurs, the glasses reflection overlaps a corneal reflection and a pupil, so that it sometimes becomes difficult to detect the corneal reflection and the pupil from the image. Thus, with a technique disclosed in the present embodiment, it is desirable to detect a line of sight of a subject who is wearing glasses.
Following, a detailed description will be given of embodiments of the line of sight detection technique according to the present disclosure with reference to the drawings. In this regard, this disclosure is not limited to the embodiments.
The detection unit 7 is a camera having a sensitivity to light of the predetermined wavelength, which is irradiated by the light source 6, for example. Accordingly, the detection unit 7 captures images of the subject so as to detect reflected light from the subject, which is irradiated from the light source 6. In this regard, the subject is a person whose line of sight is to be detected. Also, the subject represents the subject and another objects (including glasses).
In the present embodiment, near infrared light that is invisible to the subject is used as light of the predetermined wavelength. Accordingly, the camera as an example of the detection unit 7 is a near infrared light camera, the light source 6 is a light emitting diode (LED) that emits near infrared light. The images obtained by the detection unit 7 as a detection result are near infrared images. The near infrared image includes the state of the subject in brightness in accordance with the intensity of the reflection of near infrared light irradiated from the light source 6 and near infrared light (for example, natural light or fluorescent light) irradiated from another light source.
The line of sight detection device 1 detects a line of sight of the subject. In this regard, the line of sight detection device 1 is a computer including a processor that executes various kinds of processing and a memory that stores information, for example.
In the line of sight detection system 100, the light source 6 and the detection unit 7 are coupled to the line of sight detection device 1. However, the coupling mode thereof may be a mode of wireless communication in addition to a wired coupling. For example, when line of sight detection processing is started, under the control of the line of sight detection device 1, the detection unit 7 captures images of the subject at certain time intervals and outputs the captured images to the line of sight detection device 1. Also, when line of sight detection processing is started, the line of sight detection device 1 performs control so as to turn on the light source 6.
In the present embodiment, the line of sight detection device 1 performs line of sight detection processing on the near infrared image obtained from the detection unit 7 so as to detect a line of sight of the subject. For example, the line of sight detection device 1 detects a center position of the corneal reflection of the light irradiated from the light source 6 and a center position of the pupil of the subject using the near infrared image so as to detect a line of sight.
The processing result by the line of sight detection device 1 is utilized for a marketing analysis, for example. Specifically, if the detection unit 7 and the light source 6 are disposed on a shelf or a plate of goods in a store, the line of sight detection device 1 detects lines of sight of customers (subjects) who have visited the store. In the marketing analysis, an estimation is made as to which product has interested the customers from the line of sight detection result. For example, it is possible to grasp the products that have interested a large number of customers from the output (line of sight detection result) of the line of sight detection device 1.
Also, the processing result by the line of sight detection device is utilized for dangerous driving detection, for example. Specifically, if the detection unit 7 is disposed at the place capable of capturing images of a driver who sits on a driver's seat, and the light source 6 is disposed at the place capable of irradiating light onto the driver, the line of sight detection device detects the line of sight of the driver (subject). For the purpose of dangerous driving detection, estimations are made as to whether or not the driver pays attention to various directions, and whether or not there is a risk of drowsy driving from the line of sight detection result.
Further, as illustrated in
The control unit 3 changes the irradiation area of light irradiated from the light source 6 on the subject. For example, if the light source 6 is a light source including a plurality of light emitting elements, the control unit 3 controls turning on or turning off of each of the light emitting elements in order to control the number of the elements to be turned on. By controlling the number of light emitting elements to be turned on, the control unit 3 may control to change the irradiation area of light irradiated from the light source 6 on the subject. Also, if the line of sight detection system 100 is provided with an aperture disposed between the light source 6 and the subject, the control unit 3 may control the size of the aperture so as to change the irradiation area of light irradiated from the light source 6 on the subject.
The identification unit 4 identifies the line of sight position of the subject based on a change of light reflected from the subject when the irradiation area of light is changed. That is to say, the identification unit 4 detects the line of sight of the subject. For example, the identification unit 4 detects a corneal reflection and a pupil of the subject using a first image captured when the irradiation area is in a first state and a second image captured when the irradiation area is in a second state so as to identify the line of sight position of the subject.
As described later, it is assumed that the irradiation area in the first state is larger than the irradiation area in the second state. For example, if the light source 6 is a light source including an array of a plurality of light emitting elements, the first state is a state in which all the light emitting elements are turned on. Also, the second state is a state in which some of the light emitting elements are turned on (the other light emitting elements are turned off). In this regard, some of the light emitting elements to be turned on in the second state are light emitting elements that do not have a discrete positional relationship with each other, but that exist in a gathered manner, for example adjacently, or the like. Also, for example, if the line of sight detection system 100 is provided with an aperture, the first state is a state of the widest aperture. The second state is a state of a narrow aperture.
Here, in the present embodiment, when an irradiation area of light irradiated from the light source 6 onto the subject is changed, the fact that a change in the size of the glasses reflection is larger than a change in the size of the corneal reflection is used.
Here, light irradiated from the light source 6 is reflected at various places on the lens surface 11 of the glasses However, the reflected light detected by the detection unit 7 (camera) is the light reflected at an intersection point between a line connecting the middle point between the center position of the camera lens and the position of the light source 6 and the center of the spherical surface of the lens surface 11 of the glasses, and the lens surface 11 of the glasses. That is to say, when the light source 6 has the irradiation area illustrated in
in the same manner, light irradiated from the light source 6 is reflected at various places on the eyeball 10. However, the reflected light detected by the detection unit 7 (camera) is the light reflected at an intersection point between a line connecting the middle point between the center position of the camera lens and the position of the light source 6 and the center of the sphere of the eyeball 10, and the eyeball 10. That is to say, when the light source 6 has the irradiation area illustrated in
Further, in the projection plane 12 in
On the other hand, in
As described above, depending on the difference in the curvature between the glasses and the eyeball, and the difference in the radius between the glasses and the eyeball, the glasses reflection becomes significantly smaller in the state in
Referring back to
Next, a detailed description will be given of the identification unit 4.
The recognition unit 41 recognizes a subject and a specific part of the subject from images. For example, the recognition unit 41 recognizes a face of the subject from each of the first image captured when the irradiation area is in the first state and the second image captured when the irradiation area is, in the second state. In this regard, the conventional face recognition technique is used for the face recognition.
Further, the recognition unit 41 may recognize a specific part from a region (face region) recognized as a face. In a specific part recognition, the conventional recognition technique is used for recognizing parts of a human face. In the present embodiment, the recognition unit 41 searches a face region for a region matching the features of an eye that are learned in advance so as to recognize an eye. The recognition unit 41 then generates eye region information indicating an eye region in the image and stores the eye region information in the storage unit 5. In this regard, the eye region information is information including, for example information indicating a range of an eye region and information indicating a center position (or a position of the center of gravity) of the eye region.
If it is not possible to detect a corneal reflection from the first image that was captured in a state, in which the irradiation area irradiated from the light source 6 is wider, by the corneal reflection detection unit 44 described later, the glasses reflection detection unit 42 detects a glasses reflection region from the first image. For example, the glasses reflection detection unit 42 identifies a pixel region having a luminance value of a predetermined value (for example, 240) or more from the eye region in the first image based on the eye region information input from the recognition unit 41, and if the region includes pixels of a predetermined number (for example, 30 pixels) or more, the glasses reflection detection unit 42 detects the region as a glasses reflection region.
The glasses reflection detection unit 42 stores the information on the detected glasses reflection region into the storage unit 5. In this regard, the information on the glasses reflection region includes, for example information indicating the range of the glasses reflection region and information indicating the center position (or a position of the center of gravity) of the glasses reflection region.
Next, the extraction unit 43 extracts a feature point from each of the first image and the second image. For example, the extraction unit 43 extracts a corner as a feature point from the eye region in the first image using the Harris operator or the Fast operator. The extraction unit 43 then stores the position of the extracted feature point into the storage unit 5. The extraction unit 43 performs the same operations for the second image. In this regard, the extraction unit 43 extracts, as feature points, a corner of the frame of glasses, an inner corner of the eye, and the like, for example.
The corneal reflection detection unit 44 detects a corneal reflection from the first image captured in the state in which the irradiation area irradiated from the light source 6 is wider. However, if it is not possible to detect a corneal reflection from the first image, the corneal reflection detection unit 44 detects a corneal reflection from the second image captured in the state of a narrower irradiation area.
When the corneal reflection detection unit 44 detects a corneal reflection from the second image, the corneal reflection detection unit 44 uses the information of the feature point extracted by the extraction unit 43. Specifically, the corneal reflection detection unit 44 associates the feature point extracted from the first image with the feature point extracted from the second image. The corneal reflection detection unit 44 then calculates a motion vector of the associated feature points. The amount of movement and the movement direction are reflected on the motion vector when the subject moves from a point in time of capturing the first image to a point in time of capturing the second image.
For example, if a feature point (x, y) in the first image is associated with a feature point (x′, y′) in the second image, its motion vector V is expressed by the expression 1 described below.
V=(x′−x, y′−y) expression 1
The corneal reflection detection unit 44 then estimates the center position of the glasses reflection in the second image in accordance with the center position of the glasses reflection region detected from the first image and the motion vector. That is to say, the corneal reflection detection unit 44 estimates the position where the glasses reflection may occur in consideration of the movement of the subject.
Specifically, the corneal reflection detection unit 44 reflects the motion vector V on the position (Px, Py) of the glasses region in the first image and estimates the position (Px′, Py′) of the glasses region in the second image.
Px′=Px+(x′−x), Py′=Py+(y′−y) expression 2
The corneal reflection detection unit 44 then detects a set of high luminance pixels that exist at the place other than the surroundings of the estimated position of the glasses reflection from the second image and detects the set as a corneal reflection. In this regard, the surroundings of the estimated position of the glasses reflection refer to within a radius of 10 pixels, for example.
In the present embodiment, at the time of capturing the second image, the irradiation area of the light irradiated from the light source 6 to the subject is reduced so that the area in which a glasses reflection is generated is reduced in order to separate from a corneal reflection. The corneal reflection is a set of high luminance pixels similar to the glasses reflection, and thus the corneal reflection and the glasses reflection have to be distinguished.
However, it is lees effective if the same detection method of a glasses reflection region is applied to the second image as the detection method that is applied by the glasses reflection detection unit 42 to the first image. Because by reducing the irradiation area, the glasses reflection in the second image itself becomes small. Accordingly, there is a high possibility that it is not possible to detect a glasses reflection from the second image using the algorithm (for example, detecting a region of a set of 30 or more high luminance pixels) executed by the glasses reflection detection unit 42.
Thus, after the irradiation area of the light irradiated from the light source 6 is reduced, the corneal reflection detection unit 44 estimates the position of the glasses reflection in the second image in consideration of the motion vector and the glasses reflection position in the first image. The corneal reflection detection unit 44 then detects a set of high luminance pixels that exist at the place other than the surroundings of the estimated position of the glasses reflection from the second image as a corneal reflection.
In this regard, the corneal reflection detection unit 44 may estimate not only the position (the center position and the position of the center of gravity) of the glasses reflection region, but the glasses reflection region in accordance with the irradiation area. For example, if the irradiation area is made one quarter from the first state to the second state, it is estimated that a glasses reflection that is one quarter the glasses reflection region in the first image occurs at the estimated position of the glasses reflection region in the second image. The corneal reflection detection unit 44 then detects a set of high luminance pixels that exist in a region other than the estimated glasses reflection region as a corneal reflection.
Next, when a corneal reflection is detected from the first image, the pupil detection unit 45 detects a pupil from the first image. Also, when a corneal reflection is not detected from the first image and when a corneal reflection is detected from the second image, the detection unit 45 detects a pupil from the second image. In this regard, the conventional pupil detection technique is applied to the detection of a pupil. For example, the pupil detection unit 45 performs matching with a template corresponding to the shape of a pupil.
Next, if the corneal reflection could not be detected from the first image, the decision unit 46 decides the irradiation area to be irradiated from the light source 6 before the detection unit 7 captures the second image, and inputs the decided information to the control unit 3. The control unit 3 controls the irradiation area based on the control information input from the decision unit 46. For example, if the light source 6 includes a plurality of light emitting elements, the control information is the number of light emitting elements to be turned on and their positions. If the line of sight detection system 100 has an aperture, the control information is the size of the aperture.
Here, a description will be given of a method of deciding the irradiation area by the decision unit 46. First, the decision unit 46 obtains the distance between the subject or a part (eye) of the subject and the detection unit 7. The distance may be measured by a distance sensor, or may be estimated by the following method from the first image.
For example, the decision unit 46 obtains the distance (pixels) between the pupil of the right eye and the pupil of the left eye in the first image. For example, it is assumed to be understood that in the specification of a certain detection unit 7, when the distance between the subject and the detection unit 7 (camera) is 60 cm, the distance between both pupils is 30 pixels. At this time, the distance between the subject and the detection unit 7 is estimated from this association relationship and the obtained distance of both pupils from the first image. In this regard, the shorter the distance between the subject and the detection unit 7, the longer the distance between both pupils. Accordingly, the relationship between the two is inversely proportional. Also, if the distance between the subject and the detection unit 7 is obtained from the first image, the distance between the pupils may not be used, and the distance between inner corners of both eyes, or the like may be used.
The decision unit 46 then estimates the size of the corneal reflection to be originally detected in the first image in which a corneal reflection was unable to be detected from the distance between the subject and the detection unit 7. At this time, the decision unit 46 uses the association relationship between the distance of the subject and the detection unit 7, and the size of the corneal reflection. The association relationship is learned in advance, and an association table in which the association relationship is described is stored in the storage unit 5.
The decision unit 46 refers to the association table so as to estimate, for example that the size of the corneal reflection to be originally detected is “5 (pixel)” if the distance between the subject and the detection unit 7 is “50 (cm)”.
Next, the decision unit 46 decides how much the irradiation area is to be reduced in order for the glasses reflection in the first image in which a glasses reflection occurs to be reduced to the size corresponding to the corneal reflection to be originally detected. That is to say, the decision unit 46 decides to what fraction of the irradiation area (the first state) at the time of capturing the first image the irradiation area is to be reduced in accordance with the ratio between the size of the glasses reflection in the first image and the estimated corneal reflection. Here, the decided irradiation area becomes the second state.
For example, if the size of the glasses reflection in the first image is 40 pixels, and the size of the estimated corneal reflection is 5 pixels, the decision unit 46 decides the irradiation area at the time of capturing the second image to be one eighth the irradiation area at the time of capturing the first image. In this case, for example if the light source 6 includes a plurality of light emitting, elements, the decision unit 46 decides the number of light emitting elements to be turned on to be one eighth. In this regard, it is assumed that the storage unit 5 stores the number of light emitting elements included in the light source 6 in advance Also, if the number of light emitting elements to be turned on, which is decided from the ratio between the size of the glasses reflection in the first image and the estimated corneal reflection, is not an integer, the decision unit 46 ought to decide the number of light emitting elements to be turned on to be an integer by rounding down decimal places, or the like, for example.
Next, the line of sight detection unit 47 detects the line of sight position of the subject using, the center position of the corneal reflection and the center position of the pupil, which was detected from the first image or the second image. In this regard, the conventional corneal reflection method is applied to the method of deciding the line of sight position from the center position of the corneal reflection and the center position of the pupil.
Next, a description will be given of the flow of learning processing for creating an association table according to the present disclosure. In this regard, the learning processing may be executed by the line of sight detection system 100 or may be executed by another computer.
In the former case, the line of sight detection system 100 further includes a learning unit (not illustrated in the
In the latter case, another computer executes the learning processing and outputs the association relationship corresponding to the learning result to the line of sight detection system 100. The line of sight detection system 100 stores the association relationship, which is input, into the association table (the storage unit 5).
Next, the learning unit obtains an image from the detection unit 7 (Op. 2). The learning unit then obtains the distance between the subject and the detection unit 7 when the image is captured (Op. 3). In this regard, the method of obtaining the distance may be the same as the method of obtaining the distance by the corneal reflection detection unit 44, or the subject or the administrator may input the distance.
The learning unit then determines whether or not the distance between the subject and the detection unit 7 is within a predetermined range. (Op. 4) In this regard, the predetermined range is, for example 40 cm to 70 cm, and a distance suitable for the line of sight detection processing is set. If the distance between the subject and the detection unit 7 is within the predetermined range (Op. 4 YES), the learning unit detects a face region and an eye region from the image (Op. 5). Further, the learning unit determines whether or not there is no glasses reflection in the eye region (Op. 6). If there is no glasses reflection (Op. 6 YES), the learning unit detects a corneal reflection (Op. 7). Further, the learning unit stores the distance obtained in Op. 3 and the size of the corneal reflection detected in Op. 7 in association with each other in the storage unit 5 (Op. 8).
On the other hand, if the distance between the subject and the detection unit 7 is not within the predetermined range (Op. 4 NO) or if there is a glasses reflection (Op. 6 NO) or when the processing in Op.8 is completed, the learning unit determines whether or not to terminate the learning processing (Op. 9). For example, if the size of a corneal reflection in each distance has been learned, or there is an input of terminating the learning processing, the learning unit terminates the, learning processing (Op. 9 YES). On the other hand, if the learning processing is not terminated (Op. 9 NO), the learning unit obtains a new image (Op. 2). By the above learning processing, an association relationship (association table) as illustrated in
Next, a description will be given of the flow of line of sight detection processing according to the present embodiment,
First, the control unit 3 performs control so that the irradiation area irradiated from the light source 6 becomes the first state (Op. 11). For example, if the light source 6 includes a plurality of light emitting elements, the control unit 3 performs control so as to turn on all the light emitting elements. Also, if the line of sight detection system 100 has an aperture (not illustrated in
The acquisition unit 2 obtains the first image captured after the irradiation area of the detection unit 7 irradiated by the light source 6 becomes the first state (Op. 12). In this regard, before Op. 12, the control unit 3 may control the light source 6 and then may further give an instruction to capture the first image to the detection unit 7.
Next, the recognition unit 41 detects a face region from the first image, and further detects an eye region from the face region (Op. 13). For example, the recognition unit 41 executes the face recognition processing on the first image, and if a face is recognized, the eye recognition processing is executed on the face region.
Next, the corneal reflection detection unit 44 determines whether or not a corneal reflection has been detected from the eye region in the first image (Op. 14). For example, the corneal reflection detection unit 44 determines whether or not a region including a set of about 3 to 10 pixels having a luminance value of 240 or more has been detected from the eye region.
If a corneal reflection is detected (Op. 14 YES), the pupil detection unit 45 further detects a pupil from the eye region in the first image (Op. 21). The line of sight detection unit 47 then detects the line of sight direction and the position of the subject at the point in time when the first image is captured based on the center position of the corneal reflection and the center position of the pupil (Op. 22).
The line of sight detection unit 47 then determines whether or not a series of line of sight detection processing is completed (Op.23). If determined that the line of sight detection processing is completed (Op.23 YES), the series of processing is completed. On the other hand, if determined that the line of sight detection processing is to be continued (Op.23 NO), the line of sight detection device 1 repeats the processing from Op.11.
On the other hand, if a corneal reflection has not been detected (Op.14 NO), the glasses reflection detection unit 42 determines whether or not a glasses reflection has been detected in the eye region (Op.15). For example, the glasses reflection detection unit 42 determines whether or not a region including a set of 30 pixels having a luminance value of 240 or more.
If a glasses reflection has not been detected (Op.15 NO), the line of sight detection unit 47 executes the processing in Op.23. That is to say, it was not possible to detect a corneal reflection in the state in which there is no glasses reflection, and thus the line of sight detection system 100 terminates the line of sight detection processing or attempts to capture the first image once again.
On the other hand, if a glasses reflection has been detected (Op.15 YES), the glasses reflection detection unit 42 stores information indicating the position and the range of the glasses reflection region (Op.16). Next, the extraction unit 43 extracts a feature point from the first image as described above and stores the position of the feature point and the feature information related to the feature point in the storage unit 5 (Op.17). In this regard, the extraction of the feature point may be carried out with an eye region as a target, or may be carried out with a face region as a target. Also, the feature information related to the feature point is color information, luminance information, and the image information of the surroundings of the feature point. The feature information related to the feature point is used for associating the feature points between the first image and the second image.
Next, the decision unit 46 obtains the distance between the subject and the detection unit 7 (Op.18). The decision unit 46 then decides the irradiation area irradiated by the light source 6 (Op.19). For example, the decision unit 46 refers to the association table stored in the storage unit 5 and estimates the size of the corneal reflection that might have originally occurred in the first image based on the distance between the subject and the detection unit 7. Further, the decision unit 46 then decides how much the irradiation area irradiated from the light source 6 ought to be reduced from the first state based on the ratio between the size of the glasses reflection detected from the first image and the size of the estimated corneal reflection. In this regard, here the decided irradiation area becomes the second state.
The control unit 3 then performs control so that the irradiation area irradiated from the light source 6 becomes the second state in accordance with the input from the decision unit 46 (Op.20). Next, the acquisition unit 2 obtains the second image captured by the detection unit 7 from the detection unit 7 after the irradiation area irradiated from the light source 6 becomes the second state (Op.31). In this regard, in advance of Op.31, the control unit 3 may control the light source 6 and then may further instructs the detection unit 7 to capture the second image.
Next, the recognition unit 41 detects a face region from the second image and detects an eye region (Op.32). The extraction unit 43 extracts a feature point from the second image (Op.33). The corneal reflection detection unit 44 then calculates a motion vector (Op.34). For example, the corneal reflection detection unit 44 makes an association between the feature points based on the similarity between the feature information of the feature point extracted from the first image and the feature information of the feature point extracted from the second image. The corneal reflection detection unit 44 then obtains a motion vector V by the expression 1 described before using a change of positions of the associated feature points.
Next, the corneal reflection detection unit 44 estimates the position of the glasses reflection region (Op.35). For example, the corneal reflection detection unit 44 reflects the motion vector V on the position (center position) of the glasses reflection detected from the first image as the expression 2 described before so as to estimate the position (center position) of the glasses reflection in the second image.
The corneal reflection detection unit 44 then determines whether or not a corneal reflection has been detected (Op.36). For example, the corneal reflection detection unit 44 determines whether or not a set of high luminance pixels has been detected in a region other than the estimated position of the surroundings of the glasses reflection region. If a corneal reflection has been detected (Op.36 Yes), the pupil detection unit 45 detects a pupil from the second image (Op.37). Further, the line of sight detection unit 47 detects the direction and the position of the line of sight of the subject based on the center position of the corneal reflection and the center position of the pupil that are detected from the second image (Op.38).
The line of sight detection unit 47 then executes the processing of Op.23 and after that. Also, if a corneal reflection has not been detected (Op.36 NO) the line of sight detection unit 47 executes the processing of Op.23 and after that.
If a glasses reflection occurs, the line of sight detection system 100 according to the present embodiment reduces the irradiation area irradiated by the light source 6 by the above processing so as to enable reduction of the glasses reflection region. The glasses reflection then executes the line of sight detection processing on the second image captured in the state of the reduced area. Accordingly, the possibility of separating the corneal reflection and the glasses reflection becomes high, and thus it becomes highly possible to enable detection of a corneal reflection from the second image.
Further, the line of sight detection system 100 according to the present embodiment estimates the position of the glasses reflection in the second image from the position of the glasses reflection detected in the first image and the motion vector indicating the movement of the subject. Accordingly, it is possible for the line of sight detection system 100 to distinguish the target corneal reflection and the glasses reflection in the second image. Accordingly, it is possible for the line of sight detection system 100 to restrain misdetection of the line of sight, which is caused by mistakenly recognizing a glasses reflection as a corneal reflection.
In
Also, it is difficult to distinguish the glasses reflection 25 that has been reduced and the corneal reflection 26 in the image. However, with the present embodiment, it is possible to estimate the position of the glasses reflection 25 from the glasses reflection 24 in the first image and the motion vector V, and thus it is possible to estimate the position of the glasses reflection and then to correctly detect the corneal reflection.
Next, a description will be given of an example of a hardware configuration of the line of sight detection system including the line of sight detection device.
The line of sight detection device 1 included in the line of sight detection system 100 includes, as a hardware configuration, a processor 101, a read only memory (ROM) 102, a random access memory (RAM) 103, a hard disk drive (HDD) 104, a communication device 105, an input device 108, a display device 109, and a medium reading device 110. Further, the line of sight detection system 100 includes the light source 106 and the detection device 107 in addition to each of the hardware configuration included in the line of sight detection device 1. Each of the units then is mutually coupled via a bus 111. Each of the units is capable of mutually transmitting and receiving data under the control of the processor 101.
A program according to the line of sight detection processing is recorded in a recording medium which is readable by the line of sight detection system 100. The recording medium that is readable by the line of sight detection system 100 is a magnetic recording device, an optical disc, an optical magnetic recording medium, a semiconductor memory, or the like. The magnetic recording device is an HDD, a flexible disk (FD), a magnetic tape (MT), or the like.
The optical disc is a Digital Versatile Disc (DVD), a DVD-RAM, a Compact Disc-Read Only Memory (CD-ROM), a Compact Disc-Recordable/ReWritable (CD-R/RW), or the like. The optical magnetic recording medium is a Magneto-Optical disk (MO), or the like. When a program describing, the processing according to each embodiment is distributed, it is thought that for example, a portable recording medium, such as a DVD, a CD-ROM, or the like on which the program is recorded is marketed.
The medium reading device 110 of the line of sight detection system 100, which executes the program according to the present embodiment, reads the program from the recording medium on which the program is recorded. The processor 101 stores the read program into the HDD 104, the ROM 102, or the RAM 103.
The processor 101 performs overall operation control of the line of sight detection device 1. The processor includes an electronic circuit, for example a central processing unit (CPU), or the like.
The processor 101 then reads the program describing the processing according to the present embodiment from the HDD 104 and executes the program so that the processor 101 functions as the control unit 3 and the identification unit 4 in the line of sight detection device 1. The communication device 105 functions as the acquisition unit 2 under the control of the processor 101. The HDD 104 stores various kinds of information and functions as the storage unit 5 under the control of the processor 101. The various kinds of information may be stored in the ROM 102 or the RAM 103, which is capable of being accessed by the processor 101, in the same manner as the program. Further, various kinds of information that is temporarily generated and held in the process of the processing, is stored in the RAM 103, for example.
The input device 108 receives the various kinds of information. The input device 108 is, for example a keyboard or a mouse. The display device 109 displays the various kinds of information. The display device 109 is a display, for example.
In this manner, each of the functional units illustrated in
Also, the line of sight detection processing illustrated in
Here, a description will be given of variations of the present embodiment. In the present embodiment, when a glasses reflection is detected, the decision unit 46 decides the irradiation area irradiated from the light source 6, and the control unit 3 performs control so that the irradiation area of the light source becomes the second state. However, the first state and the second state may be switched over for each detection interval of the detection unit 7 as in the case of the variations. For example, in the variations, the irradiation area irradiated from the light source 6 is enlarged (the first state) and reduced (the second state) for each frame interval of a camera.
First Variation
For example, if a glasses reflection has been detected in the first image captured in the first state in the same manner as the embodiment described above, the control unit 3 performs control so that the irradiation area irradiated from the light source 6 becomes the second state. That is to say, if a glasses reflection has been detected in the first image captured in the first state, the irradiation area irradiated from the light source 6 is switched between the first state and the second state between the N-th frame and the (N+1)-th frame in the same manner as the embodiment described above.
Also, in the first variation, further if a glasses reflection has not been detected in the first image captured in the first state, the control unit 3 performs control so that the irradiation area irradiated from the light source 6 becomes a third state. In this regard, the third state is a specified state, and is a state in which the irradiation area is set to one-half the irradiation area in the first state, for example. The detection unit 7 then captures the reflected light of the light irradiated in the third state (third image). That is to say, the irradiation area irradiated from the light source 6 is switched between the first state and the third state between the N-th frame and the (N+1)-th frame.
As described above, in the variation, the irradiation area irradiated from the light source in the (N+1) frame is more reduced than that in the first state regardless of the presence or absence of a glasses reflection. However if there is a glasses reflection, the irradiation area becomes the second state, whereas if there is no glasses reflection, the irradiation area becomes the third state.
Second Variation
Further, the irradiation area irradiated from the light source 6 may be switched between the first state and the second state for each frame. However, unlike the first variation, the “second state” in the second variation is set to a specified value (the third state in the first variation). For example, the irradiation area in the first state is the maximum irradiation area, and the irradiation area in the second state is set to a specified value (for example, one-half the value in the first state).
The line of sight detection system 100 according to this variation switches the irradiation area irradiated by the light source 6 for each frame between the first state (the maximum irradiation area) and the second state (the irradiation area of one half the irradiation area in the first state). In the embodiment described above, the glasses reflection in the second image is reduced to the size of the corneal reflection that is originally detected in the first image. However, the purpose of the second variation is to reduce the glasses reflection in the second image to less than the glasses reflection in the first image.
Third Variation
In the second variation, the irradiation area in the second state is set to a specified value regardless of the distance between the subject and the detection unit 7. However, the decision unit 46 may decide the irradiation area in the second state in accordance with the distance between the subject and the detection unit 7. That is to say, in the third variation, the irradiation area is decided using the distance in the same manner as the embodiment. However, unlike the embodiment, the size of the glasses reflection in the first image is not used.
Specifically, the storage unit 5 stores in advance an association relationship between the distance between the subject and the detection unit 7, and the irradiation area to be set. The decision unit 46 then refers to the association relationship, and decides the irradiation area in the second state in accordance with the distance between the subject and the detection unit 7. That is to say, in the line of sight detection system 100 according to this variation, the irradiation area irradiated from the light source 6 is switched between the first state and the second state decided from the association relationship set in advance for each frame.
In the embodiment described above, the irradiation area is decided in order to reduce the glasses reflection in the second image to the size of the corneal reflection that is to be originally detected in the first image using the size of the glasses reflection in the first image. However, the purpose of the third variation is to reduce the glasses reflection in the second image to less than the glasses reflection in the first image.
All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the invention and the concepts contributed by the inventor to furthering the art, and are to be construed as being without limitation to such specifically recited examples and conditions, nor does the organization of such examples in the specification relate to a showing of the superiority and inferiority of the invention. Although the embodiments of the present invention have been described in detail, it should be understood that the various changes, substitutions, and alterations could be made hereto without departing from the spirit and scope of the invention.
Number | Date | Country | Kind |
---|---|---|---|
2015-209261 | Oct 2015 | JP | national |