Interaction with computing devices is a fundamental action in today's world. Computing devices, such as personal computers, tablets, and smartphones, are found throughout daily life. In addition, computing devices that are wearable, such as wearable headset devices (e.g., virtual reality (VR) headsets and augmented reality (AR) headsets), are becoming increasingly popular. The systems and methods for interacting with such devices define how they are used and what they are used for.
Advances in eye tracking technology have made it possible to interact with a computing device using a person's gaze information, such as the position on a display the user is gazing at. This information can be used for interaction solely, or in combination with a contact-based interaction technique (e.g., using a user input device, such as a keyboard, a mouse, a touch screen, or another input/output interface).
Previously proposed interaction techniques using gaze information can be found in U.S. Pat. No. 6,204,828, United States Patent Application Publication 2013/0169560, U.S. Pat. No. 7,113,170, United States Patent Application Publication 2014/0247232, and U.S. Pat. No. 9,619,020. The full specification of these patents and applications are herein incorporated by reference.
In computing devices, such as wearable devices, images of an eye may be generated and processed to detect the pupil and glints and, thereby, track the eye. In many implementations, a light source, such as one including a light emitting diode (LED), is powered on to illuminate the eye and for the imaging.
A further understanding of the nature and advantages of various embodiments may be realized by reference to the following figures. In the appended figures, similar components or features may have the same reference label. Further, various components of the same type may be distinguished by following the reference label by a dash and a second label that distinguishes among the similar components. If only the first reference label is used in the specification, the description is applicable to any one of the similar components having the same first reference label irrespective of the second reference label.
Embodiments for controlling light sources used in eye tracking are disclosed. In an example, an eye tracking system determines directly in consecutive two-dimensional images whether glints are about to fall on a pupil edge without knowledge of a three dimensional position of a cornea center. If one or more glints satisfy this condition, illuminators associated with these glints are powered off and at least a next image is generated with these illuminators in the power off state.
In particular, the eye tracking system generates multiple images showing at least a portion of the user eye illuminated by a predetermined set of illuminators of the eye tracking system. For some or all of the glints, the eye tracking system determines positions in the images of each of such glints. These positions can be relative to a pupil edge. For each of such glints, the eye tracking system also predicts a next position of the glint relative to the pupil edge based on the positions determined for the glint. Further, and for each of such glints, the eye tracking system determines, from the predetermined set, an illuminator that corresponds to the glint and determines, based on the glint's next position, whether to power off the illuminator to generate a an additional image, where this additional image is further used for the eye tracking.
These illustrative features are mentioned not to limit or define the disclosure, but to provide examples to aid understanding thereof. These and additional features may be implemented independently in various embodiments or may be combined in yet other embodiments, further details of which can be seen with reference to the following description and illustrations. Advantages offered by one or more of the various embodiments may be further understood by examining the specification or by practicing one or more of the various embodiments.
Embodiments of the present disclosure relate to, among other things, controlling light sources used for eye tracking. In an example, multiple light sources are used to illuminate an eye of a user. These light sources are referred to herein as illuminators. A predetermined set of the light sources may be powered on and multiple images are generated, each showing at least a portion of the eye. The images are processed to identify glints and the pupil edge, where each glint corresponds to a reflection from the eye of light emitted from a light source. Positions of the glints relative to the pupil edge in each image are determined and these positions over the different images are used to predict the next glint positions. If a glint is about to fall on the pupil edge or is close to the pupil edge by a certain threshold distance, the corresponding light source is selected and powered off. This approach of predictively identifying glints moving to or beyond the pupil edge and selectively powering off light sources provide many technical advantages over existing eye tracking systems.
In particular, a wearable computing device such as a VR or AR headset, may integrate an eye tracking system that uses multiple illuminators for each user eye. Different techniques are possible to use the illuminators for the eye tracking. In one prior art technique, all illuminators are powered on, thereby resulting in a large number of glints on the user eye. However, this technique may have multiple drawbacks. For example, some of the glints may occur on the pupil edge and, thus, may obstruct the pupil edge and degrade the accuracy of the eye tracking because the pupil edge may not be accurately detected. In addition, the eye tracking may also involve mapping the glints to the illuminators. However, with a large number of glints and, equivalently, illuminators, the complexity of the eye tracking increases, potentially resulting in processing latency. In another technique, a subset of the illuminators is turned on, thereby saving power consumption and generally improving the accuracy. However, this technique generally involves determining a three dimensional position of the cornea center, which can add complexity and increase the processing latency.
In comparison, embodiments of the present disclosure involve selectively powering on and off illuminators (e.g., subsets of the available light sources), thereby saving power consumption and improving the accuracy. To reduce the complexity and improve the processing latency, the embodiments also involve predicting the positions of the glints using two dimensional images without the need to determine the three dimensional position of the cornea center.
A first illuminator 111 is arranged coaxially with (or close to) the image sensor 113 so that the image sensor 113 may capture bright pupil images of the user's eyes. Due to the coaxial arrangement of the first illuminator 111 and the image sensor 113, light reflected from the retina of an eye returns back out through the pupil towards the image sensor 113, so that the pupil appears brighter than the iris surrounding it in images where the first illuminator 111 illuminates the eye. A second illuminator 112 is arranged non-coaxially with (or further away from) the image sensor 113 for capturing dark pupil images. Due to the non-coaxial arrangement of the second illuminator 112 and the image sensor 113, light reflected from the retina of an eye does not reach the image sensor 113 and the pupil appears darker than the iris surrounding it in images where the second illuminator 112 illuminates the eye. The illuminators 111 and 112 may for example, take turns to illuminate the eye, so that every first image is a bright pupil (BP) image, and every second image is a dark pupil (DP) image.
The eye tracking system 100 also comprises circuitry 120 (for example including one or more processors) for processing the images captured by the image sensor 113. The circuitry 120 may for example, be connected to the image sensor 113 and the illuminators 111 and 112 via a wired or a wireless connection. In another example, circuitry 120 in the form of one or more processors may be provided in one or more stacked layers below the light sensitive surface of the image sensor 113.
In the embodiment described with reference to
Specialized computer system 300 is shown comprising hardware elements that may be electrically coupled via a bus 390. The hardware elements may include one or more central processing units 310, one or more input devices 320 (e.g., a mouse, a keyboard, eye tracking device, etc.), and one or more output devices 330 (e.g., a display device, a printer, etc.). Specialized computer system 300 may also include one or more storage devices 340. By way of example, storage device(s) 340 may be disk drives, optical storage devices, solid-state storage devices such as a random access memory (“RAM”) and/or a read-only memory (“ROM”), which can be programmable, flash-updateable and/or the like.
Specialized computer system 300 may additionally include a computer-readable storage media reader 350, a communications system 360 (e.g., a modem, a network card (wireless or wired), an infra-red communication device, Bluetooth™ device, cellular communication device, etc.), and working memory 380, which may include RAM and ROM devices as described above. In some embodiments, specialized computer system 300 may also include a processing acceleration unit 370, which can include a digital signal processor, a special-purpose processor and/or the like.
As illustrated in
The hot mirror coating or film 510 may be applied directly onto the lens 500, or for ease of manufacturing, the hot mirror coating or film 510 may also be in the form of a separate piece which does not function as a lens. Further, and also for ease of manufacturing, a separate piece of glass or plastic may be placed atop the hot mirror, and the illuminators 520 may be affixed to this separate piece.
In an example, the number of illuminators 520 is between twenty and fifty. Each of the illuminators 520 may include an LED light on a chip of about 200 μm×200 μm. The illuminators 520 may emit infrared or near infrared light outwardly towards the eye of a wearer 540, and then reflected by the eye of the wearer back towards the lens 500. The film 510 has the properties of a hot mirror. In other words, the film 510 allows light in the visible spectrum to pass while preventing light in the infrared spectrum from passing. In this way, visible light emitted by a display 550 behind the lens 500 may pass to the wearer's eye, while most infrared light is prevented from passing.
An image sensor located behind the lens, and looking through the cut-out portion 530, captures images of the wearer's eye containing reflections of the infrared light emitted by the illuminators 520, a processing device connected to the image sensor then analyzes those images to determine a direction of the gaze of the user based on the reflections of the infrared light.
The illuminators 520 may be applied to the film 510 in multiple ways, firstly simply through glue or alternatively a possible application is by printing electronics directly onto the film 510. The illuminators 520 will further have to have communication lines applied to the film 510 such that the illuminators may receive power so as to be turned on and off.
In an example, the wearable device includes two lenses 500 with associated film 510 and illuminators 520. The film 510 may be applied to the lens through glue or some other semi-permanent substance. In a system with two lenses, light may reflect from the skin of a user into the lenses, whereby the lenses have a waveguide effect and channel the light away from the wearable device. The presence of the hot mirror 510 acts to lessen the emergence of this light.
In another example, the illuminators 520 may be placed on the side of the lens 500 (or additional pieces as previously described). Illumination from the illuminators 520 may then travel through the lens 500 to be emitted in front of the users eyes. This guiding of the light occurs in an area of the lens where the hot mirror 510 has been applied, so as to prevent illumination from being emitted towards the display 550.
Furthermore, an angled hot mirror 510 may be added in front of the display 550, and the image sensor arranged to view the hot mirror. The lens 500 is then placed in front of a user's eye with illuminators 520 located adjacent thereto. The illuminators 520 illuminate the user's eye, and due to the lens 500 having a hot mirror film or coating, as has been previously described, stray illumination from the illuminators 520 is lessened. A cutout in the hot mirror film or coating allows the image sensor to capture images of the user's eye through the angled hot mirror.
In another example, instead of using the hot mirrors 430 of
The viewing plane is the plane in which the image that the user is watching is placed. The hologram of the HOE does not need to be placed directly on the viewing plane of the user. Especially for VR headsets, a lens may be placed between the eye and the screen that places a virtual image on a comfortable distance a few meters away. In that case the HOE may be placed on the lens.
In AR devices a projector may be placed at the side of the HOE or a prism so that the projector beam is directed towards the eye. The HOE should then be placed between any physical components and the eye.
In some examples, the first HOE is adapted to direct a second portion of incident light reflected from the eye in a second angle towards the camera. This will split the image perceived by the camera into two overlapped images. This enables stereo imaging by separating the overlapped images using image analysis so that the depth of objects in the image can be determined.
As illustrated, a first image 610 and a second image 650 are generated at different points in time, with a same set of illuminators being powered on for both images 610 and 650. The first image 610 shows a user eye 612, including the iris 614, the pupil 616, the pupil edge 618, and glints 620a, 620b, 620c, and 620d corresponding to four powered on illuminators.
In an example, the position of each of the glints 620a, 620b, 620c, and 620d relative to the pupil edge is computed from the first image 610. For instance, an X,Y coordinate system is defined. The origin of this coordinate system can be at the bottom left corner of the image 610, at a particular pupil reference point 630 (e.g., at the utmost right point on the pupil edge 618, as shown in
The second image 650 may also show the user eye 612, including the iris 614, the pupil 616, the pupil edge 618, and the glints 620a, 620b, 620c, and 620d corresponding to the same four powered-on illuminators. However, because the second image 650 is generated at a different time than the first image 610, the user eye 612 may have moved relative to the illuminators and/or camera(s). Hence, at least the positions of some or all of the glints 620a, 620b, 620c, and 620d may have shifted. As illustrated in
Based on the two images 610 and 650 (and, optionally, other images similarly generated), a prediction 690 can be made of the next positions of the glints 620a, 620b, 620c, and 620d relative to the pupil edge 618. In other words, if a third image were to be generated with the same set of powered-on illuminators, the positions of the glints 620a, 620b, 620c, and 620d and, as needed, the position of the pupil reference point 630 in the third image are predicted given their corresponding positions in the first and second images 610 and 650 (and, optionally, the other images similarly generated).
Hence, glints can be detected in different images generated at different points in time. The detection can include matching the glints in the different images, as further illustrated in
As illustrated in
To illustrate, the first glint 710 is selected from the from the first plurality of glints shown in the first image. A second glint 750a and a third glint 750b are selected from the second plurality of glints shown in the second image as potential matches to the glint. Based on the first image and the second image (e.g., based on the relative positions of the selected glints in the respective images), a first distance 720 between the first glint 710 and the second glint 750a and a second distance 730 between the first glint and the third glint 750b are computed. Based on a comparison of the first distance 720 and the second distance 730, the second glint 750a is selected as matching the first glint 710 (e.g., because the first distance 720 is the smallest of the two distances 720 and 730).
In the first example 810, the predicted position of a glint is determined as a function of the distance and direction between its first and second positions determined from first and second images. This can include generating a glint vector originating at the first position and ending at the second position, and then repositioning this vector to originate at the second position. The predicted position would be the end of the glint vector originating from the second position.
As illustrated, a first position 820 of a glint is computed from a first image. A second position 830 of the glint is computed from a second image. A distance and a direction between the first position 820 and the second position 850 of the glint are determined and can be expressed as a glint vector 812. A third position 840 of the glint is estimated based on the distance and the direction, where this third position 840 corresponds to a predicted position of the glint (e.g., an estimated position of this glint in a third image to be generated with a powered-on illuminator corresponding to the glint). In particular, the third position 840 is set to be at the same equal distance and direction from the second position 830 (or some other function of the distance and direction depending, for instance, on the velocity of the eye saccade). For instance, the glint vector 812 can be repositioned to originate from the second position 830 (
To determine whether the illuminator corresponding to the glint should be turned off, the third position 840 may be compared to the pupil edge, such as a to a point on the pupil edge (e.g., a particular pupil edge reference point or the closest point on the pupil edge). This point may also have been predicted based on its positions computed from the first and second images. That relative position may include a predicted distance 816 between the third position and the pupil edge. If the third position is within a threshold distance from the pupil edge (e.g., the predicted relative position between the glint and the pupil edge, or, similarly, the predicted distance 816 between the glint and the point on the pupil edge is smaller or equal to the threshold distance), a determination is made that the illuminator should be powered off. Otherwise, the power on state of the illuminator may not be changed.
In an example, the threshold distance is a predefined value available from memory. In another example, the threshold distance includes a dynamic value defined based on the distance between the first position 820 and the second position 830 of the glint. For instance, the larger that distance is, the greater and possibly faster the eye movement is (e.g., the higher the velocity of the eye saccade is), and the larger the dynamic value becomes to avoid the potential of the glint arriving to or crossing the pupil edge.
In the second example 850, the predicted position of a glint is determined by considering an accuracy margin. In particular, the third position of the glint is predicted as described herein above in connection with the first example 810. Rather than using a distance between the third position and the pupil edge and a threshold distance, a circle 852 is generated. The circle 852 is centered at the third position. Its radius is defined based on the distance between the first and second positions (e.g., is set equal to that distance or as a function of that distance). If the circle intersects with the pupil edge, a determination is made that the illuminator should be powered off. Otherwise, the illuminator may be kept in the power on state.
The example flow starts at operation 902, where the eye tracking system selects a predetermined set of illuminators. In an example, different factors can be used for the selection and this set can be a subset from available illuminators. These factors include preferred minimum and maximum numbers of illuminators (e.g., at least two illuminators and no more than ten illuminators), a distribution of the illuminators from the available illuminators (e.g., selecting illuminators that are not too close to each other—the distance between two adjacent illuminators being larger or equal to a preferred distance), a particular pattern of illuminators (e.g., illuminators that form corners of a square or some other geometric shape), among other factors. As further illustrated in connection with remaining operations of the example flow, the predetermined set can be updated to remove and add illuminators based on the glint tracking. Hence, other usable factors can relate to the glint tracking, as further described herein below.
At operation 904, the eye tracking system powers on the predetermined set. In an example, a controller of the eye tracking system controls the power states of the illuminators and can switch power on to the illuminators from one or more power sources. As needed, the controller can also switch off power to the remaining illuminators that do not belong to the predetermined set.
At operation 906, the eye tracking system generates a first image and a second image showing at least a portion of a user eye illuminated by the predetermined set of illuminators. In an example, the first image is generated before the second image and each of these two images is generated while the predetermined set of illuminators is powered on.
At operation 908, the eye tracking system determines correspondences between glints in the two images. In an example, the eye tracking system identifies a glint in the first image and matches this glint with a glint shown in the second image. The matching can use glint patterns as described herein above in connection with
At operation 910, the eye tracking system determines a first position of the glint in the first image and a second position of the glint in the second image. In an example, each of the first position and the second position is relative to a pupil edge. These positions may be computed by using a common coordinate system in both images. In addition, the positions of the pupil edge can be determined in the two images to derive the first relative position and the second relative position.
At operation 912, the eye tracking system predicts a third position of the glint relative to the pupil edge based on the first position and the second position. In an example, the next position of the glint and the next position of the pupil edge are predicted given their positions in the two images. The third position can be derived as the predicted relative position between the glint and the pupil edge based on their predicted positions.
At operation 914, the eye tracking system determines, from the predetermined set, an illuminator that corresponds to the glint. In an example, the illuminators in the predetermined set form an illuminator pattern. The glints detected in the first image or in the second image, or the averaging of the glints between the two images (e.g., if the glint moved between the two images between the two positions, an average of the two positions is used for that glint; similar average positions can also be used for the remaining glints) form a glint pattern. The eye tracking system may match the illuminator pattern and the glint pattern to identify the correspondences between the illuminators from the predetermined set and the glints and, accordingly, identify the particular illuminator that corresponds to the glint.
At operation 916, the eye tracking system determines whether to power off the illuminator based on the third position. In an example, the third position is used to determine a predicted distance between the predicted next position of the glint and the predicted next position of the pupil edge. This predicted distance may be compared to a threshold distance, as illustrated in the first example 810 of
At operation 918, the eye tracking system has determined that the illuminator corresponding to the glint should be powered off. The eye tracking system removes an identifier of the illuminator from the predetermined set based on the determination to power it off. In this way, the eye tracking system may track which illuminator(s) should be powered off.
At operation 920, the eye tracking system powers off the illuminator based on the determination to power it off and/or the removal of its identifier from the predetermined set. In an example, the controller of the eye tracking system controls the power state of the illuminator and switches the power off from the one or more power sources to the illuminator. This operation can be performed individually for the illuminator. Alternatively, the eye tracking system may determine whether other illuminators should be powered off and other ones should be powered on, as further described in connection with the next operations, to then collectively control the power to these illuminators.
At operation 922, the eye tracking system determines whether the remaining set of powered-on illuminators is acceptable. In an example, the eye tracking system determines the remaining set from the predetermined set upon removal of the identifier of the illuminator and, optionally, the removal of identifiers of other illuminators that should also be powered off. The eye tracking system may use different factors to determine whether this remaining set is acceptable. These factors include, for instance, a size of the remaining set (e.g., the total number of remaining illuminators; if that total number is less than the minimum number, a determination is made that the remaining set is unacceptable), a distribution of illuminators from the remaining set (e.g., distances between the illuminators from the remaining set; if the distance between two adjacent remaining illuminators is too large, a determination is made that the remaining set is unacceptable), or a pattern of the illuminators (e.g., if the pattern forms a geometric shape complex to match to a glint pattern or the pattern cannot be accurately matched to the glint pattern, a determination is made that the remaining set is unacceptable). If the determination is that the remaining set is acceptable, operations 924 can follow operation 922. Otherwise, operation 926 can follow operation 922.
At operation 924, the eye tracking system generates a third image. In an example, this image is generated with the illuminator powered off, such that the glint would no longer be shown in the third image. In addition, if other illuminators from the predetermined set are powered off, their glints will also no longer be shown. Conversely, if additional illuminators are powered on (to adjust the predetermined set, per operation 926), their glints would be shown in the third image. The example flow would be repeated to continue the tracking of the glints and the controlling of the illuminators.
At operation 926, the eye tracking system adjusts the predetermined set. In an example, beyond removing identifiers of illuminators that should be powered off, identifiers of other illuminators can be added to the predetermined set. Once the adjustment is complete, the controller of the eye tracking system controls the power state of the illuminators in the adjusted set and, as applicable, switches the power on from the one or more power sources to these illuminators (and switching the power off to the illuminators not identified in the adjusted set). In an example, the eye tracking system determines additional illuminators that were excluded from the predetermined set of illuminators, selects one or more of these additional illuminators, adds their identifiers to the adjusted set, and powers them on. An additional illuminator can be selected based on at least one or more of a distance or a direction between the first position and the second position of the glint. For instance, if the glint moved to the right in the second image relative to the first image, the additional illuminator is selected be the one on the left of the illuminator (that corresponds to the glint and that was determined to be powered off), such that the new glint corresponding to the additional illuminator will likely fall inside the pupil. In another example, rather than adjusting the predetermined set, this set is emptied and the factors described herein above in connection with operation 902 are re-used to select a new predetermined set. In this example, any illuminator determined that it should be powered off can be excluded from the selection of the new predetermined set.
The disclosure has now been described in detail for the purposes of clarity and understanding. However, it will be appreciated that certain changes and modifications may be practiced within the scope of the appended claims.
The above description provides exemplary embodiments only, and is not intended to limit the scope, applicability or configuration of the disclosure. Rather, the above description of the exemplary embodiments will provide those skilled in the art with an enabling description for implementing one or more exemplary embodiments. It being understood that various changes may be made in the function and arrangement of elements without departing from the spirit and scope of the disclosure as set forth herein.
For example, any detail discussed with regard to one embodiment may or may not be present in all contemplated versions of that embodiment. Likewise, any detail discussed with regard to one embodiment may or may not be present in all contemplated versions of other embodiments discussed herein. Finally, the absence of discussion of any detail with regard to embodiment herein shall be an implicit recognition that such detail may or may not be present in any version of any embodiment discussed herein.
Specific details are given in the following description to provide a thorough understanding of the embodiments. However, it will be understood by one of ordinary skill in the art that the embodiments may be practiced without these specific details. For example, circuits, systems, networks, processes, and other elements in the disclosure may be shown as components in block diagram form in order not to obscure the embodiments in unnecessary detail. In other instances, well-known circuits, processes, algorithms, structures, and techniques may be shown without unnecessary detail in order to avoid obscuring the embodiments.
Also, it is noted that individual embodiments may be described as a process which is depicted as a flowchart, a flow diagram, a data flow diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be re-arranged. A process may be terminated when its operations are completed, but could have additional steps not discussed or included in a figure. Furthermore, not all operations in any particularly described process may occur in all embodiments. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
The term “machine-readable medium” includes, but is not limited to, transitory and non-transitory, portable or fixed storage devices, optical storage devices, wireless channels and various other mediums capable of storing, containing or carrying instruction(s) and/or data. A code segment or machine-executable instructions may represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a class, or any combination of instructions, data structures, or program statements. A code segment may be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, or memory contents. Information, arguments, parameters, data, etc. may be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, token passing, network transmission, etc.
Furthermore, embodiments of the disclosure may be implemented, at least in part, either manually or automatically. Manual or automatic implementations may be executed, or at least assisted, using machines, hardware, software, firmware, middleware, microcode, hardware description languages, or any combination thereof. When implemented in software, firmware, middleware or microcode, the program code or code segments to perform the necessary tasks may be stored in a machine-readable medium. A processor or processors may perform the necessary tasks.
As used herein, the phrase “a first thing based on a second thing,” and the like, may mean that the first thing is based solely on the second thing, or that the first thing is based on the second thing as well as one or more additional things.
Number | Name | Date | Kind |
---|---|---|---|
5844544 | Kahn et al. | Dec 1998 | A |
6204828 | Amir | Mar 2001 | B1 |
6351273 | Lemelson et al. | Feb 2002 | B1 |
7113170 | Lauper | Sep 2006 | B2 |
7164423 | Westen | Jan 2007 | B1 |
7952535 | Watanabe et al. | May 2011 | B2 |
8228315 | Starner et al. | Jul 2012 | B1 |
8386928 | Asch et al. | Feb 2013 | B1 |
8643951 | Wheeler et al. | Feb 2014 | B1 |
8793620 | Stafford | Jul 2014 | B2 |
8971570 | Raffle et al. | Mar 2015 | B1 |
9041787 | Andersson et al. | May 2015 | B2 |
9069164 | Starner et al. | Jun 2015 | B2 |
9201578 | Scott et al. | Dec 2015 | B2 |
9442567 | Scott et al. | Sep 2016 | B2 |
9507418 | Yu et al. | Nov 2016 | B2 |
9619020 | George-Svahn | Apr 2017 | B2 |
9778814 | Ambrus et al. | Oct 2017 | B2 |
9804671 | Taguchi et al. | Oct 2017 | B2 |
9830513 | Gustafsson et al. | Nov 2017 | B2 |
9958941 | Gustafsson et al. | May 2018 | B2 |
9977960 | Gustafsson et al. | May 2018 | B2 |
20010034218 | Hudecek et al. | Oct 2001 | A1 |
20020175897 | Pelosi | Nov 2002 | A1 |
20040108971 | Waldern et al. | Jun 2004 | A1 |
20040252277 | Chielewski et al. | Dec 2004 | A1 |
20050243054 | Beymer et al. | Nov 2005 | A1 |
20060069458 | Lee et al. | Mar 2006 | A1 |
20100220291 | Horning | Sep 2010 | A1 |
20100309432 | Suzuki et al. | Dec 2010 | A1 |
20100328444 | Blixt et al. | Dec 2010 | A1 |
20110032365 | Yett | Feb 2011 | A1 |
20110150334 | Du | Jun 2011 | A1 |
20110175932 | Yu et al. | Jul 2011 | A1 |
20120188148 | DeJong | Jul 2012 | A1 |
20120272179 | Stafford | Oct 2012 | A1 |
20120274745 | Russell | Nov 2012 | A1 |
20130016070 | Starner et al. | Jan 2013 | A1 |
20130021373 | Vaught et al. | Jan 2013 | A1 |
20130050432 | Perez et al. | Feb 2013 | A1 |
20130050833 | Lewis et al. | Feb 2013 | A1 |
20130083003 | Perez | Apr 2013 | A1 |
20130114850 | Publicover et al. | May 2013 | A1 |
20130169560 | Cederlund et al. | Jul 2013 | A1 |
20130228804 | Yan | Sep 2013 | A1 |
20130300636 | Cunningham et al. | Nov 2013 | A1 |
20130328762 | McCulloch et al. | Dec 2013 | A1 |
20140184550 | Hennessey et al. | Jul 2014 | A1 |
20140247232 | George-Svahn et al. | Sep 2014 | A1 |
20140300538 | Rijnders | Oct 2014 | A1 |
20140306891 | Latta et al. | Oct 2014 | A1 |
20140347623 | Inoue et al. | Nov 2014 | A1 |
20150002542 | Chan et al. | Jan 2015 | A1 |
20150042558 | Massonneau et al. | Feb 2015 | A1 |
20150061996 | Gustafsson et al. | Mar 2015 | A1 |
20150062322 | Gustafsson et al. | Mar 2015 | A1 |
20150070654 | Brown, Jr. et al. | Mar 2015 | A1 |
20150103018 | Kamin-Lyndgaard et al. | Apr 2015 | A1 |
20150130714 | Onuki | May 2015 | A1 |
20150130740 | Cederlund et al. | May 2015 | A1 |
20150138073 | Hennelly | May 2015 | A1 |
20150138079 | Lannsjö | May 2015 | A1 |
20150138244 | George-Svahn et al. | May 2015 | A1 |
20150143293 | George-Svahn et al. | May 2015 | A1 |
20150149956 | Kempinski et al. | May 2015 | A1 |
20150193018 | Venable et al. | Jul 2015 | A1 |
20150205494 | Scott et al. | Jul 2015 | A1 |
20150219899 | Mack et al. | Aug 2015 | A1 |
20150261008 | Fujii et al. | Sep 2015 | A1 |
20150268799 | Starner et al. | Sep 2015 | A1 |
20150278576 | Horesh | Oct 2015 | A1 |
20150301592 | Miller | Oct 2015 | A1 |
20150331240 | Poulos et al. | Nov 2015 | A1 |
20150340223 | Wunderer | Nov 2015 | A1 |
20150346810 | Urbach | Dec 2015 | A1 |
20160048204 | Scott et al. | Feb 2016 | A1 |
20160081547 | Gramatikov et al. | Mar 2016 | A1 |
20160085286 | Zhou | Mar 2016 | A1 |
20160179336 | Ambrus et al. | Jun 2016 | A1 |
20160189430 | Kuehne | Jun 2016 | A1 |
20160191576 | Thompson et al. | Jun 2016 | A1 |
20160320625 | Von et al. | Nov 2016 | A1 |
20170075420 | Yu et al. | Mar 2017 | A1 |
20170090562 | Gustafsson et al. | Mar 2017 | A1 |
20170090563 | Gustafsson et al. | Mar 2017 | A1 |
20170090564 | Gustafsson et al. | Mar 2017 | A1 |
20170090566 | George-Svahn et al. | Mar 2017 | A1 |
20170091549 | Gustafsson et al. | Mar 2017 | A1 |
20170177078 | Henderek et al. | Jun 2017 | A1 |
20170177079 | George-Svahn et al. | Jun 2017 | A1 |
20170316264 | Gustafsson et al. | Nov 2017 | A1 |
20190042842 | Cavin | Feb 2019 | A1 |
Number | Date | Country |
---|---|---|
2670482 | Dec 2013 | EP |
3353631 | Aug 2018 | EP |
3353632 | Aug 2018 | EP |
2010017407 | Jan 2010 | JP |
2006111744 | Oct 2006 | WO |
2017053966 | Mar 2017 | WO |
2017053971 | Mar 2017 | WO |
2017053972 | Mar 2017 | WO |
2017053974 | Mar 2017 | WO |
Entry |
---|
Daugman, J.G., “High Confidence Visual Recognition of persons by a Test of Statistical Independence”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 15, No. 11, Nov. 1993, pp. 1148-1161, 14 pages. |
Guestrin, et al., “General Theory of Remote Gaze Estimation Using the Pupil Center and Corneal Reflections”, Biomedical Engineering, IEEE Transactions, vol. 53, No. 6, pp. 1124-1133, Jun. 2006, 9 pages. |
Hamamatsu, Profile Sensor S9132 High speed frame rate sensor capable of acquiring two-dimensional projection data, Jan. 2014, pp. 1-10. |
Hillaire, et al., “Using an Eye-Tracking System to Improve Camera Motions and Depth-of-Field Blur Effects in Virtual Environments,” 2008, Virtual Reality Conference, 4 pages. |
Karitans et al., Method for Compensation of Eye Movements in Adaptive Optics, Latvian Journal of Physics and Technical Sciences vol. 47, No. 3, Jan. 1, 2010, pp. 51-56. |
Littlewort-Ford, et al., “Are your eyes smiling? Detecting Genuine Smiles with Support Vector Machines and Gabor Wavelets,” Proceedings of the 8th Joint Symposium on Neural Computation, 2001, 9 pages. |
Non-Final Office Action dated Dec. 13, 2018 in related U.S. Appl. No. 16/162,223, 15 pgs. |
Notice of Allowance dated Apr. 1, 2019 in related U.S. Appl. No. 16/162,223, 12 pgs. |
Final Office Action dated Feb. 7, 2019 in related U.S. Appl. No. 15/276,618, 14 pgs. |
Non-Final Office Action dated May 17, 2019 in related U.S. Appl. No. 15/983,772. |
Final Office Action dated May 30, 2019 in related U.S. Appl. No. 15/651,976, 18 pgs. |
Notice of Decision to Grant dated Apr. 1, 2020 in related foreign application No. KR 10-2018-7011089, all pgs. |
Non-Final Office Action dated Oct. 1, 2019 in related U.S. Appl. No. 16/097,588, all pgs. |
Office Action dated Oct. 22, 2019 in related foreign application No. KR 10-2018-7011109, all pgs. |
Office Action dated Oct. 22, 2019 in related foreign application No. KR 10-2018-7011089, all pgs. |
Office Action dated Oct. 22, 2019 in related foreign application No. KR-10-2018-7011088, all pgs. |
Notice of Allowance dated Nov. 20, 2019 in related U.S. Appl. No. 15/926,722, all pgs. |
International Preliminary Report on Patentability dated Jan. 30, 2020 in related application No. PCT/US2018/042443, all pgs. |
Number | Date | Country | |
---|---|---|---|
20200166994 A1 | May 2020 | US |