Embodiments of the invention relate to adjusting sensitivity of a camera to light that it receives from a scene that the camera images.
Time of flight (TOF) three-dimensional (3D) cameras determine distances to features in a scene by acquiring an image, conventionally referred to as a “range image”, of the scene that can be processed to determine how long it takes light to travel from the camera to the features and back to the camera. The round trip flight times of the light to, and back from the features, and the speed of light are used to determine the distances to the imaged features.
In a gated TOF-3D camera, light pulses are transmitted to illuminate a scene and the camera is shuttered on, “gated on”, following each light pulse for a relatively short exposure period to enable pixels in a photosensor of the camera to receive and register light from the light pulses reflected by features in the scene that reaches the camera. A pixel in the photosensor registers an amount of incident light it receives by generating, and accumulating, a quantity of electrons, hereinafter “photoelectrons”, in the pixel that is substantially proportional to the amount of incident light. The accumulated photoelectrons are transmitted to a readout pixel comprised in the photosensor. A readout circuit converts a charge provided by the photoelectrons transferred to the readout pixel to a voltage, hereinafter also “photo-voltage”, which provides a measure of the charge, and thereby the number of photoelectrons and the corresponding amount of incident light. A process of converting photoelectrons in a pixel to a photo-voltage is referred to as “reading” the pixel.
In order to provide reliable measurements of amount of light, photosensors are typically provided with a readout pixel that can accumulate a number of photoelectrons before it saturates that is equal to about twice a number of photoelectrons that saturates a pixel of the photosensor. A pixel saturates when a number of photoelectrons generated or received by the pixel is greater than the pixel's capacity to store photoelectrons. Photoelectrons in excess of a saturation threshold drain off to ground and cannot contribute to measurements of light.
Determination of a quantity of photoelectrons accumulated in a pixel by reading the pixel is subject to error from shot noise and read noise. Shot noise is a statistical error that is a function of the number of photoelectrons accumulated. If Ne is the number of photoelectrons, shot noise may be estimated to result in a fractional error (fractional error is % error multiplied by 100) in Ne equal to about 1/✓Ne. Read noise is independent of the number of photoelectrons. Read noise is a “device” noise that is introduced into a determination of a number of photoelectrons by the process of determining the number. If read noise is represented by a constant RE a fractional error in Ne may be estimated to be equal to about RE/Ne.
Distance to a feature in the scene imaged on a pixel of the camera photosurface is determined as a function of photo-voltage representing an amount of light that reaches the pixel from the feature during the short exposure periods, normalized to a photo-voltage representing a total amount of light reflected from the pulses that reaches the pixel from the feature. For convenience, light registered during the short exposure periods of the camera is referred to as “gated” light. An amount of light used to normalize gated light to determine distance is received during exposure periods that are substantially longer than the short exposure periods used to provide gated light, and is referred to as “ungated light”.
Typically, widths of the light pulses and the short exposure periods are very small, and may for example be less than or about equal to 20 nsec (nanoseconds). As a result, amounts of gated light registered by pixels in a gated TOF-3D camera are often relatively small, and errors in measurements of gated light, and consequently in distance measurements provided by the TOF-3D camera due to shot and read noise can be relatively large. For features of a scene that have low reflectivity and/or are relatively far from the camera, amounts of light from the features registered by pixels on which the features are imaged can be so small that shot noise error in measurements of the light generates unacceptably large errors in distance measurements to the features. If camera parameters and/or lighting are adjusted to compensate for low reflectivity and/or for relatively large distances of features in the scene, near and/or high reflectivity features in the scene may be poorly imaged because light from these features saturates pixels in the camera.
As a result, for many applications of a TOF-3D camera, the camera does not have a dynamic range (DNR) for measuring incident light sufficiently large to acquire range images that provide acceptable distance measurements for both “dim” and “bright” features in a scene. A camera DNR is typically measured as a ratio between a maximum amount of light per unit area of the camera photosensor, which the camera measures with an acceptable precision, divided by a minimum amount of light per unit area of the photosensor, which the camera measures with an acceptable precision.
Light incident on a surface per second per unit area of the surface is referred to as irradiance. A maximum amount of light incident per unit area of a photosensor that a camera measures with an acceptable precision may therefore be referred to as a maximum integrated irradiance, (MaxIr). Similarly, a minimum amount of light incident per unit area of a photosensor that a camera measures with an acceptable precision may be referred to as a minimum integrated irradiance, (MinIr). A camera DNR may conveniently be defined by an expression DNR=MaxIr/MinIr.
Aspects of an embodiment of the invention relate to providing a camera characterized by an improved DNR for acquiring an image of a scene responsive to light that the camera receives from the scene.
In an embodiment of the invention the camera comprises a photosensor having light sensitive pixels tiled into super pixels onto which the received light is imaged. Tiling a photosensor refers to subdividing the photosensor into areas that group pixels in the photosurface for convenient or advantageous reading. Tiling is generally implemented by programming
Each super pixel into which the photosensor is tiled is partitioned into a plurality of non overlapping groups of pixels, which pixel groups together include all the pixels in the super pixel. Optionally, each pixel group in a same super pixel includes a different number of pixels. A controller transfers photoelectrons generated by light from the scene and accumulated in each pixel group to a readout pixel, and determines if a photo-voltage generated responsive to a quantity of the photoelectrons transferred from the pixel group to the readout pixel indicates that the quantity saturates the readout pixel. The controller provides an estimate of an amount of light incident on the super pixel responsive to photo-voltage generated from quantities of photoelectrons from pixel groups in the super pixel that do not appear to saturate the readout pixel.
As a result of processing photoelectrons generated and accumulated in pixels of a super pixel in accordance with an embodiment of the invention, for a same given percent accuracy of measurement, a super pixel comprising N pixels of the photosensor has a MinIr that is equal to about 1/N times that of a single pixel in the photosensor. The super pixel also has a MaxIr substantially equal to a MaxIr of the single pixel. A camera, that configures its photosensor to have super pixels and determines intensities of light registered by the super pixels in accordance with an embodiment of the invention therefore has a DNR that is about N times that of a camera, hereinafter also a “conventional camera”, which operates its photosensor without super pixels.
In an embodiment of the invention, the camera photo sensor is tiled to square, 4×4 super pixels, each comprising 16 pixels of the photosensor, and each partitioned into three pixel groups comprising respectively 1, 3, and 12 pixels. Each super pixel registers about 16 times an amount of light registered by a single pixel and when processed in accordance with an embodiment of the invention has a DNR that is about 16 times that of a conventional camera. Tiling a photosensor to 4×4 super pixels is referred to as operating the photosensor in a quarter-quarter video graphic array (QQVGA) mode. In some embodiments of the invention a camera comprises a photosensor tiled into square 2×2 super pixels and may be referred to as operating in a quarter video graphics array (QVGA) mode.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
Non-limiting examples of embodiments of the invention are described below with reference to figures attached hereto that are listed following this paragraph. Identical structures, elements or parts that appear in more than one figure are generally labeled with a same numeral in all the figures in which they appear. Dimensions of components and features shown in the figures are chosen for convenience and clarity of presentation and are not necessarily shown to scale.
In the following paragraphs of the detailed description, operation of a TOF-3D camera operating in a QQVGA mode for which the camera's photosensor is tiled into 4×4 super pixels, in accordance with an embodiment of the invention is discussed with reference to
TOF-3D camera 20 comprises a light source 30 controllable to radiate a train of relatively short light pulses schematically represented by a train 32 of pulses 33 to illuminate scene 100, a photosensor 50, and an optical system represented by a lens 40. Optical system 40 collects light that is reflected from light pulses 33 back to the camera by object 102 and images the collected light onto the photosensor. The optical system also comprises a shutter (not shown) for shuttering the camera open for relatively short exposure periods to acquire an image of scene 100 responsive to gated light, and optionally for relatively long exposure periods to acquire an image of the scene responsive to ungated light. A controller 42 controls the camera components.
Photosensor 50 comprises rows 61 and columns 62 of light sensitive pixels 60, a transfer column 63 of transfer pixels 64, a readout pixel 66 and readout circuitry 68 that converts charge provided by photoelectrons accumulated in the readout pixel to photo-voltage. In accordance with an embodiment of the invention photosensor 50 is tiled into rows 71 and columns 72 of super pixels 70. The super pixels are outlined by bold lines 73. By way of example, TOF camera 20 is assumed to be operating in a QQVGA mode and each super pixel is a square 4×4 super pixel comprising sixteen pixels 60. Each super pixel 70 is optionally divided into three pixel groups 81, 82 and 83 comprising respectively 1, 3 and 12 pixels 60. The different pixels groups are shaded differently to distinguish them. Optionally, pixel groups 81, 82 and 83 in different super pixels 70 are homologous.
Let columns 62 of pixels 60 in a same column 72 of super pixels 70 be referred to, left to right in
To acquire a range image of scene 100 and determine distances to regions of object 102 controller 42 controls camera 20 to image scene 100 responsive to gated light and an image of scene 100 responsive to ungated light. A distance to a region of object 102 imaged on a super pixel 70 is determined responsive to an amount of gated light registered by the super pixel normalized to an amount of ungated light registered by the pixel.
To image scene 100 responsive to gated light, controller 42 controls the shutter to shutter camera 20 open for a short exposure period following each pulse 33 during which exposure period pixels 60 in photosensor 50 register light reflected from the light pulses that reaches the camera and is imaged on the photosensor. Light reflected from light pulses 33 and imaged by optical system 40 on photosensor 50 is schematically represented by a train 132 of reflected light pulses 133 shown in dashed lines. An image of a region of object 102 imaged on a super pixel 70 responsive to reflected light pulses 133 registered by pixels 60 during the short exposure periods is schematically represented by a shaded circle 104 on the photosensor.
Following a last pulse 33 radiated by light source 30 to illuminate scene 100, the photosurface is “read” to determine a total amount of light, in the present example gated light, registered by each super pixel 70 during the short exposure periods following light pulses 33.
In accordance with an embodiment of the invention, photosensor 50 is read by reading super pixels in a first column 72 of super pixels, and then sequentially reading super pixels in the remaining columns 72. A first column 72 of super pixels 70 is on the left hand edge of photosensor 50.
To read super pixels 70 in a column 72 photoelectrons accumulated in pixels 60 of column C1 of the super pixel column are shifted into corresponding shift pixels 64 in shift column 63 by shifting the photoelectrons in all columns 62 of photosensor 50 one column 62 “left” toward the shift column (one column to the left in
After photoelectrons in a group 81 or 82 are shifted into readout pixel 66, readout circuit 68 converts the electric charge provided by the photoelectrons to a photo-voltage, thereby emptying the readout pixel and preparing it to receive photoelectrons from a next pixel group. Controller 42 stores the photo-voltage generated by readout circuit 68 for each pixel group for use later as discussed below.
Following reading of pixel groups 81 and 82 in column C1 of the super pixel column 70, controller 42 controls photosensor 50 to shift photoelectrons accumulated in pixel groups 83 into shift pixels 64 of shift column 63 by shifting photoelectrons in all columns 62 of pixels 60 three columns 62 left, towards the shift column.
In accordance with an embodiment of the invention, controller 42 then uses the stored photo-voltages to determine a number of photoelectrons accumulated by each super pixel in column 70 in accordance with a flow diagram 200 of an algorithm shown in
Let the photo-voltages determined for pixel groups 81, 82 and 83 for a given super pixel be represented by V81, V82, and V83 respectively. Let a number of photoelectrons accumulated by pixel groups 81, 82, and 83, corresponding to photo-voltages V81, V82, and V83 be represented by V81e, V82e, and V83e. Let a number of photoelectrons in a super pixel comprising the pixel groups be represented by SPe. It is noted that for a same light intensity incident on all pixel groups of the given super pixel, it is expected that V83=12V81 and that V82=3V81.
In a decision block 202 of algorithm 200, controller 42 checks to decide if V81 indicates that a number V81e of photoelectrons in response to which V81 was generated, indicates that the number V81e saturated readout pixel 66. In an embodiment of the invention, a photo-voltage generated by readout circuit 68 responsive to charge provided by photoelectrons in readout pixel 66 is assumed to saturate the readout pixel if it is about equal to a photo-voltage, VS, which is generated when the readout pixel is filled to capacity with photoelectrons. If V81 is equal to about VS, in symbols V81≈VS, the controller proceeds to a block 204. In block 204 the controller determines that that the super pixel is saturated and that a quantity, SPe, of photoelectrons accumulated by the super pixel cannot reliably be determined.
If on the other hand V81 does not saturate readout pixel 66, controller 42 proceeds to a decision block 206 to decide whether photo-voltage V82≈VS and therefore that photoelectrons transferred to readout pixel 66 saturated the readout pixel. If V82≈VS controller 42 proceeds to a block 208 and determines that a number of photoelectrons, SPe, accumulated by the super pixel during the short gates is about 16 times V81e, that is, SPe=16(V81e). If on the other hand, V82<VS, controller 42 proceeds to a decision block 210 to determine if V83≈VS. If V83≈VS controller 42 proceeds to a block 212 and determines that SPe=4(V81e+V82e). If on the other hand V83<VS, then the controller proceeds to a block 214 and determines that SPe=(V81e+V82e+V83e).
Under the stated assumptions that object 102 is relatively far from camera 20 and not characterized by a particularly high reflectivity, it is expected that each pixel 60 on which image 104 (
The sum (V81e+V82e+V83e) is relatively large compared to a number of photoelectrons accumulated in a single pixel 60. As a result, a distance measurement based on a number of photoelectrons accumulated by a super pixel 70 in accordance with an embodiment of the invention is substantially more accurate than a distance measurement based on single pixel 60.
For example, for the QQVGA operation of camera 20, it is reasonable to assume that a number of photoelectrons, SPe, determined for a super pixel 70 will, on the average, be about 16 times larger than a number of photoelectrons accumulated in a single pixel 60. A fractional error in SPe due to shot noise will therefore be about ¼ that of a determination of a number of photoelectrons in a single pixel 60. A fractional error in SPe due to read noise will be about ⅛ that of a fractional read noise error in a number of photoelectrons in a single pixel 60. As a result, a distance measurement determined from SPe for a super pixel in accordance with an embodiment of the invention, will be characterized by a substantially smaller error than a distance determined from a number of photoelectrons accumulated in a pixel 60.
Furthermore, if a minimum number photoelectrons is required to provide a quantity of photoelectrons having an acceptable magnitude of shot noise and read noise, TOF-3D camera 20 can provide the minimum number at an irradiance of its photosensor 50 that is about 16 time smaller than an irradiance required to provide the minimum number from a single pixel 60. As a result, TOF-3D camera 20 in accordance with an embodiment of the invention is characterized by a MinIr that is about 16 times less than a MinIr of a single pixel 60. In general, a camera operating a photosensor having super pixels comprising N pixels, in accordance with an embodiment of the invention, is characterized by an improved MinIr that is 1/N that of the MinIr of a conventional camera.
Whereas for the case of discussed above for which object 102 is located relatively far from camera 20, pixels 60 in general accumulate relatively small numbers of photoelectrons responsive to light pulses 133, the pixels imaging object 102 may accumulate substantially greater numbers of photoelectrons if the object is located relatively close to the camera. Irradiance of photosensor 50 from light pulses 133 increases inversely as the square of the distance of object 102 from camera 20.
For example, if object 102 is located about 3 m (meters) from TOF-3D camera 20 in
For the situation shown in
By determining a quantity of photoelectrons for a super pixel, such as a super pixel 70, in accordance with an embodiment of the invention, responsive to quantities of photoelectrons in pixels groups that do not saturate readout pixel 66, super pixels in photosensor 50 are characterized by a MaxIr that is substantially equal to a MaxIr of a single pixel 60. TOF-3D camera 20 therefore has a DNR=MaxIr/MinIr that is about 16 times that of a conventional camera. In general, a camera operating a photosensor having super pixels comprising N pixels in accordance with an embodiment of the invention, has a DNR that is about N times that of a conventional camera.
It is noted that were a photo-voltage generated for a super pixel having N pixels responsive to photoelectrons in a readout pixel only after all the photoelectrons in the pixels of a super pixel were transferred to the readout pixel, the super pixel would have a MaxIr less than that provided by a single pixel. For example, if a quantity of Se photoelectrons saturates a pixel of the photosensor, and a quantity of SRe photoelectrons saturates the readout pixel, a MaxIr for the super pixel would be less than about (SRe/SPe)/N times a MaxIr for a single pixel. A camera operating such super pixels would have a DNR limited to no more than (SRe/SPe) times that of a conventional camera. If, as is common, Se were equal to about twice a number of photoelectrons that saturates a pixel in the photosurface, then a MaxIr available from the super pixel would be equal to about 2/N that of a single pixel and a camera operating the super pixels would have a DNR not more than twice that of a conventional camera.
Whereas in the above description embodiments of the invention refer to a TOF-3D camera, practice of the invention is not limited to TOF-3D cameras but may be used in any type of camera comprising pixels to increase a DNR of the camera. For example, super pixels and methods of determining amounts of light registered by the super pixels in accordance with an embodiment of the invention, may be used to provide an increased DNR for cameras that provide contrast images, that is pictures, of a scene.
In the description and claims of the present application, each of the verbs, “comprise” “include” and “have”, and conjugates thereof, are used to indicate that the object or objects of the verb are not necessarily a complete listing of components, elements or parts of the subject or subjects of the verb.
Descriptions of embodiments of the invention in the present application are provided by way of example and are not intended to limit the scope of the invention. The described embodiments comprise different features, not all of which are required in all embodiments of the invention. Some embodiments utilize only some of the features or possible combinations of the features. Variations of embodiments of the invention that are described, and embodiments of the invention comprising different combinations of features noted in the described embodiments, will occur to persons of the art. The scope of the invention is limited only by the claims.
Number | Name | Date | Kind |
---|---|---|---|
4627620 | Yang | Dec 1986 | A |
4630910 | Ross et al. | Dec 1986 | A |
4645458 | Williams | Feb 1987 | A |
4695953 | Blair et al. | Sep 1987 | A |
4702475 | Elstein et al. | Oct 1987 | A |
4711543 | Blair et al. | Dec 1987 | A |
4751642 | Silva et al. | Jun 1988 | A |
4796997 | Svetkoff et al. | Jan 1989 | A |
4809065 | Harris et al. | Feb 1989 | A |
4817950 | Goo | Apr 1989 | A |
4843568 | Krueger et al. | Jun 1989 | A |
4893183 | Nayar | Jan 1990 | A |
4901362 | Terzian | Feb 1990 | A |
4925189 | Braeunig | May 1990 | A |
5101444 | Wilson et al. | Mar 1992 | A |
5148154 | MacKay et al. | Sep 1992 | A |
5184295 | Mann | Feb 1993 | A |
5229754 | Aoki et al. | Jul 1993 | A |
5229756 | Kosugi et al. | Jul 1993 | A |
5239463 | Blair et al. | Aug 1993 | A |
5239464 | Blair et al. | Aug 1993 | A |
5288078 | Capper et al. | Feb 1994 | A |
5295491 | Gevins | Mar 1994 | A |
5320538 | Baum | Jun 1994 | A |
5347306 | Nitta | Sep 1994 | A |
5385519 | Hsu et al. | Jan 1995 | A |
5405152 | Katanics et al. | Apr 1995 | A |
5417210 | Funda et al. | May 1995 | A |
5423554 | Davis | Jun 1995 | A |
5454043 | Freeman | Sep 1995 | A |
5469740 | French et al. | Nov 1995 | A |
5495576 | Ritchey | Feb 1996 | A |
5516105 | Eisenbrey et al. | May 1996 | A |
5524637 | Erickson | Jun 1996 | A |
5534917 | MacDougall | Jul 1996 | A |
5563988 | Maes et al. | Oct 1996 | A |
5577981 | Jarvik | Nov 1996 | A |
5580249 | Jacobsen et al. | Dec 1996 | A |
5594469 | Freeman et al. | Jan 1997 | A |
5597309 | Riess | Jan 1997 | A |
5616078 | Oh | Apr 1997 | A |
5617312 | Iura et al. | Apr 1997 | A |
5638300 | Johnson | Jun 1997 | A |
5641288 | Zaenglein | Jun 1997 | A |
5682196 | Freeman | Oct 1997 | A |
5682229 | Wangler | Oct 1997 | A |
5690582 | Ulrich et al. | Nov 1997 | A |
5703367 | Hashimoto et al. | Dec 1997 | A |
5704837 | Iwasaki et al. | Jan 1998 | A |
5715834 | Bergamasco et al. | Feb 1998 | A |
5875108 | Hoffberg et al. | Feb 1999 | A |
5877803 | Wee et al. | Mar 1999 | A |
5913727 | Ahdoot | Jun 1999 | A |
5933125 | Fernie | Aug 1999 | A |
5980256 | Carmein | Nov 1999 | A |
5989157 | Walton | Nov 1999 | A |
5995649 | Marugame | Nov 1999 | A |
6005548 | Latypov et al. | Dec 1999 | A |
6009210 | Kang | Dec 1999 | A |
6054991 | Crane et al. | Apr 2000 | A |
6066075 | Poulton | May 2000 | A |
6072494 | Nguyen | Jun 2000 | A |
6073489 | French et al. | Jun 2000 | A |
6077201 | Cheng et al. | Jun 2000 | A |
6098458 | French et al. | Aug 2000 | A |
6100896 | Strohecker et al. | Aug 2000 | A |
6101289 | Kellner | Aug 2000 | A |
6128003 | Smith et al. | Oct 2000 | A |
6130677 | Kunz | Oct 2000 | A |
6141463 | Covell et al. | Oct 2000 | A |
6147678 | Kumar et al. | Nov 2000 | A |
6152856 | Studor et al. | Nov 2000 | A |
6159100 | Smith | Dec 2000 | A |
6173066 | Peurach et al. | Jan 2001 | B1 |
6181343 | Lyons | Jan 2001 | B1 |
6188777 | Darrell et al. | Feb 2001 | B1 |
6215890 | Matsuo et al. | Apr 2001 | B1 |
6215898 | Woodfill et al. | Apr 2001 | B1 |
6226396 | Marugame | May 2001 | B1 |
6229913 | Nayar et al. | May 2001 | B1 |
6256033 | Nguyen | Jul 2001 | B1 |
6256400 | Takata et al. | Jul 2001 | B1 |
6283860 | Lyons et al. | Sep 2001 | B1 |
6289112 | Jain et al. | Sep 2001 | B1 |
6299308 | Voronka et al. | Oct 2001 | B1 |
6308565 | French et al. | Oct 2001 | B1 |
6316934 | Amorai-Moriya et al. | Nov 2001 | B1 |
6363160 | Bradski et al. | Mar 2002 | B1 |
6384819 | Hunter | May 2002 | B1 |
6411744 | Edwards | Jun 2002 | B1 |
6430997 | French et al. | Aug 2002 | B1 |
6476834 | Doval et al. | Nov 2002 | B1 |
6496598 | Harman | Dec 2002 | B1 |
6503195 | Keller et al. | Jan 2003 | B1 |
6539931 | Trajkovic et al. | Apr 2003 | B2 |
6570555 | Prevost et al. | May 2003 | B1 |
6633294 | Rosenthal et al. | Oct 2003 | B1 |
6640202 | Dietz et al. | Oct 2003 | B1 |
6661918 | Gordon et al. | Dec 2003 | B1 |
6681031 | Cohen et al. | Jan 2004 | B2 |
6714665 | Hanna et al. | Mar 2004 | B1 |
6731799 | Sun et al. | May 2004 | B1 |
6738066 | Nguyen | May 2004 | B1 |
6765726 | French et al. | Jul 2004 | B2 |
6788809 | Grzeszczuk et al. | Sep 2004 | B1 |
6801637 | Voronka et al. | Oct 2004 | B2 |
6873723 | Aucsmith et al. | Mar 2005 | B1 |
6876496 | French et al. | Apr 2005 | B2 |
6900834 | Henderson et al. | May 2005 | B2 |
6937742 | Roberts et al. | Aug 2005 | B2 |
6950534 | Cohen et al. | Sep 2005 | B2 |
7003134 | Covell et al. | Feb 2006 | B1 |
7036094 | Cohen et al. | Apr 2006 | B1 |
7038855 | French et al. | May 2006 | B2 |
7039676 | Day et al. | May 2006 | B1 |
7042440 | Pryor et al. | May 2006 | B2 |
7050606 | Paul et al. | May 2006 | B2 |
7058204 | Hildreth et al. | Jun 2006 | B2 |
7060957 | Lange et al. | Jun 2006 | B2 |
7084973 | Ingram | Aug 2006 | B1 |
7113918 | Ahmad et al. | Sep 2006 | B1 |
7121946 | Paul et al. | Oct 2006 | B2 |
7170492 | Bell | Jan 2007 | B2 |
7184048 | Hunter | Feb 2007 | B2 |
7202898 | Braun et al. | Apr 2007 | B1 |
7222078 | Abelow | May 2007 | B2 |
7227526 | Hildreth et al. | Jun 2007 | B2 |
7259747 | Bell | Aug 2007 | B2 |
7308112 | Fujimura et al. | Dec 2007 | B2 |
7317836 | Fujimura et al. | Jan 2008 | B2 |
7348963 | Bell | Mar 2008 | B2 |
7359121 | French et al. | Apr 2008 | B2 |
7367887 | Watabe et al. | May 2008 | B2 |
7379563 | Shamaie | May 2008 | B2 |
7379566 | Hildreth | May 2008 | B2 |
7389591 | Jaiswal et al. | Jun 2008 | B2 |
7412077 | Li et al. | Aug 2008 | B2 |
7421093 | Hildreth et al. | Sep 2008 | B2 |
7430312 | Gu | Sep 2008 | B2 |
7436496 | Kawahito | Oct 2008 | B2 |
7450736 | Yang et al. | Nov 2008 | B2 |
7452275 | Kuraishi | Nov 2008 | B2 |
7460690 | Cohen et al. | Dec 2008 | B2 |
7489812 | Fox et al. | Feb 2009 | B2 |
7536032 | Bell | May 2009 | B2 |
7555142 | Hildreth et al. | Jun 2009 | B2 |
7560701 | Oggier et al. | Jul 2009 | B2 |
7570805 | Gu | Aug 2009 | B2 |
7574020 | Shamaie | Aug 2009 | B2 |
7576727 | Bell | Aug 2009 | B2 |
7590262 | Fujimura et al. | Sep 2009 | B2 |
7593552 | Higaki et al. | Sep 2009 | B2 |
7598942 | Underkoffler et al. | Oct 2009 | B2 |
7607509 | Schmiz et al. | Oct 2009 | B2 |
7620202 | Fujimura et al. | Nov 2009 | B2 |
7668340 | Cohen et al. | Feb 2010 | B2 |
7680298 | Roberts et al. | Mar 2010 | B2 |
7683954 | Ichikawa et al. | Mar 2010 | B2 |
7684592 | Paul et al. | Mar 2010 | B2 |
7701439 | Hillis et al. | Apr 2010 | B2 |
7702130 | Im et al. | Apr 2010 | B2 |
7704135 | Harrison, Jr. | Apr 2010 | B2 |
7710391 | Bell et al. | May 2010 | B2 |
7729530 | Antonov et al. | Jun 2010 | B2 |
7746345 | Hunter | Jun 2010 | B2 |
7760182 | Ahmad et al. | Jul 2010 | B2 |
7809167 | Bell | Oct 2010 | B2 |
7834846 | Bell | Nov 2010 | B1 |
7852262 | Namineni et al. | Dec 2010 | B2 |
RE42256 | Edwards | Mar 2011 | E |
7898522 | Hildreth et al. | Mar 2011 | B2 |
8035612 | Bell et al. | Oct 2011 | B2 |
8035614 | Bell et al. | Oct 2011 | B2 |
8035624 | Bell et al. | Oct 2011 | B2 |
8072470 | Marks | Dec 2011 | B2 |
20040008267 | Chen et al. | Jan 2004 | A1 |
20040079893 | Dietz et al. | Apr 2004 | A1 |
20050133687 | Bock | Jun 2005 | A1 |
20080024614 | Li et al. | Jan 2008 | A1 |
20080026838 | Dunstan et al. | Jan 2008 | A1 |
20080036893 | De Godzinsky | Feb 2008 | A1 |
20090030630 | Eitan et al. | Jan 2009 | A1 |
20100066849 | Lim et al. | Mar 2010 | A1 |
20100128109 | Banks | May 2010 | A1 |
20110052095 | Deever | Mar 2011 | A1 |
Number | Date | Country |
---|---|---|
101254344 | Jun 2010 | CN |
0583061 | Feb 1994 | EP |
08044490 | Feb 1996 | JP |
9310708 | Jun 1993 | WO |
9717598 | May 1997 | WO |
9944698 | Sep 1999 | WO |
Entry |
---|
Kanade et al., “A Stereo Machine for Video-rate Dense Depth Mapping and Its New Applications”, IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1996, pp. 196-202,The Robotics Institute, Carnegie Mellon University, Pittsburgh, PA. |
Miyagawa et al., “CCD-Based Range Finding Sensor”, Oct. 1997, pp. 1648-1652, vol. 44 No. 10, IEEE Transactions on Electron Devices. |
Rosenhahn et al., “Automatic Human Model Generation”, 2005, pp. 41-48, University of Auckland (CITR), New Zealand. |
Aggarwal et al., “Human Motion Analysis: A Review”, IEEE Nonrigid and Articulated Motion Workshop, 1997, University of Texas at Austin, Austin, TX. |
Shao et al., “An Open System Architecture for a Multimedia and Multimodal User Interface”, Aug. 24, 1998, Japanese Society for Rehabilitation of Persons with Disabilities (JSRPD), Japan. |
Kohler, “Special Topics of Gesture Recognition Applied in Intelligent Home Environments”, In Proceedings of the Gesture Workshop, 1998, pp. 285-296, Germany. |
Kohler, “Vision Based Remote Control in Intelligent Home Environments”, University of Erlangen-Nuremberg/Germany, 1996, pp. 147-154, Germany. |
Kohler, “Technical Details and Ergonomical Aspects of Gesture Recognition applied in Intelligent Home Environments”, 1997, Germany. |
Hasegawa et al., “Human-Scale Haptic Interaction with a Reactive Virtual Human in a Real-Time Physics Simulator”, Jul. 2006, vol. 4, No. 3, Article 6C, ACM Computers in Entertainment, New York, NY. |
Qian et al., “A Gesture-Driven Multimodal Interactive Dance System”, Jun. 2004, pp. 1579-1582, IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan. |
Zhao, “Dressed Human Modeling, Detection, and Parts Localization”, 2001, The Robotics Institute, Carnegie Mellon University, Pittsburgh, PA. |
He, “Generation of Human Body Models”, Apr. 2005, University of Auckland, New Zealand. |
Isard et al., “Condensation—Conditional Density Propagation for Visual Tracking”, 1998, pp. 5-28, International Journal of Computer Vision 29(1), Netherlands. |
Livingston, “Vision-based Tracking with Dynamic Structured Light for Video See-through Augmented Reality”, 1998, University of North Carolina at Chapel Hill, North Carolina, USA. |
Wren et al., “Pfinder: Real-Time Tracking of the Human Body”, MIT Media Laboratory Perceptual Computing Section Technical Report No. 353, Jul. 1997, vol. 19, No. 7, pp. 780-785, IEEE Transactions on Pattern Analysis and Machine Intelligence, Caimbridge, MA. |
Breen et al., “Interactive Occlusion and Collusion of Real and Virtual Objects in Augmented Reality”, Technical Report ECRC-95-02, 1995, European Computer-Industry Research Center GmbH, Munich, Germany. |
Freeman et al., “Television Control by Hand Gestures”, Dec. 1994, Mitsubishi Electric Research Laboratories, TR94-24, Caimbridge, MA. |
Hongo et al., “Focus of Attention for Face and Hand Gesture Recognition Using Multiple Cameras”, Mar. 2000, pp. 156-161, 4th IEEE International Conference on Automatic Face and Gesture Recognition, Grenoble, France. |
Pavlovic et al., “Visual Interpretation of Hand Gestures for Human-Computer Interaction: A Review”, Jul. 1997, pp. 677-695, vol. 19, No. 7, IEEE Transactions on Pattern Analysis and Machine Intelligence. |
Azarbayejani et al., “Visually Controlled Graphics”, Jun. 1993, vol. 15, No. 6, IEEE Transactions on Pattern Analysis and Machine Intelligence. |
Granieri et al., “Simulating Humans in VR”, The British Computer Society, Oct. 1994, Academic Press. |
Brogan et al., “Dynamically Simulated Characters in Virtual Environments”, Sep./Oct. 1998, pp. 2-13, vol. 18, Issue 5, IEEE Computer Graphics and Applications. |
Fisher et al., “Virtual Environment Display System”, ACM Workshop on Interactive 3D Graphics, Oct. 1986, Chapel Hill, NC. |
“Virtual High Anxiety”, Tech Update, Aug. 1995, pp. 22. |
Sheridan et al., “Virtual Reality Check”, Technology Review, Oct. 1993, pp. 22-28, vol. 96, No. 7. |
Stevens, “Flights into Virtual Reality Treating Real-World Disorders”, The Washington Post, Mar. 27, 1995, Science Psychology, 2 pages. |
“Simulation and Training”, 1994, Division Incorporated. |
High Dynamic Range Imaging: Spatially Varying Pixel Exposures; Nayar et al.; Published Date: 2000 http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=855857. |
Adaptive Dynamic Range Imaging: Optical Control of Pixel Exposures Over Space and Time; Nayar et al.; Published Date: 2003; Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003). |
Number | Date | Country | |
---|---|---|---|
20120287242 A1 | Nov 2012 | US |