Separation of electrical and optical components

Information

  • Patent Grant
  • 8508919
  • Patent Number
    8,508,919
  • Date Filed
    Monday, September 14, 2009
    15 years ago
  • Date Issued
    Tuesday, August 13, 2013
    11 years ago
Abstract
Embodiments related to establishing and maintaining precision alignment in an optical system are disclosed. For example, one disclosed embodiment provides an optical device comprising an outer casing and a rigid optical support disposed within an interior of the outer casing and to which a plurality of optical components are mounted. The embodiment further comprises a printed circuit board spatially separated from the optical support and the plurality of optical components, wherein one or more electrical components are mounted to the printed circuit board. The embodiment also comprises one or more electrical conductors electrically connecting the one or more electrical components to the one or more optical components.
Description
BACKGROUND

Various optical systems utilize multiple optical elements that are spaced from one another and that cooperate to produce a desired optical signal. For example, a structured light depth-sensing system utilizes a projector configured to produce a structured light pattern, and a camera configured to acquire an image of the structured light pattern as projected onto a target. A computing device may compare the acquired image to a reference image to determine a distance of the target from the camera based upon differences between the structured light pattern in the acquired image and the reference image.


Maintenance of precision alignment between optical components may help to ensure that such precision optical systems operate in a consistently reliable manner. However, current manufacturing methods used to make many optical systems may make it difficult to establish and maintain such precision alignment in a cost-effective manner.


SUMMARY

Accordingly, various embodiments are disclosed herein related to establishing and maintaining precision alignment in an optical system. For example, one disclosed embodiment provides an optical device comprising an outer casing and an optical support disposed within an interior of the outer casing and to which a plurality of optical components are mounted. The embodiment further comprises a printed circuit board spatially separated from the optical support and the plurality of optical components, wherein one or more electrical components are mounted to the printed circuit board. Further, one or more electrical conductors electrically connect the one or more electrical components to the one or more optical components.


This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. Furthermore, the claimed subject matter is not limited to implementations that solve any or all disadvantages noted in any part of this disclosure.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 shows an example embodiment of an optical device in the form of a depth sensing camera system.



FIG. 2 shows a schematic depiction of embodiments of an optical support connected to a printed circuit board via a plurality of flexible electrical connectors.



FIG. 3 shows a flow diagram depicting an embodiment of a method for assembling an optical device.





DETAILED DESCRIPTION

As described above, the establishment and maintenance of precision alignment between optical components in an optical device may be difficult to achieve in a cost-effective manner. For example, in many optical devices, optical components are mounted directly to a printed circuit board on which the electronic components of the device, such as controllers, etc., are also mounted. Such printed circuit boards may be constructed with relatively loose tolerances, such that locations of electrical components and mounting structures for optical components may vary more than desired. Relatively time-consuming and expensive optical alignment processes may be used to compensate for this. Further, printed circuit boards may not be mechanically robust, and therefore may be prone to undesirable degrees of thermal expansion, warping, and other mechanical deformations. Additionally, the presence of the electrical components on the circuit board may contribute to dimensional variation caused by thermal expansion, as varying amounts of heat output by the electrical components and associated heat sinks may cause the dimensions of the printed circuit board, and therefore the relative positions of optical components mounted to the printed circuit board, to vary with time. These factors may contribute to higher cost, lower optical yield, and lower operating performance.


Accordingly, embodiments are disclosed herein that relate to the separation of optical components and electrical components in an optical device. An example embodiment of an optical device is shown in FIG. 1 as a structured light depth-sensing camera 100. The depth-sensing camera 100 is shown in the context of a computer gaming system 102 that may be used to play a variety of different games, play one or more different media types, and/or control or manipulate non-game applications. FIG. 1 also shows a display in the form of a television 104 that may be used to present game visuals to game players, such as game player 108.


The depth-sensing camera 100 may be used in combination with software on the gaming system 102 to track one or more targets, such as game player 108, in the field of view of the depth-sensing camera 100, by comparing images of the targets taken at different times to detect motion. Gaming system 102 may then display a response to the motion on the television 104. FIG. 1 shows a scenario in which the game player 108 is tracked using the depth-sensing camera 100 so that the movements of the game player 18 may be interpreted by the gaming system 102 as controls that can be used to affect the game being executed by the gaming system 102. In other words, the game player 108 may use his movements to control the game.


The example scenario illustrated in FIG. 1 shows the game player 108 playing a boxing game that is being executed by gaming system 102. The gaming system 102 uses the television 104 to visually present a boxing opponent 110 to game player 108. Furthermore, the gaming system 102 uses the television 104 to visually present a player avatar 112 that the game player 108 controls with his movements. In one example scenario, the game player 108 can throw a punch in physical space as an instruction for the player avatar 112 to throw a punch in game space. The gaming system 102 and depth-sensing camera 100 can be used to recognize and analyze the punch of the game player 108 in physical space so that the punch can be interpreted as a game control that causes the player avatar 112 to throw a punch in game space. Likewise, other movements by the game player 108 may be interpreted as other controls, such as controls to bob, weave, shuffle, block, jab, or throw a variety of different power punches. Furthermore, some movements may be interpreted into controls that serve purposes other than controlling the player avatar 112. For example, the player may use movements to end, pause, or save a game, select a level, view high scores, communicate with a friend, etc. It will be understood that the use environment of FIG. 1 is shown for the purpose of example, and that a structured light depth-sensing camera may be used in any other suitable use environment. It will also be understood that an optical system according to the present disclosure may be used in any other suitable optical device than a structured light depth-sensing camera, including stereo depth-sensing cameras and other cameras.


Continuing with FIG. 1, the depicted depth-sensing camera 100 comprises a plurality of optical components, including a structured light projector 120, an infrared camera 122, and optionally a visible camera 124. Each of these optical components is connected to various electronics, such as a controller configured to control these components and to receive and process images from the infrared camera 122 and visible camera 124. Establishing and maintaining precision alignment of the structured light projector 120, infrared camera 122, and visible camera 124 may help to ensure proper and consistent performance by the depth-sensing camera 100. However, as mentioned above, where these optical components are mounted to the printed circuit board or printed circuit boards to which their respective electronic circuitry is mounted, such precision alignment may be difficult to establish and maintain, and may require difficult manual alignment steps during manufacturing.


Therefore, FIG. 2 shows an embodiment of an optical support 200 and separate printed circuit board 202 to which these optical components and the associated electrical and components are respectively mounted. The optical support 200 is configured to be mounted within a device outer casing (illustrated schematically by dashed line 206) in a location that is separated from the printed circuit board. In this manner, heat that is generated and dissipated by the electrical and thermal components mounted to the printed circuit board 202 may have a substantially lesser effect on the locations of the structured light projector 120, infrared camera 122, and visible camera 124 relative to one another due to the separation between the optical support 200 and the printed circuit board 202.


The use of the optical support 200 and printed circuit board 202 for separating optical and electrical components also may help to simplify manufacturing of an optical device. For example, because a printed circuit board is often designed with the mounting of electrical components in mind, the mounting of optical components to a printed circuit board may be a secondary consideration in the design of the printed circuit board, thereby leading to design compromises. Further, where spatially separated optical components (e.g. projector and camera) are mounted to different printed circuit boards, proper alignment may be difficult to achieve, and may utilize difficult and time-consuming manual alignment processes. Further, manufacturing steps such as making electrical connections between the electrical and optical components may potentially harm the alignment of the optical components.


On the other hand, the use of optical support 200 for mounting plural optical components in an optical device allows alignment of the optical components to be referenced to a common optical structure whose purpose is to maintain that alignment during manufacture and operation. To further help maintain optical alignment during manufacture, connections between electrical components 220 on the printed circuit board 202 and the optical components mounted to the optical support 200 may be made via one or more flexible conductors 230, thereby helping to ensure that optical alignment is not compromised by the conductors 230, or the process of installing and electrically connecting the conductors 230. Likewise, thermal functions also may be made secondarily to ensure that optical alignment is not compromised. Further, thermal components may interface with the optical support 200 in such a manner that the components are thermally insulated from the optical support 200 and/or are connected via interfaces that allow the components to “float” relative to the optical support. In the depicted embodiment, a heat sink 232 is shown extending from structured light projector 120, wherein the heat sink 232 interfaces with the optical support 200 via a gasket 234. Other thermal components that are associated with the electrical components 220, such as heat sink 236, may be mounted to printed circuit board 202, and therefore are spatially separated from the optical support 200, thereby reducing the effect on the optical support 200 of heat dissipated from these components. While the depicted embodiment shows a single printed circuit board coupled to a single optical support, it will be understood that other embodiments may utilize a plurality of printed circuit boards each connected to optical components on the optical support via one or more connectors. Further, some embodiments may comprise a plurality of optical supports each comprising two or more optical components.


The optical support 200 may have any suitable construction. For example, the optical support may be configured to have a relatively low coefficient of thermal expansion and high dimensional stability so that the relative alignment and positions of the optical components mounted to the optical support 200 remain stable during changes in temperature. In one example embodiment, the optical support 200 may take the form of a plate-like structure. Further, the optical support 200 may be made from any suitable material or materials, including but not limited to various metals, ceramics, composites, polymers, etc. In some embodiments, the optical support has a rigid construction to help to maintain alignment of components mounted to the optical support.



FIG. 3 shows an embodiment of a method 300 of assembling an optical device in which optical components are mounted to an optical support that is separated from a printed circuit board on which some electrical components are mounted. Method 300 comprises, at 302, mounting a plurality of optical components to an optical support. As described above, the optical support allows alignment of the optical components to be referenced to a common optical structure whose purpose is to maintain that alignment during manufacture and operation. Any suitable optical components may be mounted to the optical support. For example, in the specific embodiment of a structured light depth-sensing camera, such optical components may include a projector 308 and a first camera 304, such as an infrared camera. Further, in some embodiments, a second camera 306 also may be mounted to the optical support. The second camera may be, for example, a visible camera, a second camera for a stereo depth-sensing system, etc. It will be understood that these specific optical components are described for the purpose of example, and are not intended to be limiting in any manner.


Method 300 next comprises, at 310, positioning the optical support in an outer casing, and then at 312, fixing the position of the optical support within the outer casing so that the optical components are mounted in the casing. Likewise, method 300 also comprises, at 314, positioning a printed circuit board in the outer casing. In some embodiments, a plurality of printed circuit boards may be positioned in the outer casing, as shown at 316. Then, the position of the printed circuit board (or board) is fixed at location spaced from the optical support, as discussed above. It will be understood that the optical support and the printed circuit board may be placed and fixed in the outer container in any suitable order in a manufacturing process.


Next, as indicated at 322, method 300 comprises connecting one or more optical components to the printed circuit board via electrical conductors. In some embodiments, the conductors may be flexible, as indicated at 324. In this manner, the installation and connection of electrical components may not impact the optical components of the device to the same extent as where the optical components are mounted to a printed circuit board along with electronic components. This may help to avoid secondary alignment, fixturing and/or adjustment of optical components after mounting and device assembly, as the optical components are referenced to the optical support to which the components are designed and tolerance to.


It will be understood that the configurations and/or approaches for separating optical and electrical components in an optical device described herein are exemplary in nature, and that these specific embodiments or examples are not to be considered in a limiting sense, because numerous variations are possible. The specific routines or methods described herein may represent one or more of any number of processing strategies. As such, various acts illustrated may be performed in the sequence illustrated, in other sequences, in parallel, or in some cases omitted. Likewise, the order of any of the above-described processes is not necessarily required to achieve the features and/or results of the embodiments described herein, but is provided for ease of illustration and description.


The subject matter of the present disclosure includes all novel and nonobvious combinations and subcombinations of the various processes, systems and configurations, and other features, functions, acts, and/or properties disclosed herein, as well as any and all equivalents thereof.

Claims
  • 1. An optical device, comprising: an outer casing;an optical support fixed to a position within an interior of the outer casing;a plurality of optical components mounted to the optical support;a heat sink thermally coupled to a first optical component of the plurality of optical components, wherein at least a portion of the heat sink extends through the optical support;a gasket configured to thermally insulate the portion of the heat sink extending through the optical support;a printed circuit board separate from the optical support and fixed to a position in the outer casing that is spatially separated from the optical support and the plurality of optical components;one or more electrical components mounted to the printed circuit board; andone or more electrical conductors electrically connecting the one or more electrical components to the plurality of optical components.
  • 2. The optical device of claim 1, wherein the optical device comprises a depth-sensing camera.
  • 3. The optical device of claim 2, wherein the plurality of optical components comprises a projector, a first camera, and a second camera, and wherein the projector, the first camera, and the second camera are each mounted to the optical support.
  • 4. The optical device of claim 3, wherein the projector comprises a laser and a diffractive optical element.
  • 5. The optical device of claim 1, wherein the optical support is formed from one or more of a metal material, a ceramic material, and a polymer material.
  • 6. The optical device of claim 1, wherein the one or more of the electrical components comprises a heat sink.
  • 7. The optical device of claim 1, wherein the one or more electrical conductors are flexible.
  • 8. The optical device of claim 1, further comprising a plurality of printed circuit boards.
  • 9. A depth-sensing camera, comprising: an optical system comprising: a rigid optical support,a projector mounted to the optical support,a first camera mounted to the optical support,a second camera mounted to the optical support, anda heat sink at least partially extending through the optical support and thermally coupled to the projector, the heat sink coupled to the optical support via a gasket configured to thermally insulate the heat sink from the optical support; andan electrical/thermal system comprising: a printed circuit board separate from the optical support and spatially separated from the optical support,one or more electronic components mounted to the printed circuit board, andone or more thermal components mounted to the printed circuit board system and thermally floating relative to the optical support; anda plurality of electrical conductors electrically connecting the projector, the first camera, and the second camera respectively to the printed circuit board.
  • 10. The depth-sensing camera of claim 9, wherein the optical support is formed from one or more of a metal, ceramic, composite, and polymer material.
  • 11. The depth-sensing camera of claim 9, wherein the plurality of electrical conductors are flexible.
  • 12. The depth-sensing camera of claim 9, further comprising a plurality of printed circuit boards, wherein each printed circuit board comprises at least some of the one or more electronic components coupled to the optical system.
  • 13. The depth-sensing camera of claim 9, wherein the first camera is an infrared camera, and wherein the second camera is a visible camera.
  • 14. A method of assembling an optical device, the optical device comprising a rigid optical support, a printed circuit board, and an outer casing, the method comprising: mounting a plurality of optical components to the optical support, wherein one or more optical components of the plurality of optical components comprises a heat sink at least partially extending through the optical support, the heat sink coupled to the optical support via a gasket configured to allow the heat sink to float relative to the optical support and to thermally insulate the heat sink from the optical support;positioning the optical support in the outer casing and fixing a position of the optical support relative to the outer casing;positioning the printed circuit board in the outer casing and fixing a position of the printed circuit board in a location that is spaced from the optical support; andconnecting one or more of the plurality of optical components to the printed circuit board via one or more electrical conductors.
  • 15. The method of claim 14, wherein the device is a depth-sensing camera.
  • 16. The method of claim 15, wherein mounting a plurality of optical components to an optical support comprises mounting a projector, a first camera, and a second camera to the optical support.
  • 17. The method of claim 14, wherein fixing the printed circuit board in the outer casing in a location that is spaced from the optical support comprises fixing the printed circuit board in the outer casing in a location that is spaced from thermal components that conduct heat away from the printed circuit board.
  • 18. The method of claim 14, wherein connecting the one or more optical components to the printed circuit board comprises connecting the one or more optical components to the printed circuit board via one or more flexible connectors.
  • 19. The method of claim 14, further comprising fixing a plurality of printed circuit boards in a location spaced from the optical support, and connecting the plurality of printed circuit boards to the plurality of optical components via a plurality of electrical conductors.
US Referenced Citations (208)
Number Name Date Kind
4627620 Yang Dec 1986 A
4630910 Ross et al. Dec 1986 A
4645458 Williams Feb 1987 A
4695953 Blair et al. Sep 1987 A
4702475 Elstein et al. Oct 1987 A
4711543 Blair et al. Dec 1987 A
4751642 Silva et al. Jun 1988 A
4796997 Svetkoff et al. Jan 1989 A
4809065 Harris et al. Feb 1989 A
4817950 Goo Apr 1989 A
4843568 Krueger et al. Jun 1989 A
4893183 Nayar Jan 1990 A
4901362 Terzian Feb 1990 A
4925189 Braeunig May 1990 A
5101444 Wilson et al. Mar 1992 A
5148154 MacKay et al. Sep 1992 A
5184295 Mann Feb 1993 A
5229754 Aoki et al. Jul 1993 A
5229756 Kosugi et al. Jul 1993 A
5239463 Blair et al. Aug 1993 A
5239464 Blair et al. Aug 1993 A
5288078 Capper et al. Feb 1994 A
5295491 Gevins Mar 1994 A
5320538 Baum Jun 1994 A
5347306 Nitta Sep 1994 A
5385519 Hsu et al. Jan 1995 A
5405152 Katanics et al. Apr 1995 A
5417210 Funda et al. May 1995 A
5423554 Davis Jun 1995 A
5454043 Freeman Sep 1995 A
5469740 French et al. Nov 1995 A
5495576 Ritchey Feb 1996 A
5516105 Eisenbrey et al. May 1996 A
5524637 Erickson et al. Jun 1996 A
5534917 MacDougall Jul 1996 A
5563988 Maes et al. Oct 1996 A
5577981 Jarvik Nov 1996 A
5580249 Jacobsen et al. Dec 1996 A
5594469 Freeman et al. Jan 1997 A
5597309 Riess Jan 1997 A
5616078 Oh Apr 1997 A
5617312 Iura et al. Apr 1997 A
5638300 Johnson Jun 1997 A
5641288 Zaenglein Jun 1997 A
5682196 Freeman Oct 1997 A
5682229 Wangler Oct 1997 A
5690582 Ulrich et al. Nov 1997 A
5703367 Hashimoto et al. Dec 1997 A
5704837 Iwasaki et al. Jan 1998 A
5715834 Bergamasco et al. Feb 1998 A
5875108 Hoffberg et al. Feb 1999 A
5877803 Wee et al. Mar 1999 A
5913727 Ahdoot Jun 1999 A
5933125 Fernie Aug 1999 A
5980256 Carmein Nov 1999 A
5989157 Walton Nov 1999 A
5995649 Marugame Nov 1999 A
6005548 Latypov et al. Dec 1999 A
6009210 Kang Dec 1999 A
6054991 Crane et al. Apr 2000 A
6066075 Poulton May 2000 A
6072494 Nguyen Jun 2000 A
6073489 French et al. Jun 2000 A
6077201 Cheng et al. Jun 2000 A
6098458 French et al. Aug 2000 A
6100896 Strohecker et al. Aug 2000 A
6101289 Kellner Aug 2000 A
6128003 Smith et al. Oct 2000 A
6130677 Kunz Oct 2000 A
6141463 Covell et al. Oct 2000 A
6147678 Kumar et al. Nov 2000 A
6152856 Studor et al. Nov 2000 A
6159100 Smith Dec 2000 A
6173066 Peurach et al. Jan 2001 B1
6181343 Lyons Jan 2001 B1
6188777 Darrell et al. Feb 2001 B1
6215890 Matsuo et al. Apr 2001 B1
6215898 Woodfill et al. Apr 2001 B1
6226396 Marugame May 2001 B1
6229913 Nayar et al. May 2001 B1
6256033 Nguyen Jul 2001 B1
6256400 Takata et al. Jul 2001 B1
6283860 Lyons et al. Sep 2001 B1
6289112 Jain et al. Sep 2001 B1
6299308 Voronka et al. Oct 2001 B1
6308565 French et al. Oct 2001 B1
6316934 Amorai-Moriya et al. Nov 2001 B1
6363160 Bradski et al. Mar 2002 B1
6384819 Hunter May 2002 B1
6411744 Edwards Jun 2002 B1
6430997 French et al. Aug 2002 B1
6476834 Doval et al. Nov 2002 B1
6496598 Harman Dec 2002 B1
6503195 Keller et al. Jan 2003 B1
6539931 Trajkovic et al. Apr 2003 B2
6549288 Migdal et al. Apr 2003 B1
6570555 Prevost et al. May 2003 B1
6633294 Rosenthal et al. Oct 2003 B1
6640202 Dietz et al. Oct 2003 B1
6661918 Gordon et al. Dec 2003 B1
6681031 Cohen et al. Jan 2004 B2
6714665 Hanna et al. Mar 2004 B1
6731799 Sun et al. May 2004 B1
6738066 Nguyen May 2004 B1
6765726 French et al. Jul 2004 B2
6788809 Grzeszczuk et al. Sep 2004 B1
6801637 Voronka et al. Oct 2004 B2
6873723 Aucsmith et al. Mar 2005 B1
6876496 French et al. Apr 2005 B2
6886948 Nakano May 2005 B2
6894902 Chang May 2005 B2
6930725 Hayashi Aug 2005 B1
6937742 Roberts et al. Aug 2005 B2
6950534 Cohen et al. Sep 2005 B2
7003134 Covell et al. Feb 2006 B1
7036094 Cohen et al. Apr 2006 B1
7038855 French et al. May 2006 B2
7039676 Day et al. May 2006 B1
7042440 Pryor et al. May 2006 B2
7050606 Paul et al. May 2006 B2
7058204 Hildreth et al. Jun 2006 B2
7060957 Lange et al. Jun 2006 B2
7113918 Ahmad et al. Sep 2006 B1
7121946 Paul et al. Oct 2006 B2
7146083 Carr Dec 2006 B2
7156526 Tanaka Jan 2007 B2
7170492 Bell Jan 2007 B2
7184048 Hunter Feb 2007 B2
7202898 Braun et al. Apr 2007 B1
7214128 Kriesel May 2007 B2
7222078 Abelow May 2007 B2
7227526 Hildreth et al. Jun 2007 B2
7259747 Bell Aug 2007 B2
7287866 Yamamoto et al. Oct 2007 B2
7308112 Fujimura et al. Dec 2007 B2
7317836 Fujimura et al. Jan 2008 B2
7348963 Bell Mar 2008 B2
7349604 Clark Mar 2008 B2
7359121 French et al. Apr 2008 B2
7367887 Watabe et al. May 2008 B2
7369334 Case et al. May 2008 B2
7379563 Shamaie May 2008 B2
7379566 Hildreth May 2008 B2
7389591 Jaiswal et al. Jun 2008 B2
7391409 Zalewski et al. Jun 2008 B2
7408643 Kimba et al. Aug 2008 B2
7412077 Li et al. Aug 2008 B2
7421093 Hildreth et al. Sep 2008 B2
7430312 Gu Sep 2008 B2
7436496 Kawahito Oct 2008 B2
7450736 Yang et al. Nov 2008 B2
7452275 Kuraishi Nov 2008 B2
7460690 Cohen et al. Dec 2008 B2
7489812 Fox et al. Feb 2009 B2
7505111 Hirukawa et al. Mar 2009 B2
7536032 Bell May 2009 B2
7555142 Hildreth et al. Jun 2009 B2
7560701 Oggier et al. Jul 2009 B2
7570805 Gu Aug 2009 B2
7574020 Shamaie Aug 2009 B2
7576727 Bell Aug 2009 B2
7590262 Fujimura et al. Sep 2009 B2
7593552 Higaki et al. Sep 2009 B2
7598942 Underkoffler et al. Oct 2009 B2
7607509 Schmiz et al. Oct 2009 B2
7620202 Fujimura et al. Nov 2009 B2
7627139 Marks et al. Dec 2009 B2
7646372 Marks et al. Jan 2010 B2
7668340 Cohen et al. Feb 2010 B2
7680298 Roberts et al. Mar 2010 B2
7683954 Ichikawa et al. Mar 2010 B2
7684592 Paul et al. Mar 2010 B2
7701439 Hillis et al. Apr 2010 B2
7702130 Im et al. Apr 2010 B2
7704135 Harrison, Jr. Apr 2010 B2
7710391 Bell et al. May 2010 B2
7729530 Antonov et al. Jun 2010 B2
7746345 Hunter Jun 2010 B2
7760182 Ahmad et al. Jul 2010 B2
7782297 Zalewski et al. Aug 2010 B2
7809167 Bell Oct 2010 B2
7834846 Bell Nov 2010 B1
7852262 Namineni et al. Dec 2010 B2
7874917 Marks et al. Jan 2011 B2
RE42256 Edwards Mar 2011 E
7898522 Hildreth et al. Mar 2011 B2
7980700 Hayashi Jul 2011 B2
7986321 Zhuang et al. Jul 2011 B2
7995834 Knighton et al. Aug 2011 B1
7997736 Yoshimura et al. Aug 2011 B2
8035612 Bell et al. Oct 2011 B2
8035614 Bell et al. Oct 2011 B2
8035624 Bell et al. Oct 2011 B2
8072470 Marks Dec 2011 B2
20020037668 Tseng et al. Mar 2002 A1
20060158522 Pryor Jul 2006 A1
20060264258 Zalewski et al. Nov 2006 A1
20060291719 Ikeda et al. Dec 2006 A1
20070021208 Mao et al. Jan 2007 A1
20070023716 van der Burgt et al. Feb 2007 A1
20070195435 Theriault et al. Aug 2007 A1
20080024999 Huang Jan 2008 A1
20080026838 Dunstan et al. Jan 2008 A1
20080056561 Sawachi Mar 2008 A1
20080156619 Patel et al. Jul 2008 A1
20090016642 Hart Jan 2009 A1
20090169095 Zhuang et al. Jul 2009 A1
20090279056 Belliveau et al. Nov 2009 A1
Foreign Referenced Citations (7)
Number Date Country
201254344 Jun 2010 CN
0583061 Feb 1994 EP
0801757 Oct 2004 EP
08044490 Feb 1996 JP
9310708 Jun 1993 WO
9717598 May 1997 WO
9944698 Sep 1999 WO
Non-Patent Literature Citations (29)
Entry
Levy, et al., “A Concept for Zero-Alignment Micro Optical Systems”, retrieved at <<http://www.osti.gov/bridge/servlets/purl/14031-w1KnhU/webviewable/14031.pdf>>, Proc. SPIE vol. 3879, Micromachine Technology for Diffractive and Holographic Optics, pp. 167-173.
“Mounting & Positioning Equipment”, retrieved at <<http://www.dataoptics.com/mountposit.htm>>, Jul. 21, 2009, pp. 2.
Kanade et al., “A Stereo Machine for Video-rate Dense Depth Mapping and Its New Applications”, IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1996, pp. 196-202,The Robotics Institute, Carnegie Mellon University, Pittsburgh, PA.
Miyagawa et al., “CCD-Based Range Finding Sensor”, Oct. 1997, pp. 1648-1652, vol. 44 No. 10, IEEE Transactions on Electron Devices.
Rosenhahn et al., “Automatic Human Model Generation”, 2005, pp. 41-48, University of Auckland (CITR), New Zealand.
Aggarwal et al., “Human Motion Analysis: A Review”, IEEE Nonrigid and Articulated Motion Workshop, 1997, University of Texas at Austin, Austin, TX.
Shao et al., “An Open System Architecture for a Multimedia and Multimodal User Interface”, Aug. 24, 1998, Japanese Society for Rehabilitation of Persons with Disabilities (JSRPD), Japan.
Kohler, “Special Topics of Gesture Recognition Applied in Intelligent Home Environments”, In Proceedings of the Gesture Workshop, 1998, pp. 285-296, Germany.
Kohler, “Vision Based Remote Control in Intelligent Home Environments”, University of Erlangen-Nuremberg/Germany, 1996, pp. 147-154, Germany.
Kohler, “Technical Details and Ergonomical Aspects of Gesture Recognition applied in Intelligent Home Environments”, 1997, Germany.
Hasegawa et al., “Human-Scale Haptic Interaction with a Reactive Virtual Human in a Real-Time Physics Simulator”, Jul. 2006, vol. 4, No. 3, Article 6C, ACM Computers in Entertainment, New York, NY.
Qian et al., “A Gesture-Driven Multimodal Interactive Dance System”, Jun. 2004, pp. 1579-1582, IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan.
Zhao, “Dressed Human Modeling, Detection, and Parts Localization”, 2001, The Robotics Institute, Carnegie Mellon University, Pittsburgh, PA.
He, “Generation of Human Body Models”, Apr. 2005, University of Auckland, New Zealand.
Isard et al., “Condensation—Conditional Density Propagation for Visual Tracking”, 1998, pp. 5-28, International Journal of Computer Vision 29(1), Netherlands.
Livingston, “Vision-based Tracking with Dynamic Structured Light for Video See-through Augmented Reality”, 1998, University of North Carolina at Chapel Hill, North Carolina, USA.
Wren et al., “Pfinder: Real-Time Tracking of the Human Body”, MIT Media Laboratory Perceptual Computing Section Technical Report No. 353, Jul. 1997, vol. 19, No. 7, pp. 780-785, IEEE Transactions on Pattern Analysis and Machine Intelligence, Caimbridge, MA.
Breen et al., “Interactive Occlusion and Collusion of Real and Virtual Objects in Augmented Reality”, Technical Report ECRC-95-02, 1995, European Computer-Industry Research Center GmbH, Munich, Germany.
Freeman et al., “Television Control by Hand Gestures”, Dec. 1994, Mitsubishi Electric Research Laboratories, TR94-24, Caimbridge, MA.
Hongo et al., “Focus of Attention for Face and Hand Gesture Recognition Using Multiple Cameras”, Mar. 2000, pp. 156-161, 4th IEEE International Conference on Automatic Face and Gesture Recognition, Grenoble, France.
Pavlovic et al., “Visual Interpretation of Hand Gestures for Human-Computer Interaction: A Review”, Jul. 1997, pp. 677-695, vol. 19, No. 7, IEEE Transactions on Pattern Analysis and Machine Intelligence.
Azarbayejani et al., “Visually Controlled Graphics”, Jun. 1993, Vo1.15, No. 6, IEEE Transactions on Pattern Analysis and Machine Intelligence.
Granieri et al., “Simulating Humans in VR”, The British Computer Society, Oct. 1994, Academic Press.
Brogan et al., “Dynamically Simulated Characters in Virtual Environments”, Sep./Oct. 1998, pp. 2-13, vol. 18, Issue 5, IEEE Computer Graphics and Applications.
Fisher et al., “Virtual Environment Display System”, ACM Workshop on Interactive 3D Graphics, Oct. 1986, Chapel Hill, NC.
“Virtual High Anxiety”, Tech Update, Aug. 1995, pp. 22.
Sheridan et al., “Virtual Reality Check”, Technology Review, Oct. 1993, pp. 22-28, vol. 96, No. 7.
Stevens, “Flights into Virtual Reality Treating Real World Disorders”, The Washington Post, Mar. 27, 1995, Science Psychology, 2 pages.
“Simulation and Training”, 1994, Division Incorporated.
Related Publications (1)
Number Date Country
20110064402 A1 Mar 2011 US