An augmented reality (AR) system includes hardware and software that typically provides a live, direct or indirect, view of a physical, real world environment whose elements are augmented by computer-generated sensory information, such as sound, video and/or graphics. For example, a head mounted display (HMD) may be used in an AR system. The HMD may have a display that uses an optical see-through lens to allow a computer generated image (CGI) to be superimposed on a real-world view.
A variety of single function input devices may be used in an AR system to captures input, experience or indicate user's intent. For example, tracking input devices, such a digital cameras, optical sensors, accelerometers and/or wireless sensors may provide user input. A tracking input device may be able to discern a user's intent based on the user's location and/or movement. One type of tracking input device may be a finger tracking input device that tracks a user's finger on a computer generated keyboard. Similarly, gesture recognition input devices may interpret a user's body movement by visual detection or from sensors embedded a peripheral device, such as a wand or stylus. Voice recognition input devices may also provide user input to an AR system.
A wrist-worn input device that is used in a AR system operates in three modes of operation. In a first mode of operation, the input device is curved so that it may be worn on a user's wrist. A touch surface receives letters gestured or selections by the user.
In a second mode of operation, the input device is flat and used as a touch surface for more complex single or multi-hand interactions. The input device includes one or more sensors to indicate the orientation of the flat input device, such as portrait, landscape, one handed or two handed. The input device may include a processor, memory and/or wireless transmitter to communicate with an AR system.
In a third mode of operation, the input device receives biometric input from one or more biometric sensors. The biometric input may provide contextual information while allowing the user to have their hands free. The biometric sensors may include heart rate monitors, blood/oxygen sensors, accelerometers and/or thermometers. The biometric mode of operation may operate concurrently with either the curved or flat mode of operation.
A sticker defining one or more locations on the touch surface that corresponds a user's input, such as a character, number or intended operation, may be affixed to the touch surface. The sticker may be interchanged with different stickers based on a mode of operation, user's preference and/or particular AR experience. The sticker may be customizable as well. A sticker may include a first adhesive surface to adhere to the touch surface and a second surface that provides a user-preferred keyboard and/or keypad layout with user preferred short cut keys.
In an embodiment, an input device comprises a touch surface that receives a touch input from a user. A member is coupled to the touch surface and is curved around a wrist of the user in a first mode of operation. The member is flat in a second mode of operation. A biometric sensor also receives a biometric input from the user. A transmitter outputs a signal that represents the touch and biometric inputs.
In another embodiment, an input device used to experience augmented reality comprises a member that may be curved or extended flat. A capacitive touch surface is coupled to the member and receives a touch input from the user. A sticker is coupled to the touch surface and defines one or more locations on the touch surface that corresponds to a user's input. A biometric sensor also receives biometric input from the user. A processor executes processor readable instructions stored in memory in response to the touch and biometric input.
In still another embodiment, an AR apparatus comprises an input device and computing device that provides an electronic signal representing augmented reality information. The input device includes a member that may be curved to be worn by the user or flat. A touch surface is coupled to the member and receives touch input from the user. A biometric sensor, such as a heart rate and/or blood/oxygen sensor, also receives biometric input from the user. A processor executes processor readable instructions stored in memory in response to the touch and biometric input. A wireless transmitter outputs a wireless signal that represents the touch and biometric input. The computing device then provides the electronic signal representing augmented reality information in response to the wireless signal.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
User input in AR systems has been approached from many different directions, often times requiring many different single-function devices to capture input. These devices accomplish their goal, but are optimized for use in a single scenario that does not span a variety of scenarios in a typical day of user activity. For example, a touch device may allow for great user input when a user's hands are free, but a touch device may becomes difficult to use when a user is carrying groceries or otherwise has their hands full. The present technology supports user input through a wide range of scenarios with at least three different input modalities that allow users to accomplish their daily goals while paying attention to social and physical/functional constraints.
A wearable input device 100 is capable of operating in at least three modes operation. In a first mode of operation, wearable input device 101 may be curved (or folded) so that in may be worn on user 100 as illustrated in
Wearable input device 100 may form an open curve (like the letter “C”) or closed curve (like the letter “O) in various curved modes of operation embodiments.
In a second mode of operation, wearable input device 101 may be flat and/or rigid, as illustrated in
In a third mode of operation, wearable input device 101 receives biometric information of user 100 from one or more biometric sensors in electronic components 107 positioned on the back of wearable input device 101. In alternate embodiments, one or more biometric sensors may be positioned in other locations of wearable input device 101. The biometric information may provide contextual information to a AR system while allowing user 100 to have their hands free. The biometric sensors may include heart rate sensors, blood/oxygen sensors, accelerometers, thermometers or other type of sensor that obtains biometric information from a user 100. The biometric information may identify muscle contractions of the arm and/or movement of the arm or other appendage of user 100.
In embodiments, wearable input device 101 may be in either a flat or curved mode of operation as well as a biometric mode of operation. In still a further embodiment, wearable input device 101 may be in a biometric mode of operation and not be able to receive touch input.
Wearable input device 101 includes a member 105 that enables wearable input device 101 to be positioned in a curved or flat mode of operation. A touch surface (or layer) 104 is then positioned on member 105 to receive user 100 inputs. Touch surface 104 may be flexible and glued to member 105 in embodiments. In an embodiment, a sticker 103 that identifies where a user 100 may contact touch surface 104 for predetermined inputs is adhered to touch surface 104.
In embodiments, member 105 includes a type of material or composite that enables wearable input device 101 to be curved or extended flat during different modes of operation. For example, member 105 may include a fabric, bendable plastic/foam and/or bendable metal/alloy. In other embodiments, member 105 may include a wire frame or mesh covered with a plastic sleeve or foam. In a flat mode of operation, member 105 may be rigid or flexible in embodiments. Similarly, in a curved mode of operation, member 105 may be rigid or flexible. In an embodiment, member 105 may be a mechanical mechanism having a plurality of rib members and overlapping scales that enable a curved and flat mode of operation as described herein.
Member 105 may have a variety of geometric shapes in embodiments. While
In an embodiment, a touch surface 104 is an electronic surface that can detect the presence and location of a touch within an area. A touch may be from a finger or hand of user 100 as well as from passive objects, such as a stylus.
In various embodiments, touch surface 104 includes different touch surface technologies for sensing a touch from a user 100. For example, different touch surface technologies include resistive, capacitive, surface acoustic wave, dispersive signal and acoustic pulse technologies. Different types of capacitive touch surface technologies include surface capacitive, projected capacitive, mutual capacitive and self-capacitive technologies.
In an embodiment, touch surface 104 includes a two-dimensional surface capacitive touch surface. In an embodiment, a surface capacitive touch surface is constructed by forming a conducting material or layer, such as copper or indium tin oxide, on an insulator. A small voltage is applied to the conducting layer to produce a uniform electrostatic field. When a conductor, such as a human finger, touches the uncoated surface of the insulator, a capacitor is dynamically formed. A controller and touch surface driver software in electronics components 107 then determines the location of the touch indirectly from the change in the capacitance as measured from one or more sensors at four corners of the touch surface 104 as illustrated in
In an embodiment, sticker 103 includes a first surface providing a key or user input layout and a second surface having adhesive to affix to touch surface 104. In alternate embodiments, sticker 103 (and/or touch surface 104) may include a different type of bonding mechanism (other than adhesive) in affixing a surface having a key or user input layout to touch surface 104. For example, sticker 103 may be bonded to touch surface 104 by using a static-cling type bond, molecular bond, magnetic outer rim and/or other type of bonding mechanism. Sticker 103 includes a key layout representing locations for a user 100 to touch on surface 104 so that a predetermined AR function may be initiated, a short cut initiated and/or character input. For example, sticker 103 includes “ON” and “OFF” keys as well as “AR 100” and “MonsterPet” keys. In an embodiment, sticker 103 also includes keypad 103a having alphanumeric characters. In embodiments, a user may customize sticker 103 for functions that are often used. For example, sticker 103 includes a “MonsterPet” key that identifies a location on touch surface 104 that after touching, would create an AR monster pet for viewing in a AR system as described herein.
A user may also remove and replace sticker 103 with another sticker that may be used in a different AR application. For example, sticker 103 may be replaced with sticker that has a more detailed keypad 103a having more characters when user 100 intends to create a text message to be sent to another.
In this example embodiment, there are nine elongated rib members. It will be appreciated that more or fewer ribs may be included in alternative embodiments. Each elongated rib member is longer across its longitudinal axis (i.e., across the width of wearable input device 101) than across its latitudinal axis (i.e., from the top of wearable input device 101 to the bottom of wearable input device 101). In the illustrated embodiment, each elongated rib member is at least four times longer across its longitudinal axis than across its latitudinal axis. However, other ratios may be used.
Each elongated rib member may have a generally trapezoidal cross section. As shown with reference to elongated rib member 22a, the generally trapezoidal cross section is bounded by a top face 34a; a bottom face 36a; a left side 38a between top face 34a and bottom face 36a; and a right side 40a between top face 34a and bottom face 36a. As shown, the top face 34a opposes the bottom face 36a and the left side 38a opposes the right side 40a.
Top face 34a has a width D1 and bottom face 36a has a width D2. D1 is greater than D2, thus giving elongated rib member 22a a generally trapezoidal cross section. However, it is to be understood that one or more elongated rib members may not have perfect trapezoidal cross sections. For example, top face 34a and/or bottom face 36a may be curved, non-planar surfaces. As another example, corners between faces and sides may include bevels and/or rounded edges. These and other variations from a true trapezoidal cross section are within the scope of this disclosure.
In some embodiments, the cross section of each elongated rib member may be substantially identical to the cross sections of all other elongated rib members. In some embodiments, at least one elongated rib member may have a different size and/or shape when compared to another elongated rib member. In general, the size, shape, and number of elongated rib members can be selected to achieve a desired curved mode of operation, as described below by way of example.
Turning back to
Turning back to
In contrast, the bottom faces of neighboring elongated rib members may be spaced farther apart than the top faces when wearable input device 101 is in the flat mode of operation. As an example, top face 34a is closer to top face 34b than bottom face 36a is to bottom face 36b as illustrated in
When in a flat mode of operation, gap 46 is characterized by an angle 48 with a magnitude M1. When in the curved mode of operation, angle 48 has a magnitude M2, which is less than M1. In some embodiments, including the illustrated embodiment, the gap may essentially close when wearable input device 101 is moved into the curved mode of operation (e.g., angle 48=0 degrees). Closing each gap between neighboring elongated rib members contributes to the overall curvature of member 105 in the curved mode of operation.
A fastener such as a rivet may attach neighboring overlapping scales to an elongated rib member. For example, a rivet may be fastened through holes 54a and 50b. Similarly, a rivet may be fastened through holes 56a and 52b. Such rivets may attach both overlapping scales to the same elongated rib member (e.g., elongated rib member 22g of
In such an arrangement, the fixed holes (e.g., hole 54a and hole 56a) may be sized to closely fit the rivet so that overlapping scale 28a does not slide relative to the elongated rib member. In contrast, the slotted holes (e.g., hole 50b and hole 52b) may be sized to allow fore and aft sliding relative to the elongated rib member. In this way, each overlapping scale can be fixed to one elongated rib member and may slide relative to another elongated rib members. As such, as the gaps between neighboring elongated rib members close as wearable input device 101 moves from a flat mode of operation to a curved mode of operation the overlapping scales are able to accommodate the changing length of the bottom of wearable input device 101 as the wearable input device 101 moves from the flat mode of operation to a curved mode of operation.
The bottom flexible support may slide between the holes and the rivets. Because the bottom flexible support is not attached to the elongated rib members, the bottom flexible support may also accommodate the changing length of the bottom of wearable input device 101 as wearable input device moves from the flat mode of operation to the curved mode of operation.
The top flexible support, the bottom flexible support, and the plurality of overlapping scales may be comprised of thin sheets of a metal, such as steel. In alternative embodiments, the flexible supports and/or scales may be comprised of any material that is suitably flexible, strong, and durable. In some embodiments, one or more of the top flexible support, the bottom flexible support, and the overlapping scales may be made from plastic.
The top flexible support 24 includes a left side row of holes and a right side row of holes that extend along a longitudinal axis of member 105. Each hole in the top flexible support may be complementary to a hole in the top face of an elongated rib member. The top flexible support may be attached to an elongated rib member at each pair of complementary holes. For example, a fastener, such as a rivet, may be used to attach the top flexible support to the elongated rib members at the complementary holes. In some embodiments, the top flexible support may be attached to elongated rib members via another suitable mechanism, such as via heat stakes and/or screws. Attaching each elongated rib member to the top flexible support at two separate locations may help limit the elongated rib members from twisting relative to one another.
An elongated rib member may include one or more projections configured to mate with complementary cavities in a neighboring elongated rib member. For example,
Turning back to
Latch 80 may be located within one or more elongated rib members and/or other portions of wearable input device 101.
Latch 80 is a magnetic latch in an embodiment. While a magnetic latch is provided as a nonlimiting example of a suitable latch, it is to be understood that other latches may be used without departing from the scope of this disclosure. In the illustrated embodiment, latch 80 includes a front magnetic partner 84 and a rear magnetic partner 86 that are each attached to top flexible support 24. Latch 80 also includes an intermediate magnetic partner 88 attached to bottom flexible support 26. Intermediate magnetic partner 88 is disposed between front magnetic partner 84 and rear magnetic partner 86.
In general, the front magnetic partner and the rear magnetic partner are made of one or more materials that are magnetically attracted to the one or more materials from which the intermediate magnetic partner is made. As one example, the front magnetic partner and the rear magnetic partner may be iron that is not permanently magnetic, and the intermediate magnetic partner may be a permanent magnet (e.g., ferromagnetic iron). As another example, the front magnetic partner and the rear magnetic partner may be a permanent magnet (e.g., ferromagnetic iron), and the intermediate magnetic partner may be iron that is not permanently magnetic. It is to be understood that any combination of magnetically attractive partners may be used.
When wearable input device 101 is in a flat mode of operation, front magnetic partner 84 and intermediate magnetic partner 88 magnetically bias the plurality of elongated rib members in a flat mode of operation. In particular, front magnetic partner 84 and intermediate magnetic partner 88 magnetically attract one another. When wearable input device 101 moves from a flat mode of operation to a curved mode of operation, intermediate magnetic partner 88 moves away from front magnetic partner 84 towards rear magnetic partner 86 because the inner radius of the bottom flexible support is less than the outer radius of the top flexible support. As such, the magnetic force between front magnetic partner 84 and intermediate magnetic partner 88 works to prevent wearable input device 101 from moving from a flat mode of operation to a curved mode of operation.
When wearable input device 101 is in a curved mode of operation, rear magnetic partner 86 and intermediate magnetic partner 88 magnetically bias the plurality of elongated rib members in a curved mode of operation. In particular, rear magnetic partner 86 and intermediate magnetic partner 88 magnetically attract one another. When wearable input device 101 moves from a curved mode of operation to a flat mode of operation, intermediate magnetic partner 88 moves away from rear magnetic partner 86 towards front magnetic partner 84 because the inner radius of the bottom flexible support is less than the outer radius of the top flexible support. As such, the magnetic force between rear magnetic partner 86 and intermediate magnetic partner 88 works to prevent wearable input device 101 from moving from a curved mode of operation to a flat mode of operation.
In an embodiment, electronics components 107 are positioned on the back of wearable input device 101 as illustrated in
In an embodiment, electronic components 107 include a few electronic components and most computational tasks related to user inputs are performed externally. For example, electronic components 107 may includes a wired or wireless transmitter 602, memory 608 to store machine or processor readable instructions including a software driver 608a to read inputs from sensors 601a-d and provide an output signal to a transmitter 602 that represents touch inputs by user 100.
In embodiments, transmitter 602 may provide one or more various types of wireless and/or wired signal. For example, transmitter 602 may transmit various types of wireless signals including WiFi, Bluetooth, infrared, infrared personal area network, radio frequency Identification (RFID), wireless Universal Serial Bus (WUSB), cellular, 3G, 4G or other types of wireless signals.
In an alternate embodiment, electronic components 107 include numerous components and/or perform computational extensive tasks. In an embodiment, electronic components 107 are positioned on a flexible substrate having a plurality of electronic connections including wires or traces to transfer electronic signals between electronic components. In an embodiment, one or more electronic components 107 may be included in a single packaged chip or system-on-a-chip (SoC).
In an embodiment, electronic components 107 include one or more processors 603. Processor 603 may comprise a controller, central processing unit (CPU), graphics-processing unit (GPU), digital signal processor (DSP) and/or a field programmable gate array (FPGA). In an embodiment, memory 610 includes processor readable instructions to operate wearable input device 101. In embodiments, memory 610 includes a variety of different types of volatile as well as non-volatile memory as described herein.
In an embodiment, power supply 604 provides power or a predetermined voltage to one or more electronic components in electronic components 107 as well as touch surface 104. In an embodiment, power supply 604 provides power to one or more electronic components in electronic components 107 in response to a switch being toggled on wearable input device 101 by user 100.
In an embodiment, electronic components 107 includes inertial sensing unit 605 including one or more inertial sensors to sense an orientation of wearable input device 101, and a location sensing unit 606 to sense a location of wearable input device 101. In an embodiment, inertial sensing unit 605 includes a three axis accelerometer and a three axis magnetometer, that determines orientation changes of wearable input device 101. An orientation of wearable input device 101 may include a landscape, portrait, one hand, two-handed orientation, curved or flat orientation. Location sensing unit 606 may include one or more location or proximity sensors, some examples of which are a global positioning system (GPS) transceiver, an infrared (IR) transceiver, or a radio frequency transceiver for processing RFID data.
In an embodiment, one or more electronic components in electronic components 107 and/or sensors may include an analog interface that produces or converts an analog signal, or both produces and converts an analog signal, for its respective component or sensor. For example, inertial sensing unit 605, location sensing unit 606, touch sensors 601a-d and biometric sensors 607 may include analog interfaces that convert analog signals to digital signals.
In embodiments, one or more biometric sensors 607 may include a variety of different types of biometric sensors. For example, biometric sensors may include heart rate monitors or sensors, blood/oxygen sensors, accelerometers, thermometers or other types of biometric sensors that obtain biometric information from user 100. In an embodiment, a blood/oxygen sensor includes a pulse oximetry sensor that measures a saturation of user's hemoglobin.
In alternate embodiments, other virtual objects or virtual locations may be provided by AR system 801. For example, when user 100 picks up a book, virtual text describing reviews of the book may be positioned next to the book. In other embodiments, a virtual location at a previous time period may be displayed or provided to user 100. In an embodiment, a user 100 may select a virtual location provided by AR system 801 by touching wearable input device 101 at the defined area, such as an area defined by a “AR 100” key.
The AR system 801 may include a computing environment 804, a capture device 802, and a display 803, all in communication with each other. Computing environment 804 may include one or more processors as described herein. Capture device 802 may include a color or depth sensing camera that may be used to visually monitor one or more targets including humans and one or more other real objects within a particular environment. In one example, capture device 802 may comprise an RGB or depth camera and computing environment 804 may comprise a set-top box or gaming console. AR system 801 may support multiple users and wearable input devices.
Step 1000 illustrates determining whether a wearable input device is in a curved mode of operation or in a flat mode of operation. In an embodiment, one or more inertial sensing units 605 in electronic components 107 outputs a signal indicating an orientation. Processor 603 then may execute processor readable instructions in memory 610 to determine whether a wearable input device is in a curved or flat mode of operation.
Step 1001 illustrates determining whether a wearable input device is in a biometric mode of operation. In embodiments, a wearable input device may also be in a biometric mode of operation (receiving valid biometric information) in either a curved or flat mode of operation. In an embodiment, a biometric mode of operation does not occur when a wearable input device is in a flat mode of operation because biometric sensors are not in close proximity to skin of a user, such as a wrist. In an embodiment, biometric inputs are compared to biometric threshold values to determine whether a biometric mode of operation is available. In an embodiment, biometric threshold values stored in memory 610 are compared to biometric inputs by processor 603 and executable processor readable instructions stored in memory 610 to determine whether a biometric mode of operation is available. Biometric sensors may not be able to obtain valid biometric information because wearable input device is not in an orientation or fitted to a user such that valid sensor inputs may be obtained.
Step 1002 illustrates receiving touch inputs from a touch surface when a wearable input device is in a curved mode of operation. Step 1003 illustrates receiving touch inputs from a touch surface when a wearable input device is in a flat mode of operation. In embodiments, different key layouts may be used for the curved mode of operation and flat mode of operation. For example in a flat mode of operation, a touch surface may have many more locations that correspond to characters so that a wearable input device may be more easily used in complex two handed operations that may need multiple touches, such as forming a text message. In a curved mode of operation, a different key layout having a few larger keys or locations may be used. For example, a large key area may be identified for a favorite AR user experience or image of a user. As described herein, different key layout stickers may be adhered to a touch surface to let a user know where to touch for a particular input in different modes of operation.
Step 1004 illustrates receiving biometric inputs from biometric sensors. In an embodiment, one or more biometric sensors 607 output signals representing biometric input to processor 603 executing processor readable instructions stored in memory 610.
Step 1005 illustrates a wearable user input device performing a calculation based on the received inputs. Processor 603 executing processor readable instructions stored in memory 610 may determine or calculate a possible AR experience that a user may want to experience based on touch inputs and biometric inputs, such as heart rate. For example, if a user requests a AR experience through touch inputs that may cause excitement/fear and a heart rate exceeds a predetermined value, a wearable input device may output a calculated request for a less exciting/fearful AR experience.
In alternate embodiments, no calculations are performed in step 1005 and control proceeds to step 1006 where received inputs are transmitted to one or more AR components as described herein. In embodiments, transmitter 602 outputs a wireless or wired signal that represents the user touch and biometric inputs to an AR component, such as computing system(s) 1512 as described herein.
In step 1101, a capacitive surface receives at least one touch that represents a character input and/or gesture in an embodiment. For example, a user may touch a portion of the first sticker (attached to the capacitive surface) that corresponds to a desired character input or operation of an AR system.
In step 1102, biometric information from biometric sensors as described herein may be measured and received by the wearable input device. The biometric information may be, but not limited to, heart rate and blood information from a user wearing the wearable input device.
In step 1103, the input and biometric information may be transmitted. For example, the information may be transmitted by one or more wireless signals to one or more computing systems in an AR system.
Step 1104 illustrates receiving or attaching a second sticker that defines one or more different locations corresponding to predetermined input while the wearable input device is in a flat mode of operation. In an embodiment, the second sticker is adhered to the first sticker. In an alternate embodiment, the second sticker is adhered to at least a portion of the capacitive surface after the first sticker is removed. In an embodiment, the second sticker has a more extensive character layout so more complex multi-hand operations may be performed, such as composing and sending a text message.
In step 1105 multiple touches are received on the second sticker (attached to the capacitive surface) that represents another input information when the wearable input device is in a flat mode of operation. For example, a user may have multiple touches in forming a text message.
Step 1106 then illustrates transmitting another input information. In an embodiment another information may be transmitted by one or more wireless signals to one or more computing systems in an AR system.
Each display optical system 1514 is also referred to as a see-through display, and the two display optical systems 1514 together may also be referred to as a see-through, meaning optical see-through, AR display 1514.
Frame 1515 provides a support structure for holding elements of the apparatus in place as well as a conduit for electrical connections. In this embodiment, frame 1515 provides a convenient eyeglass frame as support for the elements of the apparatus discussed further below. The frame 1515 includes a nose bridge 1504 with a microphone 1510 for recording sounds and transmitting audio data to control circuitry 1536. In this example, the temple arm 1513 is illustrated as including control circuitry 1536 for the HMD 1502.
As illustrated in
Wearable input device 101 may communicate wired and/or wirelessly (e.g., WiFi, Bluetooth, infrared, an infrared personal area network, RFID transmission, WUSB, cellular, 3G, 4G or other wireless communication means) over one or more communication networks 1560 to one or more computer systems 1512 whether located nearby or at a remote location, other personal A/V apparatus 1508 in a location or environment. In other embodiments, wearable input device 101 communicates with HMD 1502 and/or communication network(s) by wireless signals as in
An application may be executing on a computer system 1512 which interacts with or performs processing for an application executing on one or more processors in the personal A/V apparatus 1500. For example, a 3D mapping application may be executing on the one or more computers systems 12 and the user's personal A/V apparatus 1500.
In the illustrated embodiments of
Control circuitry 1536 provide various electronics that support the other components of HMD 1502. In this example, the right temple arm 1513 includes control circuitry 1536 for HMD 1502 which includes a processing unit 15210, a memory 15244 accessible to the processing unit 15210 for storing processor readable instructions and data, a wireless interface 1537 communicatively coupled to the processing unit 15210, and a power supply 15239 providing power for the components of the control circuitry 1536 and the other components of HMD 1502 like the cameras 1613, the microphone 1510 and the sensor units discussed below. The processing unit 15210 may comprise one or more processors that may include a controller, CPU, GPU and/or FPGA.
Inside, or mounted to temple arm 1502, are an earphone of a set of earphones 1630, an inertial sensing unit 1632 including one or more inertial sensors, and a location sensing unit 1644 including one or more location or proximity sensors, some examples of which are a GPS transceiver, an IR transceiver, or a radio frequency transceiver for processing RFID data.
In this embodiment, each of the devices processing an analog signal in its operation include control circuitry which interfaces digitally with the digital processing unit 15210 and memory 15244 and which produces or converts analog signals, or both produces and converts analog signals, for its respective device. Some examples of devices which process analog signals are the sensing units 1644, 1632, and earphones 1630 as well as the microphone 1510, capture devices 1613 and a respective IR illuminator 1634A, and a respective IR detector or camera 1634B for each eye's display optical system 154l, 154r discussed below.
Mounted to or inside temple arm 1515 is an image source or image generation unit 1620 which produces visible light representing images. The image generation unit 1620 can display a virtual object to appear at a designated depth location in the display field of view to provide a realistic, in-focus three dimensional display of a virtual object which can interact with one or more real objects.
In some embodiments, the image generation unit 1620 includes a microdisplay for projecting images of one or more virtual objects and coupling optics like a lens system for directing images from the microdisplay to a reflecting surface or element 1624. The reflecting surface or element 1624 directs the light from the image generation unit 1620 into a light guide optical element 1612, which directs the light representing the image into the user's eye.
In the illustrated embodiment, the display optical system 1514 is an integrated eye tracking and display system. The system embodiment includes an opacity filter 1514 for enhancing contrast of virtual imagery, which is behind and aligned with optional see-through lens 1616 in this example, light guide optical element 1612 for projecting image data from the image generation unit 1620 is behind and aligned with opacity filter 1514, and optional see-through lens 1618 is behind and aligned with light guide optical element 1612.
Light guide optical element 1612 transmits light from image generation unit 1620 to the eye 1640 of a user wearing HMD 1502. Light guide optical element 1612 also allows light from in front of HMD 1502 to be received through light guide optical element 1612 by eye 1640, as depicted by an arrow representing an optical axis 1542 of the display optical system 1514r, thereby allowing a user to have an actual direct view of the space in front of HMD 1502 in addition to receiving a virtual image from image generation unit 1620. Thus, the walls of light guide optical element 1612 are see-through. In this embodiment, light guide optical element 1612 is a planar waveguide. A representative reflecting element 1634E represents the one or more optical elements like mirrors, gratings, and other optical elements which direct visible light representing an image from the planar waveguide towards the user eye 1640.
Infrared illumination and reflections, also traverse the planar waveguide for an eye tracking system 1634 for tracking the position and movement of the user's eye, typically the user's pupil. Eye movements may also include blinks. The tracked eye data may be used for applications such as gaze detection, blink command detection and gathering biometric information indicating a personal state of being for the user. The eye tracking system 1634 comprises an eye tracking IR illumination source 1634A (an infrared light emitting diode (LED) or a laser (e.g. VCSEL)) and an eye tracking IR sensor 1634B (e.g. IR camera, arrangement of IR photo detectors, or an IR position sensitive detector (PSD) for tracking glint positions). In this embodiment, representative reflecting element 1634E also implements bidirectional IR filtering which directs IR illumination towards the eye 1640, preferably centered about the optical axis 1542 and receives IR reflections from the user eye 1640. A wavelength selective filter 1634C passes through visible spectrum light from the reflecting surface or element 1624 and directs the infrared wavelength illumination from the eye tracking illumination source 1634A into the planar waveguide. Wavelength selective filter 1634D passes the visible light and the infrared illumination in an optical path direction heading towards the nose bridge 1504. Wavelength selective filter 1634D directs infrared radiation from the waveguide including infrared reflections of the user eye 1640, preferably including reflections captured about the optical axis 1542, out of the light guide optical element 1612 embodied as a waveguide to the IR sensor 1634B.
Opacity filter 1514, which is aligned with light guide optical element 112, selectively blocks natural light from passing through light guide optical element 1612 for enhancing contrast of virtual imagery. The opacity filter assists the image of a virtual object to appear more realistic and represent a full range of colors and intensities. In this embodiment, electrical control circuitry for the opacity filter, not shown, receives instructions from the control circuitry 1536 via electrical connections routed through the frame.
Again,
The computing environment 1754 also stores data in image and audio data buffer(s) 1799. The buffers provide memory for receiving image data captured from the outward facing capture devices 1613, image data captured by other capture devices if available, image data from an eye tracking camera of an eye tracking system 1634 if used, buffers for holding image data of virtual objects to be displayed by the image generation units 1620, and buffers for both input and output audio data like sounds captured from the user via microphone 1510 and sound effects for an application from the 3D audio engine 3704 to be output to the user via audio output devices like earphones 1630.
Image and audio processing engine 1791 processes image data, depth data and audio data received from one or more capture devices which may be available in a location. Image and depth information may come from the outward facing capture devices 1613 captured as the user moves his head or body and additionally from other physical A/V apparatus 1500, other 3D image capture devices 1520 in the location and image data stores like location indexed images and maps 3724.
The individual engines and data stores depicted in
The operating system 1790 makes available to applications which gestures the gesture recognition engine 1793 has identified, which words or sounds the sound recognition engine 1794 has identified, the positions of objects from the scene mapping engine 3706 as described above, and eye data such as a position of a pupil or an eye movement like a blink sequence detected from the eye tracking software 1796. A sound to be played for the user in accordance with the information display application 4714 can be uploaded to a sound library 3712 and identified to the 3D audio engine 3704 with data identifying from which direction or position to make the sound seem to come from. The device data 1798 makes available to the information display application 4714 location data, head position data, data identifying an orientation with respect to the ground and other data from sensing units of the HMD 1502.
The scene mapping engine 3706 is first described. A 3D mapping of the display field of view of the AR display can be determined by the scene mapping engine 3706 based on captured image data and depth data, either derived from the captured image data or captured as well. The 3D mapping includes 3D space positions or position volumes for objects.
A depth map representing captured image data and depth data from outward facing capture devices 1613 can be used as a 3D mapping of a display field of view of a near-eye AR display. A view dependent coordinate system may be used for the mapping of the display field of view approximating a user perspective. The captured data may be time tracked based on capture time for tracking motion of real objects. Virtual objects can be inserted into the depth map under control of an application like information display application 4714. Mapping what is around the user in the user's environment can be aided with sensor data. Data from an orientation sensing unit 1632, e.g. a three axis accelerometer and a three axis magnetometer, determines position changes of the user's head and correlation of those head position changes with changes in the image and depth data from the front facing capture devices 1613 can identify positions of objects relative to one another and at what subset of an environment or location a user is looking.
In some embodiments, a scene mapping engine 3706 executing on one or more network accessible computer systems 1512 updates a centrally stored 3D mapping of a location and apparatus 1500 download updates and determine changes in objects in their respective display fields of views based on the map updates. Image and depth data from multiple perspectives can be received in real time from other 3D image capture devices 1520 under control of one or more network accessible computer systems 1512 or from one or more physical A/V apparatus 1500 in the location. Overlapping subject matter in the depth images taken from multiple perspectives may be correlated based on a view independent coordinate system, and the image content combined for creating the volumetric or 3D mapping of a location (e.g. an x, y, z representation of a room, a store space, or a geofenced area). Additionally, the scene mapping engine 3706 can correlate the received image data based on capture times for the data in order to track changes of objects and lighting and shadow in the location in real time.
The registration and alignment of images allows the scene mapping engine to be able to compare and integrate real-world objects, landmarks, or other features extracted from the different images into a unified 3-D map associated with the real-world location.
When a user enters a location or an environment within a location, the scene mapping engine 3706 may first search for a pre-generated 3D map identifying 3D space positions and identification data of objects stored locally or accessible from another physical A/V apparatus 1500 or a network accessible computer system 1512. The pre-generated map may include stationary objects. The pre-generated map may also include objects moving in real time and current light and shadow conditions if the map is presently being updated by another scene mapping engine 3706 executing on another computer system 1512 or apparatus 1500. For example, a pre-generated map indicating positions, identification data and physical properties of stationary objects in a user's living room derived from image and depth data from previous HMD sessions can be retrieved from memory. Additionally, identification data including physical properties for objects which tend to enter the location can be preloaded for faster recognition. A pre-generated map may also store physics models for objects as discussed below. A pre-generated map may be stored in a network accessible data store like location indexed images and 3D maps 3724.
The location may be identified by location data which may be used as an index to search in location indexed image and pre-generated 3D maps 3724 or in Internet accessible images 3726 for a map or image related data which may be used to generate a map. For example, location data such as GPS data from a GPS transceiver of the location sensing unit 1644 on a HMD 1502 may identify the location of the user. In another example, a relative position of one or more objects in image data from the outward facing capture devices 1613 of the user's physical A/V apparatus 1500 can be determined with respect to one or more GPS tracked objects in the location from which other relative positions of real and virtual objects can be identified. Additionally, an IP address of a WiFi hotspot or cellular station to which the physical A/V apparatus 1500 has a connection can identify a location. Additionally, identifier tokens may be exchanged between physical A/V apparatus 1500 via infra-red, Bluetooth or WUSB. The range of the infra-red, WUSB or Bluetooth signal can act as a predefined distance for determining proximity of another user. Maps and map updates, or at least object identification data may be exchanged between physical A/V apparatus via infra-red, Bluetooth or WUSB as the range of the signal allows.
The scene mapping engine 3706 identifies the position and tracks the movement of real and virtual objects in the volumetric space based on communications with the object recognition engine 1792 of the image and audio processing engine 1791 and one or more executing applications generating virtual objects.
The object recognition engine 1792 of the image and audio processing engine 1791 detects, tracks and identifies real objects in the display field of view and the 3D environment of the user based on captured image data and captured depth data if available or determined depth positions from stereopsis. The object recognition engine 1792 distinguishes real objects from each other by marking object boundaries and comparing the object boundaries with structural data. One example of marking object boundaries is detecting edges within detected or derived depth data and image data and connecting the edges. Besides identifying the type of object, an orientation of an identified object may be detected based on the comparison with stored structure data 2700, object reference data sets 3718 or both. One or more databases of structure data 2700 accessible over one or more communication networks 1560 may include structural information about objects. As in other image processing applications, a person can be a type of object, so an example of structure data is a stored skeletal model of a human which may be referenced to help recognize body parts. Structure data 2700 may also include structural information regarding one or more inanimate objects in order to help recognize the one or more inanimate objects, some examples of which are furniture, sporting equipment, automobiles and the like.
The structure data 2700 may store structural information as image data or use image data as references for pattern recognition. The image data may also be used for facial recognition. The object recognition engine 1792 may also perform facial and pattern recognition on image data of the objects based on stored image data from other sources as well like user profile data 1797 of the user, other users profile data 3722 which are permission and network accessible, location indexed images and 3D maps 3724 and Internet accessible images 3726.
Device 1800 may also contain communications connection(s) 1812 such as one or more network interfaces and transceivers that allow the device to communicate with other devices. Device 1800 may also have input device(s) 1814 such as keyboard, mouse, pen, voice input device, touch input device, etc. Output device(s) 1816 such as a display, speakers, printer, etc. may also be included. These devices are well known in the art so they are not discussed at length here.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. The specific features and acts described above are disclosed as example forms of implementing the claims.