The disclosure generally relates to artificial reality systems, such as augmented reality, mixed reality, and/or virtual reality systems, and more particularly, to user interfaces in artificial reality environments.
Artificial reality systems are becoming increasingly ubiquitous with applications in many fields such as computer gaming, health and safety, industrial, and education. As a few examples, artificial reality systems are being incorporated into mobile devices, gaming consoles, personal computers, movie theaters, and theme parks. In general, artificial reality is a form of reality that has been adjusted in some manner before presentation to a user, which may include, e.g., a virtual reality (VR), an augmented reality (AR), a mixed reality (MR), a hybrid reality, or some combination and/or derivatives thereof.
Typical artificial reality systems include one or more devices for rendering and displaying content to users. As one example, an artificial reality system may incorporate a head-mounted display (HMD) worn by a user and configured to output artificial reality content to the user. The artificial reality content may include completely-generated content or generated content combined with captured content (e.g., real-world video and/or images). During operation, the user typically interacts with the artificial reality system to interact with virtual reality content in an artificial reality environment.
In general, the disclosure describes artificial reality (AR) systems and techniques for generating and presenting a virtual user interface with which users may interact using a physical peripheral device. The AR system renders, for display by an HMD, glasses or other display device, AR content in which the virtual user interface is locked to the peripheral device. That is, the AR system may render the virtual user interface having one or more virtual user interface elements at a position and pose in the artificial reality environment that is based on and corresponds to the position and pose of the physical peripheral device in the physical environment. In this way, the virtual user interface in the artificial reality environment may track the physical peripheral device. The virtual user interface elements can include, in various examples, virtual buttons, a virtual keyboard, a virtual drawing interface, a virtual selectable menu, or other user-selectable virtual user interface elements, which may be context-driven based on the current AR applications engaged by the user. The peripheral device may be a computing device having one or more presence-sensitive surfaces.
The AR systems may enable the user to interact with the peripheral device through virtual user interface elements of the virtual user interface overlaid on the peripheral device, which may be manipulated and otherwise interacted with by the user to provide input to an AR system through pose tracking of the peripheral device and image-based gesture detection and/or via one or more input devices of the peripheral device, such as a presence-sensitive surface. For example, the user may interact with the virtual user interface rendered on the physical peripheral device to perform user interface gestures with respect to virtual user interface elements. For instance, the user may press their finger at a physical location on the peripheral device corresponding to a position in the artificial reality environment at which the AR system renders a virtual user interface button of the virtual user interface. In this example, the AR system detects this user interface gesture and performs an action corresponding to the detected press of the virtual user interface button. The AR system may also, for instance, animate the press of the virtual user interface button along with the gesture.
The techniques may provide one or more technical improvements that provide at least one practical application. For example, the techniques can enable the user to provide fine-grained user inputs with respect to user interface elements rendered virtually on a physical peripheral device that provides haptic feedback, in contrast to free-floating virtual user interfaces in the artificial reality environment. This can simplify and improve the precision of gesture detection and provide a more pleasing user experience. In addition, the peripheral device may not display the user interface elements at its own display and may not even include a display. The techniques may therefore additionally reduce power consumption and simplify AR applications by eliminating a separate interface that would otherwise need to be generated and displayed, at a presence-sensitive display of a smartphone or tablet for instance, for receiving precision inputs from a user.
In some examples, an artificial reality system includes an image capture device configured to capture image data; a head-mounted display (HMD) configured to output artificial reality content; a user interface engine configured to detect a peripheral device from the image data, wherein the user interface engine is configured to generate a virtual user interface comprising one or more virtual user interface elements; and a rendering engine configured to render the artificial reality content and to render, at a user interface position locked relative to a position of the peripheral device in an artificial reality environment, the virtual user interface for display at the HMD.
In some examples, a method includes obtaining, by an artificial reality system including a head-mounted display (HMD), image data via an image capture device, the HMD configured to output artificial reality content; detecting, by the artificial reality system, a peripheral device from the image data; generating, by the artificial reality system, a virtual user interface comprising one or more virtual user interface elements; and rendering, by the artificial reality system, the artificial reality content and, at a user interface position locked relative to a position of the peripheral device in an artificial reality environment, the virtual user interface for display at the HMD.
In some examples, a non-transitory, computer-readable medium comprising instructions that, when executed, cause one or more processors of an artificial reality system, including a head-mounted display (HMD) configured to output artificial reality content, to: generate a virtual user interface comprising one or more virtual user interface elements; render the artificial reality content and, at a user interface position locked relative to a position of the peripheral device in an artificial reality environment, the virtual user interface for display at the HMD; detect a user interface gesture performed by a user at a position corresponding to one of the virtual user interface elements; and perform, in response to the user interface gesture, one or more actions associated with the one of the virtual user interface elements.
The details of one or more examples are set forth in the accompanying drawings and the description below. Other features, objects, and advantages will be apparent from the description and drawings, and from the claims.
Like reference characters refer to like elements throughout the description and figures.
In general, artificial reality system 10 uses information captured from a real-world, 3D physical environment to render artificial reality content 122 for display to user 110. In the example of
In the example shown in
In some implementations, artificial reality system 10 generates and renders virtual content items 124, 126 (e.g., GIFs, photos, applications, live-streams, videos, text, a web-browser, drawings, animations, representations of data files, or any other visible media) on a virtual surface. A virtual surface may be associated with a planar or other real-world surface (e.g., the virtual surface corresponds to and is locked to a physical planar surface, such as a wall table, or ceiling). In the example shown in
The artificial reality system 10 may render one or more virtual content items in response to a determination that at least a portion of the location of virtual content items is in the field of view 130 of user 110. For example, artificial reality system 10 may render virtual user interface 137 only if peripheral device 136 is within field of view 130 of user 110.
During operation, the artificial reality application constructs artificial reality content 122 for display to user 110 by tracking and computing pose information for a frame of reference, typically a viewing perspective of HMD 112. Using HMD 112 as a frame of reference, and based on a current field of view 130 as determined by a current estimated pose of HMD 112, the artificial reality application renders 3D artificial reality content which, in some examples, may be overlaid, at least in part, upon the real-world, 3D physical environment of user 110. During this process, the artificial reality application uses sensed data received from HMD 112, such as movement information and user commands, and, in some examples, data from any external sensors 90, such as external cameras, to capture 3D information within the real world, physical environment, such as motion by user 110 and/or feature tracking information with respect to user 110. Based on the sensed data, the artificial reality application determines a current pose for the frame of reference of HMD 112 and, in accordance with the current pose, renders the artificial reality content 122.
Artificial reality system 10 may trigger generation and rendering of virtual content items based on a current field of view 130 of user 110, as may be determined by real-time gaze tracking of the user, or other conditions. More specifically, image capture devices 138 of HMD 112 capture image data representative of objects in the real world, physical environment that are within a field of view 130 of image capture devices 138. Field of view 130 typically corresponds with the viewing perspective of HMD 112. In some examples, the artificial reality application presents artificial reality content 122 comprising mixed reality and/or augmented reality. As illustrated in
During operation, artificial reality system 10 performs object recognition within image data captured by image capture devices 138 of HMD 112 to identify peripheral device 136, hand 132, including optionally identifying individual fingers or the thumb, and/or all or portions of arm 134 of user 110. Further, artificial reality system 10 tracks the position, orientation, and configuration of peripheral device 136, hand 132 (optionally including particular digits of the hand), and/or portions of arm 134 over a sliding window of time. In some examples, peripheral device 136 includes one or more sensors (e.g., accelerometers) for tracking motion or orientation of the peripheral device 136.
In accordance with techniques of this disclosure, artificial reality system 10 presents a virtual user interface 137 with which users may interact with using a physical device, referred to as a “peripheral device.” Peripheral device 136 is a physical, real-world device having a surface on which AR system 10 overlays virtual user interface 137. That is, AR system 10 virtually renders virtual user interface 137 at a position and orientation so that virtual user interface 137 appears to be a surface of peripheral device 136 or juxtaposed with the surface of peripheral device 136. In this respect, peripheral device 136 operates as a stage for virtual content, such as virtual user interface 137. Peripheral device 136 may include one or more presence-sensitive surfaces for detecting user inputs by detecting a presence of one or more objects (e.g., fingers, stylus) touching or hovering over locations of the presence-sensitive surface. In some examples, peripheral device 136 may include an output display, which may be a presence-sensitive display. AR system 10 may cause peripheral device 136 to deactivate (i.e., turn off) the output display when rendering virtual user interface 137. AR system 10 may, in some examples, only render virtual user interface 137 when the output display is deactivated. In some examples, peripheral device 136 does not include an output display however.
In some examples, peripheral device 136 may be a smartphone, tablet computer, personal data assistant (PDA), or other hand-held device. In some examples, peripheral device 136 may be a smartwatch, smartring, or other wearable device. Peripheral device 136 may also be part of kiosk or other stationary or mobile system. Peripheral device 136 may or may not include a display device for outputting content to a screen. Peripheral device 136 may be in communication with HMD 112 and/or console 106 using one or more wired or wireless communications links (e.g., Wi-Fi, near-field communication of short-range wireless communication such as Bluetooth).
The AR system 10 renders, to HMD, glasses or other display device 112, AR content in which virtual user interface 137 is locked to a surface of peripheral device 136. That is, the AR system 10 may render virtual user interface 137 having one or more virtual user interface elements at a position and orientation in the virtual environment that is based on and corresponds to the position and orientation of the physical peripheral device 136 in the physical environment 130. For example, if the peripheral device 136 is positioned in a vertical position (referred to as “portrait mode”), the AR system 10 may render the virtual user interface in portrait mode and at a location corresponding to the position and orientation of the peripheral device 136. If the peripheral device 136 is positioned in a horizontal position (referred to as “landscape mode”), the AR system 10 may render the virtual user interface in landscape mode and at a location corresponding to the position and orientation of the peripheral device 136. In this way, the virtual user interface being rendered in the virtual environment may track the handheld physical peripheral device 136 such that peripheral device 136 appears, to the user, to be outputting virtual user interface 137 on a surface of peripheral device 136.
Virtual user interface 137 includes one or more virtual user interface elements. Virtual user interface elements may include, for instance, a virtual drawing interface 142, a selectable menu 144 (e.g., a drop-down menu), virtual buttons 146, a directional pad 148, a keyboard, or other user-selectable user interface elements, glyphs, display elements, content, user interface controls, and so forth. The particular virtual user interface elements for virtual user interface 137 may be context-driven based on the current AR applications engaged by the user.
When a user performs a user interface gesture in the virtual environment at a location that corresponds to one of the virtual user interface elements of virtual user interface 137 overlaid on the peripheral device 136, the AR system 10 detects the user interface and performs an action. For example, the user may press their finger at a physical location on the peripheral device 136 corresponding to a position in the virtual environment at which the AR system 10 renders a virtual user interface button 146 on virtual user interface 137. In this example, the AR system 10 detects this virtual button press gesture and performs an action corresponding to the detected press of a virtual user interface button 146. The AR system 10 may also, for instance, animate a press of the virtual user interface button along with the button press gesture.
AR system 10 may detect user interface gestures and other gestures using an inside-out or outside-in tracking system of image capture devices 138 and or external cameras. AR system 10 may alternatively, or in addition, detect user interface gestures and other gestures using a presence-sensitive surface. That is, a presence-sensitive interface of peripheral device 136 may receive user inputs that make up a user interface gesture.
Peripheral device 136 provides haptic feedback to touch-based user interaction by having a physical surface with which the user can interact (e.g., touch, drag a finger across, grab, and so forth). In addition, peripheral device 136 may output other indications of user interaction using an output device. For example, in response to a detected press of a virtual user interface button 146, peripheral device 136 may output a vibration or “click” noise, or peripheral device 136 may generate and output content to a display.
In some examples, the user may press and drag their finger along physical locations on the peripheral device 136 corresponding to positions in the virtual environment at which the AR system 10 renders virtual drawing interface 142 of virtual user interface 137. In this example, the AR system 10 detects this drawing gesture and performs an action according to the detected press and drag of a virtual drawing interface 142, such as by generating and rendering virtual markings at the positions in the virtual environment. In this way, AR system 10 simulates drawing or writing on peripheral device 136 using virtual user interface 137.
In some examples, the user may press their finger at a physical location on the peripheral device 136 corresponding to a position in the virtual environment at which the AR system 10 renders a virtual selectable menu 144. In this example, the AR system 10 detects this user interface gesture and performs an action according to the detected press of a virtual selectable menu 144. The virtual selectable menu 144 may comprise a drop-down menu, a pie menu, a list menu, or any menu including a plurality of selectable items. In some examples, the virtual selectable menu 144 include sub-menus, such that when a user selects an item in a main menu of the virtual selectable menu 144, the AR system 10 would render a sub-menu of one or more items corresponding to the selected item in the main menu, the one or more items further selectable by the user. In this way, AR system 10 simulates interacting with a menu on peripheral device 136 using virtual user interface 137.
In some examples, the user may press their finger at a physical location on the peripheral device 136 corresponding to a position in the virtual environment at which the AR system 10 renders a virtual directional pad 148. In this example, the AR system 10 detects this user interface gesture and performs an action according to the detected press of virtual directional pad 148. In this way, AR system 10 simulates interacting with a directional pad on peripheral device 136 using virtual user interface 137.
Accordingly, techniques of the disclosure provide specific technical improvements to the computer-related field of rendering and displaying content by an artificial reality system, which may provide one or more practical applications. For example, artificial reality systems as described herein may provide a high-quality artificial reality experience to a user, such as user 110, of the artificial reality application by generating and rendering virtual user interface 137 locked to a surface of peripheral device 136 and detecting gestures performed by user 110 with respect to peripheral device 136.
In addition, AR systems as described herein may be configured to detect gestures that provide self-haptic feedback to the user. For example, one or more fingers on each hand 132 of the user 110 may touch or approximately touch peripheral device 136 in the physical world as part an interaction with a particular virtual user interface element of virtual user interface 137 in the artificial reality content. The touch between the one or more fingers of the user's hand 132 and the peripheral device may provide the user with a simulation of the sensation felt by the user when interacting directly with a physical user input object, such as a button on a physical keyboard or other physical input device.
As another example, AR systems as described herein can enable the user to provide fine-grained user inputs with respect to virtual user interface elements rendered virtually on a peripheral device 136 that provides haptic feedback, in contrast to free-floating virtual user interfaces in the artificial reality environment. This can simplify and improve the precision of gesture detection. In addition, peripheral device 136 may not display the user interface elements at its own display and may not even include a display. The techniques may therefore additionally reduce power consumption and simplify AR applications by eliminating a separate interface that would otherwise need to be generated and displayed, at a presence-sensitive display of a smartphone or tablet for instance, for receiving precision inputs from a user. In some examples, receiving user inputs with a presence-sensitive surface may provide more precise gesture detection than image-based gesture detection techniques.
In the example of
Each of HMDs 112 concurrently operates within artificial reality system 20. In the example of
In a manner similar to the examples discussed above with respect to
As shown in
In some aspects, the artificial reality application can run on console 106, and can utilize image capture devices 102A and 102B to analyze configurations, positions, and/or orientations of hand 132B to identify input gestures that may be performed by a user of HMD 112A. Similarly, HMD 112C can utilize image capture device 138 to analyze configurations, positions, and/or orientations of peripheral device 136 and hand 132C to input gestures that may be performed by a user of HMD 112C. In some examples, peripheral device 136 includes one or more sensors (e.g., accelerometers) for tracking motion or orientation of the peripheral device 136. The artificial reality application may render virtual content items and/or user interface elements, responsive to such gestures, in a manner similar to that described above with respect to
Image capture devices 102 and 138 may capture images in the visible light spectrum, the infrared spectrum, or other spectrum. Image processing described herein for identifying objects, object poses, and gestures, for example, may include processing infrared images, visible light spectrum images, and so forth.
Virtual user interface 137 is mapped to a surface of peripheral device 136. As a result, AR system 10 renders, at a user interface position that is locked relative to a position of peripheral device 136 in the artificial reality environment, virtual user interface 137 for display at HMD 112C as part of artificial reality content 122.
In this example, HMD 112 includes a front rigid body and a band to secure HMD 112 to a user. In addition, HMD 112 includes an interior-facing electronic display 203 configured to present artificial reality content to the user. Electronic display 203 may be any suitable display technology, such as liquid crystal displays (LCD), quantum dot display, dot matrix displays, light emitting diode (LED) displays, organic light-emitting diode (OLED) displays, cathode ray tube (CRT) displays, e-ink, or monochrome, color, or any other type of display capable of generating visual output. In some examples, the electronic display is a stereoscopic display for providing separate images to each eye of the user. In some examples, the known orientation and position of display 203 relative to the front rigid body of HMD 112 is used as a frame of reference, also referred to as a local origin, when tracking the position and orientation of HMD 112 for rendering artificial reality content according to a current viewing perspective of HMD 112 and the user. In other examples, HMD 112 may take the form of other wearable head mounted displays, such as glasses or goggles.
As further shown in
In some examples, control unit 210 is configured to, based on the sensed data (e.g., image data captured by image capture devices 138, position information from GPS sensors), generate and render for display on display 203 a virtual surface comprising one or more virtual content items (e.g., virtual content items 124, 126 of
In some examples, in accordance with techniques described herein, control unit 210 is configured to, based on the sensed data, identify a specific gesture or combination of gestures performed by the user and, in response, perform an action. For example, in response to one identified gesture, control unit 210 may generate and render a specific user interface for display on electronic display 203 at a position locked relative to peripheral device 136. For example, control unit 210 can generate and render a user interface including one or more user interface elements (e.g., virtual buttons) on surface 220 of peripheral device 136. As explained herein, in accordance with techniques of the disclosure, control unit 210 may perform object recognition within image data captured by image capture devices 138 to identify peripheral device 136 and/or a hand 132, fingers, thumb, arm or another part of the user, and track movements, positions, configuration, etc., of the peripheral device 136 and/or identified part(s) of the user to identify pre-defined gestures performed by the user. In response to identifying a pre-defined gesture, control unit 210 takes some action, such as selecting an option from an option set associated with a user interface (e.g., selecting an option from a user interface menu), translating the gesture into input (e.g., characters), launching an application, manipulating virtual content (e.g., moving, rotating a virtual content item), generating and rendering virtual markings, generating and rending a laser pointer, or otherwise displaying content, and the like. For example, control unit 210 can dynamically generate and present a user interface, such as a menu, in response to detecting a pre-defined gesture specified as a “trigger” for revealing a user interface (e.g., turning peripheral device to a landscape or horizontal orientation (not shown)). In some examples, control unit 210 detects user input, based on the sensed data, with respect to a rendered user interface (e.g., a tapping gesture performed on a virtual user interface element). In some examples, control unit 210 performs such functions in response to direction from an external device, such as console 106, which may perform, object recognition, motion tracking and gesture detection, or any part thereof.
As an example, control unit 210 can utilize image capture devices 138A and 138B to analyze configurations, positions, movements, and/or orientations of peripheral device 136, hand 132 and/or arm 134 to identify a user interface gesture, selection gesture, stamping gesture, translation gesture, rotation gesture, drawing gesture, pointing gesture, etc., that may be performed by users with respect to peripheral device 136. The control unit 210 can render a virtual user interface (including virtual user interface elements) and/or a virtual surface (including any virtual content items) and enable the user to interface with the virtual user interface and/or virtual surface based on detection of a user interface gesture, selection gesture, stamping gesture, translation gesture, rotation gesture, and drawing gesture performed by the user with respect to the peripheral device, as described in further detail below.
In one example, in accordance with techniques described herein, surface 220 of peripheral device 136 is a presence-sensitive surface, such as a surface that uses capacitive, conductive, resistive, acoustic, or other technology to detect touch and/or hover input. In some examples, surface 220 of peripheral device 136 is a touchscreen (e.g., a capacitive touchscreen, resistive touchscreen, surface acoustic wave (SAW) touchscreen, infrared touchscreen, optical imaging touchscreen, acoustic pulse recognition touchscreen, or any other touchscreen). In such an example, peripheral device 136 can detect user input (e.g., touch or hover input) on surface 220. In some examples, surface 220 does not include a display and peripheral device 136 does not include a display.
In examples with a touchscreen for surface 220, peripheral device 136 may communicate detected user input to HMD 112 (and/or console 106 of
In this example, HMD 112 are glasses comprising a front frame including a bridge to allow the HMD 112 to rest on a user's nose and temples (or “arms”) that extend over the user's ears to secure HMD 112 to the user. In addition, HMD 112 of
As further shown in
In this example, HMD 112 includes one or more processors 302 and memory 304 that, in some examples, provide a computer platform for executing an operating system 305, which may be an embedded, real-time multitasking operating system, for instance, or other type of operating system. In turn, operating system 305 provides a multitasking operating environment for executing one or more software components 307, including application engine 340. As discussed with respect to the examples of
In general, console 106 is a computing device that processes image and tracking information received from cameras 102 (
In the example of
Software applications 317 of console 106 operate to provide an overall artificial reality application. In this example, software applications 317 include application engine 320, rendering engine 322, gesture detector 324, pose tracker 326, and user interface engine 328.
In general, application engine 320 includes functionality to provide and present an artificial reality application, e.g., a teleconference application, a gaming application, a navigation application, an educational application, training or simulation applications, or other application. Application engine 320 may include, for example, one or more software packages, software libraries, hardware drivers, and/or Application Program Interfaces (APIs) for implementing an artificial reality application on console 106. Responsive to control by application engine 320, rendering engine 322 generates 3D artificial reality content for display to the user by application engine 340 of HMD 112.
Application engine 320 and rendering engine 322 construct the artificial content for display to user 110 in accordance with current pose information for a frame of reference, typically a viewing perspective of HMD 112, as determined by pose tracker 326. Based on the current viewing perspective, rendering engine 322 constructs the 3D, artificial reality content which may in some cases be overlaid, at least in part, upon the real-world 3D environment of user 110. During this process, pose tracker 326 operates on sensed data received from HMD 112, such as movement information and user commands, and, in some examples, data from any external sensors 90 (
Pose tracker 326 may determine a current pose for peripheral device 136 and, in accordance with the current pose, trigger certain functionality associated with virtual content that is locked to peripheral device 136 (e.g., places a virtual content item onto a virtual surface, manipulates a virtual content item, generates and renders one or more virtual markings, generates and renders a laser pointer). In some examples, pose tracker 326 detects whether the HMD 112 is proximate to a physical position corresponding to a virtual surface (e.g., a virtual pinboard) to trigger rendering of virtual content.
User interface engine 328 is configured to generate virtual user interfaces for rendering in an artificial reality environment. User interface engine 328 generates a virtual user interface to include one or more virtual user interface elements 329, such as elements 142, 144, 146, and 148 described above with respect to
Console 106 may output this virtual user interface and other artificial reality content, via a communication channel, to HMD 112 for display at HMD 112. Rendering engine 322 receives pose information for peripheral device 136 to continually update the user interface position and pose to match that of the peripheral device 136, such as that of one of presence-sensitive surfaces 220.
Based on the sensed data from any of the image capture devices 138 or 102, presence-sensitive surfaces 220, or other sensor devices, gesture detector 324 analyzes the tracked motions, configurations, positions, and/or orientations of peripheral device 136 and/or objects (e.g., hands, arms, wrists, fingers, palms, thumbs) of the user to identify one or more gestures performed by user 110. More specifically, gesture detector 324 may analyze objects recognized within image data captured by image capture devices 138 of HMD 112 and/or sensors 90 and external cameras 102 to identify peripheral device 136 and/or a hand and/or arm of user 110, and track movements of the peripheral device 136, hand, and/or arm relative to HMD 112 to identify gestures performed by user 110. In some examples, gesture detector 324 may track movement, including changes to position and orientation, of the peripheral device 136, hand, digits, and/or arm based on the captured image data, and compare motion vectors of the objects to one or more entries in gesture library 330 to detect a gesture or combination of gestures performed by user 110. In some examples, gesture detector 324 may receive user inputs detected by presence-sensitive surface(s) of peripheral device and process the user inputs to detect one or more gestures performed by user 110 with respect to peripheral device 136.
Gesture detector 324 and gesture library 330 may be distributed, in whole or in part, to peripheral device 136 to process user inputs on peripheral device 136 to detect gestures. In such cases, presence-sensitive surface(s) 220 detects user inputs at locations of the surface. Peripheral device 136 executing gesture detector 324 can process the user inputs to detect one or more gestures of gesture library 330. Peripheral device 136 may send indications of the detected gestures to console 106 and/or HMD 112 to cause the console 106 and/or HMD 112 to responsively perform one or more actions. Peripheral device 136 may alternatively, or additionally, send indications of the user inputs at locations of the surface to console 106, and gesture detector 324 may process the user inputs to detect one or more gestures of gesture library 330.
Some entries in gesture library 330 may each define a gesture as a series or pattern of motion, such as a relative path or spatial translations and rotations of peripheral device 136, a user's hand, specific fingers, thumbs, wrists and/or arms. Some entries in gesture library 330 may each define a gesture as a configuration, position, and/or orientation of the peripheral device, user's hand and/or arms (or portions thereof) at a particular time, or over a period of time. Some entries in gesture library 330 may each define a gesture as one or more user inputs, over time, detected by presence-sensitive surface(s) 220 of peripheral device 136. Other examples of type of gestures are possible. In addition, each of the entries in gesture library 330 may specify, for the defined gesture or series of gestures, conditions that are required for the gesture or series of gestures to trigger an action, such as spatial relationships to a current field of view of HMD 112, spatial relationships to the particular region currently being observed by the user, as may be determined by real-time gaze tracking of the individual, a pose of peripheral device 136 within the current field of view of HMD 112, types of artificial content being displayed, types of applications being executed, and the like.
Each of the entries in gesture library 330 further may specify, for each of the defined gestures or combinations/series of gestures, a desired response or action to be performed by software applications 317. For example, in accordance with techniques of this disclosure, certain specialized gestures may be pre-defined such that, in response to detecting one of the pre-defined gestures, user interface engine 328 dynamically generates a user interface as an overlay to artificial reality content being displayed to the user, thereby allowing the user 110 to easily invoke a user interface for configuring HMD 112 and/or console 106 even while interacting with artificial reality content. In other examples, certain gestures may be associated with other actions, such as providing input, selecting virtual objects (including virtual content items and/or user interface elements), translating (e.g., moving, rotating) virtual objects, altering (e.g., scaling, annotating) virtual objects, making virtual markings, launching applications, and other actions.
As an example, gesture library 330 may include entries that describe a peripheral device gesture, such as user interface activation gesture, a menu scrolling gesture, a selection gesture, a stamping gesture, a translation gesture, rotation gesture, drawing gesture, and/or pointing gesture. Some of these gestures may also be performed using a virtual user interface. For example, a user may perform a drawing gesture with a peripheral device gesture (by manipulating poses of the peripheral device) or a user interface gesture (by interacting with a virtual user interface). Gesture detector 324 may process image data from image capture devices 138 to analyze configurations, positions, motions, and/or orientations of peripheral device 136 and/or a user's hand to identify a user interface gesture, selection gesture, stamping gesture, translation gesture, rotation gesture, drawing gesture, pointing gesture, etc. that may be performed by users with respect to peripheral device 136.
Gesture detector 324 may detect a user interface gesture performed by a user at a position corresponding to one of the virtual user interface elements of a virtual user interface generated by user interface engine 328. For example, rendering engine 322 may render the virtual user interface at a position in the artificial reality environment that corresponds to a position of peripheral device 136 in the physical environment. Rendering engine 322 renders user interface elements at positions within the virtual user interface that also correspond to positions of locations on the peripheral device 136. A gesture performed by a user at one of the positions of the virtual user interface elements is an indication of a gesture performed with respect to the virtual user interface element. For example, a user may perform a button press user interface gesture at a location on the peripheral device 136 that is encompassed by and overlaid by a virtual button in the artificial reality environment. In response to a detected user interface gesture that is performed by a user at a position corresponding to the virtual user interface element, the artificial reality system may perform one or more actions associated with the virtual user interface element.
Actions may include, for example, launching an application or performing some action within an application, modifying the virtual user interface, animating artificial reality content such as one of the user interface elements, closing an application, outputting artificial reality for display, configuring an application, modifying an application, or other action by console 106, HMD 112, or peripheral device 136.
The position corresponding to the virtual user interface element and at which the detected gesture was performed may be a location on a presence-sensitive surface 220. User interface engine 328 may store a mapping of locations on the presence-sensitive surface 220 to virtual user interface elements of a virtual user interface. In response to receiving an indication of a detected gesture at a location of the presence-sensitive surface 220, user interface engine 328 may map the location to the virtual user interface element. The artificial reality system may then perform one or more actions associated with the virtual user interface element.
In some examples, the position corresponding to the virtual user interface element and at which the detected gesture was performed may be a location on the peripheral device 136 that is not a presence-sensitive surface. User interface engine 328 may store a mapping of locations on peripheral device to virtual user interface elements of a virtual user interface. In response to receiving an indication of a detected gesture at a location of peripheral device 136, user interface engine 328 may map the location to the virtual user interface element. The artificial reality system may then perform one or more actions associated with the virtual user interface element.
In response to a detected gesture, virtual user interface engine 322 may modify the virtual user interface. Virtual user interface engine 322 may drop down a selection list in response to a detected selection gesture, animate a button press in response to a button press gesture, generate markings at locations in the virtual user interface, and other user interface-type animations or modifications associated with user interfaces. Rendering engine 322 renders the modified virtual user interface at the appropriate user interface position, locked to peripheral device 136.
In the example shown in
In some examples, each of processors 302, 312, 346 may comprise any one or more of a multi-core processor, a controller, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field-programmable gate array (FPGA), or equivalent discrete or integrated logic circuitry. Memory 304, 314, 344 may comprise any form of memory for storing data and executable software instructions, such as random-access memory (RAM), read only memory (ROM), programmable read only memory (PROM), erasable programmable read only memory (EPROM), electronically erasable programmable read only memory (EEPROM), and flash memory.
If peripheral device 136 includes an output display, such as a presence-sensitive surface 220 that is a touchscreen, console 106 may send a communication to peripheral device 136 directing it to deactivate (i.e., turn off) the output display. Rendering engine 322 may, in some examples, only render virtual user interface 137 when the output display for the peripheral device 136 is deactivated.
In this example, similar to
In the example of
Similar to the examples described with respect to
Similar to the examples described with respect to
Gesture detector 424 may track movement, including changes to position and orientation, of the peripheral device, hand (including digits), and/or arm based on the captured image data, and compare motion vectors of the objects to one or more entries in gesture library 430 to detect a gesture or combination of gestures performed by user 110. Gesture library 430 is similar to gesture library 330 of
Artificial reality system 10 obtains image data from one or more image capture devices (502) and processes the image data to detect a peripheral device 136 in a field of view of HMD 112 (504). Artificial reality system 10 generates a virtual user interface 137 having one or more virtual user interface elements (506). Artificial reality system 10 renders the virtual user interface 137, at a user interface position locked relative to a position of the peripheral device 136 in an artificial reality environment, virtual user interface 137 for display at HMD 112 (508). Artificial reality system 10 may also render the peripheral device, or a representation thereof (e.g., a peripheral device avatar), with the virtual user interface 137 rendered as overlaid on the rendered peripheral device representation.
Artificial reality system 10 detects a user interface gesture at a position in the artificial reality environment that corresponds to one of the virtual user interface elements (510). In response to the user interface gesture, artificial reality system 10 performs one or more actions associated with the virtual user interface element.
The HMD detects the user interface gesture performed by a user at a position corresponding to the rendered virtual elements 604. In one or more aspects, the touch gestures performed on the surface of peripheral device 136 can be detected from image data captured by an image capture device of the artificial reality system, as noted above. In one or more aspects, the surface of peripheral device 136 is a presence-sensitive surface at which the peripheral device 136 detects the touch gestures, as noted above.
In response to detecting a user interface gesture selecting or touching element 604A, the artificial reality system will perform one or more actions, such as annotate active virtual content item 616 to indicate that the user “loves” or “likes” virtual content item 616. In some examples, the “love” or “like” annotation will be reflected on virtual content item 616 (e.g., element 608 will be generated over virtual content item 616 as shown in
In response to detecting the user input (e.g., the drawing gestures) at the surface of peripheral device 136, the artificial reality system generates a modified virtual user interface comprising virtual marks 706 for display at the HMD at the locations of the surface of peripheral device 136 where the drawing gestures were performed, as illustrated in artificial content 122 of
As shown in
For example, a user may perform user interface gestures with respect to peripheral device 136 to perform various manipulations with respect to the virtual markings using peripheral device 136. The user may perform touch user interface gestures, such as a pinch-to-zoom gesture to expand or shrink the virtual markings rendered on peripheral device 136. For example, the user may press a thumb of hand 132A and a thumb of hand 132B at two positions on the virtual user interface 722 and may “pinch” the fingers closer together (to zoom out or shrink the markings or other virtual user interface elements) or “spread” the fingers further apart (to zoom in or enlarge the markings or other virtual user interface elements). The user may perform other user interface gestures to manipulate the virtual markings on peripheral device 136, such as rotating the virtual markings (e.g., placing two fingers on the virtual user interface 722 and rotating the placement of the fingers) or moving the virtual markings (e.g., placing two fingers of the virtual user interface 722 and sliding the two fingers in one direction). In other examples (not shown), the user may also perform the various manipulations to the virtual user interface as a whole or other virtual user interface elements.
In one or more aspects, the touch gestures performed on the surface of peripheral device 136 can be detected from image data captured by an image capture device of the artificial reality system, as noted above. In one or more aspects, the surface of peripheral device 136 is a presence-sensitive surface at which the peripheral device 136 detects the touch gestures, as noted above.
In response to detecting the user input (e.g., the user interface gesture), the artificial reality system may perform one or more actions such as generate a modified virtual user interface comprising modified virtual markings (e.g., enlarged virtual markings) based on the user input.
Virtual markings 706 do not exist outside of artificial reality content 122 (e.g., cannot be seen without an HMD), and are rendered at a user interface position locked relative to a position of peripheral device 136 in the artificial reality environment. In one or more aspects, virtual markings 706 are generated and rendered as the touch input is detected. For example, the virtual markings 706 may be enlarged as the thumbs of hands 132A and 132B, respectively, are spread apart.
In one or more aspects, artificial reality content items can comprise GIFs, photos, applications, live-streams, videos, text, a web-browser, drawings, animations, representations of data files, or any other visible media. In one or more aspects, the artificial reality system generates and renders a virtual user interface 802 with a virtual pointer 805 overlaid on a surface of peripheral device 136 for display at the HMD. In one or more aspects, virtual user interface 802 includes virtual elements, such as a draw element 806, move element 808, rotate element 810, scale element 812, and/or a settings element 814. In one or more aspects, virtual user interface 802 can include other combination of user interface elements. In one or more aspects, virtual user interface 802 is rendered for display at the HMD (e.g., virtual user interface 802 is overlaid on a surface of peripheral device 136 by the artificial reality system as shown in
In some examples, the rendering engine of the artificial reality system may render a directional pad to control the movement of a selected artificial reality content. For example, a user may select move element 806, select an image via the virtual pointer 804, and perform an interface gesture on the directional pad to move artificial reality content 816.
In some examples, a user may press and hold the scale element 812, move the peripheral device 136 or virtual pointer 804 to scale the artificial reality content 816, and release the scale element 812 to stop scaling the artificial reality content 816. In some examples, a user may make a first press on the scale element 812, move the peripheral device 136 or virtual pointer 804 to scale the artificial reality content 816, and make a second press on the scale element 812 to stop rotating the artificial reality content 816. In some examples, the rendering engine of the artificial reality system may generate a modified virtual user interface comprising a modified virtual user interface element including a virtual directional pad to control the scaling.
The HMD detects the interface gesture performed by a user at a position corresponding to the rendered virtual elements 894. In one or more aspects, the touch gestures performed on the surface of peripheral device 136 can be detected from image data captured by an image capture device of the artificial reality system, as noted above. In one or more aspects, the surface of peripheral device 136 is a presence-sensitive surface at which the peripheral device 136 detects the touch gestures, as noted above.
In response to detecting a gesture selecting or touching element 894A, the artificial reality system will perform one or more actions, such as annotate active virtual content item 816 to indicate that the user “loves” or “likes” virtual content item 816. In some examples, the “love” or “like” annotation will be reflected on virtual content item 816 (e.g., element 896 will be generated over virtual content item 816 as shown in
Virtual button 146 corresponds to a position on surface 220. An area of virtual button 146 in the artificial reality environment corresponds to area 146′ for virtual button 146 on surface 220. This correspondence may include, for instance, that the area of virtual button 146 in the artificial reality environment encompasses or maps to an area 146′ on surface 122 in the physical environment where virtual button 146 is virtually rendered onto a video feedthrough from HMD-mounted or other cameras and displayed to the user.
The artificial reality system may detect a user interface gesture performed by a user at area 146′, i.e., at a position corresponding to virtual button 146 on surface 200. The artificial reality system detection may detect the user interface gesture by processing inputs detected by surface 200, or by analyzing objects recognized within image data captured by image capture devices and/or sensors and external cameras to identify peripheral device 136 and/or a hand and/or arm of a user, and tracking movements of peripheral device 136, hand, and/or arm relative to an HMD to identify gestures performed by the user.
The techniques described in this disclosure may be implemented, at least in part, in hardware, software, firmware or any combination thereof. For example, various aspects of the described techniques may be implemented within one or more processors, including one or more microprocessors, DSPs, application specific integrated circuits (ASICs), field programmable gate arrays (FPGAs), or any other equivalent integrated or discrete logic circuitry, as well as any combinations of such components. The term “processor” or “processing circuitry” may generally refer to any of the foregoing logic circuitry, alone or in combination with other logic circuitry, or any other equivalent circuitry. A control unit comprising hardware may also perform one or more of the techniques of this disclosure.
Such hardware, software, and firmware may be implemented within the same device or within separate devices to support the various operations and functions described in this disclosure. In addition, any of the described units, modules or components may be implemented together or separately as discrete but interoperable logic devices. Depiction of different features as modules or units is intended to highlight different functional aspects and does not necessarily imply that such modules or units must be realized by separate hardware or software components. Rather, functionality associated with one or more modules or units may be performed by separate hardware or software components or integrated within common or separate hardware or software components.
The techniques described in this disclosure may also be embodied or encoded in a computer-readable medium, such as a computer-readable storage medium, containing instructions. Instructions embedded or encoded in a computer-readable storage medium may cause a programmable processor, or other processor, to perform the method, e.g., when the instructions are executed. Computer readable storage media may include random access memory (RAM), read only memory (ROM), programmable read only memory (PROM), erasable programmable read only memory (EPROM), electronically erasable programmable read only memory (EEPROM), flash memory, a hard disk, a CD-ROM, a floppy disk, a cassette, magnetic media, optical media, or other computer readable media.
As described by way of various examples herein, the techniques of the disclosure may include or be implemented in conjunction with an artificial reality system. As described, artificial reality is a form of reality that has been adjusted in some manner before presentation to a user, which may include, e.g., a virtual reality (VR), an augmented reality (AR), a mixed reality (MR), a hybrid reality, or some combination and/or derivatives thereof. Artificial reality content may include completely generated content or generated content combined with captured content (e.g., real-world photographs). The artificial reality content may include video, audio, haptic feedback, or some combination thereof, and any of which may be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional effect to the viewer). Additionally, in some embodiments, artificial reality may be associated with applications, products, accessories, services, or some combination thereof, that are, e.g., used to create content in an artificial reality and/or used in (e.g., perform activities in) an artificial reality. The artificial reality system that provides the artificial reality content may be implemented on various platforms, including a head mounted device (HMD) connected to a host computer system, a standalone HMD, a mobile device or computing system, or any other hardware platform capable of providing artificial reality content to one or more viewers.