This disclosure relates to optical systems such as optical systems in electronic devices having displays.
Electronic devices can include displays that provide images near the eyes of a user. Such electronic devices often include virtual or augmented reality headsets with displays having optical elements that allow users to view the displays. If care is not taken, components used to display images can be bulky and might not exhibit desired levels of optical performance. For example, scattered light can increase background noise and limit contrast associated with sensing operations performed by the displays.
An electronic device may have a display system for providing image light to an eye box. The display system may include a waveguide. A projector may generate image light. An input coupler may couple the image light into the waveguide. An output coupler may couple the image light out of the waveguide and towards the eye box.
The display system may include an optical emitter that emits infrared light. A first optical coupler may couple the infrared light into the waveguide. A second optical coupler may couple the infrared light out of the waveguide and towards the eye box. The infrared light may reflect off an eye in the eye box as reflected light. The second optical coupler may couple the reflected light into the waveguide. The first optical coupler may couple the reflected light out of the waveguide and towards an infrared camera. The infrared camera may generate sensor data based on the reflected light. Control circuitry may perform gaze tracking operations based on the sensor data.
The display system may sequentially illuminate different regions of the eye with the infrared light at different times. This may minimize infrared light scattering, which minimizes background generation and maximizes signal-to-noise ratio in the sensor data generated by the infrared camera. To illuminate the different regions, the display system may include a scanning mirror that couples the light into the waveguide at different angles at different times, the optical emitter may include an array of light sources with include sets (e.g., columns) of light sources that are sequentially activated, and/or the optical emitter may emit light at different wavelengths that are directed in different directions by diffractive gratings.
System 10 of
The operation of system 10 may be controlled using control circuitry 16. Control circuitry 16 may include storage and processing circuitry for controlling the operation of system 10. Control circuitry 16 may include storage such as hard disk drive storage, nonvolatile memory (e.g., electrically-programmable-read-only memory configured to form a solid state drive), volatile memory (e.g., static or dynamic random-access-memory), etc. Processing circuitry in control circuitry 16 may include one or more processors (e.g., microprocessors, microcontrollers, digital signal processors, baseband processors, etc.), power management units, audio chips, graphics processing units, application specific integrated circuits, and other integrated circuits. Software code may be stored on storage in control circuitry 16 and run on processing circuitry in control circuitry 16 to implement operations for system 10 (e.g., data gathering operations, operations involving the adjustment of components using control signals, image rendering operations to produce image content to be displayed for a user, etc.).
System 10 may include input-output circuitry such as input-output devices 12. Input-output devices 12 may be used to allow data to be received by system 10 from external equipment (e.g., a tethered computer, a portable device such as a handheld device or laptop computer, or other electrical equipment) and to allow a user to provide head-mounted device 10 with user input. Input-output devices 12 may also be used to gather information on the environment in which system 10 (e.g., head-mounted device 10) is operating. Output components in devices 12 may allow system 10 to provide a user with output and may be used to communicate with external electrical equipment. Input-output devices 12 may include sensors and other components 18 (e.g., image sensors for gathering images of real-world object that are digitally merged with virtual objects on a display in system 10, accelerometers, depth sensors, light sensors, haptic output devices, speakers, batteries, wireless communications circuits for communicating between system 10 and external electronic equipment, etc.).
Projectors 26 may include liquid crystal displays, organic light-emitting diode displays, laser-based displays, or displays of other types. Projectors 26 may include light sources, emissive display panels, transmissive display panels that are illuminated with illumination light from light sources to produce image light, reflective display panels such as digital micromirror display (DMD) panels and/or liquid crystal on silicon (LCOS) display panels that are illuminated with illumination light from light sources to produce image light 30, etc.
Optical systems 22 may form lenses that allow a viewer (see, e.g., a viewer's eyes at eye box 24) to view images on display(s) 20. There may be two optical systems 22 (e.g., for forming left and right lenses) associated with respective left and right eyes of the user. A single display 20 may produce images for both eyes or a pair of displays 20 may be used to display images. In configurations with multiple displays (e.g., left and right eye displays), the focal length and positions of the lenses formed by system 22 may be selected so that any gap present between the displays will not be visible to a user (e.g., so that the images of the left and right displays overlap or merge seamlessly).
If desired, optical system 22 may contain components (e.g., an optical combiner, etc.) to allow real-world light (sometimes referred to as world light) from real-world (external) objects such as object 28 to be combined optically with virtual (computer-generated) images such as virtual images in image light 30. In this type of system, which is sometimes referred to as an augmented reality system, a user of system 10 may view both real-world content (e.g., world light from object 28) and computer-generated content that is overlaid on top of the real-world content. Camera-based augmented reality systems may also be used in device 10 (e.g., in an arrangement in which a camera captures real-world images of object 28 and this content is digitally merged with virtual content at optical system 22).
System 10 may, if desired, include wireless circuitry and/or other circuitry to support communications with a computer or other external equipment (e.g., a computer that supplies display 20 with image content). During operation, control circuitry 16 may supply image content to display 20. The content may be remotely received (e.g., from a computer or other content source coupled to system 10) and/or may be generated by control circuitry 16 (e.g., text, other computer-generated content, etc.). The content that is supplied to display 20 by control circuitry 16 may be viewed by a viewer at eye box 24.
If desired, system 10 may include an optical sensor. The optical sensor may be used to gather optical sensor data associated with a user's eyes at eye box 24. The optical sensor may, for example, be a gaze tracking sensor that gathers optical sensor data such as gaze image data (gaze tracking image data or gaze tracking sensor data) from a user's eye at eye box 24. Control circuitry 16 may process the optical sensor data to identify and track the direction of the user's gaze in real time. Control circuitry 16 may perform any desired operations based on the tracked direction of the user's gaze over time.
As shown in
Infrared emitter(s) 8 may direct light 4 towards optical system 22. Optical system 22 may direct the light 4 emitted by infrared emitter(s) 8 towards eye box 24. Light 4 may enter the user's eye at eye box 24 and may reflect off portions (regions) of the user's eye such as the retina as reflected light 4R (sometimes referred to herein as reflected sensor light 4R). Optical system 22 may receive reflected light 4R and may direct reflected light 4R towards infrared sensor(s) 6. Infrared sensor(s) 6 may receive reflected light 4R from optical system 22 and may gather (e.g., generate, measure, sense, produce, etc.) optical sensor data in response to the received reflected light 4R. Infrared sensor(s) 6 may include an image sensor or camera (e.g., an infrared image sensor or camera), for example. Infrared sensor(s) 6 may include, for example, one or more image sensor pixels (e.g., arrays of image sensor pixels). The optical sensor data may include image sensor data (e.g., image data, infrared image data, one or more images, etc.). Infrared image sensor(s) 6 may pass the optical sensor data to control circuitry 16 for further processing.
If desired, waveguide 32 may also include one or more layers of holographic recording media (sometimes referred to herein as holographic media, grating media, or diffraction grating media) on which one or more diffractive gratings are recorded (e.g., holographic phase gratings, sometimes referred to herein as holograms, surface relief gratings, etc.). A holographic recording may be stored as an optical interference pattern (e.g., alternating regions of different indices of refraction) within a photosensitive optical material such as the holographic media. The optical interference pattern may create a holographic phase grating that, when illuminated with a given light source, diffracts light to create a three-dimensional reconstruction of the holographic recording. The holographic phase grating may be a non-switchable diffractive grating that is encoded with a permanent interference pattern or may be a switchable diffractive grating in which the diffracted light can be modulated by controlling an electric field applied to the holographic recording medium. Multiple holographic phase gratings (holograms) may be recorded within (e.g., superimposed within) the same volume of holographic medium if desired. The holographic phase gratings may be, for example, volume holograms or thin-film holograms in the grating medium. The grating medium may include photopolymers, gelatin such as dichromated gelatin, silver halides, holographic polymer dispersed liquid crystal, or other suitable holographic media.
Diffractive gratings on waveguide 32 may include holographic phase gratings such as volume holograms or thin-film holograms, meta-gratings, or any other desired diffractive grating structures. The diffractive gratings on waveguide 32 may also include surface relief gratings (SRGs) formed on one or more surfaces of the substrates in waveguide 32 (e.g., as modulations in thickness of a SRG medium layer), gratings formed from patterns of metal structures, etc. The diffractive gratings may, for example, include multiple multiplexed gratings (e.g., holograms) that at least partially overlap within the same volume of grating medium (e.g., for diffracting different colors of light and/or light from a range of different input angles at one or more corresponding output angles). Other light redirecting elements such as louvered mirrors may be used in place of diffractive gratings in waveguide 32 if desired.
As shown in
Optical system 22 may include one or more optical couplers (e.g., light redirecting elements) such as input coupler 34, cross-coupler 36, and output coupler 38. In the example of
Waveguide 32 may guide image light 30 down its length via total internal reflection. Input coupler 34 may be configured to couple image light 30 from projector 26 into waveguide 32 (e.g., within a total-internal reflection (TIR) range of the waveguide within which light propagates down the waveguide via TIR), whereas output coupler 38 may be configured to couple image light 30 from within waveguide 32 (e.g., propagating within the TIR range) to the exterior of waveguide 32 and towards eye box 24 (e.g., at angles outside of the TIR range). Input coupler 34 may include an input coupling prism, an edge or face of waveguide 32, a lens, a steering minor or liquid crystal steering element, diffractive grating structures (e.g., volume holograms, SRGs, etc.), partially reflective structures (e.g., louvered mirrors), or any other desired input coupling elements.
As an example, projector 26 may emit image light 30 in direction +Y towards optical system 22. When image light 30 strikes input coupler 34, input coupler 34 may redirect image light 30 so that the light propagates within waveguide 32 via total internal reflection towards output coupler 38 (e.g., in direction +X within the TIR range of waveguide 32). When image light 30 strikes output coupler 38, output coupler 38 may redirect image light 30 out of waveguide 32 towards eye box 24 (e.g., back along the Y-axis). In implementations where cross-coupler 36 is formed on waveguide 32, cross-coupler 36 may redirect image light 30 in one or more directions as it propagates down the length of waveguide 32 (e.g., towards output coupler 38 from a direction of propagation as coupled into the waveguide by the input coupler). In redirecting image light 30, cross-coupler 36 may also perform pupil expansion on image light 30 in one or more directions. In expanding pupils of the image light, cross-coupler 36 may, for example, help to reduce the vertical size of waveguide 32 (e.g., in the Z direction) relative to implementations where cross-coupler 36 is omitted. Cross-coupler 36 may therefore sometimes also be referred to herein as pupil expander 36 or optical expander 36. If desired, output coupler 38 may also expand image light 30 upon coupling the image light out of waveguide 32.
Input coupler 34, cross-coupler 36, and/or output coupler 38 may be based on reflective and refractive optics or may be based on diffractive (e.g., holographic) optics. In arrangements where couplers 34, 36, and 38 are formed from reflective and refractive optics, couplers 34, 36, and 38 may include one or more reflectors (e.g., an array of micromirrors, partial mirrors, louvered mirrors, or other reflectors). In arrangements where couplers 34, 36, and 38 are based on diffractive optics, couplers 34, 36, and 38 may include diffractive gratings (e.g., volume holograms, surface relief gratings, etc.).
The example of
The operation of optical system 22 on image light 30 is shown in
As shown in
Optical system 22 may include at least a first optical coupler 44 and a second optical coupler 65 for use in performing optical sensing for gaze tracking sensor 40 (e.g., for redirecting light 4 and reflected light 4R). Optical couplers 44 and 65 may be disposed at, on, or within waveguide 32. Optical coupler 44 may also redirect image light 30 produced by projector 26 (e.g., optical coupler 44 may also form input coupler 34, cross-coupler 36, and/or output coupler 38 of
When a user is wearing or using system 10 (
Optical coupler 44 may form an input coupler for the light 4 emitted by gaze tracking sensor 40. Optical coupler 44 may therefore couple light 4 incident upon optical system 22 from incident angles outside the TIR range of waveguide 32 into waveguide 32 (e.g., at output angles within the TIR range of the waveguide). Optical coupler 44 may also form an output coupler for the reflected light 4R received by optical system 22 after reflection off eye 58. Optical coupler 44 may therefore couple reflected light 4R incident upon optical coupler 44 at incident angles within the TIR range of waveguide 32 (e.g., after propagating along waveguide 32 via TIR) out of waveguide 32 and towards gaze tracking sensor 40 (e.g., at output angles outside the TIR range of waveguide 32).
Optical coupler 65 may form an output coupler for the light 4 propagating along waveguide 32 via TIR. Optical coupler 65 may therefore couple light 4 incident upon optical coupler 65 from incident angles within the TIR range of waveguide 32 out of waveguide 32 and towards eye box 24 (e.g., at output angles outside the TIR range of the waveguide). Optical coupler 65 may also form an input coupler for the reflected light 4R received by optical system 22 after reflection off eye 58. Optical coupler 65 may therefore couple reflected light 4R incident upon optical coupler 65 at incident angles outside the TIR range of waveguide 32 into waveguide 32 (e.g., at output angles within the TIR range of waveguide 32).
Optical coupler 44 and optical coupler 65 may each include prisms, mirrors, partial reflectors (e.g., louvered mirrors), volume holograms, surface relief gratings (SRGs), meta-gratings, waveguide facets, lenses, and/or any other desired optical coupling structures. Optical coupler 44 may include, for example, a prism such as prism 46 whereas optical coupler 65 includes one or more SRGs or volume holograms.
In the example of
In general, it may be desirable for gaze tracking sensor 40 to gather optical sensor data (images) of multiple different regions (areas or portions) 56 of eye 58 while performing optical sensing at eye box 24. The different regions may, for example, correspond to different physiological features on the retina of eye 58. These physiological features may help control circuitry 16 (
In some implementations, gaze tracking sensor 40 illuminates each of the multiple regions 56 at the same time and thus receives reflected light 4R from each of the multiple regions 56 at the same time. Each region 56 may be illuminated by a different respective optical mode of the system. For example, the system may include at least a first optical mode (propagation direction) that illuminates a first region 56-1 (as shown by arrow 54) and a second optical mode (propagation direction) that simultaneously illuminates a second region 56-2 (as shown by arrow 64).
However, not all of the light redirected by optical coupler 65 is coupled into or out of eye 58. At least some of the light from each optical mode will leak in other directions, such as towards skin 60, which will undesirably reflect or scatter the light in different directions (as optical scattering 62). For example, as shown in
To mitigate these issues, gaze tracking sensor 40 and optical system 22 may sequentially illuminate each of the multiple different regions 56 on eye 58 in series (e.g., at different times in a time-division duplexed manner). Optical system 22 and/or gaze tracking sensor 40 may, for example, include an adjustable or tunable optical component that allows gaze tracking sensor 40 and optical system 22 to sequentially illuminate each of the multiple different regions 56 on eye 58 in series. The adjustable or tunable optical component may include a scanning mirror, a selectively adjustable array of light sources, or light source(s) having variable wavelengths, as examples.
Scanning mirror 42 may overlap reflective face (surface) 48 of prism 46. Scanning mirror 42 may receive light 4 from gaze tracking sensor 40 through waveguide 32 and prism 46 and may reflect light 4 into waveguide 32 through prism 46. Similarly, scanning mirror 42 may receive reflected light 4R from waveguide 32 through prism 46 and may reflect the reflected light 4R through prism 46 and waveguide 32 towards gaze tracking sensor 40.
Scanning mirror 42 may be adjustable between multiple different orientations. In each orientation, scanning mirror 42 may reflect light 4 towards and may receive reflected light 4R from different regions 56 of eye 58. For example, as shown in
More particularly, in the first orientation, scanning mirror 42 may reflect light 4 at a first angle into waveguide 32, which propagates light 4 towards optical coupler 65. Optical coupler 65 may couple (diffract) light 4 out of waveguide 32 at the same angles at which optical coupler 65 couples (diffracts) reflected light 4R into waveguide 32 (e.g., the Bragg-matching condition of optical coupler 65 may be such that optical coupler 65 directs light 4 onto an angle that is 180 degrees opposite the angle at which it receives reflected light 4R and directs reflected light 4R onto the angle that is 180 degrees opposite the angle at which it receives light 4). Optical coupler 65 therefore receives light 4 and couples (e.g., diffracts) light 4 out of waveguide 32 and towards region 56-1 of eye 58. Light 4 reflects off region 56-1 towards optical coupler as reflected light 4R. Optical coupler 65 couples reflected light 4R into waveguide 32 such that reflected light 4R propagates along waveguide 32 and is received at scanning minor 42 (in the first orientation) on-axis with the light 4 reflected off scanning minor 42 (in the first orientation).
In the second orientation 50, scanning minor 42 may reflect light 4 at a second angle into waveguide 32, which propagates light 4 towards optical coupler 65. Optical coupler 65 receives light 4 and couples (e.g., diffracts) light 4 out of waveguide 32 and towards region 56-2 of eye 58. Light 4 reflects off region 56-2 towards optical coupler 65 as reflected light 4R. Optical coupler 65 couples reflected light 4R into waveguide 32 such that reflected light 4R propagates along waveguide 32 and is received at scanning mirror 42 (in the second orientation) on-axis with the light 4 reflected off scanning mirror 42 (in the second orientation).
In the example of
In the example of
If desired, scanning mirror 42 may be omitted and the adjustable or tunable optical component that allows gaze tracking sensor 40 and optical system 22 to sequentially illuminate each of the multiple different regions 56 may include a selectively adjustable array of light sources in infrared emitter(s) 8.
As shown in
For example, control signals CTRL may first activate each light source 74 in the first set (e.g., column) 72-1 of light sources 74 so set 72-1 emits light 4 while the other sets 72 of light sources 74 are inactive (e.g., do not emit light 4). Control signals CTRL may then activate each light source 74 in the second set (e.g., column) 72-2 of light sources 74 so set 72-2 emits light 4 while the other sets 72 of light sources 74 are inactive. Different sets 72 may be illuminated in series in this way until the Nth set 72-N is illuminated.
When sets 72 are arranged in the rectangular grid pattern of
Each set 72 may, for example, illuminate a different rectangular (column-shaped or 1D) region of the retina and scanning (selectively activating) each set 72 in series may effectively produce or paint a two-dimensional (2D) patch of illumination across each of the multiple regions 56 of eye 58. If desired, one or more of the axes of the 2D array of light sources 74 (e.g., the directions of the rows or columns of light sources 74) may be tilted with respect to one or more of the axes of system 10. While infrared sensor(s) 6 of
One or more optical diffusers such as at least a first diffuser 90 and a second diffuser 88 may be optically interposed between collimating lens 70 and infrared emitter(s) 8. First diffuser 90 may be optically interposed between second diffuser 88 and infrared emitter(s) 8. Second diffuser 88 may be optically interposed between collimating lens 70 and first diffuser 90. First diffuser 90 may be, for example, a 2D diffuser that diffuses the light 4 emitted by infrared emitter(s) 8 along both the first dimension D1 and the second dimension D2 of the array (
As shown in
If desired, the adjustable or tunable optical component that allows gaze tracking sensor and optical system 22 to sequentially illuminate each of the multiple different regions 56 may include a light source in infrared emitter(s) 8 that is adjusted to produce light 4 at different wavelengths at different times. For example, infrared emitter(s) 8 (
In these implementations, if care is not taken, light 4 and reflected light 4R will follow the same optical mode of propagation at each of the wavelengths. Optical system 22 may therefore include diffractive gratings that direct light 4 at different wavelengths to different regions 56 on eye 58 (and that direct reflected light 4 at different wavelengths from different regions 56 towards infrared sensor(s) 6). The diffractive gratings may include volume holograms in optical coupler 65, as one example. The volume holograms may be constant-pitch volume holograms if desired.
Each volume hologram 94 may be defined by a corresponding grating vector k. The grating vector k may have a direction in three-dimensional space that is normal to the plane of the fringes (e.g., lines of constant refractive index) of the hologram. The volume holograms 94 in optical coupler 65 may be constant-pitch volume holograms that have the same pitch (e.g., the same periodicity of fringes within substrate 65) but with different orientations.
Each hologram 94 may diffract a different respective wavelength of light 4 incident from the same direction onto a different respective one of the N regions 56 on eye 58. For example, optical coupler 65 may include at least a first volume hologram 94-1 defined by a first grating vector k1 and having fringes at a first orientation, a second volume hologram 94-2 defined by a second grating vector k2 and having fringes at a second orientation different from the first orientation, and an Nth volume hologram 94-N defined by an Nth grating vector kN and having fringes at an Nth orientation that is different from the first and second orientations. First volume hologram 94-1 may direct light 4 of a first wavelength and incident at a given incident angle towards region 56-1 on eye 58. Second volume hologram 94-2 may direct light 4 of a second wavelength and incident at the given incident angle towards region 56-2 on eye 58. Nth volume hologram 94-N may direct light 4 of an Nth wavelength and incident at the given incident angle towards region 56-N on eye 58. The volume holograms may conversely direct reflected light 4R from each of the regions onto the same output angle towards infrared sensor(s) 6 (
By sequentially controlling the tunable light source in infrared emitter(s) 8, different regions 56 may be illuminated with light 4 at different times. The example of
As shown in
At operation 110, infrared emitter(s) 8 may emit light 4. Optical system 22 and infrared emitter(s) 8 may sequentially illuminate N different regions 56 on eye 58 using the emitted light 4. Optical system 22 and/or infrared emitter(s) 8 may sequentially illuminate the N different regions 56 by sequentially rotating scanning mirror 42 through different orientations/angles (
At operation 112, control circuitry 16 may process the optical sensor data to identify (e.g., detect, generate, measure, sense, etc.) a gaze direction and/or other optical characteristics associated with eye 58 at eye box 24. Control circuitry 16 may, for example, detect different physiological features of eye box 24 associated with the N different regions 56 (e.g., using an object detection algorithm). Control circuitry 16 may identify the gaze direction and/or other optical characteristics associated with eye 58 based on the detected physiological features. If desired, control circuitry 16 may detect gaze by generating a gaze vector oriented in the direction of the eye's gaze. Control circuitry may track the direction of the user's gaze and/or the other optical characteristics over time.
At operation 114, control circuitry 16 may take any desired action based on the identified gaze direction and/or other optical characteristics. As one example, control circuitry 16 may adjust the image data used by projector(s) 26 (
By sequentially scanning over different regions 56 on eye 58, significant background signal due to diffuse scattering off skin 60, specular corneal reflections, and other potential sources can be eliminated from the optical sensor data gathered by infrared sensor(s) 6, thereby maximizing the SNR of the desired optical sensor data associated with regions 56. In implementations where each region 56 is simultaneously illuminated, light reflected from the skin creates a haze over the whole sensor as it is highly defocused. However, sequentially illuminating each region 56 only illuminates a single region 56 on the retina at any given time, thereby eliminating most of the haze caused by the skin and can be ignored by processing circuitry 16 when stitching images of each region 56 together to obtain a full image of the retina for use in gaze tracking.
As used herein, the term “concurrent” means at least partially overlapping in time. In other words, first and second events are referred to herein as being “concurrent” with each other if at least some of the first event occurs at the same time as at least some of the second event (e.g., if at least some of the first event occurs during, while, or when at least some of the second event occurs). First and second events can be concurrent if the first and second events are simultaneous (e.g., if the entire duration of the first event overlaps the entire duration of the second event in time) but can also be concurrent if the first and second events are non-simultaneous (e.g., if the first event starts before or after the start of the second event, if the first event ends before or after the end of the second event, or if the first and second events are partially non-overlapping in time). As used herein, the term “while” is synonymous with “concurrent.”
As described above, one aspect of the present technology is the gathering and use of information such as information from input-output devices. The present disclosure contemplates that in some instances, data may be gathered that includes personal information data that uniquely identifies or can be used to contact or locate a specific person. Such personal information data can include demographic data, location-based data, telephone numbers, email addresses, twitter ID's, home addresses, data or records relating to a user's health or level of fitness (e.g., vital signs measurements, medication information, exercise information), date of birth, username, password, biometric information, or any other identifying or personal information.
The present disclosure recognizes that the use of such personal information, in the present technology, can be used to the benefit of users. For example, the personal information data can be used to deliver targeted content that is of greater interest to the user. Accordingly, use of such personal information data enables users to have control of the delivered content. Further, other uses for personal information data that benefit the user are also contemplated by the present disclosure. For instance, health and fitness data may be used to provide insights into a user's general wellness, or may be used as positive feedback to individuals using technology to pursue wellness goals.
The present disclosure contemplates that the entities responsible for the collection, analysis, disclosure, transfer, storage, or other use of such personal information data will comply with well-established privacy policies and/or privacy practices. In particular, such entities should implement and consistently use privacy policies and practices that are generally recognized as meeting or exceeding industry or governmental requirements for maintaining personal information data private and secure. Such policies should be easily accessible by users, and should be updated as the collection and/or use of data changes. Personal information from users should be collected for legitimate and reasonable uses of the entity and not shared or sold outside of those legitimate uses. Further, such collection/sharing should occur after receiving the informed consent of the users. Additionally, such entities should consider taking any needed steps for safeguarding and securing access to such personal information data and ensuring that others with access to the personal information data adhere to their privacy policies and procedures. Further, such entities can subject themselves to evaluation by third parties to certify their adherence to widely accepted privacy policies and practices. In addition, policies and practices should be adapted for the particular types of personal information data being collected and/or accessed and adapted to applicable laws and standards, including jurisdiction-specific considerations. For instance, in the United States, collection of or access to certain health data may be governed by federal and/or state laws, such as the Health Insurance Portability and Accountability Act (HIPAA), whereas health data in other countries may be subject to other regulations and policies and should be handled accordingly. Hence different privacy practices should be maintained for different personal data types in each country.
Despite the foregoing, the present disclosure also contemplates embodiments in which users selectively block the use of, or access to, personal information data. That is, the present disclosure contemplates that hardware and/or software elements can be provided to prevent or block access to such personal information data. For example, the present technology can be configured to allow users to select to “opt in” or “opt out” of participation in the collection of personal information data during registration for services or anytime thereafter. In another example, users can select not to provide certain types of user data. In yet another example, users can select to limit the length of time user-specific data is maintained. In addition to providing “opt in” and “opt out” options, the present disclosure contemplates providing notifications relating to the access or use of personal information. For instance, a user may be notified upon downloading an application (“app”) that their personal information data will be accessed and then reminded again just before personal information data is accessed by the app.
Moreover, it is the intent of the present disclosure that personal information data should be managed and handled in a way to minimize risks of unintentional or unauthorized access or use. Risk can be minimized by limiting the collection of data and deleting data once it is no longer needed. In addition, and when applicable, including in certain health related applications, data de-identification can be used to protect a user's privacy. De-identification may be facilitated, when appropriate, by removing specific identifiers (e.g., date of birth, etc.), controlling the amount or specificity of data stored (e.g., collecting location data at a city level rather than at an address level), controlling how data is stored (e.g., aggregating data across users), and/or other methods.
Therefore, although the present disclosure broadly covers use of information that may include personal information data to implement one or more various disclosed embodiments, the present disclosure also contemplates that the various embodiments can also be implemented without the need for accessing personal information data. That is, the various embodiments of the present technology are not rendered inoperable due to the lack of all or a portion of such personal information data.
Physical environment: A physical environment refers to a physical world that people can sense and/or interact with without aid of electronic systems. Physical environments, such as a physical park, include physical articles, such as physical trees, physical buildings, and physical people. People can directly sense and/or interact with the physical environment, such as through sight, touch, hearing, taste, and smell.
Computer-generated reality: in contrast, a computer-generated reality (CGR) environment refers to a wholly or partially simulated environment that people sense and/or interact with via an electronic system. In CGR, a subset of a person's physical motions, or representations thereof, are tracked, and, in response, one or more characteristics of one or more virtual objects simulated in the CGR environment are adjusted in a manner that comports with at least one law of physics. For example, a CGR system may detect a person's head turning and, in response, adjust graphical content and an acoustic field presented to the person in a manner similar to how such views and sounds would change in a physical environment. In some situations (e.g., for accessibility reasons), adjustments to characteristic(s) of virtual object(s) in a CGR environment may be made in response to representations of physical motions (e.g., vocal commands). A person may sense and/or interact with a CGR object using any one of their senses, including sight, sound, touch, taste, and smell. For example, a person may sense and/or interact with audio objects that create 3D or spatial audio environment that provides the perception of point audio sources in 3D space. In another example, audio objects may enable audio transparency, which selectively incorporates ambient sounds from the physical environment with or without computer-generated audio. In some CGR environments, a person may sense and/or interact only with audio objects. Examples of CGR include virtual reality and mixed reality.
Virtual reality: A virtual reality (VR) environment refers to a simulated environment that is designed to be based entirely on computer-generated sensory inputs for one or more senses. A VR environment comprises a plurality of virtual objects with which a person may sense and/or interact. For example, computer-generated imagery of trees, buildings, and avatars representing people are examples of virtual objects. A person may sense and/or interact with virtual objects in the VR environment through a simulation of the person's presence within the computer-generated environment, and/or through a simulation of a subset of the person's physical movements within the computer-generated environment.
Mixed reality: In contrast to a VR environment, which is designed to be based entirely on computer-generated sensory inputs, a mixed reality (MR) environment refers to a simulated environment that is designed to incorporate sensory inputs from the physical environment, or a representation thereof, in addition to including computer-generated sensory inputs (e.g., virtual objects). On a virtuality continuum, a mixed reality environment is anywhere between, but not including, a wholly physical environment at one end and virtual reality environment at the other end. In some MR environments, computer-generated sensory inputs may respond to changes in sensory inputs from the physical environment. Also, some electronic systems for presenting an MR environment may track location and/or orientation with respect to the physical environment to enable virtual objects to interact with real objects (that is, physical articles from the physical environment or representations thereof). For example, a system may account for movements so that a virtual tree appears stationery with respect to the physical ground. Examples of mixed realities include augmented reality and augmented virtuality. Augmented reality: an augmented reality (AR) environment refers to a simulated environment in which one or more virtual objects are superimposed over a physical environment, or a representation thereof. For example, an electronic system for presenting an AR environment may have a transparent or translucent display through which a person may directly view the physical environment. The system may be configured to present virtual objects on the transparent or translucent display, so that a person, using the system, perceives the virtual objects superimposed over the physical environment. Alternatively, a system may have an opaque display and one or more imaging sensors that capture images or video of the physical environment, which are representations of the physical environment. The system composites the images or video with virtual objects, and presents the composition on the opaque display. A person, using the system, indirectly views the physical environment by way of the images or video of the physical environment, and perceives the virtual objects superimposed over the physical environment. As used herein, a video of the physical environment shown on an opaque display is called “pass-through video,” meaning a system uses one or more image sensor(s) to capture images of the physical environment, and uses those images in presenting the AR environment on the opaque display. Further alternatively, a system may have a projection system that projects virtual objects into the physical environment, for example, as a hologram or on a physical surface, so that a person, using the system, perceives the virtual objects superimposed over the physical environment. An augmented reality environment also refers to a simulated environment in which a representation of a physical environment is transformed by computer-generated sensory information. For example, in providing pass-through video, a system may transform one or more sensor images to impose a select perspective (e.g., viewpoint) different than the perspective captured by the imaging sensors. As another example, a representation of a physical environment may be transformed by graphically modifying (e.g., enlarging) portions thereof, such that the modified portion may be representative but not photorealistic versions of the originally captured images. As a further example, a representation of a physical environment may be transformed by graphically eliminating or obfuscating portions thereof. Augmented virtuality: an augmented virtuality (AV) environment refers to a simulated environment in which a virtual or computer generated environment incorporates one or more sensory inputs from the physical environment. The sensory inputs may be representations of one or more characteristics of the physical environment. For example, an AV park may have virtual trees and virtual buildings, but people with faces photorealistically reproduced from images taken of physical people. As another example, a virtual object may adopt a shape or color of a physical article imaged by one or more imaging sensors. As a further example, a virtual object may adopt shadows consistent with the position of the sun in the physical environment.
Hardware: there are many different types of electronic systems that enable a person to sense and/or interact with various CGR environments. Examples include head mounted systems, projection-based systems, heads-up displays (HUDs), vehicle windshields having integrated display capability, windows having integrated display capability, displays formed as lenses designed to be placed on a person's eyes (e.g., similar to contact lenses), headphones/earphones, speaker arrays, input systems (e.g., wearable or handheld controllers with or without haptic feedback), smartphones, tablets, and desktop/laptop computers. A head mounted system may have one or more speaker(s) and an integrated opaque display. Alternatively, a head mounted system may be configured to accept an external opaque display (e.g., a smartphone). The head mounted system may incorporate one or more imaging sensors to capture images or video of the physical environment, and/or one or more microphones to capture audio of the physical environment. Rather than an opaque display, a head mounted system may have a transparent or translucent display. The transparent or translucent display may have a medium through which light representative of images is directed to a person's eyes. The display may utilize digital light projection, OLEDs, LEDs, μLEDs, liquid crystal on silicon, laser scanning light sources, or any combination of these technologies. The medium may be an optical waveguide, a hologram medium, an optical combiner, an optical reflector, or any combination thereof. In one embodiment, the transparent or translucent display may be configured to become opaque selectively. Projection-based systems may employ retinal projection technology that projects graphical images onto a person's retina. Projection systems also may be configured to project virtual objects into the physical environment, for example, as a hologram or on a physical surface.
The foregoing is merely illustrative and various modifications can be made to the described embodiments. The foregoing embodiments may be implemented individually or in any combination.
This application claims the benefit of U.S. Provisional Patent Application No. 63/392,686, filed Jul. 27, 2022, which is hereby incorporated by reference herein in its entirety.
Number | Date | Country | |
---|---|---|---|
63392686 | Jul 2022 | US |