Near-Eye-Display (NED) systems superimpose computer-generated images (“CG images”) over a user's perspective of a physical real-world environment (referred to herein as a “real-world view”). For example, a NED system may generate composite views to enable a user to visually perceive a real-world view simultaneously with user interface (UI) menus, rendered images corresponding to multi-dimensional models (e.g. 2D and/or 3D models of virtual objects), or any other type of CG image. User perceived image quality is highly dependent on relative brightness (e.g. luminance) between the CG images and the real-world view. Various circumstances may result in inadequate brightness such that the CG images being generated by the NED system may be only faintly perceptible or even totally imperceptible to the user. For example, in instances where a brightness of the CG images is substantially lower than a brightness of the real-world view, the user may have difficulty perceiving the CG images.
In some instances, increasing CG image brightness with respect to the real-world view to make the CG images more readily perceptible may be well within the NED hardware's capability. In other instances, the NED hardware may be unable to reach a brightness level required for the CG images to become perceptible against the brightness of real-world view. In these instances, the NED would become less useful. Additionally, increasing the brightness of the CG images has numerous drawbacks, such as increasing the power draw of a NED system. Also, CG images are perceptually additive to the user's visual field and in some instances real world objects remain visible through CG imagery. Thus, CG objects do not appear solid and are sometimes described as having a “ghostly” appearance. Merely increasing the brightness of the CG images may also be ineffective at preventing these images from appearing as “ghostly images” as CG image brightness contributes to the eye's pupil response. Moreover, in instances where the CG images correspond to a rendered image of a virtual object being augmented into the real-world view, increasing the brightness of the rendered image (e.g. to overpower real world object brightness) may cause the virtual object to appear unnatural to the user. Such techniques may also present a number of inefficiencies with respect to the use of computing resources and energy resources.
It is with respect to these considerations and others that the disclosure made herein is presented.
Technologies described herein enable systems to generate dimming masks to enhance image quality of computer-generated images (“CG images”) within a real-world view. Generally described, the techniques disclosed herein enable a system to monitor eye tracking data to determine physical characteristics of a user's eye(s) (such as pupil diameter and/or gaze direction) and, based thereon, generate dimming masks with relation to CG images to decrease user perceived brightness of a real-world view (i.e. brightness of the real-world view from a user's perspective) in shaped regions where corresponding CG images are being rendered. Unlike conventional NED systems, the systems and techniques described herein are not limited to managing relative brightness between CG images and a real-world view by controlling the single variable of CG image brightness, which suffers from those drawbacks outlined above in addition to other drawbacks. Rather, the presently disclosed NED system is configured to control a perceived brightness of a real-world view with respect to CG images displayed to a user. In particular, the techniques described herein enable NED systems to dynamically alter optical properties of a transparent dimming panel to generate one or more dimming masks having a transmittance level that is less than a base transmittance of the transparent dimming panel.
As used herein, the term “contrast” may refer generally to a relationship between a luminance of one or more features of a CG image and a luminance of one or more features of a real-world view. For example, in some instances, contrast may correspond to “Weber” contrast which is defined as (I−Ib)/Ib where I represents the luminance of the one or more features of the CG image and Ib represents the luminance of the one or more features of the real-world view.
As used herein, a “transmittance level” refers generally to a proportion of incident light that propagates entirely through a physical object such as, for example, the transparent dimming panel and/or transparent display described herein. For example, a transmittance level of zero-percent corresponds to a fully opaque physical object (i.e. an object through which zero-percent of visible light is able to pass) whereas a transmittance level of one-hundred-percent corresponds to a fully transparent physical object (i.e. an object through which one-hundred-percent of visible light is able to pass). As used herein, a “base transmittance” may refer generally to a foundational level of transmittance of a physical object having transmittance level control capabilities that enable the controlled increase and/or decrease of a transmittance level of one or more regions of the physical object. Techniques described herein provide for the generation of a dimming mask(s) on a transparent dimming panel having a base transmittance wherein the dimming mask(s) may have a lower transmittance level than the base transmittance, e.g. the dimming mask(s) is more opaque than other areas of the transparent dimming panel.
For illustrative purposes, consider a scenario where an augmented reality (AR) program instructs a Near-Eye-Display (NED) device to generate a composite view by superimposing a rendered image of a virtual object over some portion of a real-world view. For example, the AR program may cause the NED device to give the appearance that a soda can (e.g. the virtual object) is resting on top of a table that actually exists in the real-world environment and, therefore, is part of the real-world view. Suppose also that the real-world environment is sufficiently bright such that achieving a suitable relative brightness between the virtual object and the actual table would require the virtual object to be rendered at a brightness level that is unnaturally high or even beyond the capability or power budget of the NED. Under these exemplary circumstances, the techniques described herein enable the NED device to render the virtual object at a brightness level that appears natural with respect to the real-world environment while achieving the desired contrast by effectively “turning down” the brightness of the real-world view. In another instance, suppose that the real-world view includes a very bright object in the region where the virtual object is to be rendered. Under these circumstances, it is desirable to block nearly all light at the specific region where the virtual object is being displayed. Accordingly, the NED device is enabled to display the virtual object at a brightness level that makes the virtual object appear to exist naturally within the real-world environment while simultaneously blocking out light from the real-world view, reducing NED power consumption and/or preventing the rendered image of the virtual object from appearing as a “ghostly image.”
As used herein, a “ghostly image” refers generally to a CG image through which one or more portions of a real-world view remain perceptible to a user to an unacceptable degree. For example, consider an example where a user is looking at a woodgrain pattern of an actual table through a NED device, and the NED device is also displaying a rendered image of a virtual soda can. In this situation, the virtual soda can may be considered a ghostly image in the event that the woodgrain pattern of the actual table remains perceptible to the user through a center region of the rendered image of the virtual soda can. With the dimming mask feature, the wood grain would not be visible and the virtual soda can would appear more natural and solid. In some instances, it may be desirable to form a drop shadow for the virtual soda can, where some wood grain is still visible through a darkened shadow shape of the virtual can. The opacity level of each dimming mask zone may be set to provide the degree of transparency desired to modulate real-world visibility. Based on the discussion herein of the user perceived penumbras that may surround a dimming mask, it will be appreciated that in various configurations a NED device may be unable to completely eliminate the occurrence of ghosting, especially around the perimeter of a dimming mask.
In some configurations, a system (e.g. a direct-view NED device) may include a transparent display to generate CG images and a transparent dimming panel to generate one or more dimming masks. The transparent dimming panel may be substantially adjacent to the transparent display. The system may receive image data and, based thereon, cause the transparent display to generate one or more CG images to create a composite view, from the perspective of a user, that includes the CG images superimposed over the real-world view. The system may also include and/or communicate with an eye tracking sensor to monitor physical characteristics of the user's eyes (e.g. a pupil size and/or a gaze direction) to determine size parameters and location parameters corresponding to one or more dimming masks. Then, the system may cause the transparent dimming panel to generate the one or more dimming masks to block light, from a particular region of the real-world view, from reaching the user's pupil(s). For example, the system may dynamically decrease, from a base transmittance, a transmittance level of one or more regions of the transparent dimming panel to block light that is reflected off (or emitted from for that matter) the real-world objects from passing through the region(s) of the transparent dimming panel. Accordingly, the system may supplement the user's perspective of the real-world view with the CG images while controlling both an actual brightness of the CG images (e.g. increasing and/or decreasing an actual luminous intensity at which the device generates the CG images) and also a user perceived brightness of the real-world view based upon real time physical characteristics of the user's eyes (e.g. reducing an amount of light transmitted from a real-world object that passes through the region(s) of the transparent dimming panel.
In some configurations, the system may determine (solely or in conjunction with other factors) opacity parameters for the dimming mask(s) based on eye tracking data that indicates physical characteristics of the user's eyes such as, for example, a current pupil diameter. For example, as described elsewhere herein, in various implementations, the system may determine a transmittance level for a dimming mask based on a negative correlation with the pupil diameter of the user's eyes. Stated alternatively, as the user's pupil diameter increases, the system may decrease the transmittance level of the dimming mask(s) whereas, in contrast, as the user's pupil diameter decreases, the system may increase the transmittance level of the dimming mask(s).
In some configurations, the system may determine size parameters for the dimming mask(s) based on eye tracking data that indicates physical characteristics of the user's eyes such as, for example, a current pupil diameter. For example, as described elsewhere herein, in various implementations, the system may determine a diameter and/or height-and-width (or any other dimension for that matter) for a dimming mask based on a positive correlation with the pupil diameter of the user's eyes. Stated alternatively, as the user's pupil diameter increases, the system may increase the size of the dimming mask(s) whereas, in contrast, as the user's pupil diameter decreases, the system may decrease the size of the dimming mask(s).
In some configurations, the system may communicate with a light sensor to obtain luminance data associated with a brightness of one or more portions of the real-world view. Based on the luminance data, the system may determine opacity parameters indicating one or more transmittance levels for a dimming mask. For example, if the brightness level of the real-world view is relatively high (e.g. due to the user being outside on a sunny day), the opacity parameters may cause the transparent dimming panel to generate a highly or even entirely opaque dimming mask to enhance contrast with a CG image. In contrast, if the brightness level of the real-world view is relatively low (e.g. due to the user being in an unlit night-time environment), the opacity parameters may cause the transparent dimming panel to generate a dimming mask with a relatively higher transmittance level(s).
In some configurations, the system may analyze image data to determine a shape of the one or more CG images and, ultimately, to dynamically tailor a dimming mask shape to the shape of the CG images. For example, continuing with the virtual soda can scenario (i.e. where the system generates the rendered image of the soda can virtual object), the system may identify a shape of the rendered image as being generally rectangular with a rounded top and a rounded bottom. Then, the system may determine shape parameters to cause a profile of the dimming mask to at least partially match the identified shape of the rendered image of the soda can virtual object. Accordingly, in various configurations, the system may selectively block only that light from the real-world environment that would negatively impact the appearance of one or more CG images, e.g. by shining light through the CG images from the perspective of the user.
In some configurations, the system may determine incident light parameters associated with a virtual object to generate an augmentation, such as a drop-shadow augmentation, that causes the user to perceive, within a composite view, one or more regions having a reduced brightness, such as a drop-shadow that is generated with respect to the virtual object. For example, continuing with the soda can virtual object scenario, in addition to identifying the shape of a rendered image of the soda can virtual object, the system may further identify at least one of a real light source corresponding to the real-world view or an augmented light source corresponding to the AR program. Then, the system may determine a drop-shadow protrusion to protrude from a dimming mask corresponding to the rendered image of the soda can virtual object in order to generate the appearance of a drop-shadow in association with the soda can virtual object. A common way to derive the drop shadow shape is by applying an affine transform to the soda can shape. In various implementations, the system may dynamically determine a transmittance level of the drop-shadow protrusion based upon a pupil diameter of the user's eye(s).
It should be appreciated that the above-described subject matter may also be implemented as a computer-controlled apparatus (e.g. a direct view and/or indirect view NED device), a computer process, a computing system, or as an article of manufacture such as a computer-readable medium. These and various other features will be apparent from a reading of the following Detailed Description and a review of the associated drawings. Among many other benefits, the techniques herein improve efficiencies with respect to a wide range of computing resources.
For instance, human interaction with a device may be improved as the use of the techniques disclosed herein enable a user to actually perceive CG images at a natural brightness level as compared to the real-world environment as opposed to ramping CG image brightness up higher than the real-world view. In addition, the techniques described herein greatly reduce power draw on NED devices as the generation of very bright CG images draws substantially more power than the generation of a dimming mask as described herein, e.g. by darkening a portion of a Liquid Crystal Display (LCD) panel. Human interaction with a NED device may further be improved as the use of the techniques disclosed herein enable a user to simultaneously view both CG images and a portion of a real-world view without light from the real-world view negatively affecting the user perceived image quality of the CG images, e.g. due to real-world light leaking through the CG images. Other technical effects other than those mentioned herein can also be realized from implementations of the technologies disclosed herein.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended that this Summary be used to limit the scope of the claimed subject matter. Furthermore, the claimed subject matter is not limited to implementations that solve any or all disadvantages noted in any part of this disclosure.
The Detailed Description is described with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The same reference numbers in different figures indicates similar or identical items.
References made to individual items of a plurality of items can use a reference number followed by a parenthetical containing a number of a sequence of numbers to refer to each individual item. References made to right-side items and left-side items can use a reference number followed by an “R” or an “L,” respectively. Generic references to the items may use the specific reference number without the sequence of numbers. For example, the items may be collectively referred to with the specific reference number preceding a corresponding parenthetical containing a sequence number.
The following Detailed Description describes technologies for generating dimming masks to enhance contrast between computer-generated images (“CG images”) and a real-world view. Generally described, the techniques disclosed herein enable a system to monitor eye tracking data to determine at least one physical characteristic of a user's eye (singular) and/or eyes (plural). Then, based on the determined physical characteristic(s), the system may generate a dimming mask with relation to a CG image to decrease user perceived brightness of a real-world view (i.e. brightness of the real-world view from a user's perspective). In particular, the dimming masks may be used to control an amount of light transmitted from one or more real-world objects that is permitted to enter one or both of the user's eyes. As described above, technologies for managing contrast by controlling a user perceived brightness of the real-world view provide benefits over conventional Near-Eye-Display (NED) systems that can only manage contrast through modifications of the actual brightness of the CG images. In particular, because light from the real-world view may interfere with a NED system's ability to adequately display CG images, it may be desirable to effectively reduce the brightness of the real-world view at specific regions of a composite view that correspond to CG images.
For illustrative purposes, consider a scenario where an AR program instructs a NED device to generate a composite view by superimposing a CG image over some portion of a real-world view. For example, the AR program may cause the NED device to display a user interface (UI) menu and/or to give the appearance that a soda can is resting on top of a table that actually exists in the real-world environment by generating a rendered image, of a soda can virtual object, over a portion of the real-world view at which the table is visible. In the event that the real-world environment is sufficiently bright to interfere with a user's view of the CG image, it may be desirable to reduce the user perceived brightness of the real-world environment.
Accordingly, the techniques described herein enable the NED device to render the CG image at an appropriate brightness level while achieving the desired contrast by effectively reducing the brightness of the real-world view at the specific region where the CG image is being displayed. Stated alternative for explanatory purposes, the dimming mask may be deployed to, in a sense, “turn down” the brightness of the real-world view from the perspective of the user. Accordingly, the NED device is enabled to filter (e.g., block anywhere from slightly greater than zero percent to one-hundred percent) out light from the real-world view to prevent CG object contrast from being impaired and/or the CG image from appearing as a “ghostly image.”
In some embodiments, the dimming mask may be generated by the display panel as an additional function. Stated alternatively, the display panel pay itself generate both of the CG images and the dimming masks. For example, one or more lenses or other optical elements may be positioned behind (e.g. distal from the user) the dimming mask and display panel or in front of (e.g. proximate to the user) the dimming mask and display panel to deliver correct images to user.
In various embodiments, the transparent display 104 may be wholly or partially transparent. For example, the transparent display 104 may have a transmittance level of one-hundred-percent, nearly one-hundred-percent, eighty-percent, or some lesser transmittance level that remains suitable for viewing a real-world environment through. The transparent display 104 can be in any suitable form such as, for example, a waveguide, prism or multi-prism assembly configured to receive a generated CG image and direct the image towards a user's eye. In various embodiments, the transparent display 104 may be configured to use one or more light sources within the device to project the CG images toward the user's eye(s) and, more particularly, toward the user's pupil(s). The transparent display 104 may include within the device any suitable light source for generating images such as, for example, an LED projection engine.
As illustrated, the head-mounted display device 100 further includes a transparent dimming panel 106 that is positioned adjacent to a side of the transparent display 104 that is situated away from the user's pupils when the head-mounted display device 100 is properly worn. In the illustrated example, the transparent dimming panel 106 includes separate right eye and left eye transparent dimming panels, labeled 106R and 106L respectively. As illustrated, the transparent dimming panel 106 is shown to be generating dimming masks 108 having a transmittance level that is at least partially decreased from a base transmittance of the transparent dimming panel. Stated alternatively, the two regions at which the dimming masks 108R (corresponding to the user's right eye) and 108L (corresponding to the user's left eye) are being generated are more opaque (e.g. absorb more light) than the remaining regions of the transparent dimming panel 106. Accordingly, in some implementations, the head-mounted display device 100 may generate the dimming masks 108 directly behind one or more CG images that are generated by the transparent display 104 to prevent light that is reflected from one or more real-world objects from passing through the transparent display 104 at the particular location at which the CG images are being generated.
In some examples, the transparent dimming panel 106 may include a single transparent dimming panel 106 that is viewable with both eyes and/or a single transparent dimming panel 106 that is viewable by a single eye only. In some examples, the periphery of the transparent dimming panel 106 may be larger than the transparent display 104. It may extend downward and/or may extend around toward the ears, e.g. the peripheral edge. Accordingly, although the techniques of the present disclosure are described mainly with reference to implementations in which CG images and dimming masks 108 are generated in front of each of a user's two eyes, implementations in which one or more images and one or more dimming masks are generated in front of only one of the user's eyes are within the scope of the present disclosure and appended claims and are contemplated. Therefore, it can be appreciated that the techniques described herein may be deployed within a single-eye Near Eye Display (NED) system (e.g. GOOGLE GLASS) and/or a dual-eye NED system (e.g. MICROSOFT HOLOLENS).
In some embodiments, the head-mounted display device 100 may further include an additional see-through optical component 110, shown in
The head-mounted display device 100 may further include an eye tracking sensor 114 that is configured to generate eye tracking data associated with one or more physical characteristics of the user's eyes. Exemplary physical characteristics include, but are not limited to, pupil size, a rate of change of pupil size, gaze direction, and/or a rate of change to a gaze direction. The eye tracking sensor 114 can be in any suitable form such as, for example, a non-contact sensor configured to use optical-based tracking (e.g. video camera based and/or some other specially designed optical-sensor-based eye tracking technique) to monitor the one or more physical characteristics of the user's eyes. The head-mounted display device 100 may further include various other components, for example speakers, microphones, accelerometers, gyroscopes, magnetometers, temperature sensors, touch sensors, biometric sensors, other image sensors, energy-storage components (e.g. battery), a communication facility, a GPS receiver, etc.
In the illustrated example, a controller 116 is operatively coupled to each of the transparent display 104, the transparent dimming panel 106, and the eye tracking sensor 114. The controller 116 may further be operatively coupled to other componentry of the head-mounted display device 100. The controller 116 includes one or more logic devices and one or more computer memory devices storing instructions executable by the logic device(s) to deploy functionalities described herein with relation to the head-mounted display device 100. The controller 116 can comprise one or more processing units 118, one or more computer-readable media 120 for storing an operating system 122 and data such as, for example, image data 124. The image data 124 may define one or more CG images and may further indicate one or more locations on the transparent display 104 to generate these CG images. The computer-readable media 120 may further include an eye tracking engine 126 configured to receive the eye tracking data from the eye tracking sensor 114 and, based thereon, determine one or more physical characteristics of the user's eyes. The computer-readable media 120 may further include a dimming engine 128 configured to determine one or more dimming parameters associated with the generation of the dimming masks 108. As discussed in more detail herein, the dimming parameters may be determined based on the image data 124 and/or one or more of the physical characteristics of the user's eyes. For example, the dimming parameters may be determined based on a pupil size of the user's eyes that is determinable by the eye tracking data as well as a location that a CG image is generated on the transparent display 104 that is determinable via the image data 124. The components of head-mounted display device 100 are operatively connected, for example, via a bus 130, which can include one or more of a system bus, a data bus, an address bus, a PCI bus, a Mini-PCI bus, and any variety of local, peripheral, and/or independent buses.
As used herein, the term “dimming parameter” may refer generally to (1) any parameter that may be used in generating a dimming mask 108 that affects (1) contrast between a CG image and a real-world view; and/or (2) an amount of ambient light transmitted from (e.g. generated by and/or reflected off) the real-world environment that reaches the user's eyes. Exemplary dimming parameters include, but are not limited to, size parameters that may at least partially control a size of one or more dimming masks 108, opacity parameters that may at least partially control a transmittance level of one or more dimming masks, location parameters that may at least partially control a location of one or more dimming masks 108 on the transparent dimming panel(s) 106, and/or shape parameters that may at least partially control a shape of one or more dimming masks 108 generated by the transparent dimming panel 106.
The processing unit(s) 118, can represent, for example, a CPU-type processing unit, a GPU-type processing unit, a field-programmable gate array (FPGA), another class of digital signal processor (DSP), or other hardware logic components that may, in some instances, be driven by a CPU. For example, and without limitation, illustrative types of hardware logic components that can be used include Application-Specific Integrated Circuits (ASICs), Application-Specific Standard Products (ASSPs), System-on-a-Chip Systems (SOCs), Complex Programmable Logic Devices (CPLDs), etc.
As used herein, computer-readable media, such as computer-readable media 120, can store instructions executable by the processing unit(s). Computer-readable media can also store instructions executable by external processing units such as by an external CPU, an external GPU, and/or executable by an external accelerator, such as an FPGA type accelerator, a DSP type accelerator, or any other internal or external accelerator. In various examples, at least one CPU, GPU, and/or accelerator is incorporated in a computing device, while in some examples one or more of a CPU, GPU, and/or accelerator is external to a computing device.
Computer-readable media can include computer storage media and/or communication media. Computer storage media can include one or more of volatile memory, nonvolatile memory, and/or other persistent and/or auxiliary computer storage media, removable and non-removable computer storage media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules, or other data. Thus, computer storage media includes tangible and/or physical forms of media included in a device and/or hardware component that is part of a device or external to a device, including but not limited to random access memory (RAM), static random-access memory (SRAM), dynamic random-access memory (DRAM), phase change memory (PCM), read-only memory (ROM), erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), flash memory, rotating media, optical cards or other optical storage media, magnetic storage, magnetic cards or other magnetic storage devices or media, solid-state memory devices, storage arrays, network attached storage, storage area networks, hosted computer storage or any other storage memory, storage device, and/or storage medium that can be used to store and maintain information for access by a computing device.
In contrast to computer storage media, communication media can embody computer-readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave, or other transmission mechanism. As defined herein, computer storage media does not include communication media. That is, computer storage media does not include communications media consisting solely of a modulated data signal, a carrier wave, or a propagated signal, per se.
Turning now to
In the illustrated implementation, the dimming masks 108 are aligned with a CG image 210 that is being generated by the transparent display 104 to cause image light 212 to propagate toward the pupil 202. In the illustrated scenario, the image light 212 is shown as originating within the optical system 200 and propagating through at least a portion of the transparent display 104 before exiting the transparent display 104 toward the pupil 202. Accordingly, in some embodiments the optical system 200 may be configured to actively generate and project the image light 212 toward the user's pupil 202. In some embodiments, however, the dimming masks 108 are driven to a transmittance level that is not fully opaque so that at least some of the incoming light 208 is allowed to pass through the dimming masks 108. For example, it is within scope of the present disclosure to deploy an at least partially transparent LCD panel to generate a CG images that are not illuminated by the system 100 but rather rely on light from the real-world environment to shine through the image to make it become visible. For example, the real-world environment may reflect (which as used herein is defined to also include emitting) through the LCD panel to act as a backlight.
In some embodiments, the transparent display 104 and/or the transparent dimming panel 106 of the optical system 200 may be positioned at a distance D from the pupil that is between 10 mm and 75 mm. In some embodiments, the transparent display 104 and/or the transparent dimming panel 106 of the optical system 200 may be positioned at a distance D from the pupil that is between 20 mm and 60 mm. In some embodiments, the transparent display 104 and/or the transparent dimming panel 106 of the optical system 200 may be positioned at a distance D from the pupil that is between 10 mm and 100 mm. In some embodiments, the transparent display 104 and/or the transparent dimming panel 106 of the optical system 200 may be positioned at a distance D from the pupil that is between 25 mm and 50 mm. In some embodiments, the transparent display 104 and/or the transparent dimming panel 106 may be positioned at a distance D from the pupil that is less than 10 mm or greater than 75 mm.
Turning now to
Turning now to
Turning now to
Turning now to
To illustrate the enhanced contrast of
Turning back now to
As illustrated, the Y-Axis corresponds to a proportion of light from a real-world environment that reaches the user's pupil at a variety of field angles corresponding to the X-Axis. For example, with reference to the line corresponding to a 6-mm dimming mask diameter, the graph 300 indicates that the user perceived transmittance is substantially zero-percent for field angles ranging from 0° to roughly 3.0°. Then, at a roughly 3.0° field angle the user perceived transmittance steeply climbs such that the user perceived transmittance from, for example, the 3.0° field angle to a 6° field angle changes from roughly zero-percent to eighty-percent. As further illustrated, the rate of change of the user perceived transmittance continually decreases until the user perceived transmittance levels out at one-hundred-percent (e.g. at roughly 8.5° for a 6-mm fully opaque dimming mask).
In the particular scenario illustrated in
To illustrate these concepts, the affected area is illustrated as being superimposed over a real-world object 310 which is illustrated as a tree. As illustrated, the blacked-out area 306 indicates that the user is unable to perceive any incoming light whatsoever from within the field angles of 0° to 3.0°. However, beginning at roughly 3.0° not all of the incoming light is blocked by the dimming mask and, therefore, the user begins to be able to faintly perceive the real-world object 310.
Turning back now to
As illustrated, the Y-Axis corresponds to a proportion of light from a real-world environment that reaches the user's pupil at a variety of field angles corresponding to the X-Axis. For example, with reference to the line corresponding to a 6-mm dimming mask diameter having fifty-percent (50%) transmissivity, the graph 300 indicates that the user perceived transmittance is fifty-percent (50%) for field angles ranging from 0° to roughly 3.0°. As in the previous graph of
Turning now to
With particular reference to
With particular reference to
Accordingly, because the user perceived transmittance at any particular field angle within the user perceived composite view 502 is dependent upon both the current pupil size in addition to a current dimming mask size, in order for the user perceived composite view 502 as shown in
Turning now to
With particular reference to
Accordingly, in the event that the pupils 202 increase from the first pupil size to a second pupil size as illustrated in
Turning now to
With particular reference to
For purposes of
Accordingly, in the event that the gaze direction changes as illustrated in
In some examples, multiple CG images may be presented to the user where each may have at least one corresponding dimming masks. In some examples, when multiple masks are in use and the CG image objects are sufficiently separated, it may be desirable to enable only the dimming masks for the CG objects that are in the line of sight based on the user's gaze angle. One benefit of the selective enabling of dimming masks with gaze angle includes, but is not limited to, preventing the user from being visually distracted.
Turning now to
It can be appreciated that the system 800 of
Under the illustrated circumstances in which the optical system 800 identifies a real light source corresponding to the real-world environment, the optical system 800 may deploy a light sensor 812 such as, for example, one or more forward facing cameras that are configured to identify one or more light sources that correspond to the real-world environment 112. The optical system 800 may then determine incident light parameters corresponding to the identified light source 804. Exemplary incident light parameters include, but are not limited to, an incident light direction 802, a luminous intensity of the incident light, and/or a color of the incident light.
Then, based on the incident light parameters, the system may determine a drop shadow protrusion for the purpose of generating an augmented drop shadow 808 in association with a rendered object 810. For example, in the illustrated scenario the rendered object 810 corresponds to a soda can virtual object that the optical system 800 is to give the appearance of resting on the actual table in front of the real-world object 206, e.g. the generic cereal box. As can be seen in each of
In one specific but non-limiting example, the optical system may vary dimming mask region size and opacity to create a narrower or wider shadow with penumbra. For example, referring to
Turning now to
At block 901, the system may receive image data associated with supplementation of a user perspective of a real-world view with at least one CG image. The image data may define the CG image(s) in addition to parameters corresponding to generating the one or more images on the transparent display 104. For example, the image data may define the UI menu shown in
As another example, with particular reference to
At block 903, the system may receive eye tracking data associated with physical characteristics of the user's eyes. For example, the system may include the eye tracking sensor 114 positioned to monitor physical characteristics of the user's eyes. Exemplary such physical characteristics include, but are not limited to, a pupil diameter and/or a gaze direction of one or more of the user's eyes. Referring specifically to
At block 905, the system may cause the transparent display 104 to generate the CG image(s) between the user's eyes 204 and the real-world environment 112. For example, as illustrated in
At block 907, the system may determine dimming parameters for at least one dimming mask based on at least one of the physical characteristics of the user's eyes and/or the image data. The dimming parameters may be associated with enhancing contrast between the CG image generated by the transparent display 104 and the real-world view. In particular, as described in more detail elsewhere herein, the dimming parameters may define at least one dimming mask that can be generated to effectively reduce brightness of at least a portion of the real-world environment 112 from the perspective of the user, i.e. in the real-world view. Stated alternatively, the dimming masks may reduce the brightness of one or more regions of the real-world view.
At block 909, determining dimming parameters for the at least one dimming mask may include determining size parameters. In some implementations, the size parameters may cause the system to generate a dimming mask that spans across substantially all of a functional area of the transparent dimming panel 106. For example, the transparent dimming panel 106 may include a functional area that has transmittance level control capabilities, i.e. a functional capability of controllably changing a transmittance level. In one exemplary embodiment, the dimming panel 106 may have a functional area having base transmittance that is highly transparent (e.g. eighty-percent (80%) transmittance or higher) and the ability of controllably decreasing the transmittance level of one or more regions of the functional area. Accordingly, under various circumstances, the size parameters may cause the system to generate a dimming mask over the entire functional area by controllably decreasing the transmittance level of the entire functional area.
In some implementations, the system may determine at least one size parameter based at least in part on the pupil diameter of the user's eyes. For example, as described with relation to
In some implementations, the system may determine at least one size parameter based at least in part on the image data. For example, under circumstances where the system is to generate a dimming mask 108 behind an entire area of a CG image that is generated by the transparent display 104, the actual size at which the transparent dimming panel 106 should generate the dimming masks 108 to achieve this goal will vary based on the size of the CG image 210 as it is generated by the transparent display 104 and perceived at the nominal focus distance.
In some implementations, the system may determine at least one size parameter based at least in part on a gaze direction of the user's eyes. For example, consider a scenario where the system is to generate dimming masks that cover substantially all of a particular quadrant of the user's vision with the exception of a portion of the quadrant that falls within a central vision area 702 as illustrated in
At block 911, determining dimming parameters for the at least one dimming masks 108 may include determining location parameters. In some implementations, the system may determine at least one location parameter based at least in part on the image data. For example, in a scenario where the system is to superimpose a dimming mask with a particular CG image, it can be appreciated that the appropriate location on the transparent dimming panel 106 to generate the dimming masks 108 will be at least partially dependent on a corresponding location in the visual field at which a corresponding CG image 210 is generated and the interpupil spacing of the user which may range from 51 mm-73 mm. In some cases, however, the interpupil spacing may be less than 51 mm or greater than 73 mm. It can be appreciated that in various implementations, the dimming mask position should be in good alignment with the viewers pupil and the CG object. In some implementations, the system may determine at least one location parameter based at least in part on the gaze direction of the user's eyes. For example, with particular reference to
At block 913, determining dimming parameters for the at least one dimming masks may include determining opacity parameters. The opacity parameters may indicate at least one transmittance level that is less than a base transmittance level of the transparent dimming panel 106. For example, under a circumstance where the base transmittance level of the transparent dimming panel 106 is ninety-percent (90%), the opacity parameters may cause the system to generate the at least one dimming masks 108 by driving one or more regions of the transparent dimming panel 106 to a relatively lesser transmittance level of, for example, twenty-percent (20%), ten-percent (10%), substantially zero-percent (0%), or any other desirable transmittance level.
In some implementations, the system may determine at least one opacity parameter based at least in part on a pupil size of the user's eyes. For example, with particular reference to
In some implementations, the system may determine at least one opacity parameter based at least in part on luminance data that indicates a luminous intensity corresponding to one or more regions of the real-world view. For example, the system may deploy a light sensor 812 to determine a brightness (e.g. a luminous intensity) of the real-world view. Then, based upon the brightness of the real-world view, the system may determine how low to set the transmittance level of the at least one dimming region. Stated alternatively, the amount to which the system effectively turns down the brightness of the real-world view may be at least partially dependent on the brightness of the real-world view to begin with.
In some implementations, the system may determine at least one opacity parameter based at least in part on the gaze direction of the user's eyes. For example, under certain circumstances it may be desirable to dynamically modify a transmittance level of a particular dimming mask based upon where that dimming mask falls within the user's vision, e.g. in terms of field angle. With particular reference to
In some implementations, the at least one opacity parameter may indicate a predetermined transmittance level for one or more dimming masks 108. For example, an opacity parameter may cause the transparent dimming panel 106 to generate a dimming mask at a particular transmittance level (e.g. fully opaque) regardless of the image data and/or various physical characteristics of the user's eyes.
At block 915, determining dimming parameters for the at least one dimming mask 108 may include determining shape parameters. The shape parameters may define a shape for the at least one dimming masks 108 by, for example, defining an outer profile of the at least one dimming mask and/or defining parameters associated with sizing, locating and/or orienting one or more predetermined shapes. Exemplary predetermined shapes include, but are not limited to, a circle shape that can be defined by a radius and a reference location, a square that can be defined by a side length and a reference location/angular orientation, a triangle that can be defined by one or more side lengths and a reference location/angular orientation, and/or a rectangle that can be defined by at least two side lengths and a reference location/angular orientation.
In some implementations, the shape parameters may be based at least partially on the image data. Determining the shape parameters may include analyzing the image data to determine a shape of at least one CG image 210. For example, the system may determine an outer profile for the UI menu of
At block 917, the system may cause a transparent dimming panel 106 to generate the dimming masks 108 between the user's eyes and the real-world environment. For example, the system may utilize the dimming parameters determined at block 907 to cause the transparent dimming panel 106 to controllably alter a transmittance level of one or more regions to enhance contrast between the real-world view and the CG image 210 generated at block 905. Upon being generated, the at least one dimming masks 108 may block at least some light that is transmitted by (e.g. either generated by or reflected off) a real-world object from passing through the transparent display and reaching a pupil of the user's eye.
While described herein in the context of near-eye display systems, the example optical systems and methods disclosed herein may be used in any suitable optical system, such as a rifle scope, telescope, spotting scope, binoculars, and heads-up display.
In some embodiments, the methods and processes described herein may be tied to a computing system of one or more computing devices. In particular, such methods and processes may be implemented as a computer-application program or service, an application-programming interface (API), a library, and/or other computer-program product.
Computing system 1000 includes a logic subsystem 1002 and a storage subsystem 1004. Computing system 1000 may optionally include a display subsystem 1006, input subsystem 1008, communication subsystem 1010, and/or other components not shown in
Logic subsystem 1002 includes one or more physical devices configured to execute instructions. For example, the logic machine may be configured to execute instructions that are part of one or more applications, services, programs, routines, libraries, objects, components, data structures, or other logical constructs. Such instructions may be implemented to perform a task, implement a data type, transform the state of one or more components, achieve a technical effect, or otherwise arrive at a desired result.
Logic subsystem 1002 may include one or more processors configured to execute software instructions. Additionally or alternatively, logic subsystem 1002 may include one or more hardware or firmware logic machines configured to execute hardware or firmware instructions. Processors of logic subsystem 1002 may be single-core or multi-core, and the instructions executed thereon may be configured for sequential, parallel, and/or distributed processing. Individual components of logic subsystem 1002 optionally may be distributed among two or more separate devices, which may be remotely located and/or configured for coordinated processing. Aspects of logic subsystem 1002 may be virtualized and executed by remotely accessible, networked computing devices configured in a cloud-computing configuration.
Storage subsystem 1004 includes one or more physical devices configured to hold instructions executable by logic subsystem 1002 to implement the methods and processes described herein. When such methods and processes are implemented, the state of storage subsystem 1004 may be transformed—e.g., to hold different data.
Storage subsystem 1004 may include removable and/or built-in devices. Storage subsystem 1004 may include optical memory (e.g., CD, DVD, HD-DVD, Blu-Ray Disc, etc.), semiconductor memory (e.g., RAM, EPROM, EEPROM, etc.), and/or magnetic memory (e.g., hard-disk drive, floppy-disk drive, tape drive, MRAM, etc.), among others. Storage subsystem 1004 may include volatile, nonvolatile, dynamic, static, read/write, read-only, random-access, sequential-access, location-addressable, file-addressable, and/or content-addressable devices.
It will be appreciated that storage subsystem 1004 includes one or more physical devices. However, aspects of the instructions described herein alternatively may be propagated by a communication medium (e.g., an electromagnetic signal, an optical signal, etc.) as opposed to being stored on a storage medium.
Aspects of logic subsystem 1002 and storage subsystem 1004 may be integrated together into one or more hardware-logic components. Such hardware-logic components may include field-programmable gate arrays (FPGAs), program- and application-specific integrated circuits (PASIC/ASICs), program- and application-specific standard products (PSSP/ASSPs), system-on-a-chip (SOC), and complex programmable logic devices (CPLDs), for example.
When included, display subsystem 1006 may be used to present a visual representation of data held by storage subsystem 1004. This visual representation may take the form of a graphical user interface (GUI). As the herein described methods and processes change the data held by the storage machine, and thus transform the state of the storage machine, the state of display subsystem 1006 may likewise be transformed to visually represent changes in the underlying data. Display subsystem 1006 may include one or more display devices utilizing virtually any type of technology. Such display devices may be combined with logic subsystem 1002 and/or storage subsystem 1004 in a shared enclosure, or such display devices may be peripheral display devices.
When included, input subsystem 1008 may comprise or interface with one or more user-input devices such as a keyboard, mouse, touch screen, or game controller. In some embodiments, the input subsystem may comprise or interface with selected natural user input (NUI) componentry. Such componentry may be integrated or peripheral, and the transduction and/or processing of input actions may be handled on-board or off-board. Example NUI componentry may include a microphone for speech and/or voice recognition; an infrared, color, stereoscopic, and/or depth camera for machine vision and/or gesture recognition; a head tracker, eye tracker, accelerometer, and/or gyroscope for motion detection and/or intent recognition; as well as electric-field sensing componentry for assessing brain activity.
When included, communication subsystem 1010 may be configured to communicatively couple computing system 600 with one or more other computing devices. Communication subsystem 1010 may include wired and/or wireless communication devices compatible with one or more different communication protocols. As non-limiting examples, the communication subsystem may be configured for communication via a wireless telephone network, or a wired or wireless local- or wide-area network. In some embodiments, the communication subsystem may allow computing system 600 to send and/or receive messages to and/or from other devices via a network such as the Internet.
The disclosure presented herein may be considered in view of the following clauses.
Example Clause A, a computer-implemented method, comprising: receiving image data indicating at least one location on a transparent display to generate at least one computer generated image (CGI); obtaining, from at least one eye tracking sensor, eye tracking data associated with at least one eye that is positioned for viewing a real-world view, the eye tracking data indicating at least a pupil diameter of the at least one eye; causing the transparent display to generate the at least one CGI at the at least one location, wherein the at least one location is positioned on the transparent display between the at least one eye and a real-world object that is visible within the real-world view; determining, based at least in part on the pupil diameter, at least one size parameter associated with at least one dimming mask for enhancing contrast between the at least CGI and the real-world view; determining, based at least in part on the at least one location of the image data, at least one location parameter associated with the at least one dimming mask; and causing, based on the at least one size parameter and the at least one location parameter, a transparent dimming panel to generate the at least one dimming mask with respect to the at least one location on the transparent display, wherein the at least one dimming mask blocks at least some light that is transmitted from the real-world object from passing through the at least one location on the transparent display toward at least one pupil of the at least one eye.
Example Clause B, the computer-implemented method of Example Clause A, further comprising: determining, based at least in part on the image data, opacity parameters that indicate at least one transmittance level that is less than a base transmittance of the transparent dimming panel; and causing a plurality of pixels of the transparent dimming panel to be driven to the at least one transmittance level.
Example Clause C, the computer-implemented method of any one of Example Clauses A through B, wherein the at least one transmittance level is further determined based on at least one of a positive relationship or an inverse relationship to the pupil diameter.
Example Clause D, the computer-implemented method of any one of Example Clauses A through C, further comprising: analyzing the image data to determine a shape of the at least one CGI; and determining, based at least in part on the shape of the at least one CGI, shape parameters to cause a profile of the at least one dimming mask to at least partially match the shape of the at least one CGI.
Example Clause E, the computer-implemented method of any one of Example Clauses A through D, further comprising: obtaining, from at least one light sensor, luminance-correlated data indicating at least a luminous intensity corresponding to the real-world view; and determining, based at least in part on the luminous intensity, opacity parameters that indicate at least one transmittance level that is less than a base transmittance of the transparent dimming panel, wherein the at least one dimming mask is driven to the at least one transmittance level.
Example Clause F, the computer-implemented method of any one of Example Clauses A through E, further comprising determining, based on the eye tracking data, a gaze direction corresponding to the at least one eye, wherein the at least one location parameter is further determined based on the gaze direction.
Example Clause G, the computer-implemented method of any one of Example Clauses A through F, wherein the at least one size parameter is further determined based on the image data.
While Example Clauses A through G are described above with respect to a method, it is understood in the context of this document that the subject matter of Example Clauses A through G can also be implemented by a device, by a system, and/or via computer-readable storage media.
Example Clause H, a Near-Eye-Display (NED) device, comprising: an eye tracking sensor to generate eye tracking data associated with at least one eye of a user; a transparent display having a first side that faces the at least one eye and a second side that faces a real-world object, the transparent display configured to cause a projection of at least one CGI outward from the first side; a transparent dimming panel that is positioned adjacent to the second side of the transparent display, the transparent dimming panel configured to generate at least one dimming mask to selectively block light from passing through at least one region of the transparent display; and at least one controller that is communicatively coupled to the eye tracking sensor, the transparent display, and the transparent dimming panel, wherein the at least one controller is configured to: receive image data that indicates at least one location on the transparent display to generate the at least one CGI; receive the eye tracking data from the eye tracking sensor, the eye tracking data indicating at least a pupil size corresponding to the at least one eye; determine for the at least one dimming mask: at least one size parameter based at least in part on the pupil size, at least one location parameter based at least in part on the at least one location, and at least one opacity parameter based at least in part on the pupil size; cause the transparent dimming panel to generate the at least one dimming mask according to the at least one size parameter, the at least one location parameter, and the at least one opacity parameter; and cause the transparent display to at least partially superimpose the at least one CGI with the at least one dimming mask to generate a composite view that includes the at least one CGI and at least a portion of the real-world view, wherein the at least one dimming mask blocks at least some light that is transmitted from the real-world object from passing through the at least one location on the transparent display.
Example Clause I, the NED device of Example Clause H, wherein the at least one dimming mask is generated directly between the real-world view and the at least one eye at a distance from at least one pupil, of the at least one eye, that is between 10 millimeters and 100 millimeters.
Example Clause J, the NED device of any of Example Clauses H through I, wherein the pupil size corresponds to a first area, and wherein the at least one size parameter causes the at least one dimming mask to mask a second area, of the transparent display, that is greater than or equal to the first area.
Example Clause K, the NED device of any of Example Clauses H through J, wherein the at least one controller is further configured to determine, for the at least one dimming mask, at least one shape parameter based at least in part on the image data.
Example Clause L, the NED device of any of Example Clauses H through K, wherein the at least one controller is further configured to: determine incident light parameters associated with at least one of a real light source corresponding to the real-world view or an augmented light source corresponding to an AR program, the incident light parameters indicating at least an incident light direction with respect to a rendered object; and based at least in part on the incident light parameters, determine, for the at least one dimming mask, a shadow protrusion to generate an augmented drop-shadow in association with the rendered object.
Example Clause M, the NED device of any of Example Clauses H through L, wherein the at least one eye comprises a first eye having a first pupil and a second eye having a second pupil, and wherein the at least one dimming mask comprises a first dimming mask disposed between the real-world object and the first pupil and a second dimming mask disposed between the real-world object and the second pupil.
Example Clause N, the NED device of any of Example Clauses H through M, wherein the at least one size parameter is further determined based on the image data.
Example Clause O, the NED device of any of Example Clauses H through N, wherein the at least one controller is further configured to monitor the eye tracking data to determine a gaze direction corresponding to the at least one eye, wherein at least one of the at least one size parameter or the at least one opacity parameter are further determined based on the gaze direction.
While Example Clauses H through O are described above with respect to a device, it is understood in the context of this document that the subject matter of Example Clauses H through O can also be implemented by a method, by a system, and/or via computer-readable storage media.
Example Clause P, a computer-implemented method, comprising: receiving image data that defines at least one CGI; monitoring a pupil diameter of at least one eye based on eye tracking data that is generated by at least one sensor; causing a transparent display to generate the at least one CGI at one or more locations, on the transparent display, that are between the at least one eye and a real-world object that is visible within a real-world view; determining at least one size parameter associated with at least one dimming mask based at least in part on the pupil diameter; and causing generation of the at least one dimming mask in accordance with the at least one size parameter to affect contrast between the at least one CGI and the real-world view, wherein the at least one dimming mask blocks at least some light that is transmitted from the real-world object from passing through the transparent display.
Example Clause Q, the computer-implemented method of Example Clause P, wherein the at least one dimming mask is at least partially aligned with the one or more locations to block the at least some light that is transmitted from the real-world object from passing through the at least one CGI at the one or more locations on the transparent display.
Example Clause R, the computer-implemented method of any one of Example Clauses P through Q, wherein the at least one dimming mask is driven to a predetermined transmittance level.
Example Clause S, the computer-implemented method of any one of Example Clauses P through R, wherein the at least one dimming mask is generated at a distance from at least one pupil, of the at least one eye, that is between 10 millimeters and 100 millimeters.
Example Clause T, the computer-implemented method of any one of Example Clauses P through S, further comprising: monitoring the eye tracking data to identify a change to the pupil diameter; and based on the change corresponding to an increase to the pupil diameter, increasing an area of the at least one dimming mask; or based on the change corresponding to a decrease to the pupil diameter, decreasing the area of the at least one dimming mask.
While Example Clauses P through T are described above with respect to a method, it is understood in the context of this document that the subject matter of Example Clauses P through T can also be implemented by a device, by a system, and/or via computer-readable storage media.
Example Clause U, a system for dynamically modifying dimming mask opacity, the system comprising: at least one sensor to generate eye tracking data associated with at least one eye; a transparent display having a first side that faces the at least one eye and a second side that faces a real-world object; a transparent dimming panel to generate at least one dimming mask at one or more locations of the transparent display, wherein the at least one dimming mask controls an amount of light, reflected off the real-world object, that passes through the one or more locations of the transparent display; and at least one controller that is communicatively coupled to the at least one sensor, the transparent display, and the transparent dimming panel, wherein the at least one controller is configured to: receive image data indicating at least one CGI; receive the eye tracking data from the at least one sensor, the eye tracking data indicating at least a pupil size corresponding to the at least one eye; determine, based at least in part on the pupil size, at least one transmittance level for the at least one dimming mask; cause the transparent dimming panel to generate the at least one dimming mask by driving at least one region, of the transparent dimming panel, to the at least one transmittance level; and cause the transparent display to generate a composite view that includes the at least one CGI and at least a portion of a real-world view, wherein the at least one dimming mask blocks at least some light that is transmitted from the real-world object from passing through the transparent display.
Example Clause V, the system of Example Clause U, wherein the at least one controller is configured to determine at least one position on the transparent dimming panel to generate the at least one dimming mask based at least in part on the image data.
Example Clause W, the system of any one of Example Clauses U through V, wherein the eye tracking data further indicates a gaze direction corresponding to the at least one eye, and wherein the at least one position is further determined based at least in part on the gaze direction.
Example Clause X, the system of any one of Example Clauses U through W, wherein the at least one CGI generated by the transparent display is at least partially aligned with the at least one dimming mask generated by the transparent dimming panel.
Example Clause Y, the system of any one of Example Clauses U through X, wherein the transparent dimming panel includes at least a functional area having transmittance level control capabilities, and wherein the at least one region corresponds to substantially all of the functional area having the transmittance level control capabilities.
Example Clause Z, the system of any one of Example Clauses U through Y, wherein the at least one controller is further configured to: determine, based on the image data, at least one profile corresponding to the at least one CGI; and determine shape parameters for the at least one dimming mask to at least partially position at least one user-perceived penumbra, corresponding to the at least one dimming mask, with respect to the at least one profile.
While Example Clauses U through Z are described above with respect to a system, it is understood in the context of this document that the subject matter of Example Clauses U through Z can also be implemented by a device, via a computer-implemented method, and/or via computer-readable storage media.
Example Clause AA, a computer-implemented method, comprising: receiving eye tracking data from at least one sensor that is positioned to monitor physical characteristics of at least one eye, wherein the eye tracking data indicates at least a pupil size and a gaze direction corresponding to the at least one eye; determining, based at least in part on the pupil size, at least one transmittance level for at least one dimming mask; determining, based at least in part on the gaze direction, at least one position on a transparent dimming panel to generate the at least one dimming mask; and causing the transparent display to generate the at least one dimming mask by driving a region, of the transparent display, that corresponds to the at least one position to the at least one transmittance level, wherein the at least one dimming mask blocks at least some light that is transmitted from a real-world object from passing through the transparent dimming panel.
Example Clause BB, the computer-implemented method of Example Clause AA, further comprising determining, based on the pupil size, at least one size parameter for the at least one dimming mask, wherein the at least one pupil corresponds to a first area and the at least one size parameter causes the at least one dimming mask to have a second area that is between one to three times the first area.
Example Clause CC, the computer-implemented method of any one of Example Clauses AA through BB, further comprising: monitoring the eye tracking data to identify a change to the gaze direction; and determining, based at least in part on the change to the gaze direction, at least one new position on the transparent dimming panel to move the at least one dimming mask to.
While Example Clauses AA through CC are described above with respect to a method, it is understood in the context of this document that the subject matter of Example Clauses AA through CC can also be implemented by a device, by a system, and/or via computer-readable storage media.
In closing, although the various techniques have been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended representations is not necessarily limited to the specific features or acts described. Rather, the specific features and acts are disclosed as example forms of implementing the claimed subject matter.