The prevalence and portability of information handling devices (e.g., smart phones, tablets, personal computers, laptop computers, etc.) allow users to use these devices to provide and accept input in a variety of locations. One way that users may interact with and receive information from a device is through the device's display. In order to properly visualize the content being portrayed on the display, the ideal brightness of the display must be set based on the environment in which it is being viewed. For example, in areas with an abundance of available light the display must be brighter to remain visible. Conversely, in a dark room the display should be relatively dim to avoid an uncomfortable glare experienced by the user.
Ambient light sensors (ALS) are often incorporated into information handling devices to measure how much light falls on the device. The device then automatically adjust the brightness of the display based upon the amount of light detected. However, ALS do not account for a user's sensitivity to light, which varies for each individual. Therefore, it would be desirable if a device better imitated the sensitivity of a user's eyes to more effectively adjust the brightness of the display.
In summary, one aspect provides a method, comprising: capturing, using a camera on a device, at least one image of a user viewing a display of the device; and adjusting, using a processor, a brightness setting of the display from a first setting to a second setting based on the at least one image.
Another aspect provides an electronic device, comprising: a camera; a processor; a memory device that stores instructions executable by the processor to: capture, using the camera, at least one image of a user viewing a display of the device; and adjust a brightness setting of the display from a first setting to a second setting based on the at least one image.
A further aspect provides a product, comprising: a storage device that stores code executable by a processor, the code comprising: code that captures, using a camera on a device, at least one image of a user viewing a display of the device; and code that adjusts, using a processor, a brightness setting of the display from a first setting to a second setting based on the at least one image.
The foregoing is a summary and thus may contain simplifications, generalizations, and omissions of detail; consequently, those skilled in the art will appreciate that the summary is illustrative only and is not intended to be in any way limiting.
For a better understanding of the embodiments, together with other and further features and advantages thereof, reference is made to the following description, taken in conjunction with the accompanying drawings. The scope of the invention will be pointed out in the appended claims.
It will be readily understood that the components of the embodiments, as generally described and illustrated in the figures herein, may be arranged and designed in a wide variety of different configurations in addition to the described example embodiments. Thus, the following more detailed description of the example embodiments, as represented in the figures, is not intended to limit the scope of the embodiments, as claimed, but is merely representative of example embodiments.
Reference throughout this specification to “one embodiment” or “an embodiment” (or the like) means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. Thus, the appearance of the phrases “in one embodiment” or “in an embodiment” or the like in various places throughout this specification are not necessarily all referring to the same embodiment.
Furthermore, the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided to give a thorough understanding of embodiments. One skilled in the relevant art will recognize, however, that the various embodiments can be practiced without one or more of the specific details, or with other methods, components, materials, et cetera. In other instances, well known structures, materials, or operations are not shown or described in detail to avoid obfuscation.
The ideal brightness of a display of an information handling device (e.g., smart phone, tablet, personal computer, laptop computer, etc., herein “device”) should be set based on the environment in which it is being viewed. In areas with an abundance of light (e.g., outdoors on a sunny day), the display must be brighter to remain visible. Conversely, in darker areas (e.g., indoors in a dark room), the display should be relatively dim so that it does not aggravate a user's eyes.
One current method to assist in improved display visibility is for a user to manually adjust the brightness settings on their device. These settings can be accessed from various locations on the device (e.g., the home screen) and may be adjusted to increase or decrease the brightness of the display. However, it is sometimes difficult to reach these settings because a user may already be in an environment that makes viewing the display difficult. In addition, it is inconvenient for users who frequently traverse between bright and dark environments to constantly have to adjust these settings.
These technical issues present difficulties for users in that constantly adjusting the brightness of a display may be difficult and burdensome. A conventional solution is to incorporate into the device one or more ambient light sensor(s) (ALS). An ALS detects the amount of light in the environment and may be used wherever the settings of a system need to be adjusted to the ambient light conditions as perceived by users. For example, an ALS is used to maintain the same display appearance in LCD screens on devices under all light conditions. When an increasing amount of light falls on the ALS, the device's display correspondingly becomes brighter to provide the user with an improved ability to view the content on the display. Conversely, when there is less light in the surrounding environment, the display dims so the brightness from the display does not aggravate a user's eyes. However, an ALS incorporated into devices does not account for a user's sensitivity to light, which varies for each individual. In addition, devices containing an ALS have difficulty in extremely bright or dark environments. For example, when outdoors on a sunny day, a user's eye may adjust to the very bright surroundings, but a device placed in the shade may not choose a bright enough setting because only a modest amount of light will fall on the ALS. In another example, in a very dark room a device's display will be the brightest object in the room so the feedback from the display that falls on the ALS may cause the display to be brighter than necessary.
Accordingly, an embodiment provides a method for automatically adjusting the brightness of a display based upon a user's pupillary dilation. In an embodiment, a forward (user) facing camera on a device may be used to capture an image of a user's eye, which may then be analyzed to measure the width of the user's pupil. The brightness of the display may then automatically be adjusted to an ideal brightness setting that corresponds with the measured pupil width or other relevant dimension related to a user's sensitivity to display screen brightness. In an embodiment, an individual user's sensitivity to light may be measured by correlating a measure of the pupillary response with ambient light measurements from the environment.
The illustrated example embodiments will be best understood by reference to the figures. The following description is intended only by way of example, and simply illustrates certain example embodiments.
While various other circuits, circuitry or components may be utilized in devices, with regard to smart phone and/or tablet circuitry 100, an example illustrated in
There are power management chip(s) 130, e.g., a battery management unit, BMU, which manage power as supplied, for example, via a rechargeable battery 140, which may be recharged by a connection to a power source (not shown). In at least one design, a single chip, such as 110, is used to supply BIOS like functionality and DRAM memory.
System 100 typically includes one or more of a WWAN transceiver 150 and a WLAN transceiver 160 for connecting to various networks, such as telecommunications networks and wireless Internet devices, e.g., access points. Additional devices 120 are commonly included. In an embodiment, a camera or other sensor may be included as an additional device 120, e.g., for imaging a user in connection with measuring eye dimensions related to display brightness, as further described herein. System 100 often includes a touch screen or touch surface 170 for data input and display/rendering. System 100 also typically includes various memory devices, for example flash memory 180 and SDRAM 190.
The example of
In
In
The system, upon power on, may be configured to execute boot code 290 for the BIOS 268, as stored within the SPI Flash 266, and thereafter processes data under the control of one or more operating systems and application software (for example, stored in system memory 240). An operating system may be stored in any of a variety of locations and accessed, for example, according to instructions of the BIOS 268. As described herein, a device may include fewer or more features than shown in the system of
Device circuitry, as for example outlined in
Referring now to
In an embodiment, a forward facing camera may be utilized. In an embodiment, the camera may be a light based-camera (e.g., an RGB camera). Light-based cameras may be used in environments that contain an amount of visible light (e.g., light with a wavelength in the 400-700 nm range). These types of cameras utilize the visible light to accurately capture the details of an image.
In darker ambient light environments, light based and ambient light cameras are not as effective because there is a scarcity of visible light. Additionally, RGB cameras perform poorly in brighter environments (e.g., outdoors on a sunny day) when the device is in shadow because they are unable to attain a good visual on the user because the user will appear very dark in contrast to the bright environment. In an embodiment, under low light conditions, the camera may be an infrared (IR) camera. An IR camera is a device that forms an image using IR radiation, which is emitted by all objects based on their temperatures. An IR camera may operate in very long wavelengths (e.g., up to 14,000 nm) which makes it possible for these cameras to accurately capture an image with or without visible illumination.
In an embodiment, alternative depth camera type arrangements may be utilized. In an embodiment, cameras having an active emitter (e.g., laser or IR emitter) may be used to capture an image in total darkness. In an embodiment, cameras having a very long exposure time may also be utilized because they work well in low light conditions. By leaving a camera's shutter open for an extended period of time, more light is absorbed, creating an exposure that captures the entire range of the digital camera sensor. In an embodiment, a variety of different types of cameras or sensors may also be utilized including, but not limited to, laser painting and specialized low-light sensors (e.g., back-illuminated sensors), which increase the amount of light captured and thereby improve low-light performance. In an embodiment, if there is more than one type of sensor or camera on board a device, the device may switch between the different types of sensors to use the sensor that would produce the most ideal image in a particular environment.
In an embodiment, the frequency of picture capture for pupil measurement may be adjusted. For example, in an embodiment, an image may be taken every three seconds. As another example, multiple images may be taken each second. In an embodiment, the frequency of image capture may be adjusted, e.g., by the user or automatically according to a policy or programmed adjustment. In an embodiment, the frequency of image capture may be dynamically adjusted based upon data received from the ALS. For example, when data associated with the ALS indicates that the device has been moved to a brighter environment, the camera may be prompted to take an image of the user's pupil in this new environment.
At 302, an embodiment may analyze the captured image to measure the width of the pupil. Rather than, or in addition to, measuring the amount of light that falls on the device's sensor, an embodiment may measure a user's biological response to the amount of light entering the eyes. In an embodiment, systems that include gaze detection and facial recognition may be used to perform the pupil measurement.
At 303, an embodiment may adjust the brightness of the display based on the pupil measurement. In an embodiment, the brightness of the display may be adjusted by reference to a pupil size, for example as stored in a data table. After the size of a pupil is measured, the measurement may be cross-referenced with a data table that contains predetermined, corresponding brightness settings for each measured pupil size. The display may be automatically adjusted to the brightness setting that corresponds to the measured pupil size. For example, when the size of a pupil is measured to be X mm wide, the brightness of the display is set to Y nits. The data table may be stored on the device, in the cloud, may be accessed online, etc.
In an embodiment, the brightness of the display may be adjusted based on an individual user's needs. As individuals age, muscles that control pupil size and reaction to light lose strength. This causes the pupils to become less responsive to changes in ambient lighting. Because of these changes, more ambient light is required to adequately perceive a particular object. For example, it is estimated that an individual in their advanced age needs as much as three times more ambient light for comfortable reading than a younger individual. Additionally, due to the sluggish response the pupils have to changes in light, certain individuals have increased difficulty when passing from a brightly lit environment to a darker one, and vice versa. For example, certain individuals are more likely to be dazed by bright sunlight and glare when emerging from a dimly lit building (e.g., a movie theater) because the dilated pupils require additional time to adjust to the new environment.
An embodiment may correlate a measure of the pupillary response with ambient light measurements from the environment to measure an individual user's sensitivity to light. In an embodiment, when new environmental light data is received by an ALS, a camera may be prompted to capture an image of the user's pupil(s) to determine pupil width. For example, if a user passes from a dark environment to a lighter environment (e.g., from a dimly lit movie theater to a bright outdoor environment), light data received by the ALS may prompt a camera to take an image of the user's pupil(s) in this new environment. Obtaining such personalized information assists in determining the optimal brightness control of the display. For example, using this measure, more accurate brightness settings may be achieved for all users in all conditions. Additionally, the rate at which the display screen brightness is adjusted may be changed based on a personalized setting. For example, a first user may have quick reaction to changing light, whereas a second user may have a slower response. Thus, an embodiment may tune the rate at which the display brightness is modified for the first and second user different.
In an embodiment, the brightness of the display may be adjusted after each instance of pupil size measurement. Referring now to
In an embodiment, the foregoing methods may be applied to virtual reality and augmented reality goggles. When in use, these goggles block out nearly all ambient light. A forward facing camera (e.g., an RGB camera) may be placed on the inner portion of the goggles to be able to capture images of a user's pupil. Measuring a user's pupil size while displaying a bright scene followed by a dark scene would provide enough data to properly adjust the brightness of the goggle display based on the user's particular sensitivity to display brightness. In an embodiment, this process requires no user interaction and is performed automatically.
The various embodiments described herein thus represent a technical improvement to conventional methods of adjusting the brightness of a display. Using the techniques described here, an embodiment provides an automated adjustment of display brightness based upon the measurement of a user's pupil size. Using this measure, more accurate brightness settings may be achieved for all users in all conditions.
As will be appreciated by one skilled in the art, various aspects may be embodied as a system, method or device program product. Accordingly, aspects may take the form of an entirely hardware embodiment or an embodiment including software that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects may take the form of a device program product embodied in one or more device readable medium(s) having device readable program code embodied therewith.
It should be noted that the various functions described herein may be implemented using instructions stored on a device readable storage medium such as a non-signal storage device that are executed by a processor. A storage device may be, for example, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a storage medium would include the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a storage device is not a signal and “non-transitory” includes all media except signal media.
Program code embodied on a storage medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, et cetera, or any suitable combination of the foregoing.
Program code for carrying out operations may be written in any combination of one or more programming languages. The program code may execute entirely on a single device, partly on a single device, as a stand-alone software package, partly on single device and partly on another device, or entirely on the other device. In some cases, the devices may be connected through any type of connection or network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made through other devices (for example, through the Internet using an Internet Service Provider), through wireless connections, e.g., near-field communication, or through a hard wire connection, such as over a USB connection.
Example embodiments are described herein with reference to the figures, which illustrate example methods, devices and program products according to various example embodiments. It will be understood that the actions and functionality may be implemented at least in part by program instructions. These program instructions may be provided to a processor of a device, a special purpose information handling device, or other programmable data processing device to produce a machine, such that the instructions, which execute via a processor of the device implement the functions/acts specified.
It is worth noting that while specific blocks are used in the figures, and a particular ordering of blocks has been illustrated, these are non-limiting examples. In certain contexts, two or more blocks may be combined, a block may be split into two or more blocks, or certain blocks may be re-ordered or re-organized as appropriate, as the explicit illustrated examples are used only for descriptive purposes and are not to be construed as limiting.
As used herein, the singular “a” and “an” may be construed as including the plural “one or more” unless clearly indicated otherwise.
This disclosure has been presented for purposes of illustration and description but is not intended to be exhaustive or limiting. Many modifications and variations will be apparent to those of ordinary skill in the art. The example embodiments were chosen and described in order to explain principles and practical application, and to enable others of ordinary skill in the art to understand the disclosure for various embodiments with various modifications as are suited to the particular use contemplated.
Thus, although illustrative example embodiments have been described herein with reference to the accompanying figures, it is to be understood that this description is not limiting and that various other changes and modifications may be affected therein by one skilled in the art without departing from the scope or spirit of the disclosure.