This application generally relates to determining oxygen levels from images of skin.
Changes in blood volume in the blood vessels of a human body relate to important physiological phenomena. For example, blood-volume pulses correspond to a person's heartbeat and blood pressure. In addition, changes in blood volume can be used to estimate oxygen levels in a person's blood. For example, changes in blood volume can provide information about oxygen saturation (SpO2), which is a measure of the percentage of oxygen-bounded hemoglobin over the total hemoglobin in a user's blood.
The oxygen saturation of a person's blood (SpO2) can be measured using an arterial blood gas test. This test requires taking a blood draw from a person's artery and must be performed in a clinical setting. The test is invasive and painful, and is not continuous in that each blood draw only provides information about a person's SpO2 at the point in time corresponding to the blood draw. In addition, the arterial blood gas test does not provide immediate results because the drawn blood must be sent to a lab for analysis.
A finger pulse oximeter is a non-invasive test that uses light to estimate a person's SpO2. However, a finger pulse oximeter requires continuous contact with a person's finger, e.g., by being clamped to the finger, and therefore the test does not provide a convenient measurements of a person's SpO2—particularly continuous measurements, as those would require a person to leave the pulse oximeter attached to their finger over time, limiting use of that hand. In addition, because a finger pulse oximeter requires contact with a person's finger, this approach can spread infections when the same oximeter is used by different people.
Non-invasive SpO2 measurements that use light rely on a ratio-of-ratios technique to estimate SpO2 in a person's blood. When light from a light source is incident on a person's skin, some of the incident light reflects off the surface of a person's skin, known as specular reflection, and some of the light passes into the person's tissue. The person's tissue also reflects some light, and some of this reflected light may travel back through the person's tissue and pass out of the person's skin, known as diffuse reflection. A light sensor can capture both light from specular reflection and light from diffuse reflection. The characteristics of specular and diffuse reflection depend in-part on the wavelength of incident light. The ratio-of-ratios is derived based on the differential absorption of oxygenated hemoglobin (HbO2) and deoxygenated hemoglobin (HbR) at two or more different wavelengths, which correlates with blood oxygen saturation.
Diffuse reflection can vary based on the amount of blood that the light interacts with.
More formally, the light Rc, for a given wavelength c, reflected off a person's skin is represented by:
R
c=ΣλcL(λ,{right arrow over (x)},t)s(λ,{right arrow over (x)},t)rc(λ,t),
where rc(λ, t) represents the sensor response for a given wavelength c; s(λ,{right arrow over (x)},t) represents the light reflectance, which is equal to =m(λ,{right arrow over (x)})b(λ,{right arrow over (x)},t); and L(λ,{right arrow over (x)},t) represents the incident light, which is equal to =l({right arrow over (x)},t){circumflex over (L)}(λ). In these equations, m(λ,{right arrow over (x)}) corresponds to specular reflection, which represents the effect from non-blood tissue (e.g., melanin, water, etc.). b(λ,{right arrow over (x)},t) corresponds to diffuse reflect, which represents the effect from blood-light interactions. {circumflex over (L)}(λ) represents the spectral distribution of the incident light and t) represents the intensity of the incident light. The time-dependent diffuse reflection can be represented as:
b(λ,{right arrow over (x)},t)=v({right arrow over (x)},t)bDC(λ,t)+Δv({right arrow over (x)},t)bAC(λ,t),
where v represents the volume of static (no-pulse) blood, Δv represents the volume of pulsatile blood; bDC(λ,t) represents the reflectance of static blood; and bAC(λ,t) represents the reflectance of pulsatile blood.
In the ratio-of-ratio technique, the effect of light intensity can be removed from the light response by:
which can be rewritten as:
The dependency on blood volume can be eliminated by:
which can be rewritten as:
where AC and DC refer to the pulsatile elements and non-pulsatile elements of the signal, at their respective subscripted wavelengths. As discussed more fully herein, S(t) is only related to blood reflectance if specular reflection is spatially invariant across the measured sample, i.e., if m(λ,{right arrow over (x)})=m(λ).
For the ratio-of-ratios technique to accurately reflect SpO2, the specular reflectance of the skin must be uniform in the measurement area. For remote PPG sensing, which attempts to measure SpO2 using non-contact light-sensing techniques, this uniformity requirement presents a significant challenge due to the non-uniformity of the human skin and due to user movement during measurement time. For example, a portion of a person's skin, such as the person's face, can have spatially varying reflectance properties due to variations in skin tissues (e.g., the presence of hair, color and thickness variations in the user's skin, moles, etc.). In addition, while movement is a non-issue in finger pulse oximetry due to the sensor being clamped to a specific portion of the user's finger, in remote PPG the person can move relative to the measurement apparatus (e.g., a camera), and this relative movement can affect the measurement results.
In addition, the ratio-of-ratios technique assumes that a sensor (e.g., camera) can independently detect two different wavelengths of light. For example, a finger pulse oximeter uses specialized sensors that detect red light and IR light, respectively, and the sensitive ranges of these sensors do not overlap. However, consumer cameras do not meet this requirement, as the camera response in different color channels (e.g., different wavelengths or ranges of wavelengths) overlaps between channels. For example,
Step 320 of the example method of
Particular embodiments may determine microregions according to the techniques described in Patent Application Publication No. 2023/0128766 which is incorporated herein by reference. However, this disclosure contemplates that any suitable technique for determining microregions of a portion of human skin may be used in step 320. Once determined, whether from one image or from multiple images, the mROIs are tracked (i.e., associated with their respective skin regions) across all of the input images.
Step 330 of the example method of
Each microregion is defined by its boundaries CX,Y, which identifies all the X,Y coordinates of all pixels within that mROI. In particular embodiments, the skin properties of an mROI may be determined based on the pixel values of the pixels within that mROI. For example, the skin properties of an mROI may be determined based on the average intensity of all pixels within the mROI for a given color channel (e.g., red, green, blue), resulting in channel-specific intensities Ir, Iy, Ib for each mROI. While this example refers to the RGB color channels, this disclosure contemplates that other color-channel representations may be used.
In the example of
Notably, this example process increases the number of pixels in each ROI (by merging similar mROIs) while maintaining the skin-homogeneity assumptions necessary to accurately determe SpO2 using the ration-of-ratios technique. The signal-to-noise (SNR) ratio in the PPG signal is proportional to the number of pixels in the ROI, and therefore more pixels is better for ensuring quality signals. However, pixels that violate the skin-homogeneity assumptions necessary for the ratio-of-ratios technique should not be included in the same ROI, and therefore the example process described above improves SNR by increasing the number of pixels in each ROI while ensuring that the skin-homogeneity assumption still apply.
In particular embodiments, the value of thresholds T1 and T2 can be customized based on the person's skin tone and/or on the ambient light intensity, as explained herein. In particular embodiments, the value of thresholds T1 and T2 can compensate for uneven light incident on the portion of the person's skin, e.g., between one side of the face and the other side of the face. For example, if the light source in the example of
Step 340 of the example method of
In the example of
Step 350 of the example method of
This estimated value of τ can then be used to generate weights for the rPPG signals for each ROI. For instance,
Particular embodiments may use an adaptive SpO2 estimation process to assess the lighting environment at the time the images were taken to determine the color pairs (i.e., the pair of color channels) used for SpO2 estimation. Then, the rPPG signals for these color channels are used to determine the second ratio within the ratio-of-ratios framework. Particular embodiments use a non-linear machine learning model, trained to estimate SpO2 values from training ratios, to estimate the SpO2 values corresponding to the determined ratios.
The lighting conditions are provided to a feature selector block, which selects the color-pairs to use for SpO2 estimation. For example, in
Particular embodiments may store the resulting SpO2 estimates for the user. Particular embodiments may display SpO2 estimates in real-time or near real-time e.g., in a health application on a user's computing device. Particular embodiments may provide these estimates to a health provider, e.g., the person's doctor. Particular embodiments may create an alert or alarm, e.g., to the user or to medical personnel, when the user's SpO2 level is determined to be below a threshold level.
Particular embodiments may repeat one or more steps of the method of
Ambient light that has a relatively stronger light intensity in longer wavelengths, which corresponds to a lower color temperature, provides relatively better discrimination power between oxygenated and deoxygenated hemoglobin, thereby providing relatively better SpO2 estimates using remote PPG. Particular embodiments may use one or more display screens in the person's environment to generate pure selected colors (e.g., pure red, pure green, etc.). For example, a display of a user's smartphone or TV may be instructed to generate a pure red image, creating red illumination, while images of the person's skin are being captured. Particular embodiments may train an SpO2 estimation model using these same light source(s). Particular embodiments may determine the distance between the user's face and the display, and use this information when making SpO2 estimates.
The color channels on a consumer RGB camera can have a relatively low color selectivity, i.e., other color components can leak into a specific channel. For example, the photons sensed by a consumer camera in the green channel include photons in neighboring blue and red channels, not just photons corresponding to the green channel. This color leaking reduces the linearity of the discrimination power between oxygenated and deoxygenated hemoglobin, complicating SpO2 estimation. Therefore, particular embodiments synchronize lighting source(s) and a camera's shutter (e.g., a smartphone camera shutter), such that the shutter is open while a red color is displayed on the display, then closes, then re-opens when a green color is displayed on the display, etc. Particular embodiments may cycle through an, e.g., red-green-blue display/shutter sequence n number of times to capture a sequence of images. Particular embodiments may assign images frames to R, G, or B corresponding to the point in the red-blue-green sequence in which those images were taken. In the rPPG extraction and processing steps (e.g., the rPPG buffering and extraction process of
Particular embodiments may remove the presence of non-controlled ambient light in a sequence of images. For example, particular embodiments may use a display screen, such as a smartphone or TV screen, to create enhanced light as described above. A camera may take an alternating sequence of images of a person under this enhanced light and without this enhanced light. The non-enhanced ambient light can then be extracted from the image sequence, and SpO2 can be determined using only the enhanced-light images.
Particular embodiments may capture a sequence of images and perform SpO2 estimates periodically (e.g., every 30 seconds, every 1 minute, every hour, etc.), for example on a schedule set by a user. Particular embodiments may perform SpO2 estimates on demand. Particular embodiments may perform SpO2 estimates passively, e.g., while a user is viewing content on their smartphone or TV or otherwise occupying an ambiently-lit space (e.g., while exercising, reading, etc.). Particular embodiments may include providing enhanced lighting conditions, such as from a device display, in order to provide better detection accuracy. Particular embodiments may capture a sequence of images while instructing a user to remain still, e.g., by keeping their face in a box displayed on a display.
This disclosure contemplates any suitable number of computer systems 1000. This disclosure contemplates computer system 1000 taking any suitable physical form. As example and not by way of limitation, computer system 1000 may be an embedded computer system, a system-on-chip (SOC), a single-board computer system (SBC) (such as, for example, a computer-on-module (COM) or system-on-module (SOM)), a desktop computer system, a laptop or notebook computer system, an interactive kiosk, a mainframe, a mesh of computer systems, a mobile telephone, a personal digital assistant (PDA), a server, a tablet computer system, or a combination of two or more of these. Where appropriate, computer system 1000 may include one or more computer systems 1000; be unitary or distributed; span multiple locations; span multiple machines; span multiple data centers; or reside in a cloud, which may include one or more cloud components in one or more networks. Where appropriate, one or more computer systems 1000 may perform without substantial spatial or temporal limitation one or more steps of one or more methods described or illustrated herein. As an example and not by way of limitation, one or more computer systems 1000 may perform in real time or in batch mode one or more steps of one or more methods described or illustrated herein. One or more computer systems 1000 may perform at different times or at different locations one or more steps of one or more methods described or illustrated herein, where appropriate.
In particular embodiments, computer system 1000 includes a processor 1002, memory 1004, storage 1006, an input/output (I/O) interface 1008, a communication interface 1010, and a bus 1012. Although this disclosure describes and illustrates a particular computer system having a particular number of particular components in a particular arrangement, this disclosure contemplates any suitable computer system having any suitable number of any suitable components in any suitable arrangement.
In particular embodiments, processor 1002 includes hardware for executing instructions, such as those making up a computer program. As an example and not by way of limitation, to execute instructions, processor 1002 may retrieve (or fetch) the instructions from an internal register, an internal cache, memory 1004, or storage 1006; decode and execute them; and then write one or more results to an internal register, an internal cache, memory 1004, or storage 1006. In particular embodiments, processor 1002 may include one or more internal caches for data, instructions, or addresses. This disclosure contemplates processor 1002 including any suitable number of any suitable internal caches, where appropriate. As an example and not by way of limitation, processor 1002 may include one or more instruction caches, one or more data caches, and one or more translation lookaside buffers (TLBs). Instructions in the instruction caches may be copies of instructions in memory 1004 or storage 1006, and the instruction caches may speed up retrieval of those instructions by processor 1002. Data in the data caches may be copies of data in memory 1004 or storage 1006 for instructions executing at processor 1002 to operate on; the results of previous instructions executed at processor 1002 for access by subsequent instructions executing at processor 1002 or for writing to memory 1004 or storage 1006; or other suitable data. The data caches may speed up read or write operations by processor 1002. The TLBs may speed up virtual-address translation for processor 1002. In particular embodiments, processor 1002 may include one or more internal registers for data, instructions, or addresses. This disclosure contemplates processor 1002 including any suitable number of any suitable internal registers, where appropriate. Where appropriate, processor 1002 may include one or more arithmetic logic units (ALUs); be a multi-core processor; or include one or more processors 1002. Although this disclosure describes and illustrates a particular processor, this disclosure contemplates any suitable processor.
In particular embodiments, memory 1004 includes main memory for storing instructions for processor 1002 to execute or data for processor 1002 to operate on. As an example and not by way of limitation, computer system 1000 may load instructions from storage 1006 or another source (such as, for example, another computer system 1000) to memory 1004. Processor 1002 may then load the instructions from memory 1004 to an internal register or internal cache. To execute the instructions, processor 1002 may retrieve the instructions from the internal register or internal cache and decode them. During or after execution of the instructions, processor 1002 may write one or more results (which may be intermediate or final results) to the internal register or internal cache. Processor 1002 may then write one or more of those results to memory 1004. In particular embodiments, processor 1002 executes only instructions in one or more internal registers or internal caches or in memory 1004 (as opposed to storage 1006 or elsewhere) and operates only on data in one or more internal registers or internal caches or in memory 1004 (as opposed to storage 1006 or elsewhere). One or more memory buses (which may each include an address bus and a data bus) may couple processor 1002 to memory 1004. Bus 1012 may include one or more memory buses, as described below. In particular embodiments, one or more memory management units (MMUs) reside between processor 1002 and memory 1004 and facilitate accesses to memory 1004 requested by processor 1002. In particular embodiments, memory 1004 includes random access memory (RAM). This RAM may be volatile memory, where appropriate Where appropriate, this RAM may be dynamic RAM (DRAM) or static RAM (SRAM). Moreover, where appropriate, this RAM may be single-ported or multi-ported RAM. This disclosure contemplates any suitable RAM. Memory 1004 may include one or more memories 1004, where appropriate. Although this disclosure describes and illustrates particular memory, this disclosure contemplates any suitable memory.
In particular embodiments, storage 1006 includes mass storage for data or instructions. As an example and not by way of limitation, storage 1006 may include a hard disk drive (HDD), a floppy disk drive, flash memory, an optical disc, a magneto-optical disc, magnetic tape, or a Universal Serial Bus (USB) drive or a combination of two or more of these. Storage 1006 may include removable or non-removable (or fixed) media, where appropriate. Storage 1006 may be internal or external to computer system 1000, where appropriate. In particular embodiments, storage 1006 is non-volatile, solid-state memory. In particular embodiments, storage 1006 includes read-only memory (ROM). Where appropriate, this ROM may be mask-programmed ROM, programmable ROM (PROM), erasable PROM (EPROM), electrically erasable PROM (EEPROM), electrically alterable ROM (EAROM), or flash memory or a combination of two or more of these. This disclosure contemplates mass storage 1006 taking any suitable physical form. Storage 1006 may include one or more storage control units facilitating communication between processor 1002 and storage 1006, where appropriate. Where appropriate, storage 1006 may include one or more storages 1006. Although this disclosure describes and illustrates particular storage, this disclosure contemplates any suitable storage.
In particular embodiments, I/O interface 1008 includes hardware, software, or both, providing one or more interfaces for communication between computer system 1000 and one or more I/O devices. Computer system 1000 may include one or more of these I/O devices, where appropriate. One or more of these I/O devices may enable communication between a person and computer system 1000. As an example and not by way of limitation, an I/O device may include a keyboard, keypad, microphone, monitor, mouse, printer, scanner, speaker, still camera, stylus, tablet, touch screen, trackball, video camera, another suitable I/O device or a combination of two or more of these. An I/O device may include one or more sensors. This disclosure contemplates any suitable I/O devices and any suitable I/O interfaces 1008 for them. Where appropriate, I/O interface 1008 may include one or more device or software drivers enabling processor 1002 to drive one or more of these I/O devices. I/O interface 1008 may include one or more I/O interfaces 1008, where appropriate. Although this disclosure describes and illustrates a particular I/O interface, this disclosure contemplates any suitable I/O interface.
In particular embodiments, communication interface 1010 includes hardware, software, or both providing one or more interfaces for communication (such as, for example, packet-based communication) between computer system 1000 and one or more other computer systems 1000 or one or more networks. As an example and not by way of limitation, communication interface 1010 may include a network interface controller (NIC) or network adapter for communicating with an Ethernet or other wire-based network or a wireless NIC (WNIC) or wireless adapter for communicating with a wireless network, such as a WI-FI network. This disclosure contemplates any suitable network and any suitable communication interface 1010 for it. As an example and not by way of limitation, computer system 1000 may communicate with an ad hoc network, a personal area network (PAN), a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), or one or more portions of the Internet or a combination of two or more of these. One or more portions of one or more of these networks may be wired or wireless. As an example, computer system 1000 may communicate with a wireless PAN (WPAN) (such as, for example, a BLUETOOTH WPAN), a WI-FI network, a WI-MAX network, a cellular telephone network (such as, for example, a Global System for Mobile Communications (GSM) network), or other suitable wireless network or a combination of two or more of these. Computer system 1000 may include any suitable communication interface 1010 for any of these networks, where appropriate. Communication interface 1010 may include one or more communication interfaces 1010, where appropriate. Although this disclosure describes and illustrates a particular communication interface, this disclosure contemplates any suitable communication interface.
In particular embodiments, bus 1012 includes hardware, software, or both coupling components of computer system 1000 to each other. As an example and not by way of limitation, bus 1012 may include an Accelerated Graphics Port (AGP) or other graphics bus, an Enhanced Industry Standard Architecture (EISA) bus, a front-side bus (FSB), a HYPERTRANSPORT (HT) interconnect, an Industry Standard Architecture (ISA) bus, an INFINIBAND interconnect, a low-pin-count (LPC) bus, a memory bus, a Micro Channel Architecture (MCA) bus, a Peripheral Component Interconnect (PCI) bus, a PCI-Express (PCIe) bus, a serial advanced technology attachment (SATA) bus, a Video Electronics Standards Association local (VLB) bus, or another suitable bus or a combination of two or more of these. Bus 1012 may include one or more buses 1012, where appropriate. Although this disclosure describes and illustrates a particular bus, this disclosure contemplates any suitable bus or interconnect.
Herein, a computer-readable non-transitory storage medium or media may include one or more semiconductor-based or other integrated circuits (ICs) (such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)), hard disk drives (HDDs), hybrid hard drives (HHDs), optical discs, optical disc drives (ODDs), magneto-optical discs, magneto-optical drives, floppy diskettes, floppy disk drives (FDDs), magnetic tapes, solid-state drives (SSDs), RAM-drives, SECURE DIGITAL cards or drives, any other suitable computer-readable non-transitory storage media, or any suitable combination of two or more of these, where appropriate. A computer-readable non-transitory storage medium may be volatile, non-volatile, or a combination of volatile and non-volatile, where appropriate.
Herein, “or” is inclusive and not exclusive, unless expressly indicated otherwise or indicated otherwise by context. Therefore, herein, “A or B” means “A, B, or both,” unless expressly indicated otherwise or indicated otherwise by context. Moreover, “and” is both joint and several, unless expressly indicated otherwise or indicated otherwise by context. Therefore, herein, “A and B” means “A and B, jointly or severally,” unless expressly indicated otherwise or indicated otherwise by context.
The scope of this disclosure encompasses all changes, substitutions, variations, alterations, and modifications to the example embodiments described or illustrated herein that a person having ordinary skill in the art would comprehend. The scope of this disclosure is not limited to the example embodiments described or illustrated herein. Moreover, although this disclosure describes and illustrates respective embodiments herein as including particular components, elements, feature, functions, operations, or steps, any of these embodiments may include any combination or permutation of any of the components, elements, features, functions, operations, or steps described or illustrated anywhere herein that a person having ordinary skill in the art would comprehend.
This application claims the benefit under 35 U.S.C. § 119 of U.S. Provisional Patent Applications 63/389,771 (filed Jul. 15, 2022) and 63/443,944 (filed Feb. 7, 2023), each of which is incorporated by reference herein.
Number | Date | Country | |
---|---|---|---|
63389771 | Jul 2022 | US | |
63443944 | Feb 2023 | US |