It is often observed that certain patients diagnosed with a disease have mild symptoms at the onset of the disease, but then experience a rapid deterioration. For example, patients diagnosed with sepsis or coronavirus disease can have mild symptoms that later lead to acute respiratory distress, multiple-organ failure, septic shock, and blood clots.
The present disclosure generally relates to eye imaging devices that include components to provide stimuli to assess disease and/or mental states. The present disclosure also relates to eye imaging devices that include one or more physical barriers to minimize exposure.
In one aspect, an example eye imaging device includes: a camera configured to capture fundus images of an eye of a patient; and a stimuli configured to stimulate the patient and assess a disease state or a mental state of the patient.
In another aspect, an example eye imaging device includes: a camera configured to capture fundus images of an eye of a patient; and a physical barrier surrounding at least a portion of the eye imaging device to minimize exposure of the eye imaging device.
The following drawing figures, which form a part of this application, are illustrative of the described technology and are not meant to limit the scope of the disclosure in any manner.
In the example illustrated in
In certain embodiments, a clinician C can use the display 108 of the eye imaging device 102 to capture and view one or more images of the patient P's eyes. Alternatively, the patient P can use the eye imaging device 102 him/herself such that assistance from the clinician C is not required to capture the one or more images of the patient P's eyes.
Accordingly, the eye imaging device 102 can be used by the clinician C or patient P to create a set of digital images of the patient P's eyes. In certain examples, the digital images of the patient P's eyes include images of the fundus. As used herein, “fundus” refers to the eye fundus and includes the retina, optic nerve, macula, vitreous, choroid, and posterior pole.
The eye imaging device 102 can be used to capture one or more images of the patient P's eyes for screening for an eye disease such as diabetic retinopathy, monitoring progression of diseases that may affect the microvasculature such as sepsis or coronavirus, and determining a disease risk assessment that predicts a likelihood of patient deterioration from the disease.
One technique for fundus imaging requires mydriasis, or the dilation of the patient's pupil, which can be painful and/or inconvenient to the patient P. The eye imaging device 102 does not require a mydriatic drug to be administered to the patient P before imaging, although the device can image the fundus if a mydriatic drug has been administered.
The eye imaging device 102 includes a camera 104 in communication with the image processor 106. The camera 104 is a digital camera including a lens, an aperture, and a sensor array. The lens of the camera 104 can be a variable focus lens, such as a lens moved by a step motor, or a fluid lens, also known as a liquid lens. In certain embodiments, the camera 104 records images of the fundus one eye at a time. In other embodiments, the camera 104 records images of both eyes substantially simultaneously. In such embodiments, the eye imaging device 102 can include two separate cameras, one for each eye.
The image processor 106 is operatively coupled to the camera 104 and is configured to communicate with the display 108 and network 110. The image processor 106 regulates the operation of the camera 104 and display 108. An example computing device that includes a processing unit such as the image processor 106, is shown in more detail in
The example eye imaging device 102 is connected to a network 110. The network 110 may include any type of wireless network, a wired network, or any communication network known in the art. For example, wireless connections can include cellular network connections and connections made using protocols such as 802.11a, b, and/or g. In other examples, a wireless connection can be accomplished directly between the eye imaging device 102 and an external display using one or more wired or wireless protocols, such as Bluetooth, Wi-Fi Direct, radio-frequency identification (RFID), or Zigbee. Other configurations are possible.
The timer 112 can be used to let the clinician C or the patient P know when an image or set of images should be taken by the eye imaging device 102. The use of the timer 112 will be described in more detail below with reference to the method 1400 of
In certain embodiments, the image processor 106 processes the one or more images captured by the camera 104 to generate an output such as a disease risk assessment that predicts a likelihood of patient deterioration from a disease based on a microvascular assessment. Thereafter, the eye imaging device 102 can utilize the network 110 to transfer the output to a server 114 that is remotely located with respect to the Patient P and clinician C.
In alternative embodiments, the image processor 106 transfers the one or more captured images to the server 114 via the network 110, and the server 114 processes the one or more images captured by the camera 104 and generates the output. In further alternative embodiments, the processing of the one or more images captured by the camera 104 is shared between the image processor 106 of the eye imaging device 102 and the server 114 such that some processing of the one or more captured images is performed by the image processor 106, while further processing of the one or more captured images is performed by the server 114.
As shown in
The housing 200 supports positional guides for the patient P such as a surface 210 on the opposite end 204 of the housing 200 that is configured to engage the patient P's head. In certain embodiments, the housing 200 may also support additional positional guides such as an optional adjustable chin rest. The surface 210 is configured to be positioned against the patient P's head and to surround both eyes of the patient P, as shown in the example of
In the example embodiment shown in
The display 108 functions to reproduce the images produced by the eye imaging device 102 in a size and format that are readable by a clinician. For example, the display 108 can be a liquid crystal display (LCD) and active matrix organic light emitting diode (AMOLED) display. The display 108 can be touch sensitive.
The housing 200 of example eye imaging device 102 is sized to be handheld. The display 108 can display images of the eye and controls for capturing those images. In some embodiments, the display 108 is a touchscreen. In some embodiments, the housing 200 additionally supports one or more user input buttons near display 108. The display 108 and user input buttons can be used to capture one or more images of the patient P's eyes. Thus, the eye imaging device 102 is capable of being configured such that the clinician C can implement one or more automatic and/or manual workflows to capture images of the patient P's eyes.
Additionally, the eye imaging device 102 can be configured to automatically perform workflows to capture one or more images of the patient P's eyes without requiring the patient P or clinician C to use the display 108 or the one or more user input buttons near display 108 to control the operation of the eye imaging device 102. Such configuration is helpful when the eye imaging device 102 is used by the patient P without assistance from the clinician C such as when the eye imaging device 102 is used to monitor the progression of a contagious disease such as a coronavirus (e.g., COVID-19) to reduce exposure to the clinician C and other caregivers within an acute care space or a medical surgical unit of a hospital.
The eye imaging device 102 can detect when the patient's P eyes are aligned with the one or two apertures 206 at the end 204 of the housing 200, such that the patient P is positioned and ready for the image capture sequence. In certain embodiments, the camera 104 of the eye imaging device 102 can detect when the patient P's eyes are aligned with the one or two apertures 206. In other embodiments, a dedicated sensor 212 can be used to detect when the surface 210 is in contact or otherwise engaged with the patient P's face. In some examples, the sensor 212 can be a pressure sensor that detects when the surface 210 is pressed against the patient P's head, or can be a light sensor that detects when the patient P's face covers the sensor 212. Alternative arrangements for the sensor 212 are possible.
As shown in
The patient support apparatus 400 further includes a right siderail assembly having at least one right siderail mounted on the right side of the frame 412 and a left siderail assembly having at least one left siderail mounted on the left side of the frame 412. In this example, the right siderail assembly includes an upper right siderail 410A and a lower right siderail 410B, and the left siderail assembly includes an upper left siderail 410C and a lower left siderail 410D.
In the embodiment shown in
In further alternative embodiments, the first end of the arm 300 can be attached directly to the frame 412 of the patient support apparatus 400 without attachment to a siderail 410A-410D. In further embodiments, the first end of the arm 300 can be attached to a headboard 413 or a footboard 414 of the patient support apparatus 400. Additional attachment locations for the arm 300 onto the patient support apparatus 400 are possible.
Like the medical surgical bed described above with respect to
As shown in
The position of the eye imaging device 102 when attached to the arm 300 can be manually adjusted by the patient P while the patient P is supported on the patient support apparatus 400, 402. For example, the patient P can manually grab the housing 200 to remove it from a storage location, and can place the housing 200, and more specifically the surface 210 up against their face as shown in
In certain embodiments, one or more user input buttons 308 can be positioned on the patient support apparatuses 400, 402, and can be used by the patient P to control the movement of the arm 300 using the motors 306 to adjust the position of the eye imaging device 102. For example, the user input buttons 308 can be positioned on a siderails 410, 430 of the patient support apparatus 400, 402. Alternatively, or in addition to the user input buttons 308, a remote control 310 can be provided for use by the patient P to control the movement of the arm 300 using the motors 306 to adjust the position of the eye imaging device 102.
In some further examples, the movement of the arm 300 and positioning of the eye imaging device 102 is automated by a controller 312 such that input from the patient P is not required, or is only partially required such as when the arm 300 is automated to place the eye imaging device 102 in front of the patient P's head, and all the patient P needs to do is to move the eye imaging device 102 a few inches to press the surface 210 against the patient P's head.
After issuing the alert, the method 500 includes an operation 504 of using the arm 300 to move the eye imaging device 102 in front of the patient P's head. In certain embodiments, the movement of the arm 300 is automated during operation 504 by the controller 312. For example, the one or more motors 306 can be controlled by the controller 312 to adjust the location of the eye imaging device 102 with respect to the patient P without requiring input from the patient.
Once the eye imaging device 102 is placed in front of the patient P's head, the method 500 includes operation 506 of alerting the patient P to grab the housing 200 and to place the eye imaging device 102 on their head, as shown in the example provided in
Next the method 500 includes operation 508 of using the camera 104 or sensor 212 to detect that the patient P's eyes are aligned with the one or two apertures 206 inside the cavity 208 formed at the end 204 of the housing 200. When it is detected that the patient P's eyes are not aligned with the one or two apertures 206 (i.e., “No” in operation 508), operation 506 can be repeated to instruct the patient P to place the eye imaging device 102 on their head.
Next, the method 500 includes an operation 510 of automatically capturing one or more images of the patient P's eyes when the camera 104 or sensor 212 detect that the patient P's eyes are aligned with the apertures 206 (i.e., “Yes” in operation 508). Advantageously, the images are captured without requiring any input from the patient P such that the patient P does not need to operate the display 108 or any user input buttons near display 108 to capture the one or more images of their eyes. Thus, the use of the eye imaging device 102 by the patient P is made easier, especially when the clinician C is unavailable to assist the patient P.
After completion of the image capture sequence, the method 500 can include an operation 512 of sending another alert to notify the patient P that the image capture sequence is complete, and to remove the eye imaging device 102 from their head. As described above, the alert can be an audible alert or any other types of notification including a notification displayed inside the cavity 208 formed at the end 204 of the housing 200 that can be viewed by the patient P while the eye imaging device 102 remains placed on the patient P's head.
Next, the method 500 can include an operation 514 of using the camera 104 or sensor 212 to detect whether the eye imaging device 102 is removed from the patient P's head. When the camera 104 or sensor 212 detects that the eye imaging device 102 has not been removed from the patient P's head (i.e., “No” in operation 514), operation 512 can be repeated to issue another alert to instruct the patient P to remove the eye imaging device 102 from their head.
Thereafter, the method 500 includes operation 516 of using the arm 300 to move the eye imaging device 102 away from the patient P's head, and to move the eye imaging device 102 to a storage location that does not interfere with the patient P while the patient P remains supported on the patient support apparatus 400, 402. For example, the arm 300 can stow the eye imaging device 102 inside the frame 412, 432, or behind a siderail 410, 430 of the patient support apparatus 400, 402 such that it does not get in the way of the patient P when it is not being used. In operation 516, the movement of the arm 300 can be automated such that the one or more motors 306 are controlled by the controller 312 to automatically adjust the location of the eye imaging device 102 with respect to the patient P without requiring input from the patient.
In certain embodiments, additional diagnostic devices may be attached to the arm 300. For example, certain diagnostic devices may require precise placement relative to the patient P, and the arm 300 can automate the movement of these devices to position them in the correct placement. Additional diagnostic devices that may be used with the arm 300 include devices used to record heart and lung sounds, and to record images the patient P's throat.
While the method 500 is described above as being performed without assistance from the clinician C, it is contemplated that in certain embodiments, the clinician C may assist the patient P in placing the eye imaging device 102 on their head. This may especially occur for patients who are too weak or deteriorated to grab and move the eye imaging device 102 even when the arm 300 is used to place the eye imaging device 102 in front of them.
As shown in
The kiosk 600 includes a digital display 606 that is also attached to the pedestal 602. The digital display 606 can be used to display instructions for a user to perform a method for disease risk assessment that uses the eye imaging device 102. In certain embodiments, the method for disease risk assessment is the method 1300 which is described in more detail below with reference to
The digital display 606 is a touchscreen that can be used by a user to enter additional information that can be used to calculate a risk score. For example, the user can use the digital display 606 to enter data such as their height, weight, body mass index (BMI), comorbidities, age, symptoms, exposure to contagious disease (e.g., COVID-19), and other similar data.
Like the kiosk 600 described above, the kiosk 700 can be placed within a hospital such as in an emergency room where the eye imaging device 102 can be used for disease risk assessment such as for sepsis or coronavirus. It is contemplated that the kiosk 700 may be used in additional medical settings such as in a primary care physician's office, a health clinic, pharmacy, long term care facility, nursing home, and the like. The kiosk 700 may also be used in non-medical settings such as in the lobby of an office building or an airport terminal.
The kiosk 700 may include a cubicle like structure 702 that includes a wall 704 that can provide privacy for the user, and may include a bench 706 for the user to sit on. Like, the kiosk 600 described above, the kiosk 700 includes a digital display 708 that can be used to display instructions that require the user to use the eye imaging device 102 to perform a method for disease risk assessment. Also, one or more speakers 710 may be embedded in the cubicle like structure 702 or may be included with the digital display 708 to provide audio instructions for the user. The digital display 708 can display a disease risk assessment based on one or more abnormalities that may be detected by the eye imaging device 102.
The kiosk 700 may include one or more input devices 712 mounted to a counter 714 that is attached to the wall 704. In certain examples, the one or more input devices 712 include a touchscreen. A user can use the one or more input devices 712 to enter additional information that can be used to calculate a risk score. For example, the user can use the one or more input devices 712 to enter data such as their height, weight, body mass index (BMI), comorbidities, age, symptoms, exposure to contagious disease (e.g., COVID-19), and other similar data.
As shown in
Next, the method 1300 includes an operation 1304 of alerting the user to use the eye imaging device 102 to capture one or more images of their eyes. The alert can be displayed on the digital displays 606, 708 or can be an audible alert generated from one or more speaker 608, 710 of the kiosks 600, 700. Additional types of alerts are possible.
In the example kiosk 600, the alert can simply request the user to press their face into the eye imaging device 102. In embodiments where the height of the pedestal 602 is adjustable, the alert can instruct the user to adjust the height of the pedestal 602 so that the eye imaging device 102 is level with their head, and thereafter press their face into the surface 210 of the eye imaging device 102. In embodiments where the eye imaging device 102 is removably attached to the pedestal 602, the alert can request the user to remove the eye imaging device 102 from the pedestal 602 and to place the eye imaging device 102 on their head.
In the example kiosk 700, the alert can instruct the user to remove the eye imaging device 102 from the wall 704 and to place it on their head. In embodiments where the eye imaging device 102 is fixed to the wall 704 of the kiosk 700, the alert can simply request the user to press their face against the surface 210 of the eye imaging device 102.
Next the method 1300 includes operation 1306 of using the camera 104 or sensor 212 to detect whether the user's eyes are aligned with the apertures 206. When the camera 104 or sensor 212 does not detect that the user's eyes are aligned with the apertures 206 (i.e., “No” in operation 1306), the method 1300 repeats operation 1304 to issue another alert that instructs the user to properly place the eye imaging device 102 on their head.
When the camera 104 or sensor 212 detects that the user's eyes are aligned with the apertures 206 (i.e., “Yes” in operation 1306), the method 1300 proceeds to an operation 1308 of capturing at least one image. The at least one image is automatically captured without requiring any input from the user such that the user does not need to operate the display 108 or any user input buttons near display 108 of the eye imaging device 102 to capture the image. This makes use of the eye imaging device 102 easier for the user.
After completion of the image capture sequence, the method 1300 includes an operation 1310 of sending another alert to notify the user that the image capture sequence is complete, and to release their head from contacting the surface 210 of the eye imaging device 102. In certain embodiments, the alert is audible. Alternatively, or in combination with an audible alert, a notification can be displayed inside the cavity 208 that can be viewed by the user while the user's face is pressed against the surface 210 of the eye imaging device 102.
Thereafter, the method 1300 includes an operation 1312 of calculating a risk score based on the information entered into the digital displays 606, 708 at operation 1302, and one or more abnormalities detected from image capture sequence performed at operation 1308. In some embodiments, the risk score is displayed on the digital displays 606, 708. In certain embodiments, in addition to displaying the risk score, the method 1300 may also display a referral for the user based on the risk score such as to seek immediate medical attention.
The method 1400 includes an initial operation 1402 of obtaining patient data. The patient data can include age, comorbidities, symptoms, complaints, lab results, prior vital signs readings, and the like. The patient data can be obtained directly from a patient such as in response to a questionnaire that allows the patient or a clinician to enter the patient data through a user input device. Alternatively, the patient data can be acquired from an Electronic Medical Record (EMR) of the patient when the identity of the patient is known.
Next, the method 1400 includes an operation 1404 of calculating an initial risk score for the patient based on the obtained patient data. The risk score represents a risk that the patient will experience a severe deterioration from a disease. A high risk score indicates that the patient P will likely experience a severe deterioration from the disease. A low risk score indicates that the patient P will not likely experience a severe deterioration from the disease.
In certain embodiments, the risk score is calculated for a virus such as COVID-19. In such embodiments, the risk score increases with age and comorbidities including, without limitation, cancer, chronic kidney disease, chronic obstructive pulmonary disease (COPD), immunocompromised state (i.e., weakened immune system), obesity (e.g., body mass index (BMI) of 30 or higher), serious heart conditions such as heart failure, coronary artery disease, or cardiomyopathies, sickle cell disease, and type 2 diabetes. Additionally, the risk score for COVID-19 may also increase with asthma (moderate-to-severe), cerebrovascular disease that affects blood vessels and bloody supply to the brain, cystic fibrosis, hypertension or high blood pressure, neurological conditions such as dementia, liver disease, pregnancy, pulmonary fibrosis (e.g., having damaged or scarred lung tissues), smoking, thalassemia, and type 1 diabetes.
In certain embodiments, the risk score is calculated for sepsis. In such embodiments, the risk score is based on age and comorbidities including, without limitation, pregnancy, weakened immune system, chronic conditions such as diabetes, lung, kidney, liver diseases, dementia, and cancer, prior hospitalizations, use of indwelling prosthetic devices such as catheters, heart valves, vascular bypass grafts, ocular lenses, artificial joints, and central nervous shunts, and malnutrition. Aspects for identifying sepsis risk and monitoring sepsis progression are described in U.S. patent application Ser. No. 16/832,672 filed on Mar. 27, 2020, U.S. patent application Ser. No. 16/847,729 filed on Apr. 14, 2020, and U.S. patent application No. 62/893,985 filed on Aug. 30, 2019, all of which are hereby incorporated by reference in their entirety.
In certain embodiments, the eye imaging device 102 automatically calculates the risk score based on the obtained patient data. In alternative examples, the eye imaging device 102 can receive the calculated risk score from another device that can transfer the calculated risk score to the eye imaging device 102 through the network 110. In some further embodiments, the eye imaging device 102 can receive the calculated risk score from a user who enters the risk score using the display 108 or one or more user input buttons near display 108.
Next, the method 1400 includes an operation 1406 of determining whether the risk score is high or low. When the risk score is low (i.e., “No” at operation 1406) the method 1400 proceeds to an operation 1408 of configuring the eye imaging device 102 to perform a first type of workflow. When the risk score is high (i.e., “Yes” at operation 1406) the method 1400 proceeds to an operation 1416 of configuring the eye imaging device 102 to perform a second type of workflow. In some embodiments, the risk score is a numerical value and the decision at operation 1406 is determined by comparing the risk score to a threshold value such that when the risk score is less than the threshold value, the risk score is determined to be low, and when the risk score exceeds the threshold value, the risk score is determined to be high.
The first type of workflow can be less complex than the second type of workflow. For example, the first type of workflow can include capturing a single image, whereas the second type of workflow can include capturing a plurality of images under different lighting conditions, filters, diopters, and the like. Each of the plurality of images from the second type of workflow may highlight different characteristics of the patient P's eyes for analysis.
As shown in
In certain embodiments, the eye imaging device 102 is connected to one or more systems via the network 110 that can identify the proper person allowed to use the eye imaging device 102 for capturing the at least one image. For example, access may be granted to certain persons such as the patient or a clinician assigned to the patient, while access may be denied for certain persons such as clinicians who are not assigned to the patient.
After completion of operation 1410, the method 1400 proceeds to an operation 1412 of analyzing the at least one image captured under the first type of workflow to determine whether there are any abnormalities. When no abnormalities are detected (i.e., “No” at operation 1412), the method 1400 terminates at operation 1414. When one or more abnormalities are detected (i.e., “Yes” at operation 1412), the method 1400 proceeds to operation 1420 of modifying the risk score based the one or more detected abnormalities. For example, the one or more detected abnormalities can increase the risk score. In certain embodiments, when one or more abnormalities are detected (i.e., “Yes” at operation 1412), the method 1400 may include presenting a manual override 1413 such that the method 1400 does not automatically proceed to operation 1420, and instead can terminate at operation 1414 when a clinician determines that they have sufficient information such that further analysis is not needed.
After completion of operation 1418, the method 1400 similarly proceeds to the operation 1420 of modifying the risk score based on data obtained from the at least one image captured under the second type of workflow. The data obtained from the at least one image captured under the second type of workflow can increase or decrease the risk score for patient deterioration. For example, when one or more abnormalities are detected from the at least one image captured under the second type of workflow, the risk score can be increased at operation 1420. Alternatively, when no abnormalities are detected from the at least one image captured under the second type of workflow, the risk score can be decreased at operation 1420.
Accordingly, the risk score determined at operation 1420 is an enhanced risk score that takes into consideration microvascular assessment from the at least one image captured under the first or second workflows. Advantageously, by adjusting the risk score to take into consideration the microvascular assessment, the enhanced risk score is more accurate in predicting patient deterioration from diseases that may affect the microvasculature. As described above, diseases that may affect the microvasculature include sepsis and COVID-19.
Next, the method 1400 proceeds to an operation 1422 of storing the enhanced risk score and the at least one image captured under the first or second types of workflow in a memory where it can be accessed by a clinician for further analysis. The memory can be on the eye imaging device 102 itself, or the memory can be remotely located from the eye imaging device 102 such as on a cloud network. The eye imaging device 102 is connected to one or more systems via the network 110 that can identify the proper person allowed to access the enhanced risk score and at least one image. For example, access may be granted to certain persons such as the patient or a clinician assigned to the patient, while access may be denied for certain persons such as clinicians who are not assigned to the patient.
Alternatively, or in addition to storing the enhanced risk score and at least one image, operation 1422 may also include sending the enhanced risk score and at least one image directly to a clinician for further analysis. In this example, the network 110 can be used to send the at least one image to the clinician, and one or more systems connected to the network 110 can be used identify the correct clinician to send the enhanced risk score and at least one image.
Next, the method 1400 proceeds to an operation 1424 of determining whether the workflows are complete. When the workflows are complete (i.e., “Yes” at operation 1424), the method 1400 ends at operation 1414. When the workflows are not complete (i.e., “No” at operation 1424), the method 1400 proceeds to an operation 1424 of setting or resetting the timer 112 (see
As described above, the timer 112 can be used to let the clinician or the patient know when an image or set of images should be taken by the eye imaging device 102. For example, the timer 112 after a predetermined period of time has expired can trigger an alert to be generated to remind the clinician or patient to take additional images using the eye imaging device 102. The alert can be generated on the eye imaging device 102 such as a flashing light or an audio instruction, or in other instances, the alert can be generated on another device such as a patient support apparatus 400, 402 (see
As an illustrative example, when the enhanced risk score is high, the timer 112 can be set to have a shorter interval of timer such that the alert will be generated more frequently. When the enhanced risk score is low, the timer 112 can be set have a longer interval of time such that the alert will be generated less frequently. Additionally, when the enhanced risk score increases from a previously calculated risk score, the interval of time set by the timer 112 for the alert can be shortened. Alternatively, when the enhanced risk score decreases from a previously calculated risk score, the interval of time set by the timer 112 for the alert can be lengthened. In certain embodiments, the timer 112 is automated by the computing device 1500 such that the time interval for the alert triggered by the timer 112 is automatically adjusted without any input from the clinician or patient, and is instead automatically updated based on the enhanced risk.
Next, the method 1400 proceeds to an operation 1428 of determining whether the condition of the patient P has changed such as whether the patient's condition has improved or deteriorated. In certain embodiments, the change in the patient's condition is determined from data obtained from the at least one image captured at operations 1410, 1418. Alternatively, or in addition to the data obtained from the at least one image captured at operations 1410, 1418, the change in the condition of the patient can be determined from updated vital signs measurements taken during performance of the method 1400 or other observations of the patient.
When the condition of the patient P is determined as not having changed (i.e., “No” at operation 1428), the method 1400 proceeds to capture additional images under the same workflow that was used to capture the images under operations 1410, 1418. When the condition of the patient P is determined as having changed (i.e., “Yes” at operation 1428), the method 1400 proceeds to an operation 1430 of reconfiguring the eye imaging device 102 to perform another type of workflow based on the changed condition that is different from the first and second types of workflows that are described above.
As an illustrative example, when the patient's condition is improving, the eye imaging device 102 can be reconfigured to perform another type of workflow that is less complex such that it captures fewer images, or less detailed images. As another illustrative example, when the patient's condition is deteriorating, the eye imaging device 102 can be reconfigured to perform another type of workflow that is more complex such that it captures a greater number of images, or images that are more detailed.
In some embodiments, the eye imaging device 102 is reconfigured at operation 1430 regardless of detected changes in the patient's condition. In such embodiments, fewer images or image types are captured to monitor the progress of that patient because fewer images or image types may be needed as the condition of the patient remains steady.
The method 1400 may repeat operations 1418-1430 as many times as needed until the workflows are complete (i.e., “Yes” at operation 1424). Each time the operations 1418-1430 are repeated, the eye imaging device 102 may automatically adjust the type of image or images it takes based on the reconfiguration performed at operation 1430. For instance, in a first pass through operations 1418-1430, a sequence of images providing a rich set of data are acquired at operation 1418, and each time the operations 1418-1430 are repeated, fewer images or fewer image types are acquired at operation 1418 as the condition of the patient progresses.
Additionally, repetition of operation 1420 provides a trend of enhanced risk scores based on updated microvascular assessments from the images acquired from the eye imaging device 102. The trended risk scores can indicate the patient's progress. Furthermore, repetition of operation 1422 can provide the clinician with continuous updates of the enhanced risk score.
The method 1400 can improve the use of the eye imaging device 102 in an acute care space or medical surgical unit by notifying a user when they should be taking an image or set of images with the eye imaging device 102 by using of the timer 112 to generate an alert. Also, the method 1400 can provide earlier detection of patient deterioration from diseases such as COVID-19 and sepsis by continuously monitoring microvascular changes in the patient's eyes. In yet other examples, the method 1400 can be used to monitor mental states, such as concussion, delirium, agitation, pain, changes in mental state, and categorization of mental state (such as objective assessment determination of Glasgow Coma Scale). Other applications are possible.
The computing device 1500 can also include a mass storage device 1514 that is able to store software instructions and data. The mass storage device 1514 is connected to the processing unit 1502 through a mass storage controller (not shown) connected to the system bus 1520. The mass storage device 1514 and its associated computer-readable data storage media provide non-volatile, non-transitory storage for the computing device 1500.
Although the description of computer-readable data storage media contained herein refers to a mass storage device, it should be appreciated by those skilled in the art that computer-readable data storage media can be any available non-transitory, physical device or article of manufacture from which the device can read data and/or instructions. The mass storage device 1514 is an example of a computer-readable storage device.
Computer-readable data storage media include volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable software instructions, data structures, program modules or other data. Example types of computer-readable data storage media include, but are not limited to, RAM, ROM, EPROM, EEPROM, flash memory or other solid-state memory technology, or any other medium which can be used to store information, and which can be accessed by the device.
The computing device 1500 may operate in a networked environment using logical connections to remote network devices, including the server 114, through the network 110, such as a local network, the Internet, or another type of network. The computing device 1500 connects to the network 110 through a network interface unit 1504 connected to the system bus 1520. The network interface unit 1504 may also be utilized to connect to other types of networks and remote computing systems.
The computing device 1500 can also include an input/output controller 1506 for receiving and processing input from a number of input devices. Similarly, the input/output controller 1506 may provide output to a number of output devices.
The mass storage device 1514 and the RAM 1510 can store software instructions and data. The software instructions can include an operating system 1518 suitable for controlling the operation of the device. The mass storage device 1514 and/or the RAM 1510 also store software instructions 1516, that when executed by the processing unit 1502, cause the device to provide the functionality of the device discussed in this document.
In addition to examples provided above, the eye image devices described herein can be used to screen for different disease states and mental states. For example, sepsis is a complex disease state caused by the dysregulated response of the immune system infection. Worldwide, 1.6 million people per year are diagnosed with sepsis, and the mortality rate ranges from 25% (severe sepsis) to over 50% (septic shock). In the United States alone, a quarter million people die from sepsis each year. This number does not include the over 150,000 deaths in the United States from infection with SARS-CoV-2 (Coronavirus), which from a pathophysiologic standpoint follows the sepsis pattern of a deadly dysregulated immune response.
A diagnostic implementation that impacts the early detection and treatment of sepsis can change the outcome metrics of sepsis. The eye imaging devices described herein can be configured to target diagnostics for assessment of early microvascular deterioration and other signs and symptoms of sepsis. For instance, microvascular deterioration caused by sepsis that may be seen with the images captured by the eye imaging device can include one or more of:
extravascular fluid/edema: will be prevalent due to the increased permeability of the damaged endothelium of the capillary walls;
microvascular clotting: caused by the hypercoagulation state of sepsis (disseminated intravascular clotting);
micro-aneurysms may be apparent: caused vessel wall weakness and sudden pressure changes;
local tissue oxygenation decreases causing pallor: occurs as oxygen transport and diffusion are disrupted by the microvascular changes;
perfusion heterogenicity: the population of perfused vessels will decrease but will be highly variable, with normally perfused and non-perfused, engorged and intermittently flowing vessels all within the same capillary beds;
visible structural changes in the vessels: vessel diameters and lumen-wall thickness ratios will change in response to dysregulated vasomotor reactions; and
blood volume flow velocities change: the vessel walls become “sticky” and the blood cells become less able to morph their shapes to accommodate the vessels.
Additional details about using the eye imaging devices to assess such microvascular deterioration is provided in U.S. Patent Application No. 63/063,593, Attorney Docket No. 10156.0167USP1, and U.S. Patent Application No. 63/063,619, Attorney Docket No. 10156.0166USP1, both filed on even date herewith, the entireties of which are hereby incorporated by reference.
Referring now to
For example,
Evaluation of a patient's mental state can inform a wide range of diagnostic tools and can guide patient care. Mental state includes a wide range of dimensions, including: Level of Consciousness (LOC); delirium; and concussion. Currently, this mental state assessment is subjective, manual, and requires trained caregiver assessment. Mental state can change rapidly, but current workflows may not allow for frequent assessments.
Integration of one or more visual and/or audible stimuli within the eye imaging device 1602 allows for an automated assessment of mental state to be performed. For instance, there is a relationship between eye movement and reaction and mental status.
The array of LEDs 1604 can be illuminated in given patterns, intensities, and/or wavelengths to provide a range of visual stimuli (varying in location, intensity, wavelength, etc.). For instance, the eye imaging device 1602 can be configured to provide photorefractive ocular screening techniques such as those described in U.S. Pat. No. 9,237,846 to Mowrey, which is hereby incorporated by reference.
The cameras 104, 1610, 1612 can be used to automate evaluation of gaze tracking in response to this stimuli. For instance, the camera 104 can be used to automate evaluation of retina size in response to stimuli. Further, the cameras 1610, 1612 can track macro eye movements in response to the stimuli. The eye imaging device 1602 can be programmed to measure various responses to those visual stimuli, including one or more of: pupil (change in pupil size, shape, or reactivity), which can be an important component of a neurological examination; and eye tracking, alignment, reflex, which are also used in neurological assessment.
In other examples, the cameras 104, 1610, 1612 can be used to capture other aspects of the eyes to aid in the assessment of disease states and/or mental states. For instance, the camera 104 can be programmed to provide conjunctiva imaging.
Like many other respiratory viruses, SARS-CoV-2 can directly inoculate the conjunctiva causing conjunctivitis. The eye imaging device 1602 can be used to image the surface of the eye before or in conjunction with retinal imaging. An algorithm can be used to analyze the surface image for abnormalities in the conjunctiva. This could include color processing for redness. The color of the sclera (whites) of the eye can indicate jaundice, which is often first detectable in the whites of the eyes and is a typically manifestation of liver disease. Liver dysfunction is one of the signs of sepsis. For example, the images could be compared against a known database of eyes on patients with healthy liver functions and those with liver dysfunction and include patient demographics.
In addition, the eye imaging device 1602 can be programmed to assess visual acuity and color perception. As the layers of the retina are affected by microvascular deterioration, and as the neurovascular coupling is consequently diminished there may be acute observable changes in visual acuity and the perception of color.
Further, the eye imaging device 1602 can be programmed to measure retinal oximetry through imaging of retinal blood vessels, which measures oxygen saturation of hemoglobin. The imaging technology is non-invasive and reproducible with low variability.
Retinal oximetry offers visible light imaging of systemic and central nervous system vessels. It senses hypoxia in cardiac and pulmonary diseases. Oximetry biomarkers have been discovered in Alzheimer's disease and multiple sclerosis and oxygen levels in the retina correspond well with brain.
In this example, the LEDs 1604 of the eye imaging device 1602 are configured to provide two different wavelengths of light (e.g., 660 nm and 940 nm) into the eye, then compare the images to determine SpO2 of the retina. The camera 104 can be configured with an imager that has multiple spectral return sensing regions.
Alternatively, the LEDs 1604 can include one or more independent red and infrared LEDs (see, e.g.,
In yet other examples, the eye imaging device 1602 is configured to provide retinal angiographic imaging optimization. Fluorescein is often used in the ophthalmological setting to assess the vascular status and function within the retina. Blood flow abnormalities including delayed retinal artery fill times and circulatory impairment at the microvascular level can be seen in septic patients using timed fluorescein contrast imaging.
The camera 104 is programmed to take real-time images. The Fluorescein is injected and then monitored for profusion over an extended period of time. The eye imaging device 1602 can be configured to provide alternate spectra (infrared or near infrared) during imaging. In some embodiments, the eye imaging device 1602 includes a hyperspectral imaging module.
In these examples, the eye imaging device 1602 can be programmed to lower the level of illumination flux (intensity of the illuminator is reduced) to minimize pupillary reaction. Further, increased exposure (time of integration is increased) and sensitivity enhancement DQE (Actively cooled CMOS, or SIPM/SPAD) can be used such that near dark imaging of the retina is possible.
Finally, ocular pressure measurement increases and decreased retinal artery filling times can be linked to changes in sepsis biomarkers, cardiac index, and APACHE II scores and thus may be a predictor of deterioration. The eye imaging device 1602 can be programmed to measure the patient's existing blood pressure, alone or in conjunction with the real-time imaging of the retina to correlate with the contraction of the heart (e.g., “thump” of the balloon).
Referring now to
In this example, the temperature sensors 1704, 1706 can be infrared sensors that are mounted around (e.g., to the left and right of) the camera 104. As the camera 104 moves from the left eye to the right eye, the temperature sensors 1704, 1706 of the eye imaging device 1702 capture infrared measurements from the inner canthus of both eyes. The eye imaging device 1702 could be programmed to adjust such inner canthus measurements and predict the patient's core temperature. Other sensors and processes for measuring temperature could also be used.
For instance, in
Further, one or more of the sensors 1804 can be programmed to measure oximetry of the surrounding tissue of the patient. For instance, the eye imaging device 1802 can be programmed to measure forehead oximetry. The retina has the highest oxygen consumption per volume in the body, and the choroid which feeds the retina has the highest blood flow per volume in the body. By capturing oximetry reading of the retina (
Specifically, the array of sensors 1804 may include reflective SpO2 sensors (one or more emitters and one or more receivers) to obtain an SpO2 reading on the forehead. The SpO2 may also be compared to the SpO2 measured from the retina. The delta may be used as an indication for perfusion issues.
Referring now to
At step 1902, a first visual stimulus is provided by the LEDs. This stimulus can vary in location color, wavelength, and/or intensity. In alternatives, audible stimuli can also be used, such as to cue action by the patient (e.g., follow light, startle, etc.).
Next, at step 1904, the time for the patient's gaze to react to the visual and/or audible stimuli is measured. Other aspects, such as pupil size, eye/gaze tracking, blinking, eye synchronization, timing of response (in relation to stimuli, e.g., lag, consistency of lag, asynchrony) can also be used.
Next, an optional second visual stimuli is provided at step 1906, and the reaction time is measured at step 1908.
Next, at step 1910, the measurements are evaluated. Certain criteria (e.g., no blink after flash of x wavelength at y intensity; or rules-based assessment, such as e.g., if time for gaze to move from LED1 to LED2 is x milliseconds, mental state categorized as y)) can be used in the assessment. In some example, artificial intelligence/machine learning is used to automate the assessment. The assessment could also include the display or playback of recorded eye movement for the caregiver. Further, the evaluation could also be performed in part or completely by an overread service. In this manner, the assessment can be localized to the eye imaging device or conducted remotely (e.g., in the cloud).
Finally, at steps 1912 and 1914, the results of the assessment are communicated to the caregiver and possible stored in the electronic medical record for the patient.
In addition to screening for different disease and mental states, the eye imaging devices can be used in different contexts and environments. Referring now to
The eye imaging devices described herein can be used in different healthcare settings, like hospitals and nursing homes. Such uses can increase the risks of pathogen spread within vulnerable patient populations. To address these concerns, the eye imaging devices can be configured to allow for ease in disinfection and minimization of pathogen spread.
For instance, referring now to
More specifically, the face cup 2110 is made up of various parts that allow for comfort, performance, and/or ease of use. A portion 2112 that interfaces with the patient's face (i.e., the surface 210) is soft silicone rubber or a similar material. A portion 2114 that interfaces with the eye imaging device 102 is made of rigid plastic that houses shunted magnets 2116 that provide a strong interface between the housing 200 and the face cup 2110 while in use, but is safe and easy enough to remove that a caregiver can detach, clean and reattach the face cup 2110 very quickly (e.g., in less than 10 seconds).
As shown, the face cup 2110 can be easily detached from the housing 200 of the eye imaging device 102 for cleaning and disinfecting. When in place on the housing 200, the magnets 2116 are received in receptacles 2006 formed in the end 204. The receptacles 2006 have corresponding magnets 2206 that attract the magnets 2116 and releasably retain the face cup 2110 on the housing 200. See
To detach the face cup 2110 for disinfection, the caregiver simply pulls the face cup 2110 away from the housing 200 with enough force to overcome the attraction of the magnets 2116, 2206. Similarly, to reattach, the face cup 2110 is moved toward the housing 200 so that the magnets 2116 are received in the receptables 2006 and the magnets 2116 are attracted to the magnets 2206.
In some examples, the face cup 2110 (e.g., the portion 2112) is made with compositions that inhibit the growth of microorganisms and/or the formation of biofilms on the surfaces. Upon detachment, the portion 2112 can be disinfected using standard disinfectants before reattachment. In yet other examples, the face cup 2110 can be enclosed within a disposable cover (see, e.g.,
Referring now to
In these examples, the physical barrier can be, for instance, one or more disposable covers that are intended for onetime use. The covers can be made of various materials, such as nitrile or polyethylene.
In the example shown in
As shown in
Referring now to
For instance, a disposable barrier sleeve 2510 is positioned to surround the pedestal 602. The sleeve 2510 includes an elastomeric end 2610 that closes off the sleeve 2510 to minimize joint and arm contamination. An interface can be provided between the end 2610 of the sleeve 2510 and the cover 2310 of the eye imaging device 102 to minimize contaminants therein.
Hands free antibacterial gel/foam dispensers and/or towelettes can be attached to the pedestal 602 and the cubicle like structure 702. In addition to covering the physical components, information about sterilization can be provided. For instance, for the pedestal 602 and the cubicle like structure 702 (see
For instance, the digital displays 606, 708 can provide instructions on disinfection before and after each use. In some instances, the eye imaging device 102 requires input and/or automatically confirms that disinfection has occurred before imaging is performed for the next patient.
The various embodiments described above are provided by way of illustration only and should not be construed to be limiting in any way. Various modifications can be made to the embodiments described above without departing from the true spirit and scope of the disclosure.
Number | Date | Country | |
---|---|---|---|
62706324 | Aug 2020 | US |