This invention relates generally to using artificial intelligence (AI) to directly determine a patient specific treatment or management of a disease that will benefit that patient, rather than manual selection based on a diagnosis. Currently, autonomous AI systems use machine learning or other optimization techniques to determine the diagnosis of a patient, and clinicians then use this AI diagnosis, in addition to other relevant patient and population information, to subjectively determine patient specific management or prescribe a patient specific treatment (also referred to herein as an “intervention”). Thus, Autonomous AI diagnosis, while capable of determining a diagnosis without human oversight, still relies on a clinician to interpret the diagnosis in terms of the patient's entire case and then decide the intervention. However, this intermediate, subjective decision step is subject to high inter- and intra clinician variability, temporal and other drift. Moreover, the interaction between artificial intelligence and the clinician is variable with often unanticipated risks, and has been known to worsen rather than improve outcome. Moreover, obtaining the highest quality reference standard (‘ground truth’) to train autonomous AI models for producing a diagnosis can be ethically problematic, and expensive. Where the reference standard is dependent on clinician expertise, such as subjective reading of images, the reference standard can be noisy. Instead, when clinical outcome, (which combines the effects of both the accuracy of the diagnostic process as well as the precision of the treatment or management selection process) rather than the interim diagnosis can be used as reference standard for training the AI, this subjective, noisy step is eliminated, with the potential to have higher performance.
Systems and methods are disclosed herein that use machine learning to output indicia of whether a patient will benefit from a specific intervention (e.g., in terms of clinical outcome), without first requiring prediction of a diagnosis for the patient. Advantageously, the disclosed systems and methods remove a need for a subjective and noisy clinician interpretation of a diagnostic output, as the disclosed machine learning models directly output whether a specific patient will benefit from a specific intervention, which may be a binary recommendation or prescription, or a likelihood or other probabilistic output. Yet further, fewer resources are needed for validation of an intervention, and patient risk of harm is lower, as validating a diagnosis from artificial intelligence typically involves expensive experts, radiation, or other potentially harmful processes for both cases and normal, and efforts in combining their decisions, while determining how much a patient benefitted is a low cost survey. Moreover, obtaining truth data as to whether a patient benefitted can be much more accurate than doing so for diagnoses, as clinical outcome is typically far easier to observe than accurately diagnosing a condition.
In an embodiment, a device receives sensor data from an electronic device that monitors a patient. The device accesses a machine learning model, the machine learning model trained using training data that pairs information about patients to labels that describe whether a particular treatment yielded a positive result. The device applies the sensor data to the machine learning model, and receives, as output from the machine learning model, data representative of a likelihood that a patient would benefit from one or more treatments.
In an embodiment, a device receives image data corresponding to an anatomy of a patient. The device applies the image data to a feature extraction model trained using training data that pairs anatomical images to an anatomical feature label, and receives, as output from the feature extraction model, scores for each of a plurality of anatomical features corresponding to the image data. The device applies the scores as input to a treatment model, the treatment model trained to output a prediction of a likelihood of efficacy of a particular treatment based on features of the patient's anatomy as described in the data structure. The devices then receives from the treatment model data representative of the predicted likelihood of efficacy of the particular treatment. Advantageously, the treatment model does not have access to the image data, thus ensuring that the treatment model is not using information from the images that might bias a determination (e.g., skin color, gender information, and so on).
The figures depict various embodiments of the present invention for purposes of illustration only. One skilled in the art will readily recognize from the following discussion that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles of the invention described herein.
Client device 100 is a device in which inputs of patient data may be provided, and where one or more recommended treatments for a patient may be output, the treatments determined by treatment determination tool 130. The term patient data, as used herein, may refer to any data describing a patient, including images of a patient's anatomy, biometric sensor data, doctor's notes, and so on. Client device 110 may run an application installed thereon, or may have a browser installed thereon through which an application is accessed, the application performing some or all functionality of treatment determination tool 130 and/or communicating information to and from treatment determination tool 130. The application may include a user interface through which a user can input data into client device 110. The user interface may be graphical, where the user can input patient data manually (e.g., through a keyboard or touch screen). The user interface may additionally or alternatively be a biometric sensor, where patient data is automatically sensed and transmitted by the application to treatment determination tool 130. The user interface may be used to access existing patient data stored in patient data 140, which may then be communicated to treatment determination tool 130 for processing.
Client device 110 may be any device capable of transmitting data communications over network 120. In an embodiment, client device 110 is a consumer electronics device, such as a laptop, smartphone, personal computer, tablet, personal computer, and so on. In an embodiment, client device 110 may be any device that is, or incorporates, a sensor that senses patient data (e.g., motion data, blood saturation data, breathing data, or any other biometric data).
Network 120 may be any data network capable of transmitting data communications between client device 110 and treatment determination tool 130. Network 120 may be, for example, the Internet, a local area network, a wide area network, or any other network.
Treatment determination tool 130 receives patient data from client device 110 and outputs a treatment from which the patient is likely to improve. Further details of the operation of treatment determination tool 130 are discussed below with reference to
Patient data 140 is a database that houses records of data of one or more patients. Patient data 140 may be hospital records, personal records of a patient, doctors notes, and so on. Patient data 140 may be co-located at client device 110 and/or at treatment determination tool 130.
Patient data retrieval module 231 receives patient data (e.g., from client device 110 and/or patient data 140). In two exemplary embodiments, the patient data may be image data corresponding to an anatomy of a patient, or may be sensor data from an electronic device that monitors a patient. While these exemplary embodiments will be the subject of detailed treatment in this disclosure, any other form of patient data may be received by patient data retrieval module 231.
Treatment determination module 232 accesses one or more machine learning models, applies the received patient data to the one or more machine learning models, and receives as output from the one or more machine learning models data representative of a predicted measure of efficacy of one or more treatments. Focusing on the embodiment where the received patient data includes image data,
As depicted in
Treatment determination module 232 applies the image data to one or more feature models 320 (e.g., accessed from feature models 241). The image data may include data from one or more images. The term feature model (or “feature extraction model”), as used herein, may refer to a model trained to identify or extract one or more features in an image. The term “feature model” and “feature extraction model” are used interchangeably herein. Where used in the singular, a “feature model” or “feature extraction model” may refer to a single model or an ensemble of two or more models. The term feature, as used herein, may refer to an object within an image. Objects may include anatomical objects (e.g., a blood vessel, an organ (e.g., optic nerve), and so on. Objects may also include biomarkers, which may be anomalies, such as lesions, fissures, dark spots, and any other abnormalities relative to a normal anatomic part of a human being. The feature model(s) may be trained using labeled training images, where the training images show at least portions of human anatomy, and are labeled with at least a score (e.g., likelihood or probability) of whether the image includes a biomarker (e.g., feature). The labels are also referred to herein as anatomical feature labels. The training images are also referred to herein as anatomical images. In an embodiment, the labels may include an identification of one or more specific biomarkers within the image. The labels may include additional information, such as other objects within the images and one or more body parts that the training image depicts. Further discussion of the structure, training, and use of feature models is disclosed in commonly-owned U.S. Pat. No. 10,115,194, filed Apr. 6, 2016, issued Oct. 30, 2018, the disclosure of which is hereby incorporated by reference herein in its entirety.
In an embodiment, data additional to the image data is input into the feature model(s) 320. The additional data may be any patient data. In such an embodiment, the training data for the feature model(s) 320 may include a pairing of images with other patient data. For example, blood pressure data, oxygen saturation data, and so on, may be included with a training image. Feature model(s) 320 output probabilities that one or more features exist in the image data, or directly output binary decisions that features do or do not exist in the image data based on the probabilities exceeding corresponding thresholds. In embodiments where the feature model(s) 320 are trained to accept additional patient data as input, probabilities may be more accurately assessed, thus enabling a more accurate assessment of whether a feature exists within input image data.
In an embodiment, image data 310 is applied to a single feature model 320 by treatment determination module 232, and the single feature model 320 outputs data representative of one or more corresponding features. As mentioned above, this data may include probabilities that the image data includes one or more features, or may include a binary determination that certain feature(s) are included in the image data.
In an embodiment, image data 310 is applied to a plurality of feature models 320 by treatment determination module 232. Returning momentarily to
As discussed above, treatment determination module 232 determines features 330 based on output from feature model(s) 320. Feature vector module 234 consolidates features 330 into feature vector 340. The term feature vector, as used herein, may refer to a data structure that includes each of the different features. The feature vector may map the different features to auxiliary information. For example, where the image data includes images corresponding to different locations of a body part, the feature vector may map the features identified from those images to their respective different locations of the body part. As an example, where the images are retinal images, and one image is taken for each quadrant of a retina, the feature vector may include four data points, the data points including respective features identified in an image of each of the four quadrants.
Treatment determination module 232 applies feature vector 340 to treatment model 350. Treatment model 350 is trained to, based on an input feature vector, output a prediction of a measure of efficacy of one or more particular treatments for a patient based on features of the patient's anatomy as described in the data structure. Treatment model 350 may be any machine learning model (e.g., deep learning model, convolutional neural network (CNN), etc.). The training data may include data manually labeled by patients, where patients indicate that they felt better or did not feel better after a treatment, and where this manual label is paired to image feature vectors, optionally with other patient data, to a label corresponding to what the patient has indicated. The labels may alternatively, or additionally, be manually labeled by doctors. In an embodiment, patient data, such as vitals or descriptors of patient health, may be monitored from a time of treatment. The patient data may be compared to data from prior to the time of treatment. The monitoring both before and after the time of treatment may be bounded by a maximum threshold amount of time. Responsive to determining that the patient data shows an improvement, the feature vectors may be labeled with the treatment and an indicator that the patient has improved. Similarly, images may be labeled showing no improvement, or a worsening condition, based on the monitored patient data. In an embodiment, a degree to which a patient improved or worsened may be labeled against images and/or feature vectors. In some embodiments, the labeling of training data might be automatic, where patient data and/or disease data and/or treatment data is labeled with outcome. These labels may alternatively be manually applied (e.g., by clinicians monitoring patient response to treatment).
Treatment determination module 232 receives, as output from treatment model 350, a prediction of a measure of efficacy (e.g., a likelihood that a treatment will be efficacious for a patient, according to a predetermined criteria therefor) of one or more particular treatments for a patient based on features of the patient's anatomy as described in the data structure. Like the output of feature model(s) 320, treatment model 350 may output probabilities corresponding to a predicted efficacy of a plurality of candidate treatments, or alternatively, may directly output identifiers of treatments having probabilities that exceed a threshold. Treatment determination module 232 determines, based on the output of treatment model 350, one or more treatments that are likely to be effective for a patient, and outputs the determined one or more treatments to a user (e.g., a doctor, a patient, or another medical clinician).
In some embodiments, rather than using a two-stage model as depicted in
Treatment determination module 232 applies sensor data 405 into treatment model 410, which may be any machine-learned model described herein (e.g., a CNN). Treatment model 410 may be trained using historical sensor data (in isolation and/or in combination with other historical sensor data) as labeled by treatment outcome. In an embodiment, the labels may be manually applied. In another embodiment, the patient's condition may be monitored before and after treatment (e.g., as bounded by a threshold period of time on either or both sides), and a label may be determined automatically from indicia of the condition improving, staying the same, or worsening following treatment. Similar to model 350, treatment model 406 outputs data representative of efficacy of treatment 415, and treatment determination module 232 determines a treatment that is likely to be effective in treating the patient therefrom and provides identifications of those determined treatment(s) to a user. In some embodiments, model 410 may be trained to output multiple treatments. In some embodiments, different ones of model(s) 410 are used to determine different treatments.
The model 506 is trained using training data that indicates input sensor data paired with labels indicating whether a patient associated with that sensor data improved or worsened after receiving the treatment associated with one of the branches of the model 506. Accordingly, the training data includes different sets of training data, each set having the same inputs (e.g., sensor data) but a different label corresponding to efficacy of a different treatment. During training, a training example from one of the sets of training data is selected, and a standard back propagation is performed through the branch (520, 530, or 540) of the model 506 corresponding to the treatment for the selected example and through the shared layers 510 of the model 506. This process is repeated, ultimately selecting training examples from each set and thereby training each of the branches 520, 530, and 540 of the model 506. In an embodiment, back propagation occurs from clinical outcome, through treatment choices, back to diagnostic choices. Back propagation may occur through any layer or branch and may selectively omit layers or branches. In some embodiments, the different training sets used to train the different branches of the multi-task model may differ substantially in terms of size. This may be a result of some treatments being used frequently among patients, and other treatments being used far less frequently, thus resulting in less training data for a given label of a given training set. The treatment determination tool 130 may sample from the different training sets by a ratio of the available training data across the different training sets. This way, the treatments for which there is relatively less training data benefit from treatments where there is relatively more, since the training data for all of the sets is used for the shared layers 510. This may enable a model to be trained for a treatment that would not otherwise have enough training data to train its own model, as the branch for that treatment is aided by the work done by the shared layers 510.
Technical benefits are thus achieved from using a multi-task model (e.g., neural network) as shown in
The above disclosures lead to patient benefit where patients suffer from any number of conditions. For example, where a patient has a sleep apnea condition, the condition may be autonomously diagnosed, where a model is trained using sleep study data such as EEG data, motion data, breathing data, and the like, where the model may output a sleep apnea diagnosis. This enables a clinician to prescribe treatment (e.g., that the patient should use a CPAP machine). Advantageously, the systems and methods disclosed herein remove the need for a diagnosis—a prediction that CPAP will help the patient is directly output without a diagnosis. The benefit from CPAP is determined with fewer computational resources and without a need for a doctor—if the patient's saturation goes up, then CPAP benefitted the patient, and thus ground truth data for using CPAP resulting in a benefit in this example is easily obtained without clinicians having to manually examine input data and decide whether a patient has sleep apnea in the first place. The systems and methods disclosed herein also has an advantage of process integrity, in that it leads to maximizing the population benefit of CPAP machines.
As another example, where a patient has an ear infection, rather than output that a patient has a “otitis media” diagnosis (which requires validation by a physician, followed by a prescription), or a specific prescription of “amoxicillin 200 mg”, the system may directly output that “patient benefits from amoxicillin.” This may be based on training data showing whether ear infections by patients with similar patient data is labeled based on whether those patients benefitted from amoxicillin at all. Again, obtaining ground truth training data for such a model requires far less cost, as the benefit to patients is self-validating. In some embodiments, for this reason, outputs of models may be dichotomous (that is, patient does or does not benefit from a particular treatment), rather than granular (e.g., specific dosage or course of treatment) which is much harder to obtain ground truth data.
The foregoing description of the embodiments of the invention has been presented for the purpose of illustration; it is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Persons skilled in the relevant art can appreciate that many modifications and variations are possible in light of the above disclosure.
Some portions of this description describe the embodiments of the invention in terms of algorithms and symbolic representations of operations on information. These algorithmic descriptions and representations are commonly used by those skilled in the data processing arts to convey the substance of their work effectively to others skilled in the art. These operations, while described functionally, computationally, or logically, are understood to be implemented by computer programs or equivalent electrical circuits, microcode, or the like. Furthermore, it has also proven convenient at times, to refer to these arrangements of operations as modules, without loss of generality. The described operations and their associated modules may be embodied in software, firmware, hardware, or any combinations thereof.
Any of the steps, operations, or processes described herein may be performed or implemented with one or more hardware or software modules, alone or in combination with other devices. In one embodiment, a software module is implemented with a computer program product comprising a computer-readable medium containing computer program code, which can be executed by a computer processor for performing any or all of the steps, operations, or processes described.
Embodiments of the invention may also relate to an apparatus for performing the operations herein. This apparatus may be specially constructed for the required purposes, and/or it may comprise a general-purpose computing device selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a non-transitory, tangible computer readable storage medium, or any type of media suitable for storing electronic instructions, which may be coupled to a computer system bus. Furthermore, any computing systems referred to in the specification may include a single processor or may be architectures employing multiple processor designs for increased computing capability.
Embodiments of the invention may also relate to a product that is produced by a computing process described herein. Such a product may comprise information resulting from a computing process, where the information is stored on a non-transitory, tangible computer readable storage medium and may include any embodiment of a computer program product or other data combination described herein.
Finally, the language used in the specification has been principally selected for readability and instructional purposes, and it may not have been selected to delineate or circumscribe the inventive subject matter. It is therefore intended that the scope of the invention be limited not by this detailed description, but rather by any claims that issue on an application based hereon. Accordingly, the disclosure of the embodiments of the invention is intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the following claims.
Number | Name | Date | Kind |
---|---|---|---|
10115194 | Niemeijer et al. | Oct 2018 | B2 |
20180240552 | Tuyl | Aug 2018 | A1 |
20210248741 | Kannengiesser | Aug 2021 | A1 |
20230148955 | Kwon | May 2023 | A1 |
Number | Date | Country |
---|---|---|
111052247 | Apr 2020 | CN |
113362944 | Sep 2021 | CN |
1916624 | Nov 2016 | EP |
WO-2023064957 | Apr 2023 | WO |
Entry |
---|
PCT Invitation to Pay Additional Fees, PCT Application No. PCT/US2022/51624, Mar. 1, 2023, 2 pages. |
Abramoff, M. D. et al. “Foundational Considerations for Artificial Intelligence Utilizing Ophthalmic Images.” Ophthalmology, Journal Pre-proof, vol. 129, No. 2, Aug. 2021, pp. 1-33. |
Abramoff, M. D. et al. “Lessons Learned About Autonomous AI: Finding a Safe, Efficacious, and Ethical Path Through the Development Process.” American Journal of Ophthalmology, vol. 214, Jun. 2020, pp. 134-142. |
PCT International Search Report and Written Opinion, PCT Application No. PCT/US2022/51624, May 25, 2023, 11 pages. |
Number | Date | Country | |
---|---|---|---|
20230178200 A1 | Jun 2023 | US |