Alzheimer's disease (AD), the most common form of dementia, is a global public health problem.1 The diagnosis of AD is complex and typically involves expensive and sometimes invasive tests not commonly available beyond highly specialized clinical settings. For example, biomarkers of amyloid-β and phosphorylated tau measured using cerebrospinal fluid (CSF), positron emission tomography (PET) scans, and plasma assays are helpful for AD diagnosis. However, these tests are not suitable for screening possible AD cases in primary care or community settings.2,3 Importantly, as new therapeutic efforts focus on early AD treatment,4-6 simple, accessible, and sensitive community-based screening tests would significantly improve population-based strategies to manage AD.
The retina is a highly accessible part of the central nervous system, with common embryology, anatomy, and physiology.7 Retinal changes in AD have been shown in histopathological studies of post-mortem specimens.8,9 This is further supported by clinical studies showing a range of retinal changes in subjects with AD, such as changes in the retinal vasculature (e.g., vessel caliber and retinopathy signs), the optic nerve, and the retinal nerve fiber layer (RNFL).10-14 These features can be non-invasively imaged using digital retinal photography, which is now widely available and affordable in primary care optometry and community settings.
Artificial intelligence (AI), particularly deep learning (DL), has been applied to retinal photographs for detecting various ophthalmic diseases (e.g., diabetic retinopathy16,17, optic disc papilledema18, glaucoma19, and age-related macular degeneration20). Furthermore, DL approaches can also detect systemic diseases based on retinal photographs (e.g., systemic biomarkers21, cardiovascular disease22,23, diabetes24, chronic kidney disease24,25, hepatobiliary diseases26). Nevertheless, the role of DL approaches in detecting patients with AD from retinal photographs has yet to be determined.
In addition, integrating DL algorithms into real-time clinical workflow has been recognized as a priority to realize the significant potential of AI for clinical diagnosis and disease risk stratification27-30. However, while many DL algorithms have shown promising results in laboratory and research settings, their performances in real-world clinical settings require further evaluations31. A major challenge is that retinal photographs captured from real-world clinical settings can have lower quality than the retinal photographs carefully curated and used specifically in DL algorithms development27,32-34, and thus, the performances of such DL algorithms are less reliable when applied clinically27,34. For example, Abramoff et al. reported while their DL algorithm achieved a sensitivity of 97% in a retrospective dataset under a laboratory setting, the performance dropped to 87.2% in a prospective study conducted in a primary care setting27. In another prospective study conducted by Beede et al.34, about 21% of retinal photographs were unsuitable for DL-based diabetic retinopathy screening because of low image quality. These were likely due to the exclusion of low-quality retinal photographs when training DL algorithms for the eye disease diagnosis17,35-38. As such, the application of these algorithms in real-world clinical settings would require the exclusion of retinal photographs of low image quality to inhibit deterioration of their diagnostic performance39-42. In addition to image-quality assessment, DL can further provide other useful information such as field-of-view and laterality-of-the-eye before disease diagnosis by subsequent DL processing algorithms. For example, DL algorithms developed for optic disc diseases (e.g., papilledema and glaucoma) should focus on optic disc-centered retinal photographs as it can work less well for macula-centered retinal photographs43-45.
Embodiments of the subject invention provide an AI-aided classification system for AD screening from retinal photographs, which includes a DL-based pre-diagnosis module for image assessment and a DL-based AD classification module with additional heatmaps for visualization. Provided embodiments of the AI system can output a pre-diagnosis image assessment (e.g., the image-quality, field-of-view, and laterality-of-the-eye) and a simple binary AD-dementia/non-demented classification based on retinal photographs. Certain embodiments can add a complimentary risk profiling tool for AD and assist physicians to identify asymptomatic individuals who are more likely to have AD in the community. Higher-risk individuals can then benefit from selective referral for more intensive and specific examinations (e.g., PET imaging, plasma assays for amyloid-β and phosphorylated tau) at highly specialized clinics for facilitating early AD diagnosis and allowing the individuals to take prevention measures, such as lifestyle modification and control of risk factors.
There is increasing evidence that a range of retinal features identified from retinal photographs is associated with AD. DL has been shown to have significant potential for eye disease detection and screening on retinal photographs in different clinical settings, particularly in primary care. Nevertheless, the role of DL approaches in detecting patients with AD from retinal photographs has yet to be determined. Besides, integrating DL algorithms into real-time clinical workflow is a priority to realize the significant potential of AI for clinical diagnosis and disease risk stratification. In addition, an automated pre-diagnosis image assessment is advantageous to streamline the application of the developed DL algorithms.
Embodiments of the subject invention provide an AI-aided classification system for AD screening from retinal photographs, which includes a DL-based pre-diagnosis module for image assessment and a DL-based AD classification module with additional heatmaps for visualization. For AD classification, embodiments provide three kinds of DL models that predict the AD-dementia/non-demented probabilities from three directions with both eyes' four images (Direction-1), both eyes' four images combining demographical information (Direction-2), and a single eye's two images (Direction-3), respectively. For the pre-diagnosis image assessment, it consists of one image pre-processing model, with three additional classification models for classifying image-quality (e.g., gradable or ungradable), field-of-view (e.g., macula-centered or optic nerve head-centered), and laterality-of-the-eye (e.g., right or left eye).
Embodiments provide a cloud-based web application and can output a pre-diagnosis image assessment (e.g., image-quality, field-of-view, and laterality-of-the-eye) and a simple binary AD-dementia/non-demented classification based on retinal photographs. The results can add a complimentary risk profiling tool for AD and assist physicians to identify asymptomatic individuals who are more likely to have AD in the community. Higher-risk individuals can then benefit from selective referral for more intensive and specific examinations (e.g., PET imaging, plasma assays for amyloid-β and phosphorylated tau) at highly specialized clinics for facilitating early AD diagnosis and allowing the individuals to take prevention measures, such as lifestyle modification and control of risk factors. It is contemplated that certain embodiments will be incorporated into retinal photography devices for automated image analysis and AD screening in different scenarios.
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
Embodiments of the subject invention provide an AD classification deep-learning module. Certain embodiments provide up to three kinds of DL models which predict the AD-dementia/non-demented probabilities from three directions with both eyes' four images (Direction-1), both eyes' four images combining demographical information (Direction-2), and single eye's two images (Direction-3), respectively.
Embodiments advantageously applied EfficientNet-b246 as the backbone for feature extractor and then a DL model was designed to integrate AD-related features from four retinal photographs for each study subject (e.g., both optic nerve head-and macula-centered fields from both eyes) (Direction-1). This model outputted subject-level detection results (e.g., AD-dementia or non-demented) accounting for AD features from both eyes' images. Second, on top of Direction-1, embodiments further trained another DL model which can additionally consider risk factors of AD (e.g., the demographical information including age, gender, presence or absence of hypertension and diabetes) (Direction-2). Finally, embodiments developed a DL model for single eye analysis as individuals can have an ungradable retinal photograph from one eye (e.g., due to severe cataract) (Direction-3).
For each AD classification model, embodiments adopted unsupervised domain adaptation with domain-specific batch normalization to address the issue of data heterogeneity and domain shift problems from different study cohorts and to improve the model generalizability. Unsupervised domain adaptation is a type of learning framework that can transfer knowledge learned from a larger number of annotated training data in the source domains to target domains with unlabeled data only. Certain embodiments provide domain-specific batch normalization as a building block for deep neural networks where the source domain and the target domain datasets can have their own separate batch normalization layer for training and extraction of hyper-parameters. This design serves to address characteristics specific to each domain that are not compatible within a single model, while retaining domain-invariant information that is common to all domains. In certain embodiments, the labelled source domain dataset was first used for training in a supervised way to generate an unsupervised domain adaption network. This unsupervised domain adaptation network was then used to generate pseudo-labels for unlabeled data in the target domain. The final classification network was subsequently trained with full supervision using labelled data from source domain and pseudo-label from target domains. Through the fusion of the domain-independent and domain-dependent knowledge learning, the provided DL models can transfer discriminative features from the labelled source domain to the unlabeled target domain (e.g., domain adaptation) and improve the classification performance on the target domain. Due to the limitation of the domain adaptation-based method, embodiments trained one model for each external dataset to obtain the prior information of unlabeled external datasets.
To better understand discriminative features between AD-dementia subjects and no dementia subjects, embodiments advantageously applied Gradient-weighted Class Activation Mapping (Grad-CAM)47 to visualize these features.
Alzheimer's disease (AD), the most common form of dementia, is a major public health and clinical challenge globally, causing a significant socioeconomic burden worldwide. Although cerebrospinal fluid (CSF) biomarkers and recent novel biomarkers include positron emission tomography (PET) scans and plasma assays for amyloid-β and phosphorylated tau showed great promise for aiding AD detection, particularly early-stage AD, these tests (e.g., CSF, PET) are not suitable for screening in routine clinical settings or communities. As early AD is now the focus of new therapeutic efforts and treatment of AD is possible, a more widely accessible screening system to identify individuals in the community, who are then referred to neurology clinics for more intensive and specific examinations to confirm AD would aid the management of AD.
The retina, a neurosensory layered tissue lining the back of the eye, has long been considered as a proxy measure to study disorders in the central nervous system (CNS), as it is an accessible extension of the brain in terms of embryology, anatomy, and physiology. Evidence of retinal pathology in AD has been shown in histopathological studies of postmortem specimens. Meanwhile, the retinal vasculature, optic nerve head (ONH), and retinal nerve fiber layer (RNFL) can be captured and assessed using retinal photography effectively and non-invasively at a relatively low cost, making it a potential ideal tool for community screening for AD. Accumulating data have shown that specific retinal features measured from retinal photographs are associated with AD, including RNFL loss, vessel caliber, vessel tortuosity, vessel fractal dimension, and retinopathy signs. A simple whole retina score, in principle, would be even more useful as a clinical screening tool for AD.
Artificial intelligence (AI), particularly deep learning (DL), can provide such a solution to facilitate the application of retinal photography for the screening of AD. DL allows an algorithm to appreciate and extract the inherent non-obvious features from training images necessary for accurate discrimination based on examples, without the need for manual engineering of discriminating features. DL has been applied in the assessment of retinal photographs for the detection of various ophthalmic diseases, including diabetic retinopathy, glaucoma, and age-related macular degeneration. However, the possibility of identifying AD from retinal photographs alone using the DL approach has yet to be determined.
Embodiments of the subject invention provide a novel AI-aided classification system for AD screening from retinal photographs with additional pre-diagnosis image assessment. Embodiments exhibit several advantages and improvements, including the following:
Embodiments provide at least two potential implementations in clinical practice:
Embodiments of the subject invention address the technical problem of detecting and screening for early-stage AD or dementia being expensive, needing excessive human processing, not being suitable for screening large populations, and requiring invasive methods.
This problem is addressed by providing digital image processing with enhanced AI, in which a deep learning method applying a combination of advanced techniques is utilized to categorize images based on the classification given during the learning process.
The transitional term “comprising,” “comprises,” or “comprise” is inclusive or open-ended and does not exclude additional, unrecited elements or method steps. By contrast, the transitional phrase “consisting of” excludes any element, step, or ingredient not specified in the claim. The phrases “consisting” or “consists essentially of” indicate that the claim encompasses embodiments containing the specified materials or steps and those that do not materially affect the basic and novel characteristic(s) of the claim. Use of the term “comprising” contemplates other embodiments that “consist” or “consisting essentially of” the recited component(s).
When ranges are used herein, such as for dose ranges, combinations and subcombinations of ranges (e.g., subranges within the disclosed range), specific embodiments therein are intended to be explicitly included. When the term “about” is used herein, in conjunction with a numerical value, it is understood that the value can be in a range of 95% of the value to 105% of the value, i.e., the value can be +/−5% of the stated value. For example, “about 1 kg” means from 0.95 kg to 1.05 kg.
The methods and processes described herein can be embodied as code and/or data. The software code and data described herein can be stored on one or more machine-readable media (e.g., computer-readable media), which may include any device or medium that can store code and/or data for use by a computer system. When a computer system and/or processor reads and executes the code and/or data stored on a computer-readable medium, the computer system and/or processor performs the methods and processes embodied as data structures and code stored within the computer-readable storage medium.
It should be appreciated by those skilled in the art that computer-readable media include removable and non-removable structures/devices that can be used for storage of information, such as computer-readable instructions, data structures, program modules, and other data used by a computing system/environment. A computer-readable medium includes, but is not limited to, volatile memory such as random access memories (RAM, DRAM, SRAM); and non-volatile memory such as flash memory, various read-only-memories (ROM, PROM, EPROM, EEPROM), magnetic and ferromagnetic/ferroelectric memories (MRAM, FeRAM), and magnetic and optical storage devices (hard drives, magnetic tape, CDs, DVDs); network devices; or other media now known or later developed that are capable of storing computer-readable information/data. Computer-readable media should not be construed or interpreted to include any propagating signals. A computer-readable medium of embodiments of the subject invention can be, for example, a compact disc (CD), digital video disc (DVD), flash memory device, volatile memory, or a hard disk drive (HDD), such as an external HDD or the HDD of a computing device, though embodiments are not limited thereto. A computing device can be, for example, a laptop computer, desktop computer, server, cell phone, or tablet, though embodiments are not limited thereto.
A greater understanding of the embodiments of the subject invention and of their many advantages may be had from the following examples, given by way of illustration. The following examples are illustrative of some of the methods, applications, embodiments, and variants of the present invention. They are, of course, not to be considered as limiting the invention. Numerous changes and modifications can be made with respect to embodiments of the invention.
Embodiment 1. An artificial intelligence (AI)-aided classification system for Alzheimer's disease (AD) screening from a source dataset comprising retinal photographs, the system comprising a deep learning (DL) model created by a process comprising:
Embodiment 2. The system according to Embodiment 1, the DL model comprising a bilateral model, a hybrid model, and a unilateral model.
Embodiment 3. The system according to Embodiment 2, wherein:
Embodiment 4. The system according to Embodiment 1, the source dataset comprising retinal photographs and demographic information derived from different centers, and the DL model comprising a feature fusion module to integrate features captured from the different centers.
Embodiment 5. The system according to Embodiment 1, the DL model comprising demographic information integrated with both eyes' four images from one subject; and demographic information integrated by bilinear transformation.
Embodiment 6. The system according to Embodiment 1, the DL model comprising EfficientNet-b2 as a backbone to extract features.
Embodiment 7. The system according to Embodiment 6, the DL model comprising a domain adaptation technique to deal with dataset discrepancies.
Embodiment 8. The system according to Embodiment 1, the DL model created by a process comprising two stages.
Embodiment 9. The system according to Embodiment 8, the two stages comprising:
Embodiment 10. The system according to Embodiment 9, wherein images from the source and target domains were fed into separate batch normalization layers in each of the first stage and the second stage, respectively.
Embodiment 11. The system according to Embodiment 10, comprising an imbalance of data between a first class with more data and a second class with less data, and comprising an over sampling for the second class.
Embodiment 12. The system according to Embodiment 11, comprising a training objective function utilizing both source dataset and target domain images.
Embodiment 13. The system according to Embodiment 12, comprising generation of heatmaps to show the significant locations which are related to the AD with a Gradient-weighted Class Activation method.
Embodiment 14. The system according to Embodiment 13, comprising pre-diagnosis image assessment and AD binary classification by a cloud-based web application.
Embodiment 15. A method for creating an artificial intelligence (AI)-aided classification system comprising a deep learning (DL) model for Alzheimer's disease (AD) screening from a source domain comprising retinal photographs, the method comprising:
Embodiment 16. The method according to Embodiment 15, comprising:
Embodiment 17. The method according to Embodiment 16, comprising:
Embodiment 18. An artificial intelligence (AI)-aided classification system for Alzheimer's disease (AD) screening from a source dataset comprising retinal photographs, the system comprising a deep learning (DL) model, the system created by a process comprising:
Embodiment 19. The system according to Embodiment 18, wherein the DL model is created by a process comprising two stages, the two stages comprising:
Embodiment 20. The system according to Embodiment 19, the system created by a process comprising:
It should be understood that the examples and embodiments described herein are for illustrative purposes only and that various modifications or changes in light thereof will be suggested to persons skilled in the art and are to be included within the spirit and purview of this application and the scope of the appended claims.
All patents, patent applications, provisional applications, and publications referred to or cited herein are incorporated by reference in their entirety, including all figures and tables, to the extent they are not inconsistent with the explicit teachings of this specification.
Following are examples that illustrate procedures for practicing the invention. These examples should not be construed as limiting. All percentages are by weight and all solvent mixture proportions are by volume unless otherwise noted.
A first exemplary embodiment provided a pre-diagnosis deep-learning module consisting of (in alternative embodiments, either comprising or consisting essentially of) one pre-processing model, with three additional models for each of the three classification tasks. As different retinal cameras captured retinal photographs with different image resolutions and degrees of view, image pre-processing was first performed to normalize the inputs to similar conditions. Image normalization was performed using the preprocessing module to standardize inputs to similar conditions. Data balancing and data augmentation were applied on the fly. During training, the pre-trained ImageNet weights were used for initial weighting. Furthermore, these assessment modules were converted into TFLite models to reduce latency inference. For all tasks, input data was randomly augmented with (−0.3, 0.3) brightness adjustment, (−0.5, 0.5) contrast adjustment, (−0.5, 0.5) saturation adjustment, (−0.1, 0.1) hue adjustment, along with 60 degrees of random rotation, 20% random translation, 10% scaling and 5 degrees of shearing. All images were augmented channel-wise with means of (0.485, 0.456, 0.406) and standard deviations of (0.229, 0.224, 0.225).
Recognizing that the tasks are different and models tend to learn different features, the inventors used EfficientNet-B048 for the image-quality and the field-of-view tasks, and MobileNetV249 for the laterality-of-the-eye task, respectively, to make use of advantages in different architectures. All the retinal photographs were used to train the image-quality assessment model. After excluding ungradable retinal photographs and off-centered retinal photographs, only gradable macula-centered and optic disc-centered retinal photographs were used to train the DL algorithms for field-of-view and laterality-of-the-eye assessments.
The inventors provided a cloud-based web application, advantageously integrating the whole process of data pre-processing, data analysis, and data output of image assessment, AD classification. The application was composed under the service-oriented-architecture (SOA) protocol, which facilitates case of maintenance. From the user perspective, no additional operation was required apart from uploading retinal photographs. The cloud-based web application also contains a by-pass function to allow manual justification on the AI-based image assessment results and ensure the success of downstream task for AD classification.
This exemplary embodiment provides a novel AI-aided classification system for AD screening from retinal photographs with additional pre-diagnosis image assessment, including the following advantageous elements:
Embodiments provide at least two potential implementations in clinical practice:
Commercial relevance and market potential for one market was estimated by using the following formula:
Where MP represents the market potential; N represents the total potential customers in Hong Kong; MS represents market share, which is the percent of consumers buying the AD screening service; P represents selling price for the screening service for each costumer, and Q represents average annual consumption. The inventors defined the potential customers as the elderly community (subjects >65 years old) in Hong Kong. Referring to statistical reports, there were around 1.3 million people aged 65 years old or above living in domestic household in Hong Kong in 2022 (https://www.statista.com/statistics/962290/hong-kong-elderly-population-in-domestic-households-by-age-group/). Regarding the market share, currently there is no existing AI-aided classification system for AD screening from retinal photographs with additional pre-diagnosis image assessment in the market. Therefore, the inventors estimated a 40% percent share of the market since the business is new. For the “P” selling price, referring to the price of retinal photography in the CUHK eye center, the inventors estimate the selling price for each subject for the screening service can be 300 HKD. A study showed rescreening every 5 years can reduce the prevalence of dementia due to AD by 50%.50 Therefore, the estimated market potential will be:
The inventors have tested the pre-diagnosis image assessment (published) and AD classification DL modules (non-published) in retrospective multi-center cohorts. The performance and heatmaps are shown in Table 1, Table 2, and
Deep-Learning-Based Pre-Diagnosis Assessment Module for Retinal photographs: A Multicenter Study (Yuen V. et al., Transl Vis Sci Technol. 2021; 10(11):16. https://doi.org/10.1167/tvst.10.11.16, which is hereby incorporated by reference in its entirety, including any tables and figures.)
The subject disclosure in this application focuses on an AI-aided classification system for Alzheimer's disease (AD) screening from retinal photographs. In contrast, related art systems teach developing and validating a deep learning-based pre-diagnosis quality control method module for retinal photographs, targeting image quality, field of view, and laterality of the eye. An exemplary and non-limiting list of advantages of the subject invention over related art systems includes the following.
Embodiments of the subject invention provide a system for AD screening using retinal photographs, whereas related art teaches systems and methods for creating an AI-driven pre-diagnosis assessment module for general eye disease detection and screening from retinal photographs.
Related art teaches the development and validation of well-established CNN architecture (EfficientNet-B0 and MobileNetV2) for prediction. However, certain embodiments of the subject invention provide new techniques to integrate features from different retinal photographs, new networks with unsupervised domain adaptation technique to address dataset shifts between the different center data, and a deep learning-based AD classification module with additional heatmaps for visualization.
Embodiments of the subject invention provide three different deep learning models to predict AD-dementia/non-demented probabilities from three directions: both eyes' four image (Direction-1, or bilateral), both eyes' four images combining demographical information (Direction-2, or hybrid), and single eye's two images (Direction-3, or unilateral).
Embodiments of the subject invention provide a complimentary risk profiling tool for AD to assist physicians in identifying asymptomatic individuals with a higher likelihood of having AD in the community.
Embodiments of the subject invention provide systems and methods to specifically target AD screening, offering numerous advantages over the related art's more general application for eye disease detection and screening.
A deep learning model for detection of Alzheimer's disease based on retinal photographs: a retrospective, multicenter case-control study (www.thelancet.com/digital-health Published online Sep. 30, 2022. https://doi.org/10.1016/S2589-7500(22)00169-8 1) which is hereby incorporated by reference in its entirety, including any tables and figures.).
There is no simple model in related art to screen for Alzheimer's disease, partly because the diagnosis of Alzheimer's disease itself is complex—typically involving expensive and sometimes invasive tests not commonly available outside highly specialized clinical settings. The inventors aimed to develop a deep learning algorithm that could use retinal photographs alone, which is the most common method of non-invasive imaging the retina to detect Alzheimer's disease-dementia.
In this retrospective, multicenter case-control study, the inventors trained, validated, and tested a deep learning algorithm to detect Alzheimer's disease-dementia from retinal photographs using retrospectively collected data from 11 studies that recruited patients with Alzheimer's disease-dementia and people without disease from different countries. The main aim was to develop a bilateral model to detect Alzheimer's disease-dementia from retinal photographs alone. The inventors designed and internally validated the bilateral deep learning model using retinal photographs from six studies. The inventors used the EfficientNet-b2 network as the backbone of the model to extract features from the images. Integrated features from four retinal photographs (optic nerve head-centered and macula-centered fields from both eyes) for each individual were used to develop supervised deep learning models and equip the network with unsupervised domain adaptation technique, to address dataset discrepancy between the different studies. The inventors tested the trained model using five other studies, three of which used PET as a biomarker of significant amyloid β burden (testing the deep learning model between amyloid β positive vs amyloid β negative).
A total of 12,949 retinal photographs from 648 patients with Alzheimer's disease and 3240 people without the disease were used to train, validate, and test the deep learning model. In the internal validation dataset, the deep learning model had 83.6% (SD 2.5) accuracy, 93.2% (SD 2.2) sensitivity, 82.0% (SD 3.1) specificity, and an area under the receiver operating characteristic curve (AUROC) of 0.93 (0.01) for detecting Alzheimer's disease-dementia. In the testing datasets, the bilateral deep learning model had accuracies ranging from 79.6% (SD 15.5) to 92.1% (11.4) and AUROCs ranging from 0.73 (SD 0.24) to 0.91 (0.10). In the datasets with data on PET, the model was able to differentiate between participants who were amyloid β positive and those who were amyloid β negative: accuracies ranged from 80.6 (SD 13.4%) to 89.3 (13.7%) and AUROC ranged from 0.68 (SD 0.24) to 0.86 (0.16). In subgroup analyses, the discriminative performance of the model was improved in patients with eye disease (accuracy 89.6% [SD 12.5%]) versus those without eye disease (71.7% [11.6%]) and patients with diabetes (81.9% [SD 20.3%]) versus those without the disease (72.4% [11.7%]).
The above results show retinal photograph-based deep learning algorithm according to an embodiment of the subject invention can detect Alzheimer's disease with good accuracy, showing its ability for screening Alzheimer's disease in a community setting.
Alzheimer's disease, the most common form of dementia, is a global public health problem.1 Diagnosis of Alzheimer's disease is complex and typically involves expensive and sometimes invasive tests not commonly available outside of highly specialized clinical settings. For example, biomarkers of amyloid β and phosphorylated tau measured through cerebrospinal fluid assessments,
PET scans, and plasma assays are helpful for Alzheimer's disease diagnosis, but these tests are not suitable for screening possible Alzheimer's disease in primary care or community settings.2 Of note, because Alzheimer's disease treatment is available,3 simple, accessible, and sensitive community-based screening tests would substantially improve population-based strategies to manage Alzheimer's disease.
The retina, a neurosensory layered tissue lining the back of the eye and directly connected to the brain via the optic nerve, has long been considered a platform to study disorders in the CNS because it is an accessible extension of the brain in terms of embryology, anatomy, and physiology.4,5 Retinal changes in Alzheimer's disease have been shown in post-mortem histopathological studies.6,7 This concept is supported by clinical studies showing a range of retinal changes in patients with Alzheimer's disease, such as changes in the retinal vasculature (e.g., vessel caliber and retinopathy signs), the optic nerve, and the retinal nerve fiber layer.5,8 These features can be non-invasively imaged using digital retinal photography, which is now widely available at a low cost in primary care optometry and community settings.
Artificial intelligence (AI), particularly deep learning, allows algorithms to extract both known and unknown features from images for accurate detection of a condition, without the need for manual identification of specific features. Deep learning has been applied to retinal photographs for detecting various ophthalmic diseases (such as diabetic retinopathy,9 optic disc papilledema,10 glaucoma,11 and age-related macular degeneration12). Furthermore, deep learning approaches can also detect systemic diseases based on retinal photographs (e.g., systemic biomarkers,13 cardiovascular disease,14 diabetes,15 chronic kidney disease,16 and hepatobiliary diseases,17). However, the role of deep learning approaches in detecting Alzheimer's disease from retinal photographs has yet to be determined in related art systems, and has only now been shown in embodiments of the subject invention.
Embodiments of the subject invention provide a novel deep learning algorithm for automated detection of Alzheimer's disease-dementia from retinal photographs alone to determine its possible use for Alzheimer's disease screening. To address this, the inventors trained, validated, and tested the deep learning models using retinal photographs from 11 clinical studies. The inventors also tested the ability of our deep learning model to differentiate patients who were amyloid β positive from those who were amyloid β negative.
In this retrospective, multicenter case-control study, the inventors trained, validated, and tested a deep learning model for detecting Alzheimer's disease from retrospectively collected retinal photographs from 648 patients with Alzheimer's disease and 3240 patients who did not have the disease. This included 11 clinical studies done at eight centers in four countries (Hong Kong Special Administrative Region, China, Singapore, the UK, and the USA; see Example 4). The inclusion and exclusion criteria for patients in each of the 11 studies are reported in Example 4. For all participants, four retinal photographs (optic nerve head-centered and macula-centered images from both eyes) were used for the model development.
This multicenter study was approved by the human ethics boards of the Joint Chinese University of Hong Kong-New Territories East Cluster Clinical Research Ethics Committee, Hong Kong Special Administrative Region, China, and local research ethics committees in each center. The 11 studies used to generate the test populations were all done according to the Declaration of Helsinki, with written informed consent obtained from each participant or their guardians. The STARD guideline was used for reporting in the current study.
The main aim of this study was to develop a bilateral deep learning model that outputted participant-level detection results (i.e., Alzheimer's disease-dementia or no dementia) accounting for Alzheimer's disease features from optic nerve head-centered and macula-centered images from both eyes. The inventors used retinal photographs from six studies with labels of either Alzheimer's disease-dementia or no dementia as primary datasets (i.e., source domain; primary 1-6; Example 4) for the development and internal validation of the deep learning model. The inventors tested the trained deep learning models with five non-overlapping studies that had labels of Alzheimer's disease-dementia or no dementia (i.e., e, target domain; testing datasets 1-5; Example 4). The image quality was labelled by three trained human graders (ARR, VTTC, and KS). Only gradable retinal photographs were used. If more than 25% of the peripheral area of the retina was unobservable due to artifacts, including the presence of foreign objects, out-of-focus imaging, blurring, and extreme illumination conditions and if the center region of the retina had significant artifacts that would affect analysis, the photograph was considered ungradable. The inter-grader reliability was high, with Cohen's K coefficients ranging from 0.868 to 0.925. If grader 2 (VTTC) and grader 3 (KS) could not make a decision as to whether an image should be included (e.g., retinal photographs with borderline quality), the senior grader (grader 1 [ARR]) made final decisions.18 The labelling of Alzheimer's disease-dementia in all studies followed the Diagnostic and Statistical Manual of Mental Disorders, 4th edition, criteria for dementia syndrome (Alzheimer's type) and National Institute of Neurological and Communicative Disorders and Stroke and the Alzheimer's Disease and Related Disorders Association criteria for probable or possible Alzheimer's disease. Retinal photographs were labelled as no dementia when the participant had no objective cognitive impairment evident in the neuropsychological assessments and no history of neurodegenerative diseases.
Three testing sets (testing set 1-3; Example 4) also included data from amyloid-PET scan examinations following intravenous 11C-Pittsburgh compound B to quantify amyloid β deposition from a series of brain regions. The retinal photographs with amyloid-PET scan available were additionally labelled as either amyloid β positive or amyloid β negative based solely on the standardized uptake value ratio with reference to the locally validated cutoff value, regardless of their clinical diagnosis. The details of the primary and testing datasets are described in Example 4.
Because the labelling input and classification output were dependent on the individual participant rather than the image, the deep learning model was designed to integrate features of Alzheimer's disease from four retinal photographs from each participant (i.e., both optic nerve head-centered and macula-centered fields from both eyes). The datasets were split at a participant level to inhibit or avoid information leakage and performance overestimation. The method consisted of four phases. In the first phase, the inventors designed a basic model, using EfficientNet-b2,19 as the backbone for feature extractor, which is based on only one single retinal photograph for the detection of Alzheimer's disease-dementia (Example 4). The inventors then proposed a bilateral model based on four retinal photographs, which learned Alzheimer's disease-related features from optic nerve head-centered and macula-centered retinal photographs from both eyes (
The inventors used unsupervised domain adaptation with domain-specific batch normalization to address data heterogeneity and domain shift problems and to improve the model performance. Unsupervised domain adaptation is a type of learning framework that can transfer knowledge learned from a larger number of annotated training data in the source domains to target domains with unlabeled data only. Domain-specific batch normalization is a building block for deep neural networks for which the source domain and the target domain datasets have their own separate batch normalization layer for training and extraction of hyper-parameters. This design addressed characteristics specific to each domain that are not compatible within a single model while retaining domain-invariant information that is common to all domains. In brief, the labelled source domain dataset was first used for training in a supervised way to generate an unsupervised domain adaption network. This unsupervised domain adaptation network was then used to generate pseudo-labels for unlabeled data in the target domain. The final classification network was subsequently trained with full supervision using labelled data from the source domain and pseudo-labelled data from the target domains. Through the fusion of the domain-independent and domain-dependent knowledge learning, the deep learning models could transfer discriminative features from the labelled source domain to the unlabeled target domain (i.e., domain adaptation) and improve the classification performance on the target domain. Due to poor model transfer capability of the domain adaptation-based method, the inventors trained one model for each testing dataset to obtain the information, such as image style distribution of unlabeled testing datasets.
Furthermore, to better understand discriminative features between patients with Alzheimer's disease-dementia and participants without the disease, the inventors used Gradient-weighted Class Activation Mapping (i.e., heatmap) to visualize the features extracted from the last convolutional layer. Details of the network architecture, training details, and objective functions were described in Example 4.
The inventors used the testing datasets to evaluate the model performance at a participant level on three aspects: clinically diagnosed Alzheimer's disease-dementia versus no dementia, individuals who were amyloid β positive versus individuals who were amyloid β negative, and individuals who had clinically diagnosed Alzheimer's disease-dementia and were amyloid β positive versus those who had no cognitive impairment and were amyloid β negative. Models were evaluated based on the following metrics from the five-fold cross validation: the area under the receiver operating characteristic curve (AUROC) and values for accuracy, sensitivity, and specificity for which the cutoff point was the largest Youden Index in each dataset.
In subgroup analyses, the inventors combined the testing 1-3 datasets and stratified individuals on the basis of the presence of eye disease from retinal photographs and diabetes diagnosis status to evaluate discriminative performance. The performance of the unilateral model was also compared between right eyes and left eyes.
A total of 5598 retinal photographs from 648 individuals with Alzheimer's disease and 7351 retinal photographs from 3240 people without the disease were used to train, validate, and test the deep learning models. The characteristics of the primary training, internal validation, and testing datasets at a participant level are reported in
In the internal validation dataset, the bilateral model had 83.6% (SD 2.5) accuracy, 93.2% (SD 2.2) sensitivity, 82.0% (SD 3.1) specificity, and an AUROC of 0.93 (0.01) for detection of Alzheimer's disease-dementia. For differentiation between patients with Alzheimer's disease-dementia and participants who did not have the disease, both bilateral and unilateral models had accuracies of more than 83% and AUROCs of more than 0.9 in internal validation (
In the subgroup analysis, the ability of the model to differentiate between people with Alzheimer's disease-dementia and those without the disease and those who were amyloid β positive from those who were amyloid β negative was improved in patients with concomitant eye disease (accuracy 89.6% [SD 12.5%]) versus those without eye disease (71.7% [11.6%];
Compared with the Hong Kong version of the Montreal Cognitive Assessment for Alzheimer's disease-dementia detection in a community-based cohort, our bilateral model's assessment of testing set 5 had higher sensitivity (100% vs 50%) and a higher AUROC (0.91 vs 0.75; Example 4).
Unsupervised domain adaptation with domain-specific batch normalization was used in the testing datasets to address the issue of data heterogeneity and domain shift problems. After domain adaptation, the model performance was generally improved, suggesting that the model also learned discriminative features from the source domain for Alzheimer's disease detection (Example 4).
In this study, the inventors developed, validated, and tested a novel, retinal photograph-based deep learning algorithm according to an embodiment of the subject invention to detect individuals with Alzheimer's disease, using an unsupervised domain adaptation deep learning technique to improve its general usability. The provided deep learning algorithm showed consistently accurate performance for differentiating between patients with Alzheimer's disease-dementia and individuals with no dementia. In particular, the performance was similar for differentiating between people who were amyloid β positive from those who were amyloid β negative. In addition, the provided deep learning algorithm had good performance in the presence of concomitant eye diseases (e.g., age-related macular degeneration), thus allowing screening in optometry and ophthalmology settings.
Embodiments of the subject invention provide the first deep learning model to detect Alzheimer's disease from retinal photographs alone. In related art, Wisely and colleagues20 proposed a deep learning system to predict Alzheimer's disease using images and measurements from multiple ocular imaging modalities (e.g., optical coherence tomography, optical coherence tomography angiography, ultra-widefield retinal photography, and retinal auto-fluorescence) and patient data. By contrast, embodiments of the subject invention predict Alzheimer's disease based on retinal photographs only, thus improving the efficiency and potential cost-effectiveness of the algorithm. The provided algorithm advantageously employs two advanced deep learning techniques: unsupervised domain adaptation and feature fusion. The novel application of these two techniques addresses two significant challenges: (1) data distribution discrepancy between training and validation and testing datasets, and (2) the integration from multiple optic nerve head-centered and macula-centered retinal photographs from both eyes. With this deep learning architecture, embodiments are transferrable to a new center without developing a new deep learning model. Retrospective data can be collected from this specific center for unsupervised domain adaptation, and the model can subsequently be refined to keep the deep learning model up to date.
To increase applicability, the inventors intentionally included retinal photographs with concomitant eye disease in the training dataset because age-associated eye conditions (e.g., age-related macular degeneration and glaucoma) are common in people older than 60 years. Excluding eyes with these conditions might also introduce selection bias because studies have shown the patients with Alzheimer's disease are more likely to have age-associated macular degeneration and glaucoma.5,21,22 The provided deep learning algorithm retained a robust ability to differentiate between people who had and did not have Alzheimer's disease, even in the presence of concomitant eye diseases. These findings suggest that Alzheimer's disease has unique retinal features that are distinguishable from other eye diseases. Furthermore, patients with type 2 diabetes are at higher risk of cognitive impairment.23 Embodiments of the provided deep learning algorithm performed well without significant interference from concomitant diabetes, and while not being bound by theory, the inventors hypothesize this can be attributed to its similarity with deep learning-based diabetic retinopathy screening.24 However, the performance of the model in participants without eye disease dropped. While not being bound by theory, the inventors hypothesize that an overlap in pathophysiological features shared between Alzheimer's disease and eye diseases might enhance the identification of Alzheimer's disease-associated features from retinal imaging.
The inventors developed a supplementary unilateral model, which can estimate the risk of Alzheimer's disease based on retinal photographs from a single eye. A unilateral model is essential for community screening of Alzheimer's disease because retinal photograph of one eye might not be assessable due to media opacity (e.g., cataract). Results suggest that the unilateral model can also reliably predict Alzheimer's disease-dementia based on unilateral retinal photographs.
The provided retinal photograph-based deep learning model addresses a current gap in Alzheimer's disease screening, in which under-diagnosis of dementia is highly prevalent.25 Early diagnosis of Alzheimer's disease relies on a complex series of cognitive tests, clinical assessments, supportive evidence from neuroimaging (e.g., PET), and cerebrospinal fluid biomarker evidence, with the definitive diagnosis only confirmed post mortem.26 Therefore, patients with Alzheimer's disease are usually diagnosed late after the onset of debilitating dementia when there has already been extensive brain neurodegeneration that might not be amenable to any disease-modifying treatment.27 Embodiments of the retinal photograph-based deep learning model provide a simple, low-cost, low labor-dependent approach to identify potential Alzheimer's disease-dementia patients in community settings with reasonable accuracy and sensitivity. In certain embodiments the identified patients can then be referred to and followed up at tertiary facilities with diagnostic evaluation and subsequent multidisciplinary managements. The detection of Alzheimer's disease based on retinal photographs can also leverage existing community eye-care infrastructure (e.g., optometry or primary care networks) that enables opportunistic Alzheimer's disease screening during routine screening for common eye diseases, such as diabetic retinopathy and glaucoma. With advances in telemedicine and the increasing popularity of non-mydriatic digital retinal cameras and smartphone-based cameras, access to retinal photography is expected to increase. Because retinal photograph-based deep learning approaches could be used for screening Alzheimer's disease-dementia, it is contemplated within the scope of certain embodiments of the subject invention to improve sensitivity and specificity when combining retinal photography with blood-based biomarkers, which have been shown to correlate with brain amyloid and tau burden—the upstream pathology of Alzheimer's disease. In addition, identifying prodromal and preclinical Alzheimer's disease and predicting progression to dementia in those with mild cognitive impairment is contemplated. Advances in retinal photograph-based deep learning model development in this direction are also contemplated for future institutional and clinical applications.
Embodiments have been proven effective across a diverse clinical sample, with datasets from multiethnic, multicountry cohorts and in different clinical settings. Embodiments have been validated in five testing datasets, three of which included amyloid-PET scan. Furthermore, the inventors used unsupervised domain adaptation with domain-specific batch normalization to address data discrepancy from different datasets, which largely improved the proposed model's generalizability and its potential feasibility in other unseen clinical settings. It is further contemplated within the scope of the subject invention that after integration with prediagnosis assessment deep learning models,18 certain embodiments can provide an integrated and comprehensive deep learning pipeline for Alzheimer's disease screening in the community.
Pathological studies suggest that clinical Alzheimer's disease diagnostic sensitivity ranges between 70.9% and 87.3%, and specificity between 44.3% and 70.8%.29 Because the labelling of training datasets are often based on clinician-derived diagnosis, the development of any deep learning algorithm can include retinal photographs from individuals incorrectly labelled as having Alzheimer's disease. Embodiments provide training and/or testing in datasets with PET imaging to mitigate this concern.
Embodiments provide a validated and tested retinal photograph-based deep learning system and method for detecting and treating Alzheimer's disease-dementia, advantageously including a unique and generalizable model useful in community and public health settings to screen for and better treat Alzheimer's disease.
The inventors used retinal photographs from 6 studies dated from 9th November 2003 to 29th September 2019 with labels of “Alzheimer's disease-dementia” and “no dementia” as primary training and validation datasets for model development. Potential for a data leak is remediated by data splitting on the subject level. For example, if one subject is grouped in the training set, all visits for that subject would only be used for training. There is no data cross between training and testing sets. These primary datasets include the following:
Primary-1: The Harmonization Cohort Study was a prospective memory clinic-based study.1,2 Participants with subjective complaints of memory problems and/or demonstrated cognitive impairment on neuropsychological assessment were recruited. Subjects with no dementia were recruited from both memory clinics and the community. All subjects were administered the clinical dementia rating (CDR) scale questionnaire, locally modified versions of the Mini Mental Status Examination (MMSE) and Montreal Cognitive Assessment (MoCA), and a standard neuropsychological battery locally validated for older Singaporeans by trained psychologists.
Primary-2: The study of Novel Retinal Imaging Biomarkers for Cognitive Decline was a prospective observational study to use retinal imaging as a novel biomarker for prognostic outcome measures of cognitive decline. Patients with Alzheimer's disease-dementia were recruited from the Cognitive Disorder Clinic or Memory Clinic of the Prince of Wales Hospital, and subjects with no dementia were recruited from an on-going community-based study in Hong Kong.
Primary-3: The Belfast study was a case-control study wherein Alzheimer's disease-dementia cases were recruited by an opportunistic strategy from the Royal Victoria Hospital, Belfast, UK.3 Subjects with no dementia were recruited from several sources: those responding to a press release on the study, friends of involved controls, carers of patients attending out-patient clinics, and patient-support groups.
Primary-4: The SEED study was a population-based study comprising adults residing in Singapore aged 40 to 80 years at baseline, from 3 major ethnic groups: Chinese, Indians, and Malays. Participants aged 60 years and older were administered the Abbreviated Mental Test (AMT) in SEED to assess cognitive function. Subjects with no dementia (i.e., screening-negative) were defined as a score >6/10 for participants with 0-6 years of formal education and >8/10 for those with more than 6 years of formal education.4,5
Primary-5: The Hong Kong Children Eye Study was a population-based cohort study of eye conditions in children of Grade 1 to Grade 3 from primary schools in Hong Kong. Eye examination and retinal photography were also performed in the parents of the study subjects.6 The Hong Kong version of MoCA was performed in a sub-group of the parental cohort for cognitive function screening.7 Subjects with no dementia (i.e., screening-negative) were defined as the lower cut-off score at 16th percentile in the age and education corrected normative data.
Primary-6: The CUHK volunteer-based cohort was aimed to recruit individuals without ocular abnormalities except for mild cataract as a control group for comparison of ocular imaging measurements with glaucoma. All subjects underwent a comprehensive ophthalmic examination, and cognitive function screening was performed using the Hong Kong version of MoCA in the cohort same as Primary-5.
The inventors obtained 5 non-overlapping datasets for testing dated from 25 Jul. 2011 to 1 Jun. 2021. Included were 3 independent and retrospectively collected datasets of retinal photographs with amyloid-PET scans examination for further assessing the discriminative ability of the deep learning algorithm between groups with Aβ-positive and Aβ-negative, in addition to the label of Alzheimer's disease-dementia/no dementia. All subjects received 11C-Pittsburgh compound B (PiB) intravenously and underwent PET imaging for quantifying the amount of Aβ deposition in a series of brain regions. In addition, the inventors obtained 2 further independent and retrospectively collected datasets, one from a clinic, and another from a community-based study for further assessing the discriminative ability of the DL algorithm between groups with Alzheimer's disease-dementia and no dementia. These 5 studies were collected from:
Testing-1: The ABRI study is a subgroup of the Harmonization Cohort study, PET-MR imaging was performed on a mMR synchronous PET/MR scanner (Siemens Healthcare GmbH) at the Clinical Imaging Research Centre of the National University of Singapore.8 PET images were obtained at 40-min post-injection. Aβ-positive was defined from visual interpretation by experts from 6 Alzheimer's disease-specific regions: frontal lobe, parietal lobe, temporal lobe, anterior cingulate, praecuneus/posterior cingulate.
Testing-2: The participants of the CU-SEEDS study were recruited from the community and from the Cognitive Disorder Clinic of the Prince of Wales Hospital, Hong Kong, to validate different biomarkers (e.g., brain MRI, plasma) for detection of Alzheimer's disease.9 PET/CT imaging was performed at the Department of Nuclear Medicine & PET of Hong Kong Sanatorium & Hospital, Hong Kong. PET images were obtained at 35 min post-injection. Aβ-positive was defined as 1) increased 11C-PIB uptake was visually observed in regions known to have amyloid-beta deposits in patients with Alzheimer's disease-dementia, e.g., frontal lobe, parietal lobe, lateral temporal lobe, posterior cingulate, praecuneus and/or caudate; and/or 2) global retention ≥1.42.
Testing-3: The participants were recruited from the Mayo Clinic Study of Aging and Alzheimer's Disease Research Centre at Mayo Clinic Rochester. In this dataset, all participants underwent PET/CT imaging. Amyloid PET imaging was performed with Pittsburgh compound B. PET images were analyzed with the institution's in-house, fully automated image-processing pipeline, where image voxel values are extracted from automatically labelled regions of interest propagated from an MRI template.10 A PiB standardized uptake value ratio for each participant was calculated as previously described.11
Testing-4: The participants were recruited from the Behavioural Neurology Clinic at Mayo Clinic Arizona and National institute of Aging Arizona Alzheimer's Disease Research Core Centre.
Testing-5: The participants aged ≥65 years from the Mr. OS and Ms. OS Hong Kong Study were mainly recruited through posting advertisement in housing estates and local community centers from August 2001 to December 2003.12 The participants were invited to follow up from 2019 to 2021 for examination including retinal photography and cognitive screening testing using the Hong Kong version of MoCA at the Jockey Club Centre for Osteoporosis Care and Control, the Chinese University of Hong Kong. Participants who were screening-positive were referred to a geriatrician for further clinical examination.13
Definition and labelling of Alzheimer's disease-dementia, cognitive impairment no dementia (CIND), and no dementia proceeded as follows. All the individuals were labelled as “Alzheimer's disease-dementia” and “no dementia” (or “cognitive impairment no dementia (CIND)”), based on a neuropsychological assessment or cognitive function screening test. Individuals with Alzheimer's disease-dementia in the Primary-1, Primary-2, Primary-3, Testing-1 to -5 fulfilled Diagnostic and Statistical Manual of Mental Disorders, 4th edition (DSM-IV) criteria for dementia syndrome (Alzheimer's type) and National Institute of Neurological and Communicative Disorders and Stroke and the Alzheimer's Disease and Related Disorders Association (NINDS-ADRDA) criteria for probable or possible Alzheimer's disease. In the Testing-1 to -3 datasets, CIND was defined as impairment on neuropsychological assessment but did not meet the criteria for dementia according to the DSM-IV, and subjects without dementia or had no cognitive impairment was defined as no objective impairment on the neuropsychological assessment. In Primary-4, Primary-5, Primary-6, Testing-4, and Testing-5, subjects with “no dementia” were selected and defined as AMT or MoCA screening-negative without any history of neurodegenerative diseases.
For deep learning model development, the inventors presented a whole framework for Alzheimer's Disease classification from retinal photographs by utilizing novel domain adaptation technique14,15. Two key points are introduced: the fusion mechanism for multiple retinal photographs and domain adaptation mechanism for domain discrepancy, respectively.
In the provided bilateral model, the classification label used for Alzheimer's disease-dementia is subject-level while the retinal photography is taken at eye-level. Therefore, one-to-one mapping between a retinal photograph and a disease-label is difficult to be guaranteed. While not being bound by theory, the inventors hypothesized that the Alzheimer's disease-related features on retinal photographs are present on either one eye or both eyes. Based on this hypothesis, the inventors provided a novel “bilateral” deep learning model to classify Alzheimer's disease through retinal photographs from both eyes.
The model first applied image pre-processing methods, including data normalization and data augmentation on the retinal photographs before model development, as the retinal photographs that were collected from different centers using different retinal cameras and imaging protocols (
The provided method consists of three phases for usage under different conditions. In the first phase, a bilateral model (“BM-Net”) for dual eye analysis took both eyes' hidden features into account. This was first designed a basic model and only used one retinal photograph for the classification network (see
In this embodiment, the provided bilateral model (BM-Net) takes the hidden features from retinal photographs in both eyes into account. Theoretically, the BM-Net made predictions with the assumption that either or both eyes have Alzheimer's disease-related retinal features are present. The inventors concatenate the features from different eyes and apply additional convolutional layers to fuse them for the final classification.
In this embodiment the inventors provided a unilateral model (“UM-Net”) for single eye analysis as individuals could have ungradable retinal photographs from one eye (e.g., due to severe cataract) but gradable retinal photographs from another eye (see
Based on the basic network, the UM-net considered the joint information of optic nerve head-centered and macula-centered retinal photographs. Similar to multi-view images, the optic nerve head-centered and macula-centered retinal photographs have intersected and non-intersected areas. This embodiment adopted the Multi-view CNN18,19 method which used a shared feature extractor ahead of information fusion. This embodiment used 3D convolution operation for feature map aggregation in the fusion layer to take the advantage of the strict imaging protocol of retinal photographs. The framework is depicted in
To explore the added values of demographic and clinical information (age, gender, yes/no hypertension, yes/no diabetes) to Alzheimer's disease classification results, the inventors further designed a hybrid model with demographic and clinical information (HM-Net) to integrate the information and high-level semantic features extracted from the deep model layers. Specifically, the inventors applied Bilinear Transformation to realize it as illustrated in
The inventors finally trained a “risk factors alone” model for Alzheimer's disease prediction. The risk factors include age, gender, yes/no diabetes, and yes/no hypertension. The inventors also utilized deep learning to model the relationship between risk factors and Alzheimer's disease. The deep learning model consisted of 3 fully-connected layers with 128, 256, and 2 nodes, respectively. After the first two fully-connected layers, the inventors added dropout layer with a ratio of 0.25. The SoftMax was employed after the output to normalize the prediction between 0-1. The inventors trained this model for 1000 epochs with a batch size of 128 on one Tesla V100 GPU. The learning rate was initially set as 0.001 and decreased by 0.1 after every 400 epochs while Adam optimizer was applied.
Retinal features on retinal photographs can appear differently due to different retinal cameras, imaging protocols, ethnicity, and ocular pathologies, etc. Such dataset discrepancy leads to the poor performance of the deep learning models on new unseen datasets. In this study, the inventors defined the training datasets as “source domain” and testing datasets as “target domain” to tackle the issue of dataset discrepancy. The inventors first utilized EfficientNet-b216 as the backbone to extract features, and then equipped with domain adaptation technique to deal with dataset discrepancy problem via a domain adaptation mechanism. Specifically, the whole framework consists of two stages. In the first stage, the inventors trained the deep models (Bilateral model or Unilateral model) with supervised learning on the source dataset with image-level annotations (Alzheimer's disease-dementia or no dementia). In the second stage, the inventors introduced a domain adaptation method by estimating the pseudo labels for the retinal photographs from the target domain using domain-specific batch normalization technique.17 According to the pseudo labels, the network can learn the domain-specific information through the multi-task learning paradigm. During training process, images from the source and target domains were fed into separate batch normalization layers (
The inventors employed cross validation on testing datasets during the domain adaptation period. Specifically, for the internal dataset, the inventors split the data into internal training and internal validation sets with a ratio of 4:1. The inventors then divided each testing dataset into 5 folds and utilized 4 folds without labels combining with the internal training set to train the model. Thus, there were 5 different models trained with different folds. The inventors tested each model on the same internal validation set and the remaining fold of each testing dataset. The final performance was the average of the five models. The unsupervised domain adaptation process was done on the “training folds” of each testing dataset.
In terms of model training, the inventors applied a series of steps to deal with the highly unbalanced dataset and the large volume of parameters. To address the data imbalance issue, the inventors used an over-sampling strategy to select an equal number of Alzheimer's disease-dementia and control subjects for the network training at each epoch. Furthermore, for each subject, the provided model was configured to randomly select 4 images across different visits (random repeating if not enough). Thus, the inventors inhibited or avoided data leakage while also maximizing the data variance in each epoch of training.
The inventors trained the network with binary cross-entropy loss in both stages for the bilateral model and unilateral model. To deal with the class imbalance problem, the inventors utilized over sampling for the class with less data, so that there was no need to apply weighted loss functions. In the first stage, the objective function can be represented as
where N is the total number of paired retinal photographs (i.e., both optic nerve head-centered and macula-centered retinal photographs in the eye are available) in the source domain, yi is the classification label of the i-th image pairs, and pi is the prediction probability of current subject which is normalized by Softmax function, and θs represents the network parameter updated for source domain dataset. The paired retinal photographs were sampled from the same patients. The number of paired retinal photographs can be different according to the unilateral model (two images) or bilateral model (four images).
In the second stage, the inventors expected the network could learn the knowledge from both source and target domains. Therefore, the inventors estimated the pseudo labels of target domain images using the network weight trained on the first stage, so that the model can be trained with full supervision L:
where M is the total number of paired retinal photographs in the target domain, yj′ is the pseudo label of j-th paired retinal photographs, and pj is the prediction probability, and θt represents the network parameter updated for target domain dataset. It is worth to note that the L1 is minimized to optimize the whole convolutional layers and batch normalization layer for the source domain, while L2 is minimized to optimize the whole convolutional layers and batch normalization layer for the target domain.
Since the pseudo labels contain noisy labels, the inventors updated pseudo labels on each epoch training in the second stage according to
where y′ is the pseudo label used for previous epoch, yc is the new calculated pseudo label in the current epoch, and y″ is the new pseudo label for the network training in the next epoch. λ is a balance coefficient and is calculated as follows.
Experimental implementation details included that the deep learning algorithm was implemented with the PyTorch library. The inventors train the bilateral model, unilateral model, and hybrid model with a batch size of 40, 60, and 40, respectively and a total of 160 epochs for domain adaptation training. The inventors utilized RAdam20 as the optimizer. The data augmentation is deployed in a random way including affine, horizontal and vertical flip, and color jitter.
To aid visualization of the features used for the classification and to better understand the discriminative features among Alzheimer's disease-dementia subjects out of non-demented subjects, the inventors used Gradient-weighted Class Activation Mapping (Grad-CAM) to visualize the discriminative features used for the classification. Since the inventors developed a multi-input model, the weighted gradients were generated from the last shared convolution layer from the basic model. Moreover, to adequately demonstrate the attention areas, the inventors used eye-wise normalization, which was used to investigate the attention on different eyes.
The training included two stages. While training in the first stage, SGD Optimizer was used with 0.001 learning rate and 0.00002 weight decay, along with the binary cross-entropy loss. In the second stage, the learning rate is decreased to 0.0001 to fine-tune the network on the target domain dataset with a hybrid loss function for both source and target domains. Specifically, the inventors used Kornia for GPU-based data augmentation to improve the data augmentation speed due to the massive parallel input images. The network was trained on two GPUs of NVIDIA Tesla V100 with CUDA v10.1.
It should be understood that the examples and embodiments described herein are for illustrative purposes only and that various modifications or changes in light thereof will be suggested to persons skilled in the art and are to be included within the spirit and purview of this application and the scope of the appended claims. In addition, any elements or limitations of any invention or embodiment thereof disclosed herein can be combined with any and/or all other elements or limitations (individually or in any combination) or any other invention or embodiment thereof disclosed herein, and all such combinations are contemplated with the scope of the invention without limitation thereto.