An example method for grading, segmenting, and analyzing lung adenocarcinoma (LUAD) pathology slides using artificial intelligence is described herein. The method includes receiving a digital pathology image of a LUAD tissue sample; inputting the digital pathology image into an artificial intelligence model; and grading, using the artificial intelligence model, the one or more tumors within the LUAD tissue sample.
The step of grading optionally includes assigning each of the one or more tumors to one of a plurality of classes. For example, the classes can include one or more of normal alveolar, normal bronchiolar, Grade 1 LUAD, Grade 2 LUAD, Grade 3 LUAD, Grade 4 LUAD, and Grade 5 LUAD. Alternatively or additionally, the step of grading includes generating graphical display data for a pseudo color map of the one or more tumors.
In some implementations, the step of grading, using the artificial intelligence model, the one or more tumors comprises assigning one or more areas within each of the one or more tumors to one of a plurality of classes on a pixel-by-pixel basis or a cell-by-cell basis.
In some implementation, the method further comprises, identifying, based at least on the pixel-by-pixel or cell-by-cell assignments, one or more genes of interest or one or more drivers of tumor progression.
In some implementations, the method further includes segmenting, using the artificial intelligence model, the one or more tumors in the digital pathology image.
In some implementations, the method further includes analyzing the one or more tumors. The step of analyzing optionally includes counting the one or more tumors. Alternatively or additionally, the step of analyzing optionally includes characterizing an intratumor heterogeneity of the one or more tumors.
In some implementations, the method further includes performing an immuno-histochemistry (IHC) analysis of the one or more tumors.
In some implementations, the artificial intelligence model is a machine learning model. For example, the machine learning model can optionally be a supervised machine learning model such as a convolutional neural network (CNN).
In some implementations, the example supervised machine learning model comprises one or more Residual Neural Network (ResNet) layers or components.
In some implementations, the supervised machine learning model further comprises one or more atrous convolutional layers and/or one or more transposed convolutional layers.
In some implementations, the digital pathology image is a hematoxylin & eosin (H&E) stained slide image. Optionally, the LUAD tissue sample is from a mouse. Alternatively, the LUAD tissue sample is optionally from a human.
An example method for integrating an immuno-histochemistry (IHC analysis) with an artificial intelligence-based LUAD tissue sample analysis is described herein. The method includes receiving a first digital pathology image of a first LUAD tissue sample, the first digital pathology image being a hematoxylin & eosin (H&E) stained slide image; inputting the first digital pathology image into an artificial intelligence model; grading, using the artificial intelligence model, one or more tumors within the first LUAD tissue sample; and segmenting, using the artificial intelligence model, the one or more tumors in the digital pathology image. Additionally, the method includes receiving a second digital pathology image comprising a second LUAD tissue sample, the second digital pathology image being an immuno-stained slide image; and identifying and classifying a plurality of positively and negatively stained cells within the second LUAD tissue sample. The method further includes co-registering the first and second digital pathology images; and projecting a plurality of respective coordinates of the positively and negatively stained cells within the second LUAD tissue sample onto the one or more tumors within the first LUAD tissue sample.
An example transfer learning method is also described herein. The method includes training a machine learning model with a dataset, where the dataset includes a plurality of mouse model digital pathology images. Each of the mouse model digital pathology images is of a respective lung LUAD tissue sample from a mouse. The method further includes receiving a digital pathology image of a LUAD tissue sample from a human; inputting the digital pathology image into the trained machine learning model; and grading, using the trained machine learning model, one or more tumors within the LUAD tissue sample from the human.
It should be understood that the above-described subject matter may also be implemented as a computer-controlled apparatus, a computer process, a computing system, or an article of manufacture, such as a computer-readable storage medium.
Other systems, methods, features and/or advantages will be or may become apparent to one with skill in the art upon examination of the following drawings and detailed description. It is intended that all such additional systems, methods, features and/or advantages be included within this description and be protected by the accompanying claims.
The components in the drawings are not necessarily to scale relative to each other. Like reference numerals designate corresponding parts throughout the several views.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art. Methods and materials similar or equivalent to those described herein can be used in the practice or testing of the present disclosure. As used in the specification, and in the appended claims, the singular forms “a,” an, “the” include plural referents unless the context clearly dictates otherwise. The term “comprising” and variations thereof as used herein is used synonymously with the term “including” and variations thereof and are open, non-limiting terms. The terms “optional” or “optionally” used herein mean that the subsequently described feature, event or circumstance may or may not occur, and that the description includes instances where said feature, event or circumstance occurs and instances where it does not. Ranges may be expressed herein as from “about” one particular value, and/or to “about” another particular value. When such a range is expressed, an aspect includes from the one particular value and/or to the other particular value. Similarly, when values are expressed as approximations, by use of the antecedent “about,” it will be understood that the particular value forms another aspect. It will be further understood that the endpoints of each of the ranges are significant both in relation to the other endpoint, and independently of the other endpoint.
As used herein, the terms “about” or “approximately” when referring to a measurable value such as an amount, a percentage, and the like, is meant to encompass variations of ±20%, ±10%, ±5%, or ±1% from the measurable value.
“Administration” of “administering” to a subject includes any route of introducing or delivering to a subject an agent. Administration can be carried out by any suitable means for delivering the agent. Administration includes self-administration and the administration by another.
The term “subject” is defined herein to include animals such as mammals, including, but not limited to, primates (e.g., humans), cows, sheep, goats, horses, dogs, cats, rabbits, rats, mice and the like. In some embodiments, the subject is a human.
The term “tumor” is defined herein as an abnormal mass of hyperproliferative or neoplastic cells from a tissue other than blood, bone marrow, or the lymphatic system, which may be benign or cancerous. In general, the tumors described herein are cancerous. As used herein, the terms “hyperproliferative” and “neoplastic” refer to cells having the capacity for autonomous growth, i.e., an abnormal state or condition characterized by rapidly proliferating cell growth. Hyperproliferative and neoplastic disease states may be categorized as pathologic, i.e., characterizing or constituting a disease state, or may be categorized as non-pathologic, i.e., a deviation from normal but not associated with a disease state. The term is meant to include all types of solid cancerous growths, metastatic tissues or malignantly transformed cells, tissues, or organs, irrespective of histopathologic type or stage of invasiveness. “Pathologic hyperproliferative” cells occur in disease states characterized by malignant tumor growth. Examples of non-pathologic hyperproliferative cells include proliferation of cells associated with wound repair. Examples of solid tumors are sarcomas, carcinomas, and lymphomas. Leukemias (cancers of the blood) generally do not form solid tumors.
The term “carcinoma” is art recognized and refers to malignancies of epithelial or endocrine tissues including respiratory system carcinomas, gastrointestinal system carcinomas, genitourinary system carcinomas, testicular carcinomas, breast carcinomas, prostatic carcinomas, endocrine system carcinomas, and melanomas. Examples include, but are not limited to, lung carcinoma, adrenal carcinoma, rectal carcinoma, colon carcinoma, esophageal carcinoma, prostate carcinoma, pancreatic carcinoma, head and neck carcinoma, or melanoma. The term also includes carcinosarcomas, e.g., which include malignant tumors composed of carcinomatous and sarcomatous tissues. An “adenocarcinoma” refers to a carcinoma derived from glandular tissue or in which the tumor cells form recognizable glandular structures. The term “sarcoma” is art recognized and refers to malignant tumors of mesenchymal derivation.
The term “artificial intelligence” is defined herein to include any technique that enables one or more computing devices or comping systems (i.e., a machine) to mimic human intelligence. Artificial intelligence (AI) includes, but is not limited to, knowledge bases, machine learning, representation learning, and deep learning. The term “machine learning” is defined herein to be a subset of AI that enables a machine to acquire knowledge by extracting patterns from raw data. Machine learning techniques include, but are not limited to, logistic regression, support vector machines (SVMs), decision trees, Naïve Bayes classifiers, and artificial neural networks. The term “representation learning” is defined herein to be a subset of machine learning that enables a machine to automatically discover representations needed for feature detection, prediction, classification, etc. from raw data. Representation learning techniques include, but are not limited to, autoencoders. The term “deep learning” is defined herein to be a subset of machine learning that that enables a machine to automatically discover representations needed for feature detection, prediction, classification, etc. in raw data using layers of processing. Deep learning techniques include, but are not limited to, artificial neural network or multilayer perceptron (MLP).
Machine learning techniques include supervised, semi-supervised, and unsupervised learning models. In a supervised learning model, the model learns a function that maps an input (also known as feature or features) to an output (also known as target or target) during training with a labeled data set (or dataset). In an unsupervised learning model, the model learns patterns (e.g., structure, distribution, etc.) within an unlabeled data set. In a semi-supervised model, the model learns a function that maps an input (also known as feature or features) to an output (also known as target or target) during training with both labeled and unlabeled data.
In one example described herein, a method for grading, segmenting, and analyzing lung adenocarcinoma (LUAD) pathology slides using artificial intelligence is provided. The method includes receiving a digital pathology image of a LUAD tissue sample. The digital pathology image can be a whole slide image (WSI) or an image field captured from a microscope. For example, in some implementations, the digital pathology image is a hematoxylin & eosin (H&E) stained slide image. Optionally, in some implementations, the LUAD tissue sample is from a mouse. Alternatively, in other implementations, the LUAD tissue sample is optionally from a human.
The method also includes inputting the digital pathology image into an artificial intelligence model. Additionally, as described herein, the digital pathology image is optionally divided into patches/tiles before being input into the artificial intelligence model. It should be understood that such artificial intelligence model is operating in inference mode. In other words, such artificial intelligence model was previously trained with a data set (or dataset) to map an input (also referred to as feature or features) to an output (also referred to as target or targets). In some implementations, the artificial intelligence model is a machine learning model. For example, the machine learning model can optionally be a supervised machine learning model such as a convolutional neural network (CNN), multilayer perceptron, or support-vector machine. An example CNN architecture is described herein.
An artificial neural network (ANN) is a computing system including a plurality of interconnected neurons (e.g., also referred to as “nodes”). This disclosure contemplates that the nodes can be implemented using a computing device (e.g., a processing unit and memory as described herein). The nodes can be arranged in a plurality of layers such as input layer, output layer, and optionally one or more hidden layers. An ANN having hidden layers can be referred to as deep neural network or multilayer perceptron (MLP). Each node is connected to one or more other nodes in the ANN. For example, each layer is made of a plurality of nodes, where each node is connected to all nodes in the previous layer. The nodes in a given layer are not interconnected with one another, i.e., the nodes in a given layer function independently of one another. As used herein, nodes in the input layer receive data from outside of the ANN, nodes in the hidden layer(s) modify the data between the input and output layers, and nodes in the output layer provide the results. Each node is configured to receive an input, implement an activation function (e.g., binary step, linear, sigmoid, tan H, or rectified linear unit (ReLU) function), and provide an output in accordance with the activation function. Additionally, each node is associated with a respective weight. ANNs are trained with a data set to maximize or minimize an objective function. In some implementations, the objective function is a cost function, which is a measure of the ANN's performance (e.g., error such as L1 or L2 loss) during training. The training algorithm tunes the node weights and/or bias to minimize the cost function. This disclosure contemplates that any algorithm that finds the maximum or minimum of the objective function can be used for training the ANN. Training algorithms for ANNs include, but are not limited to, backpropagation. It should be understood that an artificial neural network is provided only as an example machine learning model. This disclosure contemplates that the machine learning model can be another supervised learning model, semi-supervised learning model, or unsupervised learning model. Machine learning models are known in the art and are therefore not described in further detail herein.
A convolutional neural network (CNN) is a type of deep neural network that has been applied, for example, to image analysis applications. Unlike a traditional neural networks, each layer in a CNN has a plurality of nodes arranged in three dimensions (width, height, depth). CNNs can include different types of layers, e.g., convolutional, pooling, and fully-connected (also referred to herein as “dense”) layers. A convolutional layer includes a set of filters and performs the bulk of the computations. A pooling layer is optionally inserted between convolutional layers to reduce the computational power and/or control overfitting (e.g., by downsampling). A fully-connected layer includes neurons, where each neuron is connected to all of the neurons in the previous layer. The layers are stacked similar to traditional neural networks.
In some implementations, the model takes an input image of size 224×224-pixel (˜112×112 micron, 20× magnification). Although training with high-resolution images adds computational burden, it was essential to train the model with high-resolution images to capture all possible features that help classify the cells into different grades.
In some embodiments, the architecture of the GLASS-AI network is configured to classify each pixel in the input image into one of six target classes: Normal alveolar, Normal airway, Grade 1 LUAD, Grade 2 LUAD, Grade 3 LUAD, and Grade 4 LUAD. In general, the GLASS-AI network architecture consists of encoder and decoder architectures.
In some implementations, the supervised machine learning model is a convolutional neural network (CNN). For example, the supervised machine learning model can include one or more Residual Neural Network (ResNet) layers or components. Optionally, in examples described herein, the supervised machine learning model is ResNet-18, which is an 18-layer residual neural network that incorporates inputs from earlier layers to improve performance and is pretrained on a known dataset (i.e., the ImageNet dataset). It should be understood that ResNet-18 is provided only as an example and that other ResNet architectures may be used. There are multiple variations of ResNet architectures, such as ResNet16, ResNet18, ResNet50, and ResNet101, which may be used in different implementations. The reason for choosing 18-layer architecture as an encoder is to avoid the vanishing gradient problem that may occur when a network has deeper layers. Additionally, the residual neural network architecture can be modified to include one or more atrous convolutional layers. An atrous convolutional layer (sometimes referred to as a dilated convolutional layer) introduces a dilation rate parameter, which defines spacing between values in the kernel, to the convolution. Atrous convolutional layers are known in the art and therefore not described in further detail herein. Alternatively or additionally, the residual neural network architecture is optionally modified to include one or more transposed convolutional layers. A transposed convolutional layer (sometimes referred to as a fractionally strided convolutional layer) performs a convolution operation but reverts its spatial resolution. Transposed convolutional layers are known in the art and therefore not described in further detail herein. An example supervised learning model architecture is shown in
In some embodiments, decoder layers may consist of one or more components, including, but not limited to, parallel atrous spatial pyramid pooling layer(s), up-sampling layer(s), SoftMax layer(s), classification layer(s), and/or smoothing layer(s).
An example Parallel Atrous Spatial Pyramid Pooling (ASPP) may be configured to capture distinctive features, such as cell and nucleus size and shape, which helps differentiate between tumor grades that look very similar (e.g., differentiate between grade 3 and grade 4 cells). Therefore, the output of the ResNet18 is convolved with multiple parallel atrous convolutions containing different dilation rates. This ensures better capture of the image's multiscale contextual and semantic information.
An example up-sampling layer may be configured to classify each pixel in the input images, transposed convolution layer is used to up-scale the features maps to generate an output feature map with a spatial dimension equal to the input image.
An example SoftMax layer may be configured to utilize a SoftMax function that takes the up-sampled feature maps from the previous layer and assigns probabilities to each class.
An example classification layer may be configured to compute the cross-entropy loss for classification and weighted classification tasks with mutually exclusive classes. The layer infers the number of classes from the output size of the previous SoftMax layer.
An example smoothing layer may comprise a final layer added at the end of a plurality of layers to smooth predictions and minimize artifacts from image patch edges and produce smooth output labels.
In various examples, the GLASS-AI model classifies each pixel in the input image and produces an image labeled with the predicted classes. The final labeled image is smoothed by the last layer to remove artifacts and pixelation.
Additionally, the example method includes grading, using the artificial intelligence model, the one or more tumors within the LUAD tissue sample. The step of grading optionally includes assigning each of the one or more tumors to one of a plurality of classes. For example, the classes can include one or more of normal alveolar, normal bronchiolar, Grade 1 LUAD, Grade 2 LUAD, Grade 3 LUAD, Grade 4 LUAD, and Grade 5 LUAD. Alternatively or additionally, the step of grading includes generating graphical display data for a pseudo color map of the one or more tumors.
In some implementations, the method further includes segmenting, using the artificial intelligence model, the one or more tumors in the digital pathology image. Alternatively or additionally, the step of segmenting includes generating graphical display data for a segmentation map of the one or more tumors.
In some implementations, the method further includes analyzing the one or more tumors. The step of analyzing optionally includes counting the one or more tumors. Alternatively or additionally, the step of analyzing optionally includes characterizing an intratumor heterogeneity of the one or more tumors.
In some implementations, the method further includes performing an immuno-histochemistry (IHC) analysis of the one or more tumors. In other words, the artificial intelligence-based methods for studying LUAD tissues samples can be integrated with an immuno-histochemistry (IHC) analysis.
An example method for integrating with an artificial intelligence-based LUAD tissue sample analysis is described herein. The method includes receiving a first digital pathology image of a first LUAD tissue sample, the first digital pathology image being a hematoxylin & eosin (H&E) stained slide image; inputting the first digital pathology image into an artificial intelligence model; grading, using the artificial intelligence model, the one or more tumors within the first LUAD tissue sample; and segmenting, using the artificial intelligence model, the one or more tumors in the digital pathology image. Additionally, the method includes receiving a second digital pathology image comprising a second LUAD tissue sample, the second digital pathology image being an immuno-stained slide image; and identifying and classifying a plurality of positively and negatively stained cells within the second LUAD tissue sample. The method further includes co-registering the first and second digital pathology images; and projecting a plurality of respective coordinates of the positively and negatively stained cells within the second LUAD tissue sample onto the one or more tumors within the first LUAD tissue sample.
An example transfer learning method is also described herein. The method includes training a machine learning model with a dataset, where the dataset includes a plurality of mouse model digital pathology images. Each of the mouse model digital pathology images is of a respective lung LUAD tissue sample from a mouse. The method further includes receiving a digital pathology image of a LUAD tissue sample from a human; inputting the digital pathology image into the trained machine learning model; and grading, using the trained machine learning model, one or more tumors within the LUAD tissue sample from the human. Thus, as described herein, the machine learning model trained on mouse models is transferred to perform with acceptable accuracy in inference mode on digital pathology images of human tissue samples.
It should be appreciated that the logical operations described herein with respect to the various figures may be implemented (1) as a sequence of computer implemented acts or program modules (i.e., software) running on a computing device (e.g., the computing device described in
Referring to
In its most basic configuration, computing device 500 typically includes at least one processing unit 506 and system memory 504. Depending on the exact configuration and type of computing device, system memory 504 may be volatile (such as random-access memory (RAM)), non-volatile (such as read-only memory (ROM), flash memory, etc.), or some combination of the two. This most basic configuration is illustrated in
Computing device 500 may have additional features/functionality. For example, computing device 500 may include additional storage such as removable storage 508 and non-removable storage 510 including, but not limited to, magnetic or optical disks or tapes. Computing device 500 may also contain network connection(s) 516 that allow the device to communicate with other devices. Computing device 500 may also have input device(s) 514 such as a keyboard, mouse, touch screen, etc. Output device(s) 512 such as a display, speakers, printer, etc. may also be included. The additional devices may be connected to the bus in order to facilitate communication of data among the components of the computing device 500. All these devices are well known in the art and need not be discussed at length here.
The processing unit 506 may be configured to execute program code encoded in tangible, computer-readable media. Tangible, computer-readable media refers to any media that is capable of providing data that causes the computing device 500 (i.e., a machine) to operate in a particular fashion. Various computer-readable media may be utilized to provide instructions to the processing unit 506 for execution. Example tangible, computer-readable media may include, but is not limited to, volatile media, non-volatile media, removable media and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. System memory 504, removable storage 508, and non-removable storage 510 are all examples of tangible, computer storage media. Example tangible, computer-readable recording media include, but are not limited to, an integrated circuit (e.g., field-programmable gate array or application-specific IC), a hard disk, an optical disk, a magneto-optical disk, a floppy disk, a magnetic tape, a holographic storage medium, a solid-state device, RAM, ROM, electrically erasable program read-only memory (EEPROM), flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices.
In an example implementation, the processing unit 506 may execute program code stored in the system memory 504. For example, the bus may carry data to the system memory 504, from which the processing unit 506 receives and executes instructions. The data received by the system memory 504 may optionally be stored on the removable storage 508 or the non-removable storage 510 before or after execution by the processing unit 506.
It should be understood that the various techniques described herein may be implemented in connection with hardware or software or, where appropriate, with a combination thereof. Thus, the methods and apparatuses of the presently disclosed subject matter, or certain aspects or portions thereof, may take the form of program code (i.e., instructions) embodied in tangible media, such as floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium wherein, when the program code is loaded into and executed by a machine, such as a computing device, the machine becomes an apparatus for practicing the presently disclosed subject matter. In the case of program code execution on programmable computers, the computing device generally includes a processor, a storage medium readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device, and at least one output device. One or more programs may implement or utilize the processes described in connection with the presently disclosed subject matter, e.g., through the use of an application programming interface (API), reusable controls, or the like. Such programs may be implemented in a high-level procedural or object-oriented programming language to communicate with a computer system. However, the program(s) can be implemented in assembly or machine language, if desired. In any case, the language may be a compiled or interpreted language and it may be combined with hardware implementations.
Embodiments of the present disclosure provide a novel, open-source tool for the research community using a machine learning-based pipeline for grading histological cross-sections of lung adenocarcinoma in mouse models. In addition, the machine learning model uncovers a significant degree of intratumor heterogeneity that is not reported by human raters.
Preclinical mouse models of lung adenocarcinoma are invaluable for investigating molecular drivers of tumor formation, progression, and therapeutic resistance. However, histological analysis of these preclinical models requires significant time and training to ensure accuracy and consistency. To achieve a more objective and standardized analysis, deep learning was used to create GLASS-AI (Grading of Lung Adenocarcinoma with Simultaneous Segmentation by Artificial Intelligence), a histological image analysis tool that can be widely utilized by the research community to grade, segment, and analyze tumors in mouse models of lung adenocarcinoma. GLASS-AI demonstrates strong agreement with expert human raters while uncovering a significant degree of unreported intratumor heterogeneity. Integrating immunohistochemical staining with high-resolution grade analysis by GLASS-AI identified dysregulation of Mapk/Erk signaling in high-grade lung adenocarcinomas and locally advanced tumor regions. The present disclosure demonstrates the benefit of employing GLASS-AI in preclinical lung adenocarcinoma models and the power of integrating machine learning and molecular biology techniques for studying cancer progression. GLASS-AI is available from https://github.com/jlockhar/GLASS-AI.
The approval of whole slide scanners for use in clinical pathology by the U.S. Food and Drug Administration (FDA) in 2017 led to the rapid proliferation of digital pathology images in both healthcare and pre-clinical settings. Not only have whole slide images (WSIs) increased the efficiency of pathologists' workflow, but their digitization also enables collaboration among geographically distant groups. Furthermore, advances in computer vision and image processing have given rise to several applications that can assist in the histopathological analysis of WSIs, particularly in the field of oncology. These applications often utilize pre-trained convolutional neural networks (CNNs) to perform or assist with time-consuming tasks, such as nuclei segmentation1,2 histological staining analysis3, and tumor segmentation4-6. Similar machine learning approaches have been developed for more nuanced analyses, including quantifying tumor-associated or tumor-infiltrating immune cells7-9, microsatellite instability10, and prediction of patient mutational status from WSIs11,12. Machine learning models trained to classify tumors into diagnostically distinct grades using existing systems, such as Gleason score for prostate cancer13-15, have also been reported. In many of these studies, the accuracy of the machine learning model has been measured in terms of agreement with expert human raters on a sample-by-sample basis. While a suitable measure of performance, this comparison level fails to capture much of the information uncovered by the high-resolution analysis these algorithms perform.
In addition, the development of these machine learning models has been focused almost exclusively on analyzing human samples. For clinical applications, building human-focused models from observational data from human patients, like that stored in The Cancer Genome Atlas (TCGA)'s collection of WSIs and the associated molecular data16, is ideal. However, the intense focus on clinical applications has provided very few machine learning models useful for translational and basic research. Machine learning applications in pre-clinical research present an excellent opportunity to enhance and accelerate analyses of the experimental data produced from these sources.
Several mouse models of lung adenocarcinoma (LUAD) have been reported, of which the KrasLSL-G12D/+ model is the most widely used17. This well-studied model serves as a valuable baseline for studying other mutations commonly found in LUAD, such as Trp53R172H, separately or in conjunction with the activating KrasG12D mutation. These models often develop over 100 primary tumors, making a thorough analysis of these valuable specimens extremely time-consuming, even for experienced researchers.
Embodiments of the present disclosure provide GLASS-AI (Grading of Lung Adenocarcinoma with Simultaneous Segmentation by Artificial Intelligence), a machine learning pipeline for the analysis of mouse models of lung adenocarcinoma that provides a rapid means of analyzing tumor grade from WSIs. The GLASS-AI pipeline was trained on multiple genetically engineered mouse models to ensure that it generalized well. Analysis of several mouse models of LUAD revealed a high degree of accuracy, comparable to expert human raters. Furthermore, the high-resolution analysis performed by GLASS-AI revealed extensive intratumor heterogeneity that was not reported by the human raters. Alignment of these heterogeneous tumor regions with adjacent immunostained sections showed a strong correlation between tumor grade and aberrant Mapk/Erk signaling that differed between KrasG12D/+ (K), TAp73Δ/Δ; KrasG12D/+ (TK), and KrasG12D/+. Trp53R172H/+ (KP) mouse models. The GLASS-AI pipeline empowers pre-clinical research by rapid analysis of LUAD without the need for extensive training of human raters.
Developing an accurate machine learning model requires a large amount of high-quality training data. To construct a training dataset, WSIs were collected from KrasG12D/+. RosamG/mG (K) (n=4), TAp73Δ/Δ; KrasG12D/+ (TK) (n=15), and KrasG12D/+; Trp53Δ/Δ (n=14) mice 30 weeks after LUAD initiation. Slides were divided among three expert human raters who segmented and graded tumors using the Grade 1-Grade 5 scale reported by Tyler Jacks' laboratory18,19, although Grade 5 areas were not observed within the animals used for generating the training library. The WSIs were then divided into 224×224-pixel (approximately 112×112 micron) image patches and corresponding annotation patches (
In the example depicted in
Referring now to
Referring now to
Referring now to
Referring now to
After training, GLASS-AI achieved an accuracy of 88% on the patches in the final testing data set. However, the image patches used in this assessment do not entirely capture segmentation and classification accuracy due to their small size and disconnected nature. Therefore, the performance of GLASS-AI is compared against another human rater on a group of 10 complete WSIs within which a total of 1958 tumors were manually segmented and graded. After assigning a single grade to each tumor segmented based on the highest tumor grade that comprised at least 10% of the tumor's area, GLASS-AI achieved a Micro F1-score of 0.867. Examining the F1-score for each class showed a trend toward higher a higher score with increasing tumor grade (
Overall, GLASS-AI successfully recognized tumors within 1932 of the 1958 manually segmented regions as depicted in
To better compare tumor grading between GLASS-AI and the human rater, the manually annotated regions were used in combination with GLASS-AI's grading to assign the tumor grade. GLASS-AI and the human rater assigned the same grade to 1380 (85.7%) of the annotated tumors resulting in a Cohen's kappa of 0.760. It was observed that the grading agreement was high across all 4 tumor grades (
Referring now to
The initial test of the GLASS-AI pipeline was carried out on KrasG12D/+; RosamG/mG (K) and TAp73Δ/Δ; KrasG12D/+ (TK) mouse models (
After analyzing a cohort of 11 K and 13 TK mice with GLASS-AI, it was found that the number of tumors in TK mice was significantly higher than in K mice (
The distribution of individual tumor sizes was examined to determine if the increased tumor burden observed in TK mice compared to K mice was due solely to an increased tumor number. Interestingly, while the median tumor size of TK mice was found to be significantly smaller than K mice as depicted in
It is important to note that the annotations generated by the expert human raters were based on standard criteria for tumor grading, in which a tumor is assigned a single grade based on the highest grade observed that comprises at least 10-20% of the tumor area 9. However, GLASS-AI gave grades to individual pixels within the image before tumor segmentation, producing a mosaic of grades within a single tumor (
By representing each tumor as a stacked bar divided by the proportion of the tumor area made up of each grade of LUAD, the overall distribution of intratumor heterogeneity in the LUAD mouse models can be visualized (
While informative, these visual representations of tumor heterogeneity can provide only a qualitative estimation of heterogeneity in the mouse models. To overcome this shortcoming, the Shannon Diversity Index (SDI) was employed as a quantitative estimate of intratumor heterogeneity. SDI estimates the uncertainty in predicting the grade of a given square micron in a tumor given by H′=−Σi=14 pi In pi, where p is the proportion of the i-th grade from Grade 1 to Grade 4. After estimating the mean SDI from each tumor in a mouse, it was found that the TK mice had a higher overall SDI than K mice (
Referring now to
Referring now to
Referring now to
Referring now to
Referring now to
Referring now to
Referring now to
Dysfunctional Mek/Erk Signaling is Associated with Grade 4 Regions in High-Grade Tumors.
To investigate how the loss of TAp73 contributes to tumor progression and to correlate tumor grade with molecular indicators of progression, immunohistochemistry (IHC) was performed for phospho-Mek (p-Mek) and phospho-Mapk/Erk (p-Erk) on tissue sections from K and TK mouse lungs with adjacent H&E sections graded by GLASS-AI. Global and local registration on the IHC WSI were performed to ensure the highest accuracy between the H&E and IHC sections as depicted in
It was found that p-Mek and p-Erk were present in a subset of tumors in both K and TK mice but were largely absent in adjacent normal tissue, in agreement with previous reports on the Kras and Trp53 mutant mouse models18. To facilitate comparisons to these studies, a KrasLSL-G12D/+; Trp53R172H/+ (KP) mouse model depicted in
In all three mouse models, both p-Mek and p-Erk positivity increased with tumor grade, and nearly 100% of Grade 4 tumors were positively stained for both markers (
The high-resolution tumor grading produced by GLASS-AI facilitated examination of the distribution of p-Mek and p-Erk staining within regions of different grades in a single tumor. Tumors that displayed an uneven distribution of positively stained cells were determined using a likelihood-ratio G-test G=2Σi=14 Oi In
where regions of each grade containing a greater number of positive cells than expected based on the proportion of the tumor area occupied by that grade will produce a positive value while regions with lower than expected proportion of positive cells will produce a negative value. The proportion of tumors with significantly unequal distribution of either p-Mek or p-Erk was very small in Grade 3 or lower tumors. However, most Grade 4 tumors of all three mouse models displayed significantly disproportionate staining for both markers (
Based on these observations, it can be hypothesized that the enrichment of p-Mek and p-Erk staining in the high-grade LUAD of the mouse models should occur in the highest grade regions of these tumors. By examining the likelihood ratios of the individual grade regions in each tumor, it was found that K mice displayed the most robust enrichment of p-Mek staining in Grade 3 areas, but such a clear trend was not present in either the TK or KP tumors (
Referring now to
Applying machine learning models to digitized WSIs will likely revolutionize how these data are analyzed. Not only can computer vision assist clinicians by providing rapid screening of images, but the higher resolution analysis performed by machine learning models can uncover features that go unnoticed or unreported by human raters. Preclinical studies will also benefit from employing these machine learning models in analysis pipelines by facilitating rapid, reproducible analysis. In accordance with the present disclosure, a purpose-built neural network for grading lung adenocarcinomas in mouse models that provides an unparalleled identification and analysis of tumor grade heterogeneity is provided.
Tumor heterogeneity has been implicated in the progression of many cancer types, including non-small cell lung cancers23,24. Increased intratumor heterogeneity has been linked to decreased overall survival23,25,26, poor response to therapy27, and even increased metastasis28 This heterogeneity is presumed to arise from the clonal evolution of tumor cells within a neoplasm29,30. Typically, tumor heterogeneity is estimated using bulk molecular analyses, such as RNAseq or copy number variation. Previous studies have utilized bulk sample analyses correlated with histomorphological features to predict spatial heterogeneity of molecular markers31,32. However, recent studies have begun using spatially sensitive techniques29 or multi-region sampling33. Combining these approaches with high-resolution analysis from machine learning pipelines like GLASS-AI may provide an unprecedented understanding of cancer development, progression to metastasis, and treatment response through information derived from spatial genomics, transcriptomics, and proteomics correlated with tumor phenotype.
The recent development of commercially available spatial transcriptomics platforms is a promising step forward in correlating molecular and histological analyses. Some groups have already begun developing machine learning applications utilizing these technologies34. However, these platforms are currently focused on fresh-frozen specimens rather than the FFPE samples typically used for histological analyses in both mouse and human LUAD. Further improvement of these technologies to enable the use of FFPE archival tissues would significantly enhance our understanding of the molecular drivers of tumor progression and heterogeneity and allow the prediction of molecular features from routine histological preparations. This ability to accurately predict molecular markers from simple histology images could be used to flag specimens for further molecular characterization and even provide increased diagnostic and therapeutic choices to clinics without regular access to these molecular techniques. Given the rapid pace of advancement in this field, it seems likely that the first clinical applications of this technique will be realized in the near future.
KrasLSL-G12D/+; RosamTmG/mTmG (K), Tap73fltd/fltd; KrasLSL-G12D/+ (TK), and KrasLSL-G12D/+; RosamTmG/mTmG; Trp53LSL-R172H/+ (KP), and KrasLSL-G12D/+; Trp53fl/fl mice were generated on a C57BL/6 background. Between 8-10 weeks of age, mice were intratracheally instilled with 7.5×107 PFU of adenovirus containing Cre recombinase under the control of a CMV promoter, as previously described19. Mice were euthanized 30 weeks after infection, and lungs were collected, fixed overnight in formalin, and embedded in paraffin for further processing. All procedures were approved by the Institutional Animal Care and Use Committee (IACUC) at the University of South Florida.
Formalin-fixed paraffin-embedded (FFPE) lung tissue blocks were sectioned at 4-micron thickness by the Tissue Core at Moffitt Cancer Center. Hematoxylin and eosin (H&E)-stained sections were prepared by the Tissue Core immediately after sectioning. Immunostaining of mouse lung sections was performed overnight at 4° C. in humidified chambers with antibodies against p-Mek1/2 (Ser221) (Cell Signaling Technology Cat #2338, RRID: AB_490903; 1:200) or p-Mapk (Erk1/2) (Thr202/Tyr204) (Cell Signaling Technology Cat #4370, RRID: AB_2315112; 1:400) in 2.5% normal horse serum. The IHC signal was developed using DAB after conjugation with ImmPRESS HRP Horse anti-rabbit IgG PLUS polymer kit (Vector Laboratories Cat #MP-7801). Nuclei were counterstained by immersing the slides in Gill's hematoxylin for 1 minute (Vector Laboratories Cat #H-3401).
Whole slide images (WSIs) were generated from H&E and immunostained slides using an Aperio ScanScope AT2 Slide Scanner (Leica) at 20× magnification with a resolution of 0.5 microns/pixel. To improve the consistency of the pipeline on H&E slides with various staining intensities, staining was normalized using Non-Linear Spline Mapping35. WSIs of immunostained sections were co-registered to adjacent H&E-stained sections by a combination of global and local co-registration in MATLAB. The global co-registration was achieved by first applying a rigid co-registration to the whole slide of IHC and aligned to the H&E slide. After the initial rigid alignment, the co-registration was further refined by applying an affine transformation to the IHC slide to ensure tissues were adequately aligned in both slides. The affine co-registration step was lightly applied using only a few iterations to avoid undesired deformation. Local co-registration was then performed by manually aligning tumor regions identified by the pipeline in the H&E image to tumor regions in the IHC slide. WSIs were then divided into 224×224-pixel patches before analysis by GLASS-AI.
GLASS-AI was written in MATLAB using the Parallel Processing, Deep Learning, Image Processing, and Computer Vision toolboxes. The standalone applications for Windows and Mac were built using the MATLAB Compiler. The network architecture of GLASS-AI was based on ResNet1820; an 18-layer residual network pre-trained on the ImageNet dataset36. An atrous convolution layer and atrous spatial pyramid pooling layer were added after the final convolutional layer to improve context assimilation in the model. The latent features were then processed with transposed convolution and up-scaling before classification. Finally, after classification, a smoothing layer was added to minimize artifacts from image patch edges. An overview of the network architecture and hyperparameters of GLASS-AI are provided herein.
To construct the training dataset, WSIs from 33 mice (KrasG12D/+ n=4, TAp73Δ/Δ; KrasG12D/+ n=15, KrasG12D/+; Trp53Δ/Δ, n=14) were manually annotated by three expert raters who segmented and graded each tumor within 11 of the WSIs each. The annotated WSIs were then divided into 224×224-pixel images and corresponding label patches. Patches were then grouped by the annotated class (Normal alveolar, Normal airway, Grade 1 LUAD, Grade 2 LUAD, Grade 3 LUAD, and Grade 4 LUAD) that was most abundant within each patch, however, all the annotations present within these patches was left intact (i.e., a patch that was predominantly Grade 3 could still contain Normal Alveolar and Grade 4 LUAD annotated pixels). 6,000 patches were selected for each class from the respective patch group and split 60/20/20 for training, validation, and testing of the machine learning model after ensuring that patches from an individual slide were only present within a single split. Because each image patch could contain varying amounts of each target class, the area of each of the six target classes in each library was balanced via data augmentation by shifting, skewing, and/or rotating patches in which the underrepresented class was the most abundant class present. Using MATLAB Deep Learning Toolbox and 2 NVIDIA P2000 GPUs, the model was set to train for 20 epochs using adaptive moment estimation on 128-patch minibatches with an initial learning rate of 0.01.
Data were analyzed using the statistical tests indicated in the figure legends using GraphPad Prism 9 software. p<0.05 was considered statistically significant.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.
This application claims priority to and incorporates by reference herein U.S. Patent Application Ser. No. 63/282,214, entitled ARTIFICIAL INTELLIGENCE-BASED METHODS FOR GRADING, SEGMENTING, AND/OR ANALYZING LUNG ADENOCARCINOMA PATHOLOGY SLIDES, the contents of which is hereby incorporated by this reference in its entirety as if fully set forth herein.
This invention was made with government support under Grant nos. R35CA197452 and T32CA233399-03 awarded by the National Institutes of Health/National Cancer Institute. The government has certain rights in the invention.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2022/050865 | 11/23/2022 | WO |
Number | Date | Country | |
---|---|---|---|
63282214 | Nov 2021 | US |