The invention relates to the application of methods of image processing, computer vision, machine learning and deep learning to create new algorithms for the detection of specific types of cells in Whole Slide Images (WSI) obtained by scanning the biopsies with a digital scanner.
In pharma research and medical diagnosis, the detection and quantification of specific types of cells, e.g. lymphocytes, is important. The usual practice is that the pathologist views the slide under a microscope and roughly estimates the number and density of the cells of interest. The availability of high resolution digital scanners for pathology that produce digitized WSI allows the development of state of the art Computer Vision and Deep Learning methods for cell detection and quantification. Different applications require the detection of different cells. Each new cell detection algorithm usually requires two major efforts: the first is the annotation of the cells of interest by an expert pathologist, and the second is the development of specific computer vision and deep learning algorithms tailor made for the detection of the specific cells of interest. Both efforts require dedicated expert teams and resources.
The invention provides a tool to be used by pathologists that allows them to create new algorithms for specific cell detection. The invention also provides a tool for rapid annotation of image patches taken from WSI, as well as a visualization tool for the cells detected.
The ability to automatically detect certain types of cells in pathology images and to localize them is of significant interest to a wide range of pharma research and clinical practices. Cell detection is a common task that is routinely performed by pathologists, who examine slides under a microscope and provide an estimation of the quantity and density (or other attributes) of the cells based on their empirical assessments. These assessments are generally time consuming and tedious and are prone to fatigue induced errors.
For example, the presence of tumor-infiltrating lymphocytes (TILs), have become a central research topic in oncology and pathology. Immunohistochemical staining (IHC) is a technique that allows to target specific cell types, including lymphocytes, by attaching a colored label to a specific antigen in (subcompartment of) a cell. In this way, immune cells can be distinguished from other type of cells.
Accurate detection and assessment of presence of lymphocytes in cancer could potentially allow for the design of new biomarkers that can help monitor the rapid progression of a tumor. Moreover, automated tools to quantify the immune cells density and their localization in the proximity of tumor cells might help to predict the presence and development of metastases and overall survival of cancer patients. In addition, it allows personalized treatments that can significantly benefit the patients.
Given the very large amount of lymphocytes (≈100,000) in a single cancer tissue specimen, manual assessment at whole-slide image level is a very tedious, time-consuming, and therefore unfeasible task. Moreover, manual assessment suffers from intra- and inter-observer variability. Consequently, a method for automatic detection and quantification of immune cells is of great research and clinical interest.
Moreover, once a cell detection capability is available various quantitative attributes such as cellular morphology, size, shape and texture can be calculated.
The task of cell detection is a very popular topic in digital pathology. Computer-aided methods provide faster image analysis and can significantly improve the objectivity and reproducibility of cell detection. Moreover, the basic science researchers and clinician scientists can be released from boring and repeated routine efforts. Several approaches have been proposed for automatic cell detection on different types of digitized microscopical specimens and for various types of stained specimens. In many cases, detection algorithms are based on morphological operations, region growing, analysis of hand-crafted features and image classifications.
Cell detection and localization constitute several challenges. First, target cells are surrounded by clutters represented by complex histological structures like capillaries, adipocytes, collagen etc. In many cases, the size of the target cell is small, and consequently, it can be difficult to distinguish from the aforementioned clutter. Second, the target cells can appear very sparsely (only in tens), moderately densely (in tens of hundreds) or highly densely (in thousands) in a typical WSI. Additionally, significant variations in the appearance among the targets can also be seen. Moreover, due to the enormous variability (cell types, stains and different microscopes) and data complexity (cell overlapping, inhomogeneous intensities, background clutters and image artifacts), robust and accurate cell detection is usually a difficult problem that requires a dedicated R&D effort of experienced algorithms developers.
Cell detection methods have evolved from employing hand-crafted features to deep learning-based techniques. Traditional computer vision based cell detection systems adopt classical image processing techniques, such as intensity thresholding, feature detection, morphological filtering, region accumulation, and deformable model fitting. Deep neural networks recently have been applied to a variety of computer vision problems, and have achieved better performance on several benchmark vision datasets. The most compelling advantage of deep learning is that it has evolved from fixed feature design strategies towards automated learning of problem-specific features directly from the training data. By providing massive amount of training images and problem-specific labels, users do not have to go into the elaborate procedure for the extraction of features. Instead, a deep neural network (DNN) is subsequently optimized using a mini-batch gradient descent method over the training data, so that the DNN allows autonomic learning of implicit relationships within the data.
In order to develop deep learning neural network based cell detection algorithms it is required to first annotate thousands of cells within WSI and then develop a specific cell detection deep learning algorithm. Then, there should be a dedicated R&D effort for the development of the neural network for the detection of the specific cells. This is a major effort that is not readily available for every pathology lab.
This invention is not another cell detection algorithm. This invention provides a “do it yourself” tool for pathologists in order to create new cell detection algorithms suited to the problem at hand without a need to annotate thousands of cells and without the need for a dedicated research and development effort.
In this invention the use of the following components represent a novel contribution to the state of the art and constitute a Cell Detection Studio Framework:
The present invention is a method for automated detection of a cells categories in histological specimens, comprising: providing a specimen-stained slide; obtaining a scanned image of the slide with a digital scanner; detecting all cells in the slide; generating image patches that contain various types of cells; annotating those image patches according to categories; creating a cell classifier for the cell categories annotated; apply the generated algorithm on new whole slide images; detect centers and contours of cell categories of interest; generate a report on various categories of cells attributes.
The subject matter regarded as the invention is particularly pointed out and distinctly claimed in the concluding portion of the specification. The invention, however, both as to organization and method of operation, together with objects, features, and advantages thereof, may best be understood by reference to the following detailed description when read with the accompanying drawings in which:
In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the invention. However, it will be understood by those skilled in the art that the present invention may be practiced without these specific details. In other instances, well-known methods, procedures, and components have not been described in detail so as not to obscure the present invention.
The resultant image is then submitted to a generic cell detection algorithm A2 that aims to detect all cells in the slide, of any category. A detailed description of the generic cell detector is given in the text for
The result of A2 is a list of all the centers and contours of the cells present in the slide. The image patches extraction module A3 aims to extract crops that surround every cell selected. The size of the image patches can be set as a parameter and its default value is 32. The image patches created are then submitted to interactive image annotation using a GUI (graphical user interface) application A4. Each time a single image crop is presented to the annotator, and using a keyboard press or mouse click, or touch screen tap, the annotator chooses one of the possible categories, each containing a specific type of cell or background. A classification CNN is trained on all the available annotated image patches to create a cell categories classifier A5. Online learning is used to update the cell classification neural network as more annotations become available as described in the text for
The input to the generic cell detection block is patches from whole slide images. The output of the generic cell detection block is contours and centers of all detected cells in each patch. The generic cell detection can be any method for nuclei detection. One method for doing this is using a neural network for semantic segmentation, e.g Unet. In this case Unet is trained on a dataset of cells and outputs the body and the contour segmentation for each cell. The dataset is consisted of annotated cells where each cell has a marked polygon around its border.
In case of a segmentation network based on Unet, the architecture is consisted of an Encoder and a Decoder. The Encoder has 5 convolutional blocks, each with a kernel of 3×3 and a stride of 2. The Decoder has 5 convolutional blocks, each with a kernel of 3×3, and an up sampling ratio of 2. Each decoder block performs concatenation with features from a layer from the Encoder. The last layer in the decoder has an output size of 3: One for the background class, one for the cell body class, and one for the cell border class.
The input is digital slides that contain cells of categories that are of interest to the user C1. First, the generic cell detection algorithm whose generation is described in the text for
The CNN architecture relies on the standard VGG16 architecture (other variants could be equivalently used), with 7 convolutional blocks (each containing a Convolutional layer, followed by a Rectified Linear Unit and a Batch Normalization unit), and then 3 fully connected layer, the first two of them followed by ReLU layers, and dropout layers. The last layer in the network outputs a score for the presence of the cell category in the input image patch (and equivalently a score for the lack of presence of a cell category in the input image patch).
Unbalanced data is a common situation where the number of instances of one category is significantly smaller than the number of instances of another category. In order to obtain a robust network there should be enough examples of each category. We therefore add data balancing methodology for effective active learning F7. The data balancing methods can be one of the following: we rank the cells inversely proportional to their existence. We duplicate image patches that belong to the least frequent category. We can also add data balancing using weighting. The weight is inversely proportional to the proportion of least frequent category. Another approach is to add data balancing using the following weighting: Weight=E*A−B*(N-E)*Pminority where: *E=Entropy(class proportion)
Once we have enough examples of each category than we can move to the usual approach of active learning. The image patches are then ranked according to active learning methodology F8 as was described in the text for