Quantitative information about objects within an image can provide information critical to identification, decision making and classification. For example, characterization of single or multiple biological cells from microscope images can help determine therapeutic strategies for patients, or aid with the identification of a person in a large crowd of people.
There are a variety of segmentation methods available that can be used to isolate and analyze objects of an image. However, these methods can be time-consuming, as they require significant user inputs and adjustments of image processing parameters, and biased, as they are often prone to both user error and variable interpretation of object boundaries.
Additionally, while prior image segmentation techniques allow for segmentation of components in a single image, they do not allow for automated processing of multiple images. Many raw images require pre-processing and adjustment before segmentation can effectively be used to locate objects of interest in the field of view. Even when the images seem to be very similar, the properties of objects in one image may dictate the need for very different processing and parameter values than those required by another image.
Once the image has been segmented, an additional problem is that of determining the properties of objects that have been segmented. While the human eye quickly recognizes patterns across images, automated means of identifying and classifying objects often are unable to capture complex patterns because of their reliance on a small set of metrics, metrics not optimized for a particular application, or metrics that are considered without regard to an object's local environment and communication with other objects.
Furthermore, there is currently no optimized and automatic way to search for objects of interest within images, as commercial image searches so far have focused on whole image searches.
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee. A more complete understanding of this disclosure may be acquired by referring to the following description taken in combination with the accompanying figures.
The inventors have identified a need for a system which would allow users to automatically segment and classify objects in one or more images, determine object properties, identify how objects are connected to each other, and match object features and morphology with object network connectivity and object communication. Additionally, the inventors have identified a need for an image search system which allows users to search for specific objects and object features within an image, rather than requiring them to search for an entire image.
While methods, apparatuses, and computer-readable media are described herein by way of example, those skilled in the art recognize that methods, apparatuses, and computer-readable media for automatic image segmentation, classification, and analysis are not limited to the embodiments or drawings described. It should be understood that the drawings and description are not intended to be limited to the particular form disclosed. Rather, the intention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the disclosure. Any headings used herein are for organizational purposes only and are not meant to limit the scope of the description or the claims. As used herein, the word “may” is used in a permissive sense (i.e., meaning having the potential to) rather than the mandatory sense (i.e., meaning must). Similarly, the words “include,” “including,” and “includes” mean including, but not limited to.
The disclosed system addresses the unmet need for an automated, optimized method to identify, characterize and match objects within images. Methods, apparatuses and computer-readable media are described for automated and adaptive image segmentation into objects, automated determination of object properties and features, automated determination connectivity between objects, mapping of object morphology and characteristics with object connectivity and communication, and automated searching and visual presentation of objects within images. The system disclosed herein allows for classifying and matching individual objects within an image in a manner that can be specified as independent of object orientation and size and identifying community structures in an image. Using the disclosed system, objects within single or multiple images can be compared and ranked for similarity in shape, features and connectivity.
Furthermore, the methods disclosed herein can be utilized for biological applications, such as classifying responses of human vascular cells to stimuli, in order to improve regenerative medicine strategies.
Of course, the steps shown in
Although many of the examples used throughout this specification refer to cells and other biological structures, the methods, apparatuses, and computer-readable media described herein can be utilized in diverse settings and for a variety of different applications. For example, the images can be taken from a video of people in some settings, such as a shopping mall, and the segmentation can be used to identify individual persons. In this case, the persons can be the image objects and the analysis can focus on the dynamics of person-to-person interaction within the particular setting. Within the biological arena, the images can correspond to an image of a biopsy and the system can be used to produce the identification and morphological metric sets for similar cancerous or benign cells and a measure of how they are connected. Other applications include predictions of the movement of vehicles, animals, or people over time. For example, the image objects can be cars on a highway, and the system can be used to model and analyze car-to-car connections and traffic patterns. Another application is that of predicting and tracking the presence of animals in particular region, such as a forested region.
Referring to
Image segmentation can be performed using the watershed method for simultaneous segmentation of all of the objects in an image.
As flooding continues, the outline of the rising waterline will follow the rising contours of the map. During this process, it may be possible for separate, growing bodies of water to meet. If the two bodies originated from different original element markers, this junction region will define a boundary between unique objects in the image. On the other hand, the areas will unite to form a single body if they do not both originate from watershed starting points, or markers.
The flooding proceeds until all regions of the topography have been covered and the basins have been flooded to their edges. Finally, these edges, which can either be cell or image boundaries, are used define and isolate individual components of the image. The edges are shown in part f of
A comparison of the watershed image segmentation technique and manual object identification in an image is shown using the example of cells in
While the watershed method lends itself nicely to simultaneous segmentation of all of the components in a single image, it is difficult to adapt for automated processing of multiple images. Many raw images require pre-processing and adjustment before the algorithm can effectively locate objects of interest in the field of view. Even when the images seem to be very similar, the properties of objects in one image may dictate the need for very different processing and parameter values than those required by another image. Even when staining and imaging conditions are tightly controlled, the properties of elements in one image may dictate the need for very different parameter values than those required by another image. For a further discussion of staining and imaging techniques for cell cultures, refer to “Predicting Endothelial Cell Phenotypes in Angiogenesis” authored by Ryan D T, Hu J, Long B L, and Qutub A A and published in Proceedings of the ASME 2013 2nd Global Congress on NanoEngineering for Medicine and Biology (NEMB2013), Feb. 4-6, 2013, Boston, Mass., USA, the contents of which are herein incorporated by reference in their entirety.
The present system provides an automated version of the watershed algorithm designed to execute image processing and perform segmentation for groups of images, eliminating the need for user input or adjustment for each image. The output of the watershed segmentation algorithm takes the form of masks, or binary representations of the area of the individual image components. These masks have the potential to be either too large or too small, and over represent or under represent the actual areas of the individual objects, respectively. The size, and accuracy, or these masks largely depends on the grayscale threshold value used to create a binary representation of the original image that aids in watershed implementation. The present system utilizes an adaptive threshold evaluation process that selects the optimal threshold value for segmentation by comparing a baseline binary representation of the original image and its objects to the areas of the generated component masks. The system iterates through the segmentation process by decreasing or increasing the grayscale threshold value until an acceptable area ratio between the baseline and the masks is reached, at which time the resulting masks are saved and the process moves on to another image in the queue. By automatically selecting the optimal threshold value, the process circumvents the need for manual input with each image that previously prevented automated processing of large image sets.
The system also incorporates improved methods for fine-tuning the generated masks that are not possible with traditional, single executions of the process. For instance, in many images, it can be difficult to discern ownership of borders between adjacent objects. For example, in biological cell images, cytoskeletal components can appear indistinguishable, bound via junctions. Alternatively, in images of humans, contact (i.e. hugging) can create similar problems when attempting to distinguish which features (i.e. clothing, limbs, etc.) belong to which individual.
In order to improve the potential for accurate segmentation, two watershed segmentation executions can be used in sequence. The first iteration can create masks of an underlying component feature that can serve as a baseline representation of the overall shape or area, but which typically does not extend to the true edges of the object. For example, in biological cell images, microtubules (a particular cytoskeletal element) do not always extend to the periphery of the cell, and are easily distinguishable for association with a particular cell. The resulting masks from this initial segmentation subsequently serve as the markers for the second iteration, which employs the final image. Since the initial masks will typically take up much of the actual cell area, the masks generated with final iteration only extend the borders slightly and refine them to match the actual contours of the image objects.
Additionally, the system includes the ability to output images to visualize the final, optimal masks for user review and reference. The program can also actively display during execution the effects of the grayscale threshold value adjustments on image pre-processing steps, as well as the generated mask areas. The user can also choose to create a movie to visualize in real-time the adaptive threshold value adjustments and their effects on mask generation and fine-tuning.
The adaptive, automated watershed segmentation system disclosed herein provides a method for segmenting images and identifying and isolating its individual components. It can be used with cells, such as human umbilical vein endothelial cells—HUVECs, but it amenable to other cell types, as well as co-cultures and three-dimensional assays. The system can also be useful in other types of image analysis, such as in the evaluation of micro-scale properties of biomaterials (i.e. collagen scaffold fibers), as well as applications requiring isolation of vehicles or human individuals from an image, such as for criminal investigations.
The system can be used to execute image processing and perform segmentation for large groups of images by eliminating the need for user input or adjustment for each image. This goal is accomplished by evaluating the accuracy of segmentation attempts associated with specific image pre-processing and watershed segmentation parameter values (i.e. grayscale threshold value), and adjusting these values accordingly in an attempt to find the optimal conditions required for effective segmentation. This prevents the need for user input and parameter adjustment, as well as biased boundary interpretation and segmentation evaluation, associated with many current segmentation techniques.
As explained earlier, watershed segmentation can involve many pre-processing steps.
The first step can be the pre-processing of original image to prepare for watershed segmentation, which can include one or more of the following steps:
(a) selecting and defining markers of individual image objects,
(b) histogram equalization of an image, such as the original image,
(c) 2-D Gaussian filtering an image, such as the image produced by step (b),
(d) global image thresholding of an image, such as the image produced in step (c) with grayscale threshold value to create a binary image,
(e) removal of small objects in the binary image produced in step (d),
(f) generation of a finalized template for watershed segmentation by imposing the minimum of the combination the following:
(g) generation of a baseline binary image for area comparison via global thresholding of (c) with grayscale threshold value, which can determined by Otsu's method.
The second step can be the comparison of total mask area in the segmented image to the white area it shares with the baseline image, including one or more of the following steps:
The first and second steps can then be repeated with the masks generated from the first iteration serving as the markers of the individual objects for the next segmentation cycle.
An output file can be generated, such as a TIFF file, with each layer representing a binary mask of an individual object in the image. Visualizations of segmentation effectiveness, segmentation iterations, and other similar information can also be output. The adaptive image segmentation is described in greater detail in Ryan, previously incorporated by reference.
The user can define an area ratio value (between the baseline representation and the generated masks) that can serve as a threshold for designating acceptable segmentations. While a single ratio value will typically be suitable for all images of a particular type or set, this value can also be adjusted by the user when switching to different image set types. Alternatively, this value can be learned by the system based on previous image databases and image types. By analyzing sample image sets of cell types and determining appropriate area ratio value adjustments for optimal segmentations for these sets, the ratio can be automatically adapted when moving among image types. This adaptation can be a function of properties of the objects (i.e. cells) in the image set that are unique from objects in other image sets.
Referring to
The specific metrics will now be described in greater detail. Note that the descriptions below assume actin and microtubules or vinculin are stained using DAPI, but any other makers or stains can be substituted. These are illustrative but not inclusive metrics.
Exemplary contouring metrics can include:
Exemplary texturing metrics are described below:
Σp2
−Σp+log2(p)
Exemplary polarity metrics are described below:
Exemplary intensity, area, and shape metrics are described below:
Exemplary adhesion site metrics are described below:
Exemplary actin fiber alignment metrics are described below:
Compares both the number of angle peaks and the percent of fibers aligned at each peak to the COI (cell of interest) fiber alignment metrics. The following equation defines how closely the fiber alignment in a patterned cells matches the COI. The lower the value the closer the match. For each original peak α0 in the cell of interest with its associated fraction of pixels ω0 (fractional area under the under for the peak), and all comparison peaks in the patterned cells, αi and their respective fractional weights ωi:
Returning to
The process and system disclosed herein allows for the determination of connectivity and graph-based metrics which are means of measuring communication across objects (e.g., cell-cell communication, person-to-person interactions).
Users can define cutoff distances and/or a minimum number of shared pixels to seed the initial connectivity analysis. Alternatively, these values can be determined intelligently through domain specific analysis. Additionally, although the graphs shown in
Referring back to
The system disclosed herein utilizes imaging, image analysis, and clustering to automatically categorize and define distinct cellular phenotypes or states. Users of the method and system disclosed can automatically categorize and define cellular states, or phenotypes, on a large scale and subsequently assign cells to these phenotypes based on their morphological responses to angiogenic stimuli.
Returning again to
Image or Object Search: The system can be used to perform an image search. For example, an image file can dropped into a folder or database, objects in the image can then characterized as described above, and the closest matches to an overall image and individual object matches can be returned by comparing feature sets and network connectivity. Unlike existing image searches, objects within multiple images can be compared and ranked for similarity in shape, features and connectivity. The image search can also be optimized for biological images, such as cells and tissues.
Merging of Objects: To assist in the interpretation of image classification, the system can be used to visualize an “average object” for each type of component in the image. To accomplish this, the system can align each segmented object in the same direction and overlay either all of the objects or a designated number of objects from each group or cluster in a single image, such as shown in
The basic steps used to perform the merge process can be described as follows:
Generating a merged representation of similarly grouped objects allows users to visualize shared physical properties and represent the general appearance of an average object of a determined category. In cellular imaging, this is useful in visualizing common physical properties associated with identified morphological phenotypes, and how these features differ among the different phenotype groups. While generating average values of each metric used to quantify cells for all of the cells within a phenotype group can help represent the “average cell”, generating a visual representation of the average cell helps users better identify similar cells in images and associate them with particular phenotypes. This could be useful in the future in assessing the effectiveness of efforts to reproduce identical features; in cells or other applications such as biomaterials. Any deviations from a desired layout in the “average object” can represent an instance where the optimal solution was not reached.
As discussed earlier, this system can be used to classify responses of human vascular cells to stimuli, in order to improve regenerative medicine strategies. This system and method can also be applied to other areas, for example, to develop biomarkers of leukemia and to assess leukemic cells response to drugs, or to characterize the functional response of human neurons and neural stem cells to different microenvironments.
Users of the systems and methods disclosed herein can provide (such as by uploading or through some user interface) an image (.JPG, .TIF, .PNG) to a folder, GUI element, application, website, mobile app, or database, and the system can then automatically perform the steps described above.
One or more of the above-described techniques can be implemented in or involve one or more computer systems.
With reference to
A computing environment may have additional features. For example, the computing environment 1500 includes storage 1540, one or more input devices 1550, one or more output devices 1560, and one or more communication connections 1590. An interconnection mechanism 1570, such as a bus, controller, or network interconnects the components of the computing environment 1500. Typically, operating system software or firmware (not shown) provides an operating environment for other software executing in the computing environment 1500, and coordinates activities of the components of the computing environment 1500.
The storage 1540 may be removable or non-removable, and includes magnetic disks, magnetic tapes or cassettes, CD-ROMs, CD-RWs, DVDs, or any other medium which can be used to store information and which can be accessed within the computing environment 1500. The storage 1540 may store instructions for the software 1580.
The input device(s) 1550 may be a touch input device such as a keyboard, mouse, pen, trackball, touch screen, or game controller, a voice input device, a scanning device, a digital camera, remote control, or another device that provides input to the computing environment 1500. The output device(s) 1560 may be a display, television, monitor, printer, speaker, or another device that provides output from the computing environment 1500.
The communication connection(s) 1590 enable communication over a communication medium to another computing entity. The communication medium conveys information such as computer-executable instructions, audio or video information, or other data in a modulated data signal. A modulated data signal is a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media include wired or wireless techniques implemented with an electrical, optical, RF, infrared, acoustic, or other carrier.
Implementations can be described in the general context of computer-readable media. Computer-readable media are any available media that can be accessed within a computing environment. By way of example, and not limitation, within the computing environment 1500, computer-readable media include memory 1520, storage 1540, communication media, and combinations of any of the above.
Of course,
Having described and illustrated the principles of our invention with reference to the described embodiment, it will be recognized that the described embodiment can be modified in arrangement and detail without departing from such principles. It should be understood that the programs, processes, or methods described herein are not related or limited to any particular type of computing environment, unless indicated otherwise. Various types of general purpose or specialized computing environments may be used with or perform operations in accordance with the teachings described herein. Elements of the described embodiment shown in software may be implemented in hardware and vice versa.
In view of the many possible embodiments to which the principles of our invention may be applied, we claim as our invention all such embodiments as may come within the scope and spirit of the following claims and equivalents thereto.
This application claims priority to U.S. Provisional Application No. 61/865,642, filed Aug. 14, 2013, the disclosure of which is hereby incorporated in its entirety.
This invention was made with government support under Grant Number CBET-1150645 awarded by the National Science Foundation. The government has certain rights in the invention.
Number | Date | Country | |
---|---|---|---|
61865642 | Aug 2013 | US |