System and method for automation of surgical pathology processes using artificial intelligence

Information

  • Patent Grant
  • 12014830
  • Patent Number
    12,014,830
  • Date Filed
    Monday, April 18, 2022
    2 years ago
  • Date Issued
    Tuesday, June 18, 2024
    2 months ago
  • CPC
    • G16H50/20
    • G16H15/00
    • G16H30/40
    • G16H70/60
  • Field of Search
    • CPC
    • G16H50/20
    • G16H70/60
    • G16H15/00
    • G16H30/40
  • International Classifications
    • G16H50/20
    • G16H15/00
    • G16H30/40
    • G16H70/60
    • Term Extension
      241
Abstract
This invention provides a system for automation of a pathology process, which includes a processor having trained artificial intelligence (AI) modules operating in association therewith, adapted to receive image data from camera images of whole tissue acquired by a camera assembly and whole slide images (WSIs) of inked and segmented tissue samples. A mask produces image results for tissue with holes and free of holes, and a filter provides filtered image results to the AI modules, detecting tumors and macroarchitecture features. A quality assessment process produces quality score outputs for tumors and macroarchitecture features. A report generator provides reports with one or more parameters to a user via an interface. More particularly, the report generator automatically creates a pathology report, having a written description and pictorial diagram relative to the gross images of the tissue integrating the outputs of the AI modules used to analyze the whole slide digital images.
Description
FIELD OF THE INVENTION

This invention relates to pathology systems and methods that employ automated processes, and more particularly to tissue slide analysis.


BACKGROUND OF THE INVENTION

Measurement of tissue size, determination of grossing approach, and tissue inking are all currently performed manually by highly reimbursed specialized histotechnicians, physician assistants, or pathology residents. Analysis of tissue quality is performed by the pathologists but is not reflected in the pathology report and because of its subjective nature results in significant variation.


Following the resection of solid tumors, the tissue is removed and (a) measured, (b) processed, (c) grossed, (d) inked, (e) sectioned, (f) stained, (g) slides read by a pathologist, (h) pathology report is created. By way of background, a typical pathology report 100 is shown in FIG. 1. This exemplary report 100 provides a Diagnosis section 100, the Specimens Submitted 112 (e.g. location on the body), Clinical Information 114, and Specimen Processing details 116 related to (e.g.) quantity and size, tissue description, inking, etc. In a current implementation, each of steps (a)-(h) are typically performed manually, and there are no standardized quality control mechanisms in place outside of the pathologist reading the slides to ensure that high quality tissue is available to the pathologist and automated reagent checks in processors and autostainers (which do not directly assess tissue quality). This process occurs hundreds of thousands of times per year in surgical pathology laboratories across the United States, requiring significant manual input from staff. Whole slide imaging is becoming a more commonly accepted method to digitize the slides allowing subsequent interpretation by the pathologist. Digitization of slides has allowed the field of pathology to start developing and integrating machine learning and artificial intelligence approaches to act as diagnostic aids for the pathologist. The majority of research and efforts currently are focused around clinical diagnosis however in addition to diagnosis there are tissue characteristics that must be taken into account. Additionally, these machine learning and AI approaches are aimed at increasing the efficiency of the pathologist, however digitized efforts have not focused on the upstream process from removed tissue to digitized slide, only on the quality and speed of scanning.



FIG. 1A hence shows, an arrangement 101 that is part of an automated slide scanning and analysis system and method, such as shown in commonly assigned U.S. patent application Ser. No. 16/679,133, entitled SYSTEM AND METHOD FOR ANALYZING CYTOLOGICAL TISSUE PREPARATIONS, filed Nov. 8, 2018, the teachings of which are incorporated by reference. The depicted arrangement 101 provides a user-based microscopic analysis system/device 111, that includes an image acquisition module of known construction, including an imaging sensor S that can convert received light, via magnifying optics O, from a slide or other surface 113 containing a tissue sample or other smear of cellular matter 115, that can be prepared using conventional cytological techniques and substances from a human or animal subject. Note that the sensor S can be part of a generalized “camera” or “camera assembly” that also includes optics O, and which should be taken broadly to mean any mechanism that can convert an imaged scene into image data define (e.g.) in the form of black and white or color pixels of a given array (e.g. N×M) size. The acquired, magnified image data 120 is stored and/or transmitted over a network (e.g. the Internet and/or a WAN/LAN) to a computing system 130 that can be instantiated in one or more servers and/or a cloud environment. The server/cloud system 130 includes one or more processors (that can be local and/or distributed 140, and associated data storage (that can be local and/or distributed) 150. As described below, the data 150 can consist of training data, based upon prior images and associated diagnosis. Other data can also be stored as appropriate to carry out the functions of the system and method. The processor(s) 140 carry out an associated image process 160 that can include the various functional modules described hereinbelow. One exemplary module includes vision tools 162, which can be a variety of edge detectors, color analyzers, blob analyzers and other commercially available machine vision applications used to resolve objects within an image and apply pattern recognition thereto. The process(or) 160 further includes a neural net/deep learning-based training process(or) 164 that carries out various training tasks and steps as described below, including the identification and classification of expert-based image data on various cell conditions in an overall library of cell images. Additionally, the process(or) 160 includes a neural net/deep learning-based runtime process(or) 166 that applies the training data and associated classifiers to the acquired image data of a slide 112 to determine associated diagnostic data. A user interface device 170, in the form of a computer, having a screen/touchscreen 172, keyboard 174, mouse 176, etc. is shown linked to the server/cloud 130 via an appropriate network, such as the Internet, LAN or WAN. This device 170 can be any acceptable data handling appliance, such as a smartphone, tablet, laptop, PC or server, and is capable of manipulating (e.g.) web page information that can be used to communicate with the cloud/server and associated processes/modules. This device 170 can be one of many, used by clinicians and other interested parties, who desire to obtain diagnostic data from the system and method. The computing/interface device 170 can also facilitate upload of local image data 180 of slides via a link 182 with the microscope/acquisition device 111.


It is generally desirable to use machine learning and artificial intelligence to automate the routine but very necessary parts of the surgical pathology report while significantly improving efficiency of the surgical pathology lab thereby reducing turnaround time and decreasing cost while increasing quality.


Tissue Size

Recording tissue size is a critical piece of information in the pathology report. Tissue size dictates how the piece of tissue is grossed in terms of the number of tissue blocks and cassettes that are needed to adequately assess the tissue. The number of cassettes and subsequent slides per cassette influence the amount of total tissue that is analyzed by the pathologist and the amount of time that is required to produce and read the slides. This is a balance. Currently the amount of tissue analyzed (compared to the total volume of tissue removed) is not well reflected in the pathology report. If insufficient tissue is analyzed serious adverse consequences can occur, for instance, a false negative margin resulting in tumor recurrence in the future. Therefore maximizing the amount of tissue that can be placed into a cassette/onto a slide in an efficient manner decreases the number of required cassettes while increasing the proportion of tissue analyzed. This results in an overall increase in efficiency, and a more regimented, reproducible process. Note that in the report 100 of FIG. 1 there are a total of 10 cassettes utilized in the grossing of the tissue resulting (see box 117 in FIG. 1), can be decreased by knowing the optimal grossing and inking scheme which are dependent on accurately calculating the size of the piece of tissue removed. Additionally, accurate calculation of the piece of tissue removed will aid in quantifying the percentage of margin analyzed by the pathologist and develop a body of quantitative body of data that can inform more advanced prognostication.


1. Tissue Inking

Tissue ink is used for at least purposes, namely (a) to orient the tissue, and (b) to define whether all of the tissue is contained within the tissue section, i.e. there is not any tissue missing and ink spans the complete border of the tissue. Orientation of the tissue is required when evaluating specimen margins either in real-time (from the operating room (OR), or in Mohs Surgery), or post-operatively when tissue is being analyzed in the event that there is tumor at the margin which would require additional treatment. See the diagram 200 of an exemplary procedure in FIG. 2 in which the tissue sample 210 is divided into quarters 212, and such is diagrammed/mapped 214 by the practitioner. An exemplary, detailed diagram 220 with notations regarding the excision site, and associated features is also shown. Tissue inking is also used to map the tumor, so as to allow planning of additional removal. Currently, there is no standardization across institutions, or mandate by regulatory agencies (CAP, AJCC) regarding the grossing and inking process, and the ink colors are interpreted by the pathologist (typically according to institutional norms). Notably, current tissue mapping and ink interpretation is performed manually, which takes time and creates the possibility for error, particularly in situations when the individual inking the tissue and pathologists reading the slides are different (which is virtually universally the case in modern, sub-specialized pathology labs).


2. Slide Quality

The ability to make an accurate histologic diagnosis or fully identify tissue margins for the presence or absence of a tumor relies on the high quality tissue sections (a very commonly overlooked factor, even for experienced pathologists). Holes or defects (e.g. hole/nick 210 in FIG. 2) in the tissue can represent areas where tumor has fallen out during tissue processing and must be recognized and taken into account by the pathologists. Production of high quality tissue sections is not trivial and can vary based on the type of tissue processing (frozen versus permanent section). Currently, slides are read by pathologists however there is no standardized method to report tissue section quality, beyond reporting on processing issues such as understaining and blade chatter. This is particularly relevant when working with fresh frozen tissue as the ability to obtain complete tissue sections can be challenging based on the nature of the tissue, for example, tissue containing significant fat components can be more challenging to cut/excise without (free of) tearing. However, the ability to use fresh frozen tissue for analysis tasks allows for real-time diagnosis or margin analysis allowing for improved patient care by decreasing the need for additional procedures (the so-called intraoperative frozen section). This also has significant cost benefits as it eliminates the need for tissue fixation, transport, and processing, all of which involve significant labor and include the use of toxic chemicals which must be both handled in special areas and then disposed of in a safe manner. In addition, to quality, quantifying the amount of tissue analyzed may provide important information in instances where tumor recurrence occurs or there are downstream pathologic questions.



FIG. 3 shows inked tissue samples, by way of example, in which holes 310 are outlined in (e.g.) brown. Other features, such as tumor features 320 are outlined as shown, and described further below. Entry of information pertaining to tissue size and appearance, grossing, and inking are all manually inputted into the system by highly reimbursed specialists including histotechnicians and physician assistants. The repetitive nature of the surgical pathology lab make it desirable in an optimal setting to automate processes through the use of artificial intelligence and indeed very simple improvements such as keyboard shortcuts and voice dictation are very well received and have been shown to improve efficiency and reduce errors.


SUMMARY OF THE INVENTION

This invention overcomes disadvantages of the prior art by providing a system and method that effectively automates the early/preliminary and vital part of the overall pathology procedure that is typically repetitive and labor intensive and fundamentally determines which tissue is actually provided to diagnostic slides used by downstream pathologists and other practitioners in treatment of a condition, such as but not limited to skin cancer. These tools, including ink recognition, can provide additional information to the pathologist when analyzing digital pathology images. Significantly, this system and method aids in supporting a vital practice guideline, which is to avoid a diagnosis without sufficient information. This system organizes automated input and output data as above to generate a pathology report that can be efficiently and clearly interpreted.


In an illustrative embodiment, a system and method for automation of a pathology process is provided, which includes a processor having trained artificial intelligence (AI) modules operating in association therewith, adapted to receive image data from camera images of whole tissue acquired by an camera assembly and whole slide images (WSIs) of inked and segmented tissue samples. A mask produces image results for tissue with holes and free of holes, and a filter provides filtered image results to the AI modules, which thereby detects tumors and macroarchitecture features. A quality assessment process produces quality score outputs for tumors and macroarchitecture features. A report generator then provides reports with one or more parameters to a user via an interface. More particularly, the report generator automatically creates a pathology report consisting of a written description and pictorial diagram relative the images of the tissue. Illustratively, a tissue determination process can determine a size and a description of tissue automatically based upon characteristics in the images acquired by the camera assembly. Additionally, a tissue grossing and inking process can automatically generate a grossing and inking scheme based upon a user input of desired percentage of tissue margins analyzed and the tissue size. The AI modules can include a tumor inflammation convolutional neural network (CNN), a macroarchitecture hole CNN, a tumor detection graphical neural network (GNN) and a macroarchitecture detection GNN. Additionally, the filters can include at least one of a Sobel filter and a gradient-based filter. An ink detection and orientation process can operate on the filtered images of the tissue samples and delivers results thereof to the report generator. The report generator can receive results from a quality assessment process acting on results from the a tumor detection GNN and a macroarchitecture detection GNN. Illustratively, the macroarchitecture can be defined by at least one of holes, fat, edges, dermis and epidermis information. A visual generation and 3D stitching process can provide results to the report generator, and can receive information from a cell nuclei detection process. Outputs of the aforementioned CNN/GNNs are thereby used to automatically generate a pathology report.





BRIEF DESCRIPTION OF THE DRAWINGS

The invention description below refers to the accompanying drawings, of which:



FIG. 1 is a diagram showing a typical surgical pathology report of interest in association with the system and method according to embodiments herein;



FIG. 1A is an exemplary, generalized computing arrangement including an artificial intelligence process and associated data-handling processes/modules for use with the system and method herein;



FIG. 2 is a diagram showing an exemplary of tissue grossing, inking, mounting, and mapping process;



FIG. 3 is diagram showing a series of images of exemplary tissue sections with holes outlined thereon;



FIG. 4 is a flow diagram showing a generalized procedure for automation of tissue size, appearance description and grossing;



FIG. 5 is a diagram showing an overview of inking scheme and associated procedures;



FIG. 6 is a diagram showing exemplary images and results associated with automated ink detection system associated with FIG. 5;



FIG. 7 is a diagram showing exemplary images associated with results of an artifact finding process and connection of holes to adjacent subregions using TDA;



FIG. 8 is a diagram showing WSI annotated images to define appropriate tissue layers or components and areas missing;



FIG. 9 is an exemplary bar graph showing slide quality score based upon the processes of the system and method;



FIG. 9A is an exemplary printout of the numerical scores according to the graph of FIG. 9; and



FIG. 10 (and FIG. 10 (Continued)) is a flow diagram showing the operation of quality/ink assessments and automated report generation in a digital pathology ecosystem according to the system and method.





DETAILED DESCRIPTION

With further reference to the generalized computing environment described in FIG. 1A above, a system and method for automation of surgical processes using artificial intelligence and/or deep learning is now described.


1. Tissue Size

With reference to the flow diagram 400 of FIG. 4, images of tissue 410 are captured on a defined grid 420, and photographed from a standard distance. This is used to (a) calculate tissue size (length×width), (b) volume added in 3-D if necessary via trapezoid, and (c) B-spline fitting between serial sections of detected tissue. Additionally, an image recognition feature is used to define the gross appearance of the tissue—for example a “tan ellipse with central biopsy site”. A schematic 430 of the tissue sample is shown and individual features 440 are broken out based upon inking. The tissue size can be used to define a grossing and inking scheme that will subsequently as defined below inform the exact piece of tissue on the slide. Note that normally this specimen would be grossed and placed into e.g. five (5) different cassettes. In this example the case can be placed entirely into less cassettes. As the size of the tissue increases this can be significant important to maximize efficiency in the surgical pathology laboratory and reduce the chance of error. The predefined inking system will be interpreted histologically as per the tissue ink detection described below. This technology can be applied to any solid tissue.


2. Tissue Orientation via Ink Detection System

The system and method includes a process that determines the ink color scheme for a piece of tissue 510 that has been surgically removed. The system and method thereby defines the grossing and inking scheme (procedure 200) based on the size, shape of the tissue, and % of margin desired to be analyzed. The tissue is then inked 512 accordingly (See the process 500 of FIG. 5). Slides are created and scanned resulting in digital whole slide images (WSI) 520 (See also FIG. 6). The system and method further provides an automated digital ink identification system that selectively identifies the location and color of ink on the edges 522 of a piece of tissue. The ink detection system is based on color thresholding in the HSV space and a series of morphometric operations—for example, Sobel and gradient-based filters can operate (arrow 540) to find the tissue edge 522; and a series of binary opening, closing and nearest-neighbor labeling operations, followed by removal of small connected components) that extract ink only on the edge of the tissue. It is notable that the process is capable of automatically estimating the orientation, placement and depth of the tissue section based on series of spatial statistics (See box 610 showing the results of the system and method's automated ink detection system) that are calculated on the locations of ink colors including but not limited to: red, yellow, green, blue, black, orange and purple inks (512) for each tissue section in accordance with best practices.


3. Use of Convolutional/Graph Neural Networks to Assess Slide Quality

Slide quality is assessed by the system and method by identifying tissue artifacts, tears and holes which may cause the surgical pathologist to unknowingly miss diagnostically important regions of tissue, weighted by both their prognostic significance and contextualized by information presented in adjacent sections (3D). First, tissue is detected using a combination of advanced filtration techniques/morphometric operations, which also serves to identify and flag potential holes and tears in the tissue. Since these areas of low intensity imaging can be confused with regions of fat, wispy dermis, edema, cystic cavities or lumina, the computing procedure separately applies a set of deep learning algorithms to distinguish these holes from other possible conflating regions by informing prediction from contextual cues in the surrounding tissue macroarchitecture.


With reference to the color imagery diagram 700 of FIG. 7, the system and method applies two separate Convolutional Neural Network/Graph Neural Network (CNN/GNN) models to: (a) label fat and holes 710, dermis and epidermis 720, and tumor regions and (b) predict the presence of malignant tumor and differentiate it from benign and inflammatory tissue. Expert annotations from whole slide images (WSIs) contribute development/training of the CNN/GNN models used by the system and method. Reference is made to the images 800 of FIG. 8, in which WSI annotation 810, 812 and 814 (each defining a differing color) is used to define appropriate tissue layers or components and areas missing.


With predictions produced by the system and method, persistence homology and Mapper subroutines from topological data analysis are used to both (a) process the shape of the tissue for overall metrics of missing tissue and (b) develop/generate a score for each section for how much the tissue artifacts intermingle into prognostically important regions of tissue (See box 730 in FIG. 7 and the graph 900 of FIG. 9). By way of example, the graph divides the results (bars) into three sections 910920 and 930 and based upon dermis 902, epidermis 914 and subcutaneous tissue 916. In this example, the score is logarithmically presented between 0 and 10. The hole quality score 952 and tumor quality score 954 are each depicted numerically in the printout 950 of FIG. 9A. After calculating the intermingling between the artifacts, tumor, fat, squamous regions, epidermis and dermis regions, the amount of intermingling of both holes and tumors with these tissue regions in serial sections or levels of the same tissue are compared stereotactically to capture the importance of missing a hole in the current section as a function of the quality and malignancy of its neighboring sections incorporating three dimensional information.


The system and method can employ Topological Data Analysis (TDA) so as to reveal key relationships in the data (in this case, morphological and spatial information encoded in each of the WSI subimages) by collapsing irrelevant structures and predicting how tissue is expected to be distributed versus what is observed. The Mapper process/algorithm, a smart TDA clustering technique, decomposes the WSI into overlapping Regions of Interest (ROI) that are representative of different tissue subcompartments, and forms weighted connections between ROIs to portray shared information content and important functional relationships. The system and method applies these processes/algorithms to calculate the intermingling of the hole with the surrounding tissue architecture for where artifacts are located in the slide with respect to important architectural components (importance assignment). The system and method further uses area calculations for quantitating total amount of holes (amount of bad quality), and compares TDA measurements across sections for assess the likelihood of a tumor being where the hole is in the current section (importance assignment; e.g. if hole in dermis of present section but adjacent section contains tumor in dermis). The quality of a particular tissue may be assessed given the following mathematical relationship:







Quality

section


i


=




region


{

fat
,
dermis
,


}





importance
region

*

(


holes
region

+




j


{

0
,
1
,
2
,


}





tumor

region
,

section


j



*


[

λ
*

distance


section


i

,

section


j




]


-
1





)







Where the importance of each region and the weight given to adjacent sections at a particular distance is determined via expert knowledge. Note these approaches can be applied to any tissue type assuming CNN/GNN trained on expert truth and/or on specialized staining patterns.


5. Operational Procedure of a Digital Pathology System

Reference is made to FIG. 10 (which is continued at branch A on a second sheet), which shows the generalized operation of an automated pathology procedure 1000 according to an illustrative embodiment. The procedure 1000 provides serial section WSIs 1010 that can define various sections (1 and 2) in exemplary slides (A and B). The inked/segmented images in the slides are scanned and divided into images based upon tissue masks, so as to provide results with no holes 1020 and results with holes 1022. These images 1020, 1022 are provided to a filtering process 1030 to derive a final masked image for no holes 1040 and holes 1042. Both image results are delivered to a preprocess step 1050 and the final mask results for holes 1042 is also generated and provided to the ink detection and orientation process 1052, which is represented by oriented image 1054. The preprocess step 1050 delivers results to a trained tumor inflammation CNN 1060 and a macroarchitecture hole CNN 1062. These CNNs thereby provide results to a graphical creation step 1070. This step delivers results to a tumor detection GNN 1080 and macroachitecture detection GNN 1082, which, in turn provides results to a quality assessment step 1086. This allows generation of score graphs 1087 for each section (e.g. 1 and 2) as also shown in FIG. 9. A nuclei detection sub-process 1088 is applied to the results and can be used in conjunction with a visual generation step 1090, which provides results to a 3D stitching step 1092. The results of the steps 1090 and 1092 are used to provide report displays 1094 as part of report generation (step 1096). Such report generation can also include use of results from the above-described ink detection and orientation process 1052, as shown.


6. Conclusion

It should be clear that the above-described system and method provides an effective and useful tool for automating the overall pathology process for a range of WSI samples. It provides the practitioner with objectively scored results that assist in making further determinations relative to diagnosis of conditions, such as cancer. The techniques herein can be refined continuously by further training of applicable AI algorithms and can be applied to an increasing range of medical conditions and parts of the body. Results can be delivered in a manner that provides revenue to the system operators, and can be delivered to any region or device via existing communications networks, including highly remote areas that may lack sophisticated equipment or facilities.


The foregoing has been a detailed description of illustrative embodiments of the invention. Various modifications and additions can be made without departing from the spirit and scope of this invention. Features of each of the various embodiments described above may be combined with features of other described embodiments as appropriate in order to provide a multiplicity of feature combinations in associated new embodiments. Furthermore, while the foregoing describes a number of separate embodiments of the apparatus and method of the present invention, what has been described herein is merely illustrative of the application of the principles of the present invention. For example, the use of various colors and color-codings in images is exemplary of a wide range of possible indicia for distinguishing between characteristics in a display or physical structure. Grayscale shading and/or use of non-visible wavelengths, or other characteristics, can be used to distinguish such items, in a manner clear to those of skill. Also, as used herein, various directional and orientational terms (and grammatical variations thereof) such as “vertical”, “horizontal”, “up”, “down”, “bottom”, “top”, “side”, “front”, “rear”, “left”, “right”, “forward”, “rearward”, and the like, are used only as relative conventions and not as absolute orientations with respect to a fixed coordinate system, such as the acting direction of gravity. Additionally, where the term “substantially” or “approximately” is employed with respect to a given measurement, value or characteristic, it refers to a quantity that is within a normal operating range to achieve desired results, but that includes some variability due to inherent inaccuracy and error within the allowed tolerances (e.g. 1-2%) of the system. Note also, as used herein the terms “process” and/or “processor” should be taken broadly to include a variety of electronic hardware and/or software based functions and components. Moreover, a depicted process or processor can be combined with other processes and/or processors or divided into various sub-processes or processors. Such sub-processes and/or sub-processors can be variously combined according to embodiments herein. Accordingly, this description is meant to be taken only by way of example, and not to otherwise limit the scope of this invention.

Claims
  • 1. A system for automation of a pathology process comprising; a processor having a neural network, having trained artificial intelligence (AI) modules operating in association therewith, adapted to receive image data from images of whole tissue acquired by an camera assembly and whole slide images (WSIs) of inked and segmented tissue samples;a mask that produces image results for tissue with holes and free of holes;a filter that provides filtered image results to the AI modules, which thereby detects tumors and macroarchitecture features;a quality assessment process that produces quality score outputs for tumors and macroarchitecture features;a report generator that provides reports with one or more parameters to a user via an interface;a tissue determination process that determines a size and a description of tissue automatically based upon characteristics in the images acquired by the camera assembly;a tissue grossing and inking process that automatically generates a grossing and inking scheme based upon a user input of desired percentage of tissue margins analyzed and the tissue size; andan ink detection and orientation process that operates on the filtered images of the tissue samples and delivers results thereof to the report generator, andwherein the macroarchitecture features are defined by holes, fat, edges, dermis and epidermis information.
  • 2. The system as set forth in claim 1 wherein the AI modules include a tumor inflammation convolutional neural network (CNN), a macroarchitecture hole CNN, a tumor detection graphical neural network (GNN) and a macroarchitecture detection GNN.
  • 3. The system as set forth in claim 1 wherein the report generator receives results from the quality assessment process, and wherein the quality assessment process acts on results from a tumor detection GNN and a macroarchitecture detection GNN.
  • 4. The system as set forth in claim 1, further comprising, a visual generation and 3D stitching process that provides results to the report generator.
  • 5. The system as set forth in claim 1, wherein the report generator automatically creates a pathology report consisting of a written description and pictorial diagram relative to the images of the tissue.
  • 6. The system as set forth in claim 2 wherein the filters include at least one of a Sobel filter and a gradient-based filter.
  • 7. The system as set forth in claim 3 wherein the report generator automatically creates a pathology report consisting of a written description and pictorial diagram relative to the images of the tissue.
  • 8. The system as set forth in claim 4 wherein the visual generation process receives information from a cell nuclei detection process.
  • 9. The system as set forth in claim 8 wherein the report generator automatically creates a pathology report consisting of a written description and pictorial diagram relative to the images of the tissue.
  • 10. A method for automation of a pathology process comprising the steps of: providing a processor with a neural network having artificial intelligence (AI) modules operating in association therewith, which receive image data from camera images of whole tissue acquired by an camera assembly and whole slide images (WSIs) of inked and segmented tissue samples;producing image results for tissue with holes and free of holes with a mask;providing filtered image results to the AI modules, and detecting tumors and macroarchitecture features therewith;generating quality score outputs for tumors and macroarchitecture features;generating a report with one or more parameters to a user via an interface;determining a size and a description of tissue automatically based upon characteristics in the images acquired by the camera assembly;automatically generating a grossing and inking scheme based upon a user input of desired percentage of tissue margins analyzed and the tissue size; andoperating an ink detection and orientation process on the filtered images of the tissue samples that delivers results thereof to the report generator, andwherein the macroarchitecture features are defined by holes, fat, edges, dermis and epidermis information.
  • 11. The method as set forth in claim 10, wherein the step of generating the report includes automatically creating a pathology report consisting of a written description and pictorial diagram relative to the images of the tissue.
  • 12. The system as set forth in claim 10, further comprising, defining the AI modules to include a tumor inflammation convolutional neural network (CNN), a macroarchitecture hole CNN, a tumor detection graphical neural network (GNN) and a macroarchitecture detection GNN.
  • 13. The method as set forth in claim 10, further comprising, wherein the step of generating includes receiving results from a quality assessment process acting on results from a tumor detection GNN and a macroarchitecture detection GNN.
  • 14. The method as set forth in claim 10, further comprising, wherein the step of generating includes operating a visual generation and 3D stitching process that provides results.
  • 15. The method as set forth in claim 12, wherein the step of providing filtered image results includes operating at least one of a Sobel filter and a gradient-based filter.
  • 16. The method as set forth in claim 14, further comprising, receiving by the visual generation process, information from a cell nuclei detection process.
  • 17. The method as set forth in claim 16, wherein the step of generating the report includes automatically creating a pathology report consisting of a written description and pictorial diagram relative to the images of the tissue.
RELATED APPLICATION

This application claims the benefit of co-pending U.S. Provisional Application Ser. No. 63/176,333, entitled SYSTEM AND METHOD FOR AUTOMATION OF SURGICAL PATHOLOGY PROCESSES USING ARTIFICIAL INTELLIGENCE, filed Apr. 18, 2021, the teachings of which are expressly incorporated herein by reference.

US Referenced Citations (86)
Number Name Date Kind
3975762 van den Bosch Aug 1976 A
5075214 Connor Dec 1991 A
5218529 Meyer Jun 1993 A
5544650 Boon Aug 1996 A
5784162 Cabib Jul 1998 A
5976885 Cohenford Nov 1999 A
5991028 Cabib Nov 1999 A
6146897 Cohenford Nov 2000 A
6463438 Veltri Oct 2002 B1
6690817 Cabib Feb 2004 B1
7693334 Ogura Apr 2010 B2
8351676 Dai Jan 2013 B2
9518982 Sood Dec 2016 B2
9786050 Bhargava Oct 2017 B2
9983195 King May 2018 B2
10013760 Bhargava Jul 2018 B2
10013781 Gammage Jul 2018 B1
10521901 Ikemoto Dec 2019 B2
10935773 Johnson Mar 2021 B2
11504103 Johnson Nov 2022 B2
11508045 Amthor Nov 2022 B1
11526984 Barnes Dec 2022 B2
11614610 Johnson Mar 2023 B2
11620751 Sarkar Apr 2023 B2
11621058 Gurcan Apr 2023 B2
11631171 Leng Apr 2023 B2
11633146 Leng Apr 2023 B2
11672425 Pyun Jun 2023 B2
11675178 Wirch Jun 2023 B2
11681418 Wirch Jun 2023 B2
11684264 Bryant-Greenwood Jun 2023 B2
11751903 Knowlton Sep 2023 B2
11751904 Knowlton Sep 2023 B2
11759231 Knowlton Sep 2023 B2
11776124 Behrooz Oct 2023 B1
11776681 Godrich Oct 2023 B2
11783603 Stumpe Oct 2023 B2
20010018659 Koritzinsky Aug 2001 A1
20030026762 Malmros Feb 2003 A1
20060210153 Sara Sep 2006 A1
20070127022 Cohen Jun 2007 A1
20070135999 Kolatt Jun 2007 A1
20080015448 Keely Jan 2008 A1
20080166035 Qian Jul 2008 A1
20080273199 Maier Nov 2008 A1
20080319324 Maier Dec 2008 A1
20090002702 Maier Jan 2009 A1
20090024375 Kremer Jan 2009 A1
20090319291 Noordvyk Dec 2009 A1
20110080581 Bhargava Apr 2011 A1
20110182490 Hoyt Jul 2011 A1
20110286654 Krishnan Nov 2011 A1
20120034647 Herzog Feb 2012 A1
20120052063 Bhargava Mar 2012 A1
20120092663 Kull Apr 2012 A1
20120143082 Notingher Jun 2012 A1
20120200694 Garsha Aug 2012 A1
20120212733 Kodali Aug 2012 A1
20120226644 Jin Sep 2012 A1
20120290607 Bhargava Nov 2012 A1
20130022250 Nygaard Jan 2013 A1
20140235487 McDevitt Aug 2014 A1
20140270457 Bhargava Sep 2014 A1
20140336261 Chin Nov 2014 A1
20150268226 Bhargava Sep 2015 A1
20150374306 Gelbman Dec 2015 A1
20160042511 Chukka Feb 2016 A1
20160272934 Chander Sep 2016 A1
20160335478 Bredno Nov 2016 A1
20170160171 Tsujikawa Jun 2017 A1
20170169567 Chefd'hotel Jun 2017 A1
20170322124 Barnes Nov 2017 A1
20170358082 Bhargava Dec 2017 A1
20170372471 Eurèn Dec 2017 A1
20180232883 Sethi Aug 2018 A1
20190188446 Wu Jun 2019 A1
20200302603 Barnes Sep 2020 A1
20200372235 Peng Nov 2020 A1
20200372635 Veidman Nov 2020 A1
20200394825 Stumpe Dec 2020 A1
20210103797 Jang Apr 2021 A1
20210295507 Nie Sep 2021 A1
20220146418 Bauer May 2022 A1
20230249175 Linnes Aug 2023 A1
20230279512 Masters Sep 2023 A1
20230394716 de Haan Dec 2023 A1
Foreign Referenced Citations (1)
Number Date Country
2020180755 Sep 2020 WO
Non-Patent Literature Citations (8)
Entry
Akbar, 2019, Scientific Reports, pp. 1-9.
Arunachalam, 2019, PLOS ONE, pp. 1-19.
Corvo, 2017 IEEE Workshop on Visual Analytics in Healthcare, pp. 77-83.
McCann, IEEE Signal Processing Magazine, 2015, pp. 78-87.
Rivenson, 2020, BMEF, pp. 1-11.
Taieb, 2019, ArXiv, pp. 1-58.
Grala, 2009, pp. 587-592.
U.S. Appl. No. 16/679,133, entitled System and Method for Analyzing Cytological Tissue Preparations, Louis J. Vaickus, filed Nov. 8, 2018.
Related Publications (1)
Number Date Country
20220375604 A1 Nov 2022 US
Provisional Applications (1)
Number Date Country
63176333 Apr 2021 US