Embodiments of this disclosure relate to methods and apparatus configured to provide training in automated diagnostic systems.
Automated diagnostic systems analyze (e.g., test) biological specimens, such as whole blood, blood serum, blood plasma, urine, interstitial liquid, cerebrospinal liquid, and the like, in order to identify analytes or other constituents in the specimens. The specimens are usually contained within specimen containers (e.g., specimen collection tubes) that can be transported via automated track systems to various pre-processing modules, pre-screening modules, and analyzers (e.g., including immunoassay and clinical chemistry) within such automated diagnostic systems.
In some systems, the pre-processing modules can carry out processing on the specimen or specimen container, such as de-sealing, centrifugation, aliquoting, and the like, all prior to analysis by one or more analyzers. In some systems, the pre-screening may be used to characterize specimen containers and/or the specimens. Characterization may be performed by an artificial intelligence (AI) model and may include a segmentation operation, which may identify various regions of the specimen containers and/or specimens. Characterization of the specimens using the AI model may include an HILN process that determines a presence of an interferent, such as hemolysis (H), icterus (I), and/or lipemia (I), in a specimen to be analyzed or determining that the specimen is normal (N) and can thus be further processed.
After pre-processing and/or pre-screening, the specimens are analyzed by one or more analyzers of the automated diagnostic system. Measurements may be performed on the specimens via photometric analyses such as, fluorometric absorption and/or emission analyses. Other measurements may be used. The measurements may be analyzed to determine amounts of analytes or other constituents in the specimens.
Over time, components of the systems may change. For example, imaging devices and illumination sources used during imaging may change. In some embodiments, the specimen containers may also change over time. The AI model(s) may not be adequately trained to characterize the components and specimen containers that have changed over time. Thus, the above-described analysis using the AI models may be erroneous.
Based on the foregoing, improved methods of training AI models for use in automated diagnostic systems are sought.
According to a first aspect, a method of characterizing a specimen container or a specimen in an automated diagnostic system is provided. The method includes capturing an image of a specimen container containing a specimen using an imaging device; characterizing the image using a first AI model; determining whether a characterization confidence of the image is below a pre-selected threshold; and retraining the first AI model with at least the image having the characterization confidence below the pre-selected threshold to a second AI model, wherein the retraining includes data selected from one or more of a group of: non-image data and text data.
According to another aspect, a method of characterizing a specimen in an automated diagnostic system is provided. The method includes capturing an image of the specimen using an imaging device; characterizing the image using a first AI model to determine a presence of at least one of hemolysis, icterus, or lipemia; determining whether a characterization confidence of the determination of the presence of at least one of hemolysis, icterus, or lipemia is below a pre-selected threshold; and retraining the first AI model with at least the image having the characterization confidence below the pre-selected threshold to a second AI model, wherein the retraining includes data selected from one or more of a group of: non-image data, and text data.
According to another aspect, an automated diagnostic system is provided. The automated diagnostic system includes an imaging device configured to capture an image of a specimen container containing a specimen; and a computer configured to: characterize the image using a first AI model; determining whether a characterization confidence of the image is below a pre-selected threshold; and retrain the first AI model with at least the image having the characterization confidence below the pre-selected threshold to a second AI model, wherein the retraining includes data selected from one or more of a group of: non-image data and text data.
Still other aspects, features, and advantages of this disclosure may be readily apparent from the following description and illustration of a number of example embodiments and implementations, including the best mode contemplated for carrying out the invention. This disclosure may also be capable of other and different embodiments, and its several details may be modified in various respects, all without departing from the scope of the disclosure. This disclosure is intended to cover all modifications, equivalents, and alternatives falling within the scope of the claims.
The drawings, described below, are for illustrative purposes, and are not necessarily drawn to scale. Accordingly, the drawings and descriptions thereof are to be regarded as illustrative in nature, and not as restrictive. The drawings are not intended to limit the scope of the disclosure in any way.
Automated diagnostic systems described herein analyze (e.g., test) biological specimens to determine the presence and/or concentrations of analytes in the specimens. In some embodiments, the systems may perform one or more pre-screening analyses on the specimens. In some embodiments, the systems may perform pre-screening analyses on specimen containers. The analyses may be performed using artificial intelligence (AI) models as described herein.
The AI models described herein may be implemented as machine learning, neural networks, and other AI algorithms. The AI models may be trained to characterize images or portions of captured images. Characterizing images includes identifying items in one or more portions of an image. For example, a first or initial AI model may be trained to characterize items in images that are expected to be captured by the system, such as specimens and specimen containers. In some embodiments, a large dataset of images of items that are to be characterized may be captured in different configurations, such as different views and/or different lighting conditions, and may be used to train a first AI model. One or more algorithms or programs may be used to check characterization confidences that trained the first AI model. In some embodiments, the characterization confidences may be in the form of a value (e.g., between 1 and 100) or a percentage. A low characterization confidence may be indicative of inadequate characterization, i.e., the AI model has not been adequately trained to recognize the specimen or specimen container.
Over time, the items being characterized may change. In some embodiments, the conditions under which the images are captured also may change. These changed items and conditions may be due to hardware changes (e.g., updates), software changes, changes in specimen containers, labeling changes on the specimen containers, changes in assays, and other changes. The first AI model may not be able to characterize these changed items or may not be able to characterize the items under the changed conditions. For example, images used to train the first AI model may not include every variation of a specimen container and/or a specimen received in a system. For example, the sizes, types, and characteristics of specimen containers may change over time, resulting in incorrect or low characterization confidences. Accordingly, the first AI model may have to be updated to a second AI model in order to be able to properly characterize the changed items and conditions.
Conventional Automated diagnostic systems do not provide for easy and/or automatic updates to AI models. Rather, the process for updating AI models may be relatively cumbersome. For example, in some embodiments, the deficiencies in the first AI model may not be identified until the system fails. Troubleshooting the system failure may be required to determine that the first AI model is not adequate. Once the first AI model inadequacies are identified, data pertaining to the above-described changes is collected and transmitted to engineering teams. This data is then used by the engineering teams to update or retrain the first AI model. The conventional update processes are very expensive and time consuming. Methods and apparatus disclosed herein provide for improved updating and/or replacing of first AI models with second AI models.
The AI models described herein may be used in systems that include a quality check module. A quality check module performs pre-screening of specimens and/or specimen containers based on images captured in the quality check module. The prescreening may use an AI model as described herein to characterize images of specimen containers and/or specimens. The pre-screening characterization may include performing segmentation and/or interferent (e.g., HILN—hemolytic, icteric, lipemic, normal) identifications on the captured images. The segmentation determination may identify various regions (areas) in the image of the specimen container and specimen, such as a serum or plasma portion, a settled blood portion, a gel separator (if used), an air region, one or more label regions, a type of specimen container (indicating, e.g., height and width or diameter), and/or a type and/or color of a specimen container cap.
As described above, the interferent identified by the AI models may include hemolysis (H), icterus (I), or lipemia (L). The degree of hemolysis may be quantified using the AI model by assigning a hemolytic index (e.g., H0-H6 in some embodiments and more or less in other embodiments). The degree of icterus may be quantified using the AI model by assigning an icteric index (e.g., I0-I6 in some embodiments and more or less in other embodiments). The degree of lipemia may be quantified using the AI model by assigning a lipemic index (e.g., L0-L4 in some embodiments and more or less in other embodiments). In some embodiments, the pre-screening process may include determination of an un-centrifuged (U) class for a serum or plasma portion of a specimen that has not been centrifuged.
An HILN network implemented by an AI model may be or include a segmentation convolutional neural network (SCNN) that receives as input one or more captured images of fractionated specimens contained in specimen containers. The SCNN may include, in some embodiments, greater than 100 operational layers including, e.g., BatchNorm, ReLU activation, convolution (e.g., 2D), dropout, and deconvolution (e.g., 2D) layers to extract features, such as simple edges, texture, and parts of the serum or plasma portion and label-containing regions of images. Top layers, such as fully convolutional layers, may be used to provide correlation between the features. The output of the layer may be fed to a SoftMax layer, which produces an output on a per pixel (or per superpixel (patch) —including n×n pixels) basis concerning whether each pixel or patch includes HIL or is normal.
In some embodiments, only an output of HILN may be provided by the SCNN. In other embodiments, the output of the SCNN may include multiple classes of HILN, such as greater than 20 classes of HILN, so that for each interferent present, an estimate of the level (index) of the interferent can also be obtained. Other numbers of classes of each of HIL may be included in the SCNN. The SCNN may also include a front-end container segmentation network (CSN) to determine a specimen container type and a specimen container boundary. Other types of HILN networks may be used in quality check modules.
In some embodiments, an initial set of training images used to train an initial or first AI model may be compiled by operating a newly-installed automated diagnostic analysis system for a given period of time (e.g., one or two weeks). Captured image data of specimens received in the newly installed system may be forwarded to a database/server (which may be local and/or a part of the newly installed system or it may be a cloud-based server). The image data may be annotated (e.g., annotated manually and/or annotations generated automatically) to create the initial set of annotated training images. The set of annotated training images may then be used to train the initial or first AI model in the HILN network of the quality check module of the automated diagnostic analysis system.
During operation of the automated diagnostic analysis system, images of specimens having characterizations generated by the HILN network that are determined to be incorrect or have low characterization confidences (e.g., low confidence levels) may not be automatically forwarded to an analyzer of the automated diagnostic analysis system, but may be stored for further review. For example, the images of specimens having characterizations determined to be incorrect or having low characterization confidences may be stored (and encrypted in some embodiments) in a database/server. The training updates (e.g., training of the second AI model) may be based at least in part on the incorrect or low confidence characterizations that stored in the database/server.
In some embodiments, the training updates may be based at least in part on manual annotations and/or automatically generated annotations of the captured images of the specimens having characterizations determined to be incorrect or have low confidence. The training updates may be forwarded to the HILN network for incorporation therein via retraining of the first AI model to generate a retrained second AI model. In some embodiments, a report or prompt of the availability of one or more training updates may be provided to a user to allow the user to decide when and if the training updates are to be incorporated into the HILN network. In other embodiments, training updates may be automatic.
In some embodiments, the initial set of training images and/or the training updates, each of which is software based, may be provided to an automated diagnostic analysis system (and the HILN network in particular) as a retrained model via the internet or by using a physical media (e.g., a storage device containing programming instructions and data).
Some embodiments of systems disclosed herein can provide continuous training updates of AI models that may be automatically incorporated into the system via retraining and/or AI model replacement on a frequent or regular basis, such as, e.g., upon meeting or exceeding a threshold number of incorrect or low characterization confidences. Other criteria may be used to automatically incorporate training updates into the systems. In some embodiments, training updates may be incorporated into a system by a user at the discretion of the user, such as via a user prompt.
Further details of characterization apparatus and methods including updating AI models will be further described with reference to
Reference is now made to
The specimen containers 102 may be received at the system 100 in one or more racks 104 provided at a loading area 106. The specimen containers 102 may be transported throughout the system 100, such as to and from modules 108 and instruments 110 on a track 112 by carriers 114.
Processing of the specimens and/or the specimen containers 102 may include preprocessing or pre-screening of the specimens and/or the specimen containers 102 prior to analysis by one or more of the modules 108 configured as analyzer modules, which may be referred to herein as analyzers. The system 100 may also include one or more instruments 110, wherein each instrument may include one or more modules, such as preprocessing modules and/or analyzer modules. In the embodiment of
In some embodiments, the first module 116A may be a preprocessing module, for example, that processes the specimen containers 102 and/or specimens located therein prior to analyses by analyzer modules. The second module 116B and the third module 116C may be analyzer modules that analyze specimens as described herein. Other embodiments of the instruments 110 may be used for other purposes in the system 100.
In the embodiment of
The modules implemented as analyzer modules of the modules 108 and the instruments 110 may be any combination of any number of clinical chemistry analyzers, assaying instruments, and/or the like. The term “analyzer” as used herein includes a device used to analyze a specimen for chemistry or to assay for the presence of, amount, or functional activity of a target entity (e.g., an analyte), such as DNA or RNA, for example. Analytes commonly tested for in analyzer modules include enzymes, substrates, electrolytes, specific proteins, drugs of abuse, and therapeutic drugs.
Additional reference is made to
The specimen container 202 may be provided with at least one label 222 that may include identification information 222I (i.e., indicia) thereon, such as a barcode, alphabetic characters, numeric characters, or combinations thereof. The identification information 222I may include or be associated data provided by a laboratory information system (e.g., LIS 131—
The identification information 222I may be darker (e.g., black) than the label material (e.g., white paper) so that the identification information 222I can be readily imaged. The identification information 222I may indicate, or may otherwise be correlated, via the LIS or other test ordering system, to a patient's identification as well as tests to be performed on the specimen 216. The identification information 222I may be provided on the label 222, which may be adhered to or otherwise provided on an outside surface of the tube 218. In some embodiments, the label 222 may not extend all the way around the specimen container 202 or along a full length of the specimen container 202.
The specimen 216 may include a serum or plasma portion 216SP and a settled blood portion 216SB contained within the tube 218. A gel separator 216G may be located between the serum or plasma portion 216SP and the settled blood portion 216SB. Air 224 may be provided above the serum and plasma portion 216SP. A line of demarcation between the serum or plasma portion 216SP and the air 224 is defined as the liquid-air interface (LA). A line of demarcation between the serum or plasma portion 216SP and the gel separator is defined as a serum-gel interface (SG). A line of demarcation between the settled blood portion 216SB and the gel separator 216G is defined as a blood-gel interface (BG). An interface between the air 224 and the cap 220 is defined as a tube-cap interface (TC).
The height of the tube (HT) is defined as a height from a bottom-most part of the tube 218 to a bottom of the cap 220 and may be used for determining tube size (e.g., tube height and/or tube volume). A height of the serum or plasma portion 216SP is HSP and is defined as a height from a top of the serum or plasma portion 216SP at LA to a top of the gel separator 216G at SG. A height of the gel separator 216G is HG and is defined as a height between SG and BG. A height of the settled blood portion 216SB is HSB and is defined as a height from the bottom of the gel separator 216G at BG to a bottom of the settled blood portion 216SB. HTOT is a total height of the specimen 216 and equals the sum of HSP, HG, and HSB. The width of the cylindrical portion of the inside of the tube 218 is W. Preprocessing performed in one or more of the preprocessing modules 108 and/or instruments 110 may measure or calculate at least one of the above-described dimensions.
The embodiment of
Referring again to
The computer 124, by way of the programs 124C, may control movement of the carriers 114 to and from the loading area 106, about the track 112, to and from the modules 108 and the instruments 110, and to and from other modules and components of the system 100. One or more of the modules 108 or instruments 110 may be in communications with the computer 124 through a network, such as a local area network (LAN), wide area network (WAN), or other suitable communication network, including wired and wireless networks. In some embodiments, the operation of some or all of the above-described modules 108 and/or instruments 110 may be performed by the computer 124.
One or more of the programs 124C may be artificial intelligence (AI) models or algorithms that process and/or analyze image data and other data as described herein. The other data may include non-image data (510—
The first AI model 130A and the second AI model 130B may be implemented as one or more of the programs 124C or an algorithm that is stored in the memory 124B and executed by the processor 124A. In some embodiments, the first AI model 130A and the second AI model 130B may be executed remotely from the system 100. The first AI model 130A and the second AI model 130B may be implemented as various forms of artificial intelligence, including, but not limited to, neural networks, including convolutional neural networks (CNNs), deep learning, regenerative networks, and other machine learning and artificial intelligence algorithms. Accordingly, the first AI model 130A and the second AI model 130B may not be simple lookup tables. Rather, the first AI model 130A and the second AI model 130B are trained to recognize (e.g., characterize) a variety of different images. A lookup table, on the other hand, is only able to identify images that are specifically in the lookup table.
In some embodiments, the computer 124 may be coupled to a computer interface module (CIM) 126. The CIM 126 and/or the computer 124 may be coupled to a display 128. The CIM 126, in conjunction with the display 128, enables a user to access a variety of control and status display screens and to input data into the computer 124. These control and status display screens may display and enable control of some or all aspects of the modules 108 and/or instruments 110 used for preparation, pre-screening, and analysis of specimen containers 102 and/or the specimens located therein. Thus, the CIM 126 may be adapted to facilitate interactions between a user and the system 100. The display 128 may be configured to display a menu including icons, scroll bars, boxes, and buttons through which the operator may interface with the system 100. The menu may include a number of functional elements programmed to display and/or operate functional aspects of the system 100. In some embodiments, the display 128 may include a graphical user interface that enables a user to instruct the computer 124 to update the first AI model 130A as described herein.
As described herein, the modules 108 and the instruments 110 may perform analyses on the specimen containers 102 and/or the specimens (e.g., specimen 216—
In some embodiments, specimens and/or specimen containers 102 are front illuminated in one or more of the modules 108 and/or the instruments 110. Images of the reflected light from the specimen containers 102 and/or the specimens are captured by one or more imaging devices and converted to image data that is processed as described herein. In some embodiments, images of light transmitted through the specimens and/or the specimen containers 102 is captured and converted to image data that is processed as described herein. In some embodiments, chemicals are added to the specimens to cause the specimens to fluoresce and emit light under certain conditions. Images of the emitted light may be captured and converted to image data that is processed as described herein.
The first AI model 130A may be trained by a first validation dataset. The first validation dataset is data collected and used to train and/or verify the first AI model 130A. In some embodiments, the first validation dataset may include data that is verified by various testing or analyses mechanisms. In some embodiments, the first validation dataset may include data that was used for regulatory approval of the system 100 and/or similar systems. For example, in some embodiments, the first validation dataset may include data that may be collected across multiple systems that may be identical or similar to the system 100. In some embodiments, the first validation dataset may be compressed and/or encrypted. In some embodiments, the first validation dataset and/or the first AI model 130A and the second AI model 130B may be stored and/or executed remotely, such as in a cloud. The ground truth for the first validation data set may come from secondary resources, such as a gold standard device and/or data based on the gold standard device. In other embodiments, the gold truth may be automatically generated using an existing trained system or by self-supervision.
Over time, changes, in the data processed by the system 100, including, for example, data generated by the modules 108 and/or the instruments 110, may occur. These changes include, for example, hardware changes (e.g., updates), software changes, changes in the specimen containers 102, changes in the labels (e.g., label 222) and/or barcodes (e.g., information 2201) affixed to the specimen containers 102, assay protocols, and other changes. These changes may not be able to be characterized (e.g., identified) by the first AI model 130A, so the system 100 may have to be updated to a new AI model (e.g., the second AI model 130B). The methods of updating the system 100 to the second AI model 130B are described herein.
Additional reference is made to
The method 300, in 302, includes capturing an image. The image may be captured using an imaging device (e.g., one or more of the imaging devices 440,
The method 300 includes, in 304, characterizing the image using the first AI model 130A. Different characterizations are described in greater detail. In some embodiments, the characterization may include identifying one or more items in the captured image, such as the specimen container 202 (
The method 300, in 306, includes determining confidence of the characterization (characterization confidence) performed in 304. The characterization confidence may be a score or probability that the first AI model 130A characterized or correctly identified items in the captured image. Various known techniques may be used to determine the characterization confidence as described herein. In some embodiments, the characterization confidence may be zero if the characterization was not able to characterize or identify one or more items in the captured image.
The method 300 includes, in 308, determining whether the characterization confidence is above a pre-established threshold. If the confidence is above the pre-established threshold, processing proceeds to 308 where the first AI model 130A is used to characterize the captured image and future captured images. In some embodiments, the pre-established threshold may be 0.7 on a scale between zero and 1.0. This pre-established threshold provides a likelihood that the characterization is correct. In other embodiments, the pre-established threshold may be 0.9 on a scale between zero and 1.0. This pre-established threshold provides more confidence that the characterization is correct.
If, in 308, the determination is made that the confidence is not above the pre-selected threshold, the system or other device generates the second AI model 130B (
The processing from 308 proceeds to 312 where sensor data from at least one or more non-image sensors and/or text data are received. In some embodiments, the data is received in one of the programs 124C that may generate the second AI model 130B. In other embodiments, the data is received in one or more other devices that train the second AI model 130B. This data is used to train the second AI model 130B or update the first AI model 130A to the second AI model 130B. In some embodiments, the second AI model 130B may be the same as the first AI model 130A, but trained using the data described herein. Accordingly, the second AI model 130B is trained to characterize items that are different than items the first AI model 130A is trained to characterize. In some embodiments, the data used to train the second AI model 130B includes at least some of the data used to train the first AI model 130A, so the second AI model 130B may characterize at least some of the items that the first AI model 130A was trained to characterize. In some embodiments, a user of the system 100 may be prompted to train the second AI model 130B. The user may then initiate the training such as by the CIM 126 (
The non-image sensors may include, for example, temperature sensors, acoustic sensors, humidity sensors, liquid volumes sensors, vibration sensors, current sensors, and other sensors related to the operation of the system 100. The text data may include tests being performed (e.g., assay types), patient information (e.g., age, symptoms, etc.), date of the test, time of the test, system logs (e.g., system status), label information from the specimen containers 102 (e.g., data from the label 222—
The method 300 may proceed to 314 where the second AI model 130B (
In some embodiments, updating the first AI model 130A may include updating the model capacity (e.g., adding residual layers) or model weights. The model weights determine which data samples are used for backpropagation by the AI model. In some embodiments, the AI model may include a deep network, such as a variational auto-encoder, that can be trained to determine if data provided is out of a training manifold or within an original training manifold. In embodiments where the second AI model 130B replaces the first AI model 130A, the second AI model 130B may be trained as described above.
The data used to train the second AI model 130B may be referred to as the sampling data. The sampling data incorporated into the second AI model 130B may be selected to avoid divergence of the second AI model 130B. In divergence, the second AI model 130B will perform worse than the first AI model 130A that was trained on the first validation dataset, which may be a gold standard or ground truth. Divergence may be indicative as either underfitting or “catastrophic forgetting.” Underfitting may be identified by the second AI model 130B not able to identify or characterize items in the new data. Catastrophic forgetting may be identified as the second AI model 130B overfitted to the new data, wherein the second AI model 130B is not able to characterize items in in the first validation dataset. Neither underfitting nor catastrophic forgetting may be acceptable because underfitting restricts the range of improvements that can be made and catastrophic forgetting (e.g., overfitting) may no longer meet the requirements of the regulatory clearance obtained based on the first validation dataset.
Based on the foregoing, in some embodiments, the first AI model 130A may only be updated when the updates are likely to help the system 100. In some embodiments, outliers may exist in the sample data that may cause the second AI model to degenerate, such as by underfitting or catastrophic forgetting as described above. The problems may be avoided by having access to a validation dataset on which the performance of the second AI model 130B may be evaluated and if a divergence occurs, the second AI model 130B may be rollbacked to an older AI model, such as the first AI model 130A. In some embodiments, the first AI model 130A may be updated continuously.
In some embodiments, training the second AI model 130B may include validating the second AI model 130B using a validation dataset as described above. The validation dataset may be data correlating the captured images to certain characterizations. In other embodiments, the validation dataset may be based on data received from other sources, such as other systems or data sets generated specifically to validate the second AI model 130B.
The second AI model 130B may be validated in 316. For example, the captured images and/or other images having characteristics similar to the captured images may be characterized using the second AI model 130B. Characterization confidences performed using the second AI model 130B may be determined. If the characterization confidences are below a pre-selected threshold, the second AI model 130B may not be trained correctly. In such situations, the second AI model 130B may be trained further or replaced with the first AI model 130A. If the characterization confidences are greater than the pre-selected threshold, the method 300 may proceed to 318 where the second AI model 130B is used to characterize images as described herein.
The method 300 will now be described implemented in modules 108 and/or the instruments 110 of the system 100 (
Additional reference is made to
The quality check module 132 may include a housing (not shown) that may at least partially surround or cover the track 112 to minimize outside lighting influences. The specimen container 202 may be located inside the housing at an imaging location 442 during the image-capturing sequences. The housing may include one or more doors (not shown) to allow the carrier 214 to enter into and/or exit from the quality check module 132. In some embodiments, a ceiling (not shown) of the housing may include an opening that allows the specimen container 202 to be loaded into the carrier 214 by a robot (not shown).
The quality check module 132 may include one or more imaging devices 440. The imaging devices 440 are referred to individually as a first imaging device 440A, a second imaging device 440B, and a third imaging device 440C. The imaging devices 440 may be configured to capture images of the specimen container 202 and specimen 216 at the imaging location 442 from multiple viewpoints (e.g., viewpoints labeled 1, 2, and 3). While three imaging devices 440 are shown in
The images of the specimen 216 and/or the specimen container 202 may be captured while the specimen container 202 is residing in the carrier 214 at the imaging location 442. The field of view of the multiple images obtained by the imaging devices 440 may overlap slightly in a circumferential extent. Thus, in some embodiments, portions of the images may be digitally added to arrive at a complete image of the serum or plasma portion 216SP (
The imaging devices 440 may be any suitable devices configured to capture digital images. In some embodiments, each of the imaging devices 440 may be a conventional digital camera capable of capturing pixelated images, a charged coupled device (CCD), an array of photodetectors, one or more CMOS sensors, or the like. The sizes of the captured images may be about 2560×694 pixels, for example. In other embodiments, the imaging devices 440 may capture images having sizes of about 1280×387 pixels, for example. The captured images may have other sizes.
The quality check module 132 may include one or more light sources 444 that are configured to illuminate the specimen container 202 and/or the specimen 216 during image capturing. In the embodiment of
In addition to the imaging devices 440, the quality check module 132 may include one or more non-image sensors. Non-image sensors are sensors that may be used by a module, such as the quality check module 132 to generate data, other than image data, related to the operation of the module. Image data is data representative of a captured image and may include a plurality pixel values. Instruments also may include non-image sensors. The embodiment of the quality check module 132 depicted in
The current sensor 450 may measure current drawn by one or more components of a module, such as the quality check module 132, and generate current data. In the embodiment of
The vibration sensor 452 may measure vibration of a module, such as the quality check module 132, and/or one or more components in the module and generate vibration data. In the embodiment of
The humidity sensor 454 may measure humidity in a module, such as the quality check module 132, and generate humidity data. In some embodiments, the humidity sensor 454 may measure humidity in the location of the system 100 (
The temperature sensor 456 may measure temperature in a module, such as the quality check module 132, and generate temperature data. In some embodiments, the temperature sensor 456 may measure temperature of a component within a module or within the system 100 (
The acoustic sensor 458 may measure noise in a module, such as the quality check module 132, and may generate acoustic data. In some embodiments, the acoustic sensor 458 may measure ambient noise proximate the system 100 (
In some embodiments, the characterizations associated with data generated by sensors in the quality check module 132 may include determining a presence of and/or an extent or degree of hemolysis (H), icterus (I), and/or lipemia (L) contained in the specimen 216. In some embodiments, the characterization may determine whether the specimen 216 is normal (N). If the specimen 216 is found to contain low amounts of H, I and/or L, so as to be considered normal (N), the specimen 216 may continue on the track 112 where the specimen 216 may be analyzed (e.g., tested) by the one or more of the modules 108 and/or the instruments 110. Other pre-processing operations may be conducted on the specimen 216 and/or the specimen container 202.
In some embodiments, in addition to detection of HILN, the characterization may include segmentation of the specimen container 202 and/or the specimen 216. From the segmentation data, post processing may be used for quantification of the specimen 216 (i.e., determination of HSP, HSB, HTOT, HG, W, and/or possibly a determination of location of TC, LA, SG, and/or BG). In some embodiments, characterization of the physical attributes (e.g., size-height and/or width) of the specimen container 202 may be performed at the quality check module 132. From these characterizations, the size of the specimen container 202 may be calculated. Moreover, in some embodiments, the quality check module 132 may also determine the type of the cap 220 (
Additional reference is made to
In some embodiments, the database 504 may be resident in the memory 124B (
Data used for the training updates 508 may be stored in the database 504. In some embodiments, the data may be stored in different databases that are collectively referred to as the database 504. The database may include non-image data 510 from the non-image sensors and text data 512. In some embodiments, the non-image data 510 may be generated at the same time the image data is generated. Likewise, the text data 512 may correlate to the image data. For example, the text data 512 may be related to specimens and/or specimen containers captured to generate the image data. The data used for the for the training updates 508 may also include original data used to train the first AI model 130A. Accordingly, the updated AI model (the second AI model 130B) may also be configured to characterize images that the first AI model 130A was trained to characterize.
In some embodiments, the database 504 may include image data 514 of images having low characterization confidences. For example, referring to
The training updates 508 may access the database 504 to obtain data to train the AI model(s) on a regular basis, upon detection of low characterization confidence, or by an action of a user. In some embodiments, the training updates 508 may be applied immediately into the second AI model 130B. In such embodiments, the training updates 508 may be performed as data is received into the database 504. In some embodiments, a user of the system 100 (
In some embodiments, the training updates 508 may be initiated automatically, such as without user initiation. For example, the HILN network 502 or other component may determine the characterization confidences. If the characterization confidences are below a pre-established threshold, such as the pre-established threshold in 308 of
In some embodiments, the training updates 508 may be performed locally, such as in the computer 124 (
In some embodiments, the HILN network 502 or a program receiving data from the HILN network 502 may generate a performance report related to the second AI model 130B. In some embodiments, the performance report may be in compliance with a regulatory process and may expedite regulatory approval of the updated HILN network 502. In some embodiments, the performance report may highlight improvement of the second AI model 130B over the first AI model 130A.
The HILN network architecture 500 may be implemented in a quality check module 132 and controlled by the computer 124 (
One or more images (e.g., multi-viewpoint images) may be captured by at least one of the plurality of imaging devices 440 as represented at functional block 522. The raw image data for each of the captured images may be processed and consolidated as described in US Pat. App. Pub. 2019/0041318 to Wissmann et al. titled “Methods And Apparatus For Imaging A Specimen Container and/or Specimen Using Multiple Exposures” to process the raw image data as represented at functional block 524. In some embodiments, the raw image data may include a plurality of optimally exposed and normalized image data sets (hereinafter “image data sets”) in functional block 524. The processing in functional block 524 may produce the image data 514 (i.e., pixel data). The image data of a captured image data set of the specimen 216 (and specimen container 202) may be provided as input to the HILN network 502 in the HILN network architecture 500 in accordance with one or more embodiments. In some embodiments, the image data 514 may be raw image data.
The HILN network 502 may be configured to perform characterizations, such as segmentation and/or HILN determinations, on the image data 514 using the first AI model 130A and/or the second AI model 130B. The first AI model 130A is used in situations where the first AI model 130A has not been updated to or replaced by the second AI model 130B. In some embodiments, the segmentations and HILN determinations may be accomplished by a segmentation convolutional neural network (SCNN). Other types of HILN networks may be employed to provide segmentation and/or HILN determinations.
In some embodiments, the HILN network architecture 500 may perform pixel-level classification and may provide a detailed characterization of one or more of the captured images. The detailed characterization may include separation of the specimen container 202 from the background and a determination of a location and content of the serum or plasma portion 216SP of the specimen 216. In some embodiments, the HILN network 502 can be operative to assign a classification index (e.g., HIL or N) to each pixel of the image based on local appearances of each pixel. Pixel index information can be further processed by the HILN network 502 to determine a final HILN classification index for each pixel.
In some embodiments, the classification index may include multiple serum classes, including an un-centrifuged class, a normal class, and multiple classes/subclasses. In some embodiments, the classification may include 21 serum classes, including an un-centrifuged class, a normal class, and 19 HIL classes/subclasses, as described in greater detail below.
One challenge to determining an appropriate HILN classification index for a specimen 216 undergoing pre-screening at the quality check module 132 may result from the small appearance differences within each sub-class of the H, I, and L classes. That is, the pixel data values of adjacent sub-classes can be very similar. To overcome these challenges, SCNN, which may be implemented by an AI model of the HILN network 502, may include a very deep semantic segmentation network (DSSN) that includes, in some embodiments, more than 100 operational layers.
To overcome appearance differences that may be caused by variations in specimen container type (e.g., size, shape, and/or type of glass or plastic material used in the container), the HILN network 502 may also include a container segmentation network (CSN) at the front end of the DSSN and implemented by the first AI model 130A or the second AI model 130B. The CSN may be configured to determine an output container type that may include, for example, a type of the specimen container 202 indicating height HT and width W (or diameter), and/or a type and/or color of the cap 220. In some embodiments, the CSN may have a similar network structure as the DSSN, but shallower (i.e., with fewer layers). The DSSN may be configured to determine output boundary segmentation information 520, which may include locations and pixel information of the serum or plasma portion 216SP, the settled blood portion 216SB, the gel separator 216G, the air 224, and the label 222.
The HILN determination of a specimen characterized by the HILN network 502 may be a classification index 528 that, in some embodiments, may include an un-centrifuged class 522U, a normal class 522N, a hemolytic class 522H, an icteric class 522I, and a lipemic class 522L. In some embodiments, the hemolytic class 522H may include sub-classes H0, H1, H2, H3, H4, H5, and H6. The icteric class 522I may include sub-classes I0, I1, I2, I3, I4, I5, and I6. The lipemic class 522L may include sub-classes L0, L1, L2, L3, and L4. Each of the hemolytic class 522H, the icteric class 522I, and/or the lipemic class 522L may have, in some embodiments, other numbers of fine-grained sub-classes.
The captured images, the non-image data 510, and the text data 512 of imaging having incorrect or low confidence characterizations may be forwarded (and encrypted in some embodiments) to the database 504. Various algorithms and/or techniques may be used to identify incorrect characterizations and/or determine characterization confidences (i.e., predictions of the accuracy of characterization determinations). In some embodiments, the incorrect or low confidence determinations are determinations that the HILN determinations are incorrect or have low probabilities of being correct. Incorrect characterizations, as used herein, mean that the HILN determinations are improper because the corresponding characterization confidences are low. In embodiments where there is a low characterization confidence determination of the HILN class or class index, the low characterization confidence may be based on using a characterization confidence level of less than 0.9, for example. This characterization confidence level limit may be pre-selected by a user or determined based on regulatory requirements.
In some embodiments, the incorrect or low characterization confidence determination is a determination that the segmentation determination is incorrect or has low confidence. In particular, the incorrect determination of the segmentation may involve identification of a region, such as the serum or plasma portion 216SP, the gel separator 216G, or the settled blood portion 216SB that has low characterization confidence or that is not in certain order with respect to one or more other regions. For example, the cap 220 may be expected to be on top of the serum or plasma portion 216SP and the gel separator 216G may be below the serum or plasma portion 216SP. If the relative positioning is not met, then an error may have occurred during the segmentation.
A determination of low characterization confidence in the segmentation or other process may involve reviewing a probability score for each segmented pixel (or collection of pixels in a superpixel—e.g., a collection of pixels, such as 11 pixels). If the probability score indicating a particular classification (e.g., serum or plasma portion 216SP) of a region of pixels has too much disagreement, then that segmented pixel would likely be a candidate for low characterization confidence. The probability scores of the pixels (or superpixels) of a region that has been segmented (e.g., region classified at serum or plasma portion 216SP) can be aggregated to determine if the region contains too much disagreement. In this case, the region would likely be a candidate for a low characterization confidence if the characterization confidence level is less than the pre-selected value (e.g., 0.9). Other suitable aggregated characterization confidence levels for a region can be used.
Based on the incorrect or low characterization confidences, the computer 124 (
The training updates 508 may be incorporated into the HILN network 502 (e.g., via the Internet or a physical media). In some embodiments, the training updates 508 may be an update applied to the first AI model 130A and in other embodiments, the training updates 508 may generate a new AI model (e.g., the second AI model 130B) as described herein. The incorporation of the training updates 508 into the HILN network 502 may be automatic under the control of the computer 124 (
Additional reference is made to
Additional reference is made to
While the disclosure is susceptible to various modifications and alternative forms, specific method and apparatus embodiments have been shown by way of example in the drawings and are described in detail herein. It should be understood, however, that the particular methods and apparatus disclosed herein are not intended to limit the disclosure but, to the contrary, to cover all modifications, equivalents, and alternatives falling within the scope of the claims.
This application claims the benefit of U.S. Provisional Patent Application No. 63/219,343, entitled “METHODS AND APPARATUS PROVIDING TRAINING UPDATES IN AUTOMATED DIAGNOSTIC SYSTEMS” filed Jul. 7, 2021, the disclosure of which is hereby incorporated by reference in its entirety for all purposes.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2022/073474 | 7/6/2022 | WO |
Number | Date | Country | |
---|---|---|---|
63219343 | Jul 2021 | US |