The present disclosure relates to methods and apparatus for use in biological specimen testing, and, more particularly, to methods and apparatus for characterizing a specimen container in biological specimen testing.
Automated testing systems may conduct immunoassays or clinical chemistry analysis to identify an analyte or other constituent in a specimen such as blood serum, blood plasma, urine, interstitial liquid, cerebrospinal liquids, and the like. For convenience and safety reasons, these specimens are almost universally contained within specimen containers (e.g., blood collection tubes), which may be capped with a colored cap. Some of the specimen is removed from the specimen container and is subjected to analysis via an assay and/or clinical chemistry analysis. The reactions during the assays or clinical chemistry analysis generate various changes that may be read and/or manipulated to determine a concentration of an analyte or other constituent contained in the specimen, that may, in some embodiments, be suggestive of a patient's disease state.
Improvements in automated testing technology have been accompanied by corresponding advances in pre-analytical sample preparation and handling operations such as centrifugation of specimen containers to separate sample constituents, cap removal (de-capping) to facilitate specimen access, aliquot preparation, and quality checks, which may be used to identify specimen container dimensions, such as height and width, and/or the presence of an interferent such as hemolysis, icterus, or Lipemia (HIL) or the presence of an artifact, such as a clot, bubble, or foam. Such pre-analytical devices may be part of a laboratory automation system (LAS). LAS may automatically transport specimens in specimen containers to one or more pre-analytical sample processing stations on a track, so that various pre-processing operations can be performed thereon prior to performing the analysis.
The LAS may handle a number of different specimens contained in barcode-labeled specimen containers (e.g., tubes). The barcode label may contain an accession number correlated to demographic information from a Laboratory Information System (LIS) along with test orders and other desired information. An operator or robot may place the labeled specimen containers onto the LAS system, which may automatically route the specimen containers along the track for pre-analytical operations, and all prior to subjecting the specimen to an assay or clinical chemistry analysis by one or more analyzers that may be coupled to or part of the LAS.
In such testing systems, the specimen containers presented for analysis may be of varying sizes, such as of differing height and differing widths (e.g., diameters) and identification thereof is desired.
According to a first aspect, the disclosure is directed at a method of determining a location of a specimen container on a track. The method includes providing a calibration object on the track; providing an initially calibrated image capture device adjacent to the track; moving the calibration object to at least two different longitudinal positions along the track including a first longitudinal position and a second longitudinal position, the first longitudinal position being different than the second longitudinal position; capturing a first image with the image capture device with the calibration object located at the first longitudinal position; capturing a second image with the image capture device with the calibration object located at the second longitudinal position; and determining a three-dimensional path trajectory of a center location along the track based at least upon the first image and the second image.
According to another aspect, a characterization apparatus is provided. The characterization apparatus includes a calibration object moveable on a track, a calibrated image capture device located adjacent to the track, and a computer coupled to the calibrated image capture device, the computer configured and operable to cause: the calibration object to move to at least two different longitudinal positions along the track including a first longitudinal position and a second longitudinal position, wherein the second longitudinal position is different from the first longitudinal position, capture a first image with the calibrated image capture device with the calibrated object located at the first longitudinal position, capture a second image with the calibrated image capture device with the calibrated object located at the second longitudinal position, and determine a three-dimensional path trajectory of a center location along the track based at least upon the first image and the second image. A 3D center location of a specimen container stopped anywhere within an imaging area can be determined based on the three-dimensional path trajectory of a center location.
In another aspect, a specimen testing apparatus is provided. The testing apparatus includes a track; specimen carriers moveable on the track, the specimen carriers configured to carry specimen containers; and one or more characterization apparatus arranged around the track, each of the one or more characterization apparatus, comprising: a calibrated image capture device adjacent to the track; and a computer coupled to the calibrated image capture device and configured to: determine a three-dimensional path trajectory of a center location along a segment of the track based at least upon a first image and a second image taken of a calibration object at an imaging area; cause a specimen container carried by a carrier on the track to move to the an imaging area; cause imaging of the specimen container within the imaging area to obtain an container image; determining a center plane between lateral edges of the specimen container; and back-project the center plane to find an intersection point between the center plane and the three-dimensional path trajectory, wherein the intersection point is a three-dimensional center of the specimen container at the position of the center plane.
Still other aspects, features, and advantages of the present disclosure may be readily apparent from the following description by illustrating a number of example embodiments and implementations, including the best mode contemplated for carrying out the present disclosure. The present disclosure may also be capable of other and different embodiments, and its several details may be modified in various respects, all without departing from the scope of the present disclosure. Accordingly, the drawings and descriptions are to be regarded as illustrative in nature, and not as restrictive. The disclosure is to cover all modifications, equivalents, and alternatives falling within the scope of the claims.
The drawings, described below, are for illustrative purposes, and are not necessarily drawn to scale. The drawings are not intended to limit the scope of the disclosure in any way. Like numerals are used throughout the drawings to denote like elements.
FIG. 3D illustrates a lateral image of a specimen container at position along the track in the image area from which the 3D center location of the specimen container at that position can be determined according to one or more embodiments of the disclosure.
Because of tolerance buildups and variability, the exact location of the three-dimensional (3D) center location of the specimen container at locations along the track and specifically at the locations in front of various pre-analytical operations and analyzers may be unknown. Because of the difficulties encountered in determining the exact location of the specimen container in 3D (a center location thereof) and/or the size or type of specimen container, there is an unmet need for methods and apparatus adapted to readily and accurately determine such center location, as well as such sizes.
In particular, at one or more pre-analytical stages, it may be desirable to obtain the various a specimen containers 3D center location and specimen container sizes, since this information can help inform the pre-analytical apparatus (e.g., centrifuge, decapper, aspirator, etc.) of the alignments that should be followed. Furthermore, this acts as a fail-safe in the event of an unsupported tube geometry being introduced into the specimen testing system. In another aspect, when handling of the specimen containers by a robot, knowing the specimen container size and 3D center location can help the robot grippers be properly aligned to the specimen container in 3D space and thus avoid or minimize collisions there between. Further, knowing the specimen container size and 3D center location can aid in lowering pipettes for aspiration at the correct location to avoid specimen container/pipette collisions, and/or faulty aspirations.
In some prior quality check modules configured to assess container size, levels of specimen, and quality of the specimen such as determining levels of HIL therein, three cameras are provided that aid in providing a full 360-degree view the specimen contained in the specimen container. Therefore, it is possible in such conventional quality check modules to reconstruct the specimen container geometry in 3D space, once the cameras have been appropriately calibrated. Once such reconstruction using input from a plurality of cameras is complete, the height and width (e.g., diameter) can be determined fairly accurately.
However, in some systems, having three cameras to accomplish multi-view imaging is impractical from a cost standpoint. Accordingly, the present disclosure, in some embodiments, provides methods, apparatus, and systems that are capable of measuring a geometry of a specimen container and also computing 3D center coordinates of a 3D center location of the specimen container using just one image capture device (e.g., using a single camera). Moreover, in some LAS systems, knowing the location on the 3D center location of the specimen container within a quality check module may not translate into accurate location of the center location at other locations about the track, as accurate track positioning is challenging because of tolerance stack-ups and installation variations. By using a single-image capture device in a characterization apparatus in conjunction with the track that enables motion of the specimen container about the track, a simple and effective method and apparatus is achieved to determine 3D center locations and sizes of specimen containers at any desired location along the track in the LAS.
In some existing testing systems, the specimen container geometry is measured at a quality check module using just one of the multiple cameras in either of the following two ways. In a first method, a specimen container (e.g., tube) of known geometry (such as a cylindrical calibration tool) is moved to a pre-determined position on the track within the quality check module then captures an image. The height (HT) and width (W) are measured in pixels. When a differently sized specimen container is encountered, the differently sized specimen container is moved to the exact same position on the track as before and the height HT and width W thereof can be derived proportionally based on the previously obtained image measurement in pixels. While this method can fairly accurately derive height HT and width W, it cannot however provide a precise 3D center location estimate for the specimen container.
Thus, in a first broad aspect, embodiments of the present disclosure provide characterization methods, characterization apparatus, and specimen testing systems configured and capable of being operated (operable) to determine a 3D center location of the specimen container, as well as physical dimensions of a specimen container, such as width W, and height HT.
In one or more embodiments, the characterization method neither requires the specimen container to move to an exact pre-determined position each time, nor does it require prior computer aided drafting (CAD) information about the geometry of the track. Furthermore, the present characterization method derives a precise estimation of the 3D co-ordinates of the center location of the specimen container (that can be used for robot gripper and pipette alignment tasks) without the need for extremely tight tolerances in the mechanical setup. Finally, the present characterization method does not strictly require the track to be parallel to the image capture device (e.g., camera), and it can even handle cases where the track is slightly slanted or even curved as long as there is no overlapping point on the track along each line of sight from the image capture device.
Knowing the width W of the specimen container can be used for further quantification (e.g., volume) of the various portions of the specimen, such quantification of the volume of the serum or plasma portion, settled blood portion, or both. Height HT of the specimen container may be used by the any robotic system in the testing system to establish a home height for a pipette of a liquid aspiration system in order to minimize specimen container-pipette collisions when moving the pipette to accomplish an aspiration. Knowing the exact 3D center location of the specimen container also allows picking up of the specimen container with robotic grippers while avoiding specimen container-robot gripper collisions. Further, HT and W and the exact 3D center location may be used so to locate and appropriately separate jaws of any robot grippers such that the grippers may appropriately grasp the specimen containers.
According to the disclosure, the characterization method can use a single image capture device (e.g., camera) located along the track at any location where it is useful to know the 3D center location and/or size of the specimen container. For example, a characterization apparatus can be implemented at a location within a quality check module, at an centrifuge station, such as at an a centrifuge pick location thereof, at an aliquoter aspiration/dispense location, at any other robot pick and/or place location on a track, at an analyzer location, or at any other suitable location where a robotic pick or place operation is repeatedly occurring.
The characterization method involves first mapping out the track path at a desired area of interest in three-dimensional (3D) space. For example, the characterization method can involve taking multiple images of the same calibration object (e.g., a calibration tool) at various longitudinal positions along the track at an imaging area, and from those images determine a 3D trajectory of the center along the track. This is done for a calibration object of known geometry, and using this trajectory, the calibration method can map a calibration object's center in three-dimensional space. The present method is applicable to any image capture devoice (e.g., camera) and track setup, such as in automated diagnostics equipment.
In particular, the track carries the specimens in specimen containers to various locations for analysis (e.g., analytical testing or assaying) on carriers and the other locations about the track may use the geometrical dimensions (W and HT) and the 3D center location determined from the quality check module, albeit it may not be fully accurate at those locations. For more accurate 3D center location, one or more characterization apparatus may be included at other locations about the track. Following pre-screening at a quality check module, chemical analysis or assaying may take place on a suitable analyzer. The term “analyzer” is used herein to mean a clinical chemistry analyzer, and/or an assaying instrument, and/or the like. In one embodiment, the quality check module may be provided on the track so that the specimen container may be characterized for dimensions while resident on the track, such as on an input lane of the track or elsewhere along the track.
Further details of inventive characterization methods, characterization apparatus, and testing systems including one or more characterization apparatus will be described with reference to
As shown in
Again referring to
Loading area 105 may serve a dual function of also allowing offloading of the specimen containers 102 from the carriers 122 after processing. Robot grippers of the robot 124 may be configured to grasp the specimen containers 102 from the one or more racks 104 and move and load the specimen containers 102 onto the carriers 122, generally one per carrier 122. In some embodiments, robot 124 can be configured to remove specimen containers 102 from the carriers 122 upon completion of testing. The robot 124 may include one or more (e.g., least two) robot arms or components capable of X and Z (perpendicular to the X-Y plane), Y and Z, X, Y, and Z, or r and theta motion, wherein the robot 124 may be equipped with robotic grippers adapted to pick up and place the specimen containers 102 by grasping the sides thereof. However, any suitable type of robot 124 may be used.
Upon being loaded onto track 121 by robot 124, the specimen containers 102 carried by carriers 122 may progress to a centrifuge station 125 (e.g., an automated centrifuge configured to carry out fractionation of the specimen 212). A characterization apparatus 101 may be provided at a location adjacent to the centrifuge station 125 and track 121, such as where a load/unload robot 126 may pick up the specimen container 102 from the carrier 122 and place it into a centrifuge of the centrifuge station 125. Knowing the exact 3D center location of the specimen container 102 this location (or at any other location about the track) where the carrier 122 stops helps to avoid robot gripper-container collisions that may spill specimen 212 or break the specimen container 102, as the robot knows the exact position of the center location of the specimen container 102 at that location. As will be recognized, a characterization apparatus 101, as described herein, can be used at any location where the 3D center location is desired to be known. For example, a characterization apparatus 101 may be positioned at loading area 105, quality check module 130 (using one or the image capture devices thereof), aliquoting station 131 so as to avoid pipette-container collisions that may spill specimen 212, break the specimen container 102, or pipette, and at one or more of the analyzers 106, 108, 110 to avoid pipette-container collisions or robot-container collisions. Characterization apparatus 101 may be positioned at other locations.
In the depicted embodiment, the carrier 122 can be carried on the track 121 by a cart 324, for example. Cart 324 may be programmed, commanded, or otherwise forced to stop at desired locations along the track 121. Carrier 122 may be removable from the cart 324 and may include any suitable means for registration to the cart 324, such as multiple pins registering in holes, for example. This positions the carrier 122 on the cart 324 in a fixed orientation. In some embodiments, the cart 324 may include an onboard drive motor, such as a linear motor, that is configured to move the specimen container 102 about the track 121, while stopping at desired locations along the track 121 according to programmed instructions. Carriers 122 may each include a holder 122H (
The calibration object 325, as best shown in
Characterization apparatus 101 further includes a calibrated image capture device 328 located at a position adjacent to the track 121, such as along a lateral side thereof. The calibrated image capture device 328 can be calibrated by any suitable means to obtain intrinsic properties (e.g., focal length, image center, skew, and lens distortion coefficients) of the image capture device 328, such as by using standard, automated, calibration techniques (e.g., camera-calibration techniques). These calibration techniques typically involve using printed planar targets (e.g. Hoffman marker grid or checkerboard pattern) of known dimensions and applying iterative refinement techniques to determine the intrinsic parameters of the image capture device 328. Knowing the intrinsic properties of the image capture device 328 is a prerequisite to any 3D imaging task, as it enables estimating a scene's structure in Euclidean space while removing at least some inaccuracies caused by any lens distortion (e.g., that may stem from imperfect lens manufacturing).
Calibrated image capture device 328 can be any combination of focusing lens system and one or more sensors. For example, the calibrated image capture device 328 may be conventional digital camera (e.g., color or monochrome camera), or a charged coupled device (CCD), an array of photodetectors, one or more CMOS sensors, or the like coupled with any suitable focusing lens system. For example, calibrated image capture device 328 may be configured to capture images at multiple different imaging locations (including location A and second location B) along the track 121. The calibrated image capture device 328 may be a device capable of capturing a digital image (i.e., a pixelated image) at the multiple different imaging locations. The image resolution of each image may be about 0.5 MP or more, such as from 0.5MP to 10 MP, for example. Other pixel resolutions may be used. Calibrated image capture device 328 may be a high-speed image capture device, and although desirable to stop the calibration object 325 at first location A and second location B and the carrier 122 at an imaging location, if the speed is fast enough, the images may be taken while the carrier 122 or base 331 and calibration object 325 are still moving.
Characterization apparatus 101 further includes a computer 123 coupled to the calibrated image capture device 328, such as by a USB cable or the like. The computer 123 is configured and operable to cause the calibrated image capture device 328 to capture lateral images of the calibration object 325 at the multiple imaging locations (A and B) along the track 121. As the imaging takes place, the calibration object 325 may be illuminated. For example, the illumination of the calibration object 325 may be by one or more light sources 330A, 330B, such as light panels described in US201/0041318, providing illumination. Light sources may be positioned relative the calibration object 325 so as to illuminate the faces 325A, 325B thereof. For example, light panels may provide front lighting and can be positioned in front of the calibration object 325, and may comprise multiple light sources 330A, 330B positioned on either lateral side of the calibrated image capture device 328, for example. Other positioning and forms of light sources may be used.
In particular, the computer 123, through drive signals to cart 324, can causes the calibration object 325 to move to at least two different longitudinal positions along the track 121 within the imaging area 335 (e.g., a wide angle viewing area) including at least a first longitudinal position A and a second longitudinal position B (position shown dotted) as shown in
According to the method, a three-dimensional path trajectory of a center location 340 along the track 121 is determined based at least upon the first image and the second image. The center location 340 can be at any predetermined height on the calibration object 325 and is determinable in relationship to the imaged location of two or more of the calibrated patterns 325P. In particular, one or more additional images may be taken if the path is other than linear, such as along a curve of the track 121. With the intrinsic camera parameters computed, such as focal length, image center, skew, and possibly the lens distortion coefficients (for more precision), the method can compute a relative extrinsic pose of the 3D center location 350 of the calibration object 325 with respect to the calibrated image capture device 328 for at least the first image and the second image. An algorithm like Perpsective-n-Point can be used to compute the relative extrinsic pose of the 3D center location 350 of the calibration object 325 with respect to the calibrated image capture device 328 for each image, such as for the first image and second image, and any other image that is captured. Perspective-n-Point is the problem of estimating the pose of a calibrated image capture device (e.g., camera) given a set of n three-dimensional points in the world and their corresponding 2D projections in the image. The pose of the image capture device 328 consists of 6 degrees-of-freedom, which are made up of the rotation (roll, pitch, and yaw) and 3D translation (X, Y, Z) of the image capture device 328 with respect to the world. Given a set of n 3D points in a world reference frame and their corresponding 2D image projections as well as the calibrated intrinsic parameters, the 6 DOF pose of the image capture device 328 in the form of its rotation and translation with respect to the world can be determined as follows:
spcK[R|T]pw
where pw=[x y z 1]T is the homogeneous world point, pc=[u V 1]T is the corresponding homogeneous image point, K is the matrix of intrinsic parameters of the image capture device 328, where fx and fy are the scaled focal lengths, γ is the skew parameter, which is sometimes assumed to be 0, and (u0, v0) is the principal point, s is a scale factor for the image point, and R and T are the desired 3D rotation and 3D translation of the image capture device (extrinsic parameters) that are being calculated. This leads to the following equation for the model:
Optionally, P3P can be used when there are n=3 points, or EPnP for n≥4 points. RANSAC can be used if there are outliers.
Now that the three-dimensional path trajectory of a center location 340 along the track 121 is determined, in a next phase, the exact location of the 3D center location 350 of any specimen container 102 that is stopped within the imaging area 335 of the characterization apparatus 101 can be obtained. One particular advantage is that the stopping location of the carrier 122 need not be exact within the imaging area 335, as the method can determine the 3D center location 350 anywhere within the imaging area 335 as long as the sides and top of the specimen container 102 can be viewed/imaged therein. The imaging area 335 is the area that can be imaged by the calibrated image capture device 328. The image area 335 can be at least as tall as the expected specimen containers 102 and can be a wide angle as disclosed herein.
Once the three-dimensional path trajectory of a center location 340 along the track 121 is determined, the 3D center location 350 of the specimen container 102 at the imaging location within the imaging area 335 where the specimen container 102 is imaged is determined. As shown in
When presented with a specimen container 102 in the carrier 122 at the imaging location 333 of the characterization apparatus 101, the method can capture the container image 336 and estimate the 3D center location 350 of the specimen container 102 in the first two dimensions (2D) from the captured container image 336. In the image space, the method first computes the center of the specimen container 102 in the X dimension by determining the locations of a first edge 341 and second edge 342 in pixel space, such as at the same height as the center location 340, wherein the center point trajectory is designated by line 344. The edges 341, 342 can be found by any edge finding routine, such as by raster scanning across the trajectory path 344 to find the abrupt changes in light intensity above a preset threshold value. Raster scanning one or more times above and/or below the trajectory path 344 can be used to confirm that the edges 341, 342 are indeed edges when abrupt intensity changes are found at a same X location in pixel, such as space above and/or below. Once the location of vertical edges 341 and 342 are determined, the center point in 2D space (in the X-Y plane) along trajectory path 344 can be found by adding the two dimensions and dividing by two. The determined 2D centerline is shown as center plane 346. Intersection of trajectory path 344 and center plane 346 includes the 2D center point. The 2D center point can then be mapped to 3D space, i.e., mapped to the closest point on the 3D trajectory (in the Z dimension—into and out of the paper) to determine the 3D center location 350. The 3D center location of 350 is computed by using the 2D image coordinates from the imaging of the specimen container 102 and projecting it into the Z-dimension using the intrinsic parameters of the calibrated image capture device 328. A way to think of this is drawing a line from the center of the image capture device 328 through 350 (in 3D Euclidean space) that extends until infinity, and finding the closest point on the 3D trajectory of 340 (i.e. 344 in 3D Euclidean space) that intersects with it. If there is no intersection, we may choose the point on 344 (in 3D Euclidean space) that minimizes the distance to the Z-projection of 3D center location 350. This point on trajectory path 344 is the 3D center location 350 of the specimen container 102. Thus, robot 126 can accurately know the 3D center location 350 of the specimen container 102 that it will use to pick the specimen container 102 so as to place the specimen container 102 into the centrifuge of the centrifuge station 125, and return the specimen container 102 to the carrier 122 after fractionation.
After fractionation by the centrifuge, the specimen 212 may include, as best shown in
As was indicated above, the carriers 122 may move on to a quality check module 130. Optionally, the centrifugation may occur previously and the specimens 212 contained in specimen containers 102 may be loaded directly into a quality check module 130 that is located at the loading area 105, such as part of an input lane, for example. The quality check module 130 is configured and adapted to automatically determine/characterize physical attributes of the specimen container 102 containing the specimen 212 to be processed by the specimen testing apparatus 100. Characterization may include characterizing tube size, cap type, and/or cap color. Once characterized, the specimen 212 may be further characterized to determine the depth and/or volume of the specimen 212, screened for hemolysis, icterus, or Lipemia (HIL), and/or a presence of one or more artifacts, such as a clot, bubble, or foam. If found to contain no HIL and/or no artifact(s), the specimens 212 may continue on the track 121 and then may be analyzed in the one or more analyzers (e.g., first, second and third analyzers 106, 108, and/or 110) before returning each specimen container 102 to the loading area 105 for offloading.
In some embodiments, quantification of physical attributes of the specimen container 102 may take place at the quality check module 130 (i.e., determining height HT, width W, cap color, cap type, and or tube type). In some embodiments, quantification of the specimen 212 may also take place at the quality check module 130 and may involve determination of HSB, HSP, HTOT, and may determine a vertical location of SB and LA.
The specimen testing apparatus 100 may include a number of sensors 116 at one or more locations around the track 121. Sensors 116 may be used to detect a location of specimen containers 102 along the track 121 by means of reading the identification information 215 (
Embodiments of the present disclosure may be implemented using a computer interface module (CIM) 145 that allows for a user to easily and quickly access a variety of control and status display screens. These control and status screens may describe some or all aspects of a plurality of interrelated automated devices used for preparation and analysis of specimens 212. The CIM 145 may be employed to provide information about the operational status of a plurality of interrelated automated devices as well as information describing the location of any specimen 212 as well as a status of screening or tests to be performed on, or being performed on, the specimen 212. The CIM 145 may be adapted to facilitate interactions between an operator and the specimen testing apparatus 100. The CIM 145 may include a display screen adapted to display a menu including icons, scroll bars, boxes, and buttons through which the operator may interface with the specimen testing apparatus 100. The menu may comprise a number of function buttons programmed to display functional aspects of the specimen testing apparatus 100.
With reference to
In addition to the specimen container quantification, other detection methods may take place on the specimen 212 contained in the specimen container 102 at the quality check module 430. For example, the quality check module 430 may be used to quantify the specimen 212, i.e., determine certain physical dimensional characteristics of the specimen 212 (e.g., a physical location of LA, SB, and/or determination of HSP, HSB, and/or HTOT, and/or a volume of the serum or plasma portion and/or a volume of the settled blood portion.
Again referring to 4A and 4B, the quality check module 430 in an inexpensive form may include a single (one and only one) calibrated image capture device 328 (e.g., single conventional digital camera such as a color or monochrome camera), or a lens system coupled with a charged coupled device (CCD), an array of photodetectors, a CMOS sensor, or the like. For example, a single calibrated image capture device 328 may be configured to capture an image of a specimen container 102 and specimen 212 at an imaging location 333 from a single viewpoint. In this embodiment, the specimen container 102 may be positioned in a rotational orientation so that a clear image of the specimen 212 is possible, such as by a user or a robot determining an unobstructed orientation (orientation unobstructed by label 218) and then inserting the specimen container in the a carrier 122 in that orientation.
This embodiment of the quality check module 430 comprising a single image capture device 328, in addition to determining the geometrical attributes of the specimen container 102 (e.g., width W and height HT), may be used to prescreen for HIL such as is described in U.S. Pat. No. 10,816,538 to Kluckner et al. entitled “Methods and Apparatus for Detecting an Interferent in a Specimen” and/or prescreening for the presence of an artifact, such as is described in U.S. Pat. No. 10,746,665 to Kluckner et al. entitled “Methods and Apparatus for Classifying an Artifact in a Specimen. For example, backlighting using a backlighting source 400C, such as a panelized light source may be used to perform HIL pre-screening.
In one or more embodiments, the characterization method of determining a 3D center location 350 may be undertaken using characterization apparatus 101 as a subcomponent of the quality check module 430. Characterization apparatus 101 includes one or more lighting sources 300A, 300B, calibrated image capture device 328, and calibration object 325 as described above in
In operation, each of the front-lighted and back-lighted images captured by the quality check module 430 may be triggered and captured responsive to a triggering signal. The triggering signal may be generated by the computer 123 and provided in communication lines coupled to the computer 123. Each of the captured images may be processed according to one or more embodiments of the characterization method provided herein. In particular, image processing may be used to determine the width W, height HT. In addition a cap color and cap type may be determined using known methods. Moreover prescreening for HIL and/or the presence of an artifact may be determined, such as using a back-lighted image provided by backlighting with light source 400C.
To improve discrimination, more than one wavelength spectra may be used. Multi-spectral images may then be captured by the image capture device 328. Each of the color spectral images (represented by a nominal wavelength with some relatively narrow wavelength band) is captured, one after another, at one or more exposures (e.g., 4-8 or more exposures). Each exposure may for a different length of time. The spectral images may be taken in any order, such as red at multiple exposures, green at multiple exposures, and blue at multiple exposures. For the detection method, transmittance images may be computed, wherein each transmittance image (for each of R, G, and B illumination) can be computed from optimally-exposed images. The optimally-exposed images may be normalized by their respective per-pixel intensity.
In one or more embodiments, the characterization apparatus 101 and quality check module 430 may include a housing 345 that may at least partially surround or cover the track 121 and provide a closed or semi-closed environment for image capture, such as exterior light influences may be minimized. The specimen container 102 may be located inside the housing 345 during each image capture. Housing 345 may include one or more doors to allow the carrier 122 to enter and/or exit the housing 345. In some embodiments, the ceiling may include an opening to allow a specimen container 102 to be loaded into a carrier 122 stationed inside the housing 345 by a robot (e.g., robot 124) including moveable robot grippers from above, such as when the characterization apparatus 101 and/or quality check module 430 is located at the loading area 105. In cases where front lighting is used and no backlighting (e.g.,
The images may then be further processed to determine segmentation 550 in the manner described in U.S. Pat. No. 10,816,538 to Kluckner et al. and entitled “Methods And Apparatus For Detecting An Interferent In A Specimen” and US2019/0041318 to Wissmann et al. entitled “Methods And Apparatus For Imaging A Specimen Container Using Multiple Exposures.” Other suitable segmentation methods based on artificial intelligence, such as convolutional neural networks (CNN's) may be used. In some embodiments, the images from front-lighting may be best used for segmentation 550. Likewise, the images captured using back-lighting may be best used for HILN classification 552 and/or artifact detection 556 using methods described above.
Liquid quantification 554 may also be carried out following segmentation 550. Quantifying the liquid may involve the determination of certain physical dimensional characteristics of the specimen 212 such as a physical location of LA, SB, and/or determination of HSP, HSB, and/or HTOT, and/or a volume of the serum or plasma portion and/or a volume of the settled blood portion. The identification may be accomplished by selecting the pixels at these demarcation areas and averaging their location values in pixel space to obtain a value for LA and SB. From this information, the volume of the serum or plasma portion 212SP may be determined by using the width W and the cross sectional shape of the specimen container 102. Correlation from pixel space to mechanical measurements may be accomplished by using any suitable calibration method to calibrate pixel space in pixels to mechanical space in mm.
Further characterization of the specimen container 102 may also be accomplished according to the characterization method such as determination of the 3D center location 350. As discussed above, the 3D path trajectory is first determined using a 3D path trajectory determination 551, followed by determination of the 3D center location 350 in 3D center location determination block 553. Tube type detection 558, cap type detection 560, and cap color detection 562 may be achieved based on processing the images from image capture device 328 using conventional methods.
The method 600 further includes, in block 606, moving the calibration object to at least two different longitudinal positions along the track 121 including a first longitudinal position (e.g., longitudinal position A of
Once the three-dimensional path trajectory (three-dimensional path trajectory 344) is known within the imaging area 335, it may be used to determine the 3D center location (e.g., 3D center location 350) of any specimen container 102 brought into the imaging area on carrier 122.
As described in
As part of the edge finding block 706, the method 700 can include identifying a width W of the specimen container 102. Pixel width can simply be converted to distance in mm based upon the calibration of the image capture device 328. Height HT of the specimen container 102 may be determined using a similar edge finding routine wherein the top of the tube 213 at TC is determined. Edge finding may be by segmentation or otherwise looking for transitions in light intensity above a threshold within an area in the imaging area 335 where the tube-cap interface TC may be expected to be located.
In some embodiments, once the specimen container 102 has been given a characterization of size, such as from width W and height HT, a volume of the specimen 212 may be obtained. The inner width may be determined such as by using a lookup table based upon the size of the specimen container 102. The inner width may be used to accurately calculate volume of the serum or plasma portion 212SP and/or the volume of the settled blood portion 212 SB based on the location of the serum-blood interface SB and the liquid-air interface LA obtained from segmentation, for example.
Accordingly, based on the foregoing it should be apparent that the characterization methods 600, 700 carried out by the characterization apparatus 101, which may be included in a quality check module 130, 430 or may be a stand-alone characterization apparatus 101, may result in a rapid characterization of the 3D trajectory path 344 and the 3D center location 350 of the specimen container 102. Physical attributes of the specimen container 102 such as tube size (W and HT), cap type, and cap color can also be obtained using the characterization apparatus 101. In some embodiments including back-lighting, such as shown in
While the disclosure is susceptible to various modifications and alternative forms, specific apparatus embodiments and methods thereof have been shown by way of example in the drawings and are described in detail herein. It should be understood, however, that it is not intended to limit the disclosure to the particular apparatus or methods disclosed but, to the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the scope of the claims and their equivalents.
This application claims the benefit of U.S. Provisional Patent Application No. 63/148,529, entitled “METHODS AND APPARATUS ADAPTED TO IDENTIFY 3D CENTER LOCATION OF A SPECIMEN CONTAINER USING A SINGLE IMAGE CAPTURE DEVICE” filed Feb. 11, 2021, the disclosure of which is incorporated by reference in its entirety for all purposes.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US22/70605 | 2/10/2022 | WO |
Number | Date | Country | |
---|---|---|---|
20240133908 A1 | Apr 2024 | US |
Number | Date | Country | |
---|---|---|---|
63148529 | Feb 2021 | US |