The present disclosure relates generally to the field of medical imaging, and specifically to medical laser speckle imaging.
Laser speckle imaging (LSI) is a promising technique for clinical assessment of tissue perfusion, due to the convenience of the imaging without requiring use of a contrast agent. Tissue perfusion LSI is performed by illuminating a subject tissue area with coherent light and acquiring an image of the laser speckle pattern produced. The speckle pattern signal will fluctuate if there is motion in the subject area, and such fluctuations can be used to detect local motion of blood cells to generate an image of perfusion in the tissue. However, a common issue with the technique is that gross movement of the subject tissue area and/or the LSI device used for imaging introduces errors, or speckle motion artifacts, in the LSI signal, which can make quantitative or comparison applications impractical. The speckle motion artifacts are distinct from blurring or image registration artifacts such as may happen in standard, non-LSI, photography. Because LSI measures motion (blood motion), any motion between the camera and the subject modifies the measured signal. In general, an introduced motion increases the measured LSI signal, with signal change referred to as the speckle motion artifact.
A previous attempt to use a reference patch to provide a measure of the speckle motion artifact in the LSI signal relied on manual designation of the patch location in the acquired images, which may be impractical during routine clinical use and especially if using a handheld LSI device. Furthermore, the previous attempt appeared to teach use of a patch comprising an aluminum layer, which may cause subject discomfort, especially for prolonged patch attachment such as for multiple hours to days for follow-up imaging after clinical treatment.
Other previous work has attempted to reduce speckle motion artifacts in LSI by averaging image frames over several seconds, or by detecting the amount of motion between frames by tracking a fiducial marker on the subject and discarding image frames exceeding a threshold of acceptable motion, but these attempts introduce limitations or discontinuities to the image frame rate that may be undesirable or inadequate in some clinical applications.
Therefore, there is a need for novel and improved modifications to the LSI technique that can facilitate effective and convenient compensation for subject-device motion.
According to some embodiments, a device may be used to facilitate detection of gross movement between the subject and the imaging device during laser speckle imaging. According to some embodiments, the device can be placed within an imaging field of view, near tissue of interest so that both the device and the tissue of interest are imaged simultaneously. According to some embodiments, the device may include a speckle target region that is nontransparent to illumination light in a wavelength range and that generates a speckle pattern when illuminated with light in that range. The speckle pattern generated by the speckle target region may be representative of gross subject-device motion, and thus, the signal measured in this region can constitute a speckle signal motion value that may be used as a reference value for blood flow or perfusion measurements. According to some embodiments, the device can include one or more markers that may facilitate automated visual recognition of the patch location and orientation within images. Thus, the device according to various embodiments can facilitate effective and convenient compensation for subject-device motion.
According to some embodiments, a device for providing reference information in laser speckle medical imaging of tissue of a subject includes a support that includes a marker for image recognition and further includes a speckle target region that is nontransparent to light in a first illumination wavelength range, and that generates a speckle pattern when illuminated with light in the first wavelength range, wherein the support is configured to be attachable to tissue of the subject.
In any of these embodiments, the support can further includes a window region that is transparent to light in the first wavelength range.
In any of these embodiments, the marker can be detectable under near infrared illumination and not detectable under visible light illumination.
In any of these embodiments, the device can further include a coding region including visually coded information.
In any of these embodiments, the coding region can include a barcode.
In any of these embodiments, the coding region can include a QR code.
In any of these embodiments, the visually encoded information can include a unique identity for the device.
In any of these embodiments, the first wavelength range can be in the near infrared range of the spectrum.
In any of these embodiments, the device can further include multiple material layers, wherein one of the material layers forms the part of the speckle target region that generates the speckle pattern, and another of the material layers forms the part of the speckle target region that is nontransparent to light in the first illumination wavelength range.
According to some embodiments, a method for providing reference information for medical imaging of a subject includes placing a device on the subject, the device comprising a support, wherein the a support comprises at least one marker for automated image recognition and a speckle target region that is nontransparent to light in a first illumination wavelength range and that generates a speckle pattern when illuminated with light in the first wavelength range; illuminating the subject, in a first imaging region including the device, with a light within the first illumination wavelength range; acquiring, with an imaging device, a speckle image of the illuminated imaging region; extracting a speckle signal motion value corresponding to the speckle image signal in the speckle target region; adjusting the speckle image by reference to the speckle signal motion value; and displaying the adjusted speckle image to a user on a display.
In any of these embodiments, adjusting the speckle image signal obtained for the subject can include subtracting the speckle signal motion value from the speckle image signal.
In any of these embodiments, adjusting the speckle image signal obtained for the subject can include normalizing the speckle image signal with respect to the speckle signal motion value.
In any of these embodiments, adjusting the speckle image signal obtained for the subject can include scaling the speckle image signal by a scaling factor and subtracting an offset value, the scaling factor and offset value being based on the speckle signal motion value.
In any of these embodiments, the method can further include acquiring a sequence of speckle images of illuminated imaging regions that overlap the first illuminated imaging region; calculating the relative positions and orientations of the sequence of speckle images based on the relative positions and orientations of the at least one markers; and correcting for relative motion between the imaging device and the subject by aligning the sequence of speckle images based on their relative positions and orientations, through one or more image transformations of translation, rotation, and scaling.
In any of these embodiments, the method can further include extracting a sample speckle image signal from a sample region that is in a fixed, predetermined position relative to the device.
In any of these embodiments, the device can further include a window region that is transparent to light in the first wavelength range, wherein the sample region is within the window region.
In any of these embodiments, the device can further include a coding region including visually coded information.
In any of these embodiments, the coding region can include a barcode.
In any of these embodiments, the coding region can include a QR code.
In any of these embodiments, the visually coded information can include a unique identification for the device.
In any of these embodiments, the first wavelength range can be in the near infrared range of the spectrum.
According to some embodiments, a method of assembling a device for providing reference information in medical imaging, wherein the device is attachable to tissue of an imaging subject, can include providing, on a first layer, a first speckle target region that is nontransparent to light in a first illumination wavelength range; providing, on a second layer, a second speckle target region that generates a speckle pattern when illuminated with light in the first wavelength range; arranging the second layer on top of the first layer, wherein the first speckle target region is aligned with the second speckle target region; and attaching the first layer to the second layer.
In any of these embodiments, can include providing a coding region including visually coded information on the first material layer, the second material layer, or both.
In any of these embodiments, the method can include providing a marker for automated image recognition on the first material layer, the second material layer, or both.
In any of these embodiments, the first speckle target region can absorb at least 90 percent of light intensity in the first illumination wavelength range.
In any of these embodiments, the method can further include providing, on the first material layer, a first window region that is transparent to light in the first wavelength range; and providing, on the second material layer, a second window region that is transparent to light in the first wavelength range, wherein the second window region is located such that it is aligned with the first window region when the first speckle target region is aligned with the second speckle target region.
In any of these embodiments, the first illumination wavelength range can be in the near infrared.
According to some embodiments, a device for providing reference information in medical imaging of a subject can include a support comprising a first speckle target region that absorbs at least 90 percent of light intensity in a first illumination wavelength range and a second speckle target region, aligned with the first, that generates a speckle pattern when illuminated with light in the first wavelength range, wherein the support is configured to be attachable to tissue of the subj ect.
In any of these embodiments, the support can further include a window region that is transparent to light in the first wavelength range.
In any of these embodiments, the support can further include a marker for automated image recognition.
In any of these embodiments, the marker can be detectable under near infrared illumination and not detectable under visible light illumination.
In any of these embodiments, the support further include a coding region including visually coded information.
In any of these embodiments, the coding region can include a barcode.
In any of these embodiments, the coding region can include a QR code.
In any of these embodiments, the visually coded information can include a unique identity for the device.
In any of these embodiments, the first wavelength range can be in the near infrared range of the spectrum.
In any of these embodiments, the support can include multiple material layers, and wherein a first one of the material layers includes the first speckle target region and a second one of the material layers includes the second speckle target region.
According to some embodiments, a method for compensating for speckle motion artifacts in medical laser speckle imaging of a target tissue area, includes providing a series of laser speckle images of the target, the series of images acquired by a laser speckle imaging device; calculating a measure of relative motion between a first image in the series and a second image in the series; and correcting the image data of the second image to compensate for speckle motion artifacts by scaling the image data by a scaling factor and by subtracting an offset value from the image data, wherein the scaling factor and the offset value are based on the calculated measure of relative motion.
In any of these embodiments, the series of images can include at least 30 images.
In any of these embodiments, the series of images can be acquired during a continuous period of time.
In any of these embodiments, the continuous period of time can be at least 2 seconds.
In any of these embodiments, calculating the measure of relative motion can be based on a measured image signal from any one of devices described above, wherein the device is attached to tissue in the target tissue area.
In any of these embodiments, calculating the measure of relative motion can be based on measurements by a motion sensor mounted to the laser speckle imaging device, the measurements having been acquired simultaneously with the acquisition of the series of images.
In any of these embodiments, calculating the measure of relative motion can be based on calculating, for each image in the series of laser speckle images, a representative value for the laser speckle image signal within a region of interest.
In any of these embodiments, calculating the measure of relative motion can include determining the minimum representative value among the representative values for all images in the series of laser speckle images; and calculating the difference between the representative value for each image in the series of laser speckle images and the minimum representative value; wherein the measure of relative motion for each image in the series of laser speckle images is based on the calculated difference for that image.
In any of these embodiments, the region of interest can include about one quarter of the total number of pixels in each image in the series of laser speckle images, and can include pixels at the center of each image and excludes pixels at the edges of each image.
In any of these embodiments, calculating the measure of relative motion can be based on calculating, for each image in the series of laser speckle images, a first representative value for the laser speckle image signal within a first region of interest and a second representative value within a second region of interest.
In any of these embodiments, calculating the measure of relative motion can include determining a minimum first representative value among the first representative values for all images in the series of laser speckle images; calculating a first difference between the first representative value for each image in the series of laser speckle images and the minimum first representative value; determining the minimum second representative value among the second representative values for all images in the series of laser speckle images; and calculating a second difference between the second representative value for each image in the series of laser speckle images and the minimum second representative value; wherein the measure of relative motion for each image in the series of laser speckle images is based on the calculated first and second differences for an image.
In any of these embodiments, calculating the measure of relative motion can further include calculating an intermediate representative value in an intermediate region of interest between the first and second regions of interest, wherein the intermediate representative value is interpolated based on the first and second representative values; determining the minimum intermediate representative value among the intermediate representative values for all images in the series of laser speckle images; and calculating a third difference between the intermediate representative value for each image in the series of laser speckle images and the minimum intermediate representative value.
In any of these embodiments, each representative value can be a mean value of pixel signal values within the respective region of interest.
In any of these embodiments, each representative value can be a percentile value of pixel signal values within the respective region of interest.
In any of these embodiments, each representative value can be the 20th percentile value of pixel signal values within the respective region of interest.
In any of these embodiments, signal values below a first threshold value can be excluded from the calculation of each representative value.
In any of these embodiments, pixel signal values above a second threshold value can be excluded from the calculation of each representative value.
In any of these embodiments, the method can further include acquiring a second series of laser speckle images of the target, the second series of images acquired by the laser speckle imaging device subsequently to acquisition of the first and second images, wherein correcting the image data of the second image is performed while acquiring the second series of images.
In any of these embodiments, the scaling factor and the offset used for correcting the image data of the second image to compensate for speckle motion artifacts can satisfy the following equations: SF=1/(1+A×i) and O=SF×(B×i+C), wherein SF is the scaling factor, A is a first calibration constant, i is the measure of relative motion for the second image, 0 is the offset, B is a second calibration constant, and C is a third calibration constant.
According to some embodiments, a method for compensating for speckle motion artifacts in medical laser speckle imaging of a target tissue area, includes providing a series of laser speckle images of the target, the series of images acquired by a laser speckle imaging device; calculating a measure of relative motion between a first image in the series and a second image in the series, wherein calculating the measure of relative motion is based on calculating, for each image in the series of laser speckle images, a first representative value for the laser speckle image signal within a first region of interest; and correcting the image data of the second image to compensate for speckle motion artifacts by subtracting an offset value from the image data, wherein the offset value is based on the calculated measure of relative motion.
In any of these embodiments, the series of images can include at least 30 images.
In any of these embodiments, the series of images can be acquired during a continuous period of time.
In any of these embodiments, the continuous period of time can include at least 2 seconds.
In any of these embodiments, calculating the measure of relative motion can include determining the minimum first representative value among the first representative values for all images in the series of laser speckle images; and calculating the difference between the first representative value for each image in the series of laser speckle images and the minimum first representative value; wherein the measure of relative motion for each image in the series of laser speckle images is based on the calculated difference for that image.
In any of these embodiments, the first region of interest can include about one quarter of the total number of pixels in each image in the series of laser speckle images, and can include pixels at the center of each image and excludes pixels at the edges of each image.
In any of these embodiments, calculating the measure of relative motion can be further based on calculating, for each image in the series of laser speckle images, a second representative value for the laser speckle image signal within a second region of interest.
In any of these embodiments, calculating the measure of relative motion can include determining the minimum first representative value among the first representative values for all images in the series of laser speckle images; calculating a first difference between the first representative value for each image in the series of laser speckle images and the minimum first representative value; determining the minimum second representative value among the second representative values for all images in the series of laser speckle images; and calculating a second difference between the second representative value for each image in the series of laser speckle images and the minimum second representative value; wherein the measure of relative motion for each image in the series of laser speckle images is based on the calculated differences for that image.
In any of these embodiments, calculating the measure of relative motion can further include calculating an intermediate representative value in an intermediate region of interest between the first and second regions of interest, wherein the intermediate representative value is interpolated based on the first and second representative values; determining the minimum intermediate representative value among the intermediate representative values for all images in the series of laser speckle images; and calculating a third difference between the intermediate representative value for each image in the series of laser speckle images and the minimum intermediate representative value.
In any of these embodiments, each representative value can be a mean value of pixel signal values within the respective region of interest.
In any of these embodiments, each representative value can be a percentile value of pixel signal values within the respective region of interest.
In any of these embodiments, each representative value can be the 20th percentile value of pixel signal values within the respective region of interest.
In any of these embodiments, pixel signal values below a first threshold value can be excluded from the calculation of each representative value.
In any of these embodiments, pixel signal values above a second threshold value can be excluded from the calculation of each representative value.
In any of these embodiments, the method can further include acquiring a second series of laser speckle images of the target, the second series of images acquired by the laser speckle imaging device subsequently to acquisition of the first and second images, wherein correcting the image data of the second image is performed while acquiring the second series of images.
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawings(s) will be provided by the Office upon request and payment of the necessary fee. Features will become apparent to those of ordinary skill in the art by describing in detail exemplary embodiments with reference to the attached drawings in which:
Reference will now be made in detail to implementations and embodiments of various aspects and variations of the invention, examples of which are illustrated in the accompanying drawings. Although several exemplary variations of the systems and methods are described herein, other variations of the systems and methods may include aspects of the systems and methods described herein combined in any suitable manner having combinations of all or some of the aspects described.
In the following description of the various embodiments, reference is made to the accompanying drawings, in which are shown, by way of illustration, specific embodiments that can be practiced. It is to be understood that other embodiments and examples can be practiced, and changes can be made without departing from the scope of the disclosure.
In addition, it is also to be understood that the singular forms “a,” “an,” and “the” used in the following description are intended to include the plural forms as well, unless the context clearly indicates otherwise. It is also to be understood that the term “and/or” as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items. It is further to be understood that the terms “includes, “including,” “comprises,” and/or “comprising,” when used herein, specify the presence of stated features, integers, steps, operations, elements, components, and/or units but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, units, and/or groups thereof.
Certain aspects of the present disclosure include process steps and instructions described herein in the form of an algorithm. It should be noted that the process steps and instructions of the present disclosure could be embodied in software, firmware, or hardware and, when embodied in software, could be downloaded to reside on and be operated from different platforms used by a variety of operating systems. Unless specifically stated otherwise as apparent from the following discussion, it is appreciated that, throughout the description, discussions utilizing terms such as “processing,” “computing,” “calculating,” “determining,” “displaying,” “generating” or the like, refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system memories or registers or other such information storage, transmission, or display devices.
The present disclosure in some embodiments also relates to a device and/or system for performing the operations herein. This device and/or system may be specially constructed for the required purposes, or it may comprise a general purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a non-transitory, computer readable storage medium, such as, but not limited to, any type of disk, including floppy disks, USB flash drives, external hard drives, optical disks, CD-ROMs, magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, application specific integrated circuits (ASICs), or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus. Furthermore, the processors referred to in the specification may include a single processor or may be architectures employing multiple processor designs for increased computing capability.
The methods, devices, and systems described herein are not inherently related to any particular computer or other apparatus. Various general-purpose systems may also be used with programs in accordance with the teachings herein, or it may prove convenient to construct a more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will appear from the description below. In addition, the present invention is not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein.
Referring to
Optics 142 such as, for example, an engineered diffuser and/or lens system, may be used to shape or otherwise modify illumination of a region on the tissue 150. In some variations, the image sensor 110 may be a CMOS or CCD type color image sensor and may acquire light from the illuminated part of the object or region of interest through a lens or objective 112. In some variations, the image sensor 110 may be a CMOS or CCD type grayscale image sensor. A filter 113 (e.g., a long-pass or band-pass filter) may be used to filter light so as to only acquire the wavelength of the coherent illumination source. In the variation shown in
The control unit 210 may control the image acquisition, and in some variations, may pulse coherent light from the coherent light source 141 (e.g., laser). In some variations, the control unit 210 may control or adjust the exposure time of the image sensor 110. The exposure time should be sufficiently long such that the moving speckle pattern blurs the speckle image, but short enough such that the expected moving speeds can be differentiated. In some variations, the exposure time may be in the range of about 1 ms to about 10 ms. In some variations, the exposure time is at least 1 ms, at least 2 ms, at least 3 ms. In some variations, the exposure time is less than 20 ms, less than 10 ms, less than 8 ms. These exposure times may be desirable for measuring the perfusion in human skin. Furthermore, the exposure time may be fixed or adjustable depending on the object or region of interest.
The acquired speckle images may be transferred to the processing unit 220, which may calculate the perfusion images. In some variations, the calculation process may include calculating a contrast image and converting the contrast image to a perfusion image. The contrast image is calculated by dividing the mean pixel values by the standard deviation of the pixel values in a small area (kernel) of the image. The kernel size may be, for example, 7×7 to 15×15 pixels. The process is calculated for many kernels in the image. Generally, the contrast image may be converted to a perfusion image by relying on an inversely proportional relationship between the square of the speckle contrast and the flow parameters (e.g., speed, concentration, etc.) of particles (e.g., red blood cells in capillaries). One skilled in the art will appreciate that the systems, devices, and methods described herein shall not be limited to the details of the processing algorithm.
In some embodiments the processing algorithm is used as follows: Speckle contrast (K) is defined in terms of the standard deviation (σs) and the mean speckle intensity fluctuations I. While the contrast and mean intensity could also be determined in time-domain, the processing in spatial-domain has advantages in reducing motion artefacts. The relation between speckle contrast and spatial variance of the time-averaged speckle pattern is:
where g(τ) is the normalized autocorrelation function, τ is the lag time, and T is the camera exposure time. The exposure time is often selected in the range of 1-10 ms. Assuming a Lorentzian spectrum the relationship can be rewritten as:
where τc is the correlation time and
is proportional to the mean velocity V. This can be further approximated to:
While V has a linear response to velocity within the assumed range limited by the exposure time, it also depends on the concentration of the moving scatterers vs. the static scatterers (i.e. the concentration of blood). Thus V is often also referred as LSI value, flow or perfusion.
The perfusion image may be shown on the user interface 230. The user interface 230 may include, for example, a display and input means such as mouse, keyboard, buttons, or touch screen. In some variations, the sequence of some or all of the optics elements such as the aperture 111, the lens 112, the filter 113, or a combination thereof may be rearranged.
An adhesive patch may be used to provide reference information during laser speckle imaging, and may facilitate detection of gross movement between the subject and the imaging device. The patch may also facilitate image registration between successive image frames in one session, or between images from separate imaging sessions, by serving as a highly visible fiducial marker. The patch may be placed within the imaging field of view and near to the tissue of interest, so that both may be imaged simultaneously. The size of the patch may be any size that is appropriate for a particular application such that the patch is large enough to be reliably imaged but not so large that it obscures imaging of the tissue of interest.
A portion of the patch 300 may comprise a speckle target region 320 that is nontransparent to illumination light from an imaging device in a first wavelength range and that generates a speckle pattern when illuminated with light in that range. In some embodiments, the first wavelength range may be in the near infrared. In some variations, the first wavelength range may be about 750 to 850 nm. Because the fixed adhesive patch does not feature any moving parts, the only motion that the laser speckle signal corresponding to the speckle target region 320 will be representative of is gross subject-device motion, and thus the signal measured in this region can constitute a speckle signal motion value that may be used as a reference value for blood flow or perfusion measurements. For example, the laser speckle signal of an image acquired by the imaging device may be adjusted by reference to the laser speckle signal of the speckle target region 320, the speckle signal motion value, such as by subtracting the speckle signal motion value and/or by normalizing the speckle image signal with respect to the speckle signal motion value. In some embodiments, the speckle target region 320 may comprise multiple unconnected areas on the patch 300.
A portion of the patch 300 may comprise a data region 350 that may contain encoded data. For example, data region 350 may contain visually encoded data such as a barcode or a QR code. In some variations, the visually encoded data in data region 350 may be visible when illuminated by light in the visible wavelength range and/or by light in the near infrared range. In some variations, the visually encoded data in data region 350 may be visible when illuminated by light in the near infrared range, and not visible when illuminated by light in the visible wavelength range.
The patch may also contain Near Field Communication (NFC) means, such as RFID, and data storage capability for NFC data. The NFC data may contain the same or different data as the data region 350 or it may replace the data region.
A portion of the patch 300 may comprise a skin reference region 330 which may provide a window region through which to obtain a reference measure of the LSI signal for comparison with measures of the LSI signal in other subject areas of interest. In some embodiments, a skin reference region 330 may be formed by a cutout or hole formed in a region of the patch 300. In some embodiments, a skin reference region 330 may be formed by a material that is transparent to light in the first wavelength range. In some variations, a region that is outside of the boundaries of the patch 300 and that is located at a predetermined location relative to the patch 300 may be used as a skin reference region.
The skin reference region 330 may be used for indicating a consistent reference zone for intra-individual relative imaging such as the intra-individual relative imaging disclosed in WO/2013/160861.
In some embodiments, an adhesive patch for use with laser speckle imaging may comprise multiple material layers. For example,
In some embodiments, a patch 400 may be partially constructed and/or manufactured at or before the time of use by combining a speckle generating layer 420 with an illumination blocking layer 440. For example, an illumination blocking layer 440 may be placed on a subject and then a speckle generating layer 420 may be placed on the illumination blocking layer 440 to form a patch 400. As another example, a speckle generating layer 420 may be placed on an illumination blocking layer 440 to form a patch 400, and then the patch may be placed on a subject. In some variations, the illumination blocking layer 440 may have two adhesive surfaces. In some variations, the illumination blocking layer 440 may have an adhesive surface and the speckle generating layer 420 may have an adhesive surface.
In some embodiments, an adhesive patch may include an adhesive that is suitable and biocompatible for multiple days of use, such as is used in typical dermal or transdermal patches, so that the patch may be used in one or more follow-up imaging sessions after an initial imaging session. For example, the patch may be used to assist with imaging and perfusion assessment during a surgical session such as a surgery involving a skin flap, and may also be used during imaging in one or more post-operative follow-up sessions tracking progress of the skin flap and surrounding tissue. Use of the same patch for multiple imaging sessions may facilitate precise image registration and/or alignment between images obtained in separate sessions, by using the patch as a clearly visible fiducial marker. Use of the same patch may also facilitate automatic recognition of the subject identity and the subject's imaging history and other notes associated with the subject and triggering instructions in the image processor, including, for example, automatically loading the same imaging settings as used in a previous imaging session, and recording images to the same imaging history as used in a previous imaging session. Such automatic recognition and triggering of image processor instructions may be facilitated by recognition of a unique patch identification that may be stored in the patch data region 350.
In some variations the unique patch identification is a serial number. In yet another variation the unique patch identification is a numerical or alpha-numerical code which is unique and encoded in such a way that every code can only be generated with specific know-how (e.g. by use of a symmetrical or asymmetrical encryption or signature). Such unique patch identification could be used as payment confirmation for a pay-per-use model.
One or more other techniques of speckle motion artifact detection may be used alternatively to, or in combination with, use of an adhesive patch for such detection. For example, image processing may be performed to process multiple image frames, calculate a measure of relative motion (e.g. motion between the camera and the subject) occurring between image frames, and estimate the speckle motion artifacts due to such motion. As a further example, motion sensors such as accelerometers and/or gyroscopes may be used to measure motion parameters of the imaging device and to estimate the speckle motion artifacts due to such motion.
In some embodiments, a measure of relative subject-device motion between image frames may be based on calculation of representative values within a region of interest (ROI) in the images.
In some embodiments, in order to calculate a measure of relative subject-device motion between images in a series of images, a representative value may be calculated for each image in the series and the minimum representative value among the representative values for all images in the series may be determined. The difference between the representative value for each image in the series and the minimum representative value may then be calculated and used as the basis for the measure of relative subject-device motion for each image. In some embodiments, the number of images included in the series of images for these calculations may represent 1, 2 or 3 seconds of image acquisition (with a frame rate of 30 fps, this corresponds to at least about 30, at least about 60, or at least about 90 images). In some variations the frame rate may be 60 fps or 120 fps.
If it can be assumed that the operator tries to keep the camera stable, at least one frame in the series of images is typically quite stable, thus the difference between the representative values allows determining the LSI value without motion, or with minimal motion, for at least one frame. This information can then be used to approximate the motion for any other frame by comparing representative values. One example is given in
In some embodiments, the calculation of a representative value within a ROI may be performed on each acquired single image frame in order to have the highest possible temporal resolution for characterizing a measure of relative subject-device motion, while one or more single image frames may subsequently be averaged for display.
In some embodiments, the representative value may be chosen to yield a value of the baseline perfusion in the ROI. For example, a percentile value, mode, mean, or any other statistic with regard to the distribution of LSI values for the pixels within the ROI may be used. With regard to a typical histogram of such a distribution of LSI values, such as displayed in
In some embodiments, it may be useful to exclude some pixels in the ROI from the calculation of the representative value, in order to better represent the baseline perfusion in the ROI. For example, pixel values below a first threshold value may be excluded from calculation of the representative value. For example, the pixel values before the first threshold value may include pixel values at or near zero and/or other relatively low pixel values. Additionally or alternatively, pixel values above a second threshold value may be excluded from calculation of the representative value. For example, the pixel values above the second threshold value may include saturated pixel vales and/or other relatively high pixel values.
In one example the representative value may be calculated as the 20th percentile value of the LSI value V within a centered ROI that covers ¼ of the total image before any time-domain filtering is performed. In other examples the representative value may be calculated as other percentile values, for example, the 25th percentile value of the LSI value V within a centered ROI that covers ¼ of the total image before any time-domain filtering is performed.
In some embodiments, multiple ROIs may be specified within each image, and a measure of relative subject-device motion may be calculated for each ROI, so that different relative motion in different parts of the image may be separately corrected for. Instead of calculating a single representative value for each image as described above, representative values may be calculated for each ROI within the image. After calculating the representative values in the multiple ROIs, one or more intermediate representative values may be calculated in one or more intermediate ROIs between the multiple ROIs by interpolation based on the representative values in the multiple ROIs and the relative positions of the multiple ROIs and the one or more intermediate ROIs.
In some embodiments the series of images may be first registered to align showing the same target area before the calculation of relative motion is performed. Registration methods are known to those skilled in the art. The registration may be based on feature detection, patch detection or motion/gyro sensors.
In case a user of a handheld laser speckle imaging device wishes to save an image of subject tissue near to a particular time point while imaging, the best candidate image for saving may be selected automatically from a series of recently acquired images based on a measure of motion or speckle motion artifacts, such as any of the measures described herein, so as to select the candidate image suffering from the smallest amount of speckle motion artifacts. Such a method for automatically selecting the best candidate image may be implemented without, or in combination with, methods of speckle motion artifact compensation on the series of acquired images, such as any of the methods of speckle motion artifact compensation described herein. In some embodiments, one or more of the images in the series of recently acquired images may be calculated from an average of a number of nearby images, in order to reduce speckle motion artifacts and/or noise.
In one embodiment the selection of the motion optimized frames 520 is done by calculating the mean representative value for any sequence of N consecutive frames and selecting the sequence of N consecutive frames which has the mean representative value. The candidate image for saving is then calculated as the pixel-by-pixel time-domain averaged image of the selected N consecutive frames. The choice of N may be made depending on the frame-rate. For 30fps, it may be chosen in the range of 1 to 30 frames, for example.
One or more of the methods and systems described herein for calculating a measure of relative subject-device motion between laser speckle images may be combined with a method of speckle motion artifact compensation in order to compensate for speckle motion artifacts in the images.
Use of an adhesive patch such as any of those described herein may permit correction for speckle motion artifacts in the laser speckle image signal, because the adhesive patch may provide an image signal region that is purely due to motion between the imaging device and the subject surface. In some embodiments, it may be assumed that the measured LSI signal over regions with exposed subject skin includes a sum of a first LSI signal due to motion of blood cells in perfused tissue beneath the skin and a second LSI signal due to motion between the imaging device and the subject skin. It may also be assumed that the measured LSI signal over a speckle target region on an adhesive patch is primarily composed of only the LSI signal due to motion between the imaging device and the subject skin. Therefore, to compensate for speckle motion artifacts that may distract from the blood cell LSI signal of interest, the measured LSI signal over a speckle target region on an adhesive patch may be subtracted from the entire LSI signal. In some embodiments, if a material is used on the exposed surface of the target speckle region of the patch that results in a different signal contribution from gross subject-device motion than would be measured on subject tissue, then calibration may be performed in order to determine an appropriate calibration factor to scale the target speckle region signal before subtracting the calibrated result from the subject tissue signal. For example, calibration may be performed by placing a patch in the same imaging field of view as a tissue region known to be free of any perfusion or blood flow, such as a cadaver specimen, and measuring the LSI signal on the target speckle region of the patch and in the tissue region during subject-device motion.
Output from one or more of the methods described herein for calculating a measure of relative subject-device motion between image frames may also be used to compensate for speckle motion artifacts by subtracting an offset value from the image data of each image, with the offset value being based on the calculated measure of motion for each image. For example, the offset value may represent an estimate of the LSI signal due to subject-device motion.
Additionally or alternatively to subtracting an estimate of the LSI signal due to motion from the measured LSI signal to correct for speckle motion artifacts, other operations may be performed on the measured LSI signal based on a measure of relative subject-device motion between laser speckle images to correct for speckle motion artifacts. For example, in some applications it may be observed that speckle motion artifacts result in a scaling of laser speckle image data in combination with subtracting an offset from the laser speckle image data, and in these cases correction of the laser speckle image data may be performed by scaling the image data by a scaling factor and by subtracting an offset value from the image data based on a calculated measure of relative subj ect-device motion.
For example, the following equations demonstrate how a scaling factor (SF), and offset (O) value may be calculated for a given measure of relative subject-device motion (i). The speckle motion artifact (m), in the image may be assumed to be comprised of a motion offset component (mo), and a motion scaled component (ms), which are equivalent to a function of the true perfusion value (y) multiplied by the measure of relative subject-device motion (i).
m=mo+ms=f(y)·i
The true perfusion value (y), may be calculated from the measured perfusion value (p), by subtracting the device offset and the speckle motion artifact.
y=p−DO−m=p−DO−f(y)·i
f(y) may be approximated as a linear function.
f(y)=Ay+B
Then, the real perfusion value (y) may be rewritten as:
The scaling factor (SF) may be calculated as:
The offset (0) to be subtracted may be calculated as:
O=SF·(B·i+DO)
The constant parameters A, B and DO may be experimentally calibrated. DO may be determined, for example by measuring the perfusion value in a very stable setup and imaging a completely static target without any perfusion. DO may be taken as the lowest response that the system can measure, and it largely depends on the optical setup and the image sensor noise. A and B may be tuned such that in an image that has both high and low perfused regions, both the high and low perfused regions are stable despite introduced motion. Images may be acquired in stable and motion-introduced conditions. All those images may be processed to calculate the real perfusion value (y). A and B can be tuned so that both high and low perfused regions keep their real perfusion value in all those images. In one approach B may be calibrated first and A may then be used to further tune the correction. In some embodiments, for example, A may be set to around 0.00016, B may be set to around 0.78, and DO may be set to around 150.
Considering that the scaling factor (SF) and offset (O) are each functions of the measure of motion (i), the motion corrected perfusion (y) for any given frame can then be calculated from the measured perfusion (p) as:
y=SF(i)·p−O(i)
The measure of motion (i) may be determined in different ways as described in this invention. If the motion is determined by finding the minimum representative value, the measure of motion (i) is the difference between the representative value of the current frame (ra) and the minimum representative value (rm). In that case the measure of motion (i) is calculated as:
i=(ra−rm)
The example above should not limit the exact way of how the motion compensation is performed. In a generic approach, the real value (y) is a function of the measured perfusion (p) and the measure of motion (i).
y=f(p,i)
In some embodiments, any combination of one or more of the methods described herein for calculating a measure of relative subject-device motion and correcting image data to compensate for speckle motion artifacts may be performed in real time. For example, in some variations a subsequent second series of images may be acquired by the laser speckle imaging device while calculations for correcting the image data from a first series of images are being performed. As another example, in some variations after performing calculations for correcting the image data from a series of images the series may be modified by removing one or more images from the series and acquiring one or more new images, adding the new images to the modified series, and then correcting the modified series of images. In such real time image correction embodiments, some computational effort may be saved by reusing representative values already calculated for any image frames that are carried forward from a first series of images to a second or to a modified series of images.
In some embodiments, any combination of one or more of the methods described herein for calculating a measure of relative subject-device motion and correcting image data to compensate for speckle motion artifacts may be performed in post-processing after a sequence of images has been acquired and saved.
In some embodiments, any combination of one or more of the methods described herein for calculating a measure of relative subject-device motion and correcting image data to compensate for speckle motion artifacts may be performed to correct the image data on a pixel by pixel basis for each pixel in the image data.
In some embodiments, any combination of one or more of the methods described herein for calculating a measure of relative subject-device motion and correcting image data to compensate for speckle motion artifacts may be performed to correct a subset of pixels within the image data.
In some embodiments the image data may be registered to compensate for any displacement that accrued due to motion. Registration methods are known to those skilled in the art. The registration can be based on feature detection, patch detection or motion/gyro sensors.
The methods, embodiments, and systems described herein may be used for blood flow imaging, tissue perfusion imaging, or a combination thereof, which may be performed during an invasive surgical procedure, a minimally invasive surgical procedure, a non-invasive surgical procedure, a non-invasive and non-surgical clinical procedure and/or assessment, or a combination thereof. Examples of invasive surgical procedures which may involve blood flow and tissue perfusion include a cardiac-related surgical procedure (e.g., CABG on pump or off pump) or a reconstructive surgical procedure. An example of a non-invasive or minimally invasive procedure includes wound (e.g., chronic wound such as for example pressure ulcers) treatment and/or management. In this regard, for example, a change in the wound over time, such as a change in wound dimensions (e.g., diameter, area), or a change in tissue perfusion in the wound and/or around the peri-wound, may be tracked over time with the application of the methods and systems. An example of a non-invasive and non-surgical clinical assessment includes wound assessment, for example to assist with perfusion imaging in and around the wound to facilitate ascertaining what course of treatment may or may not be required.
It will be appreciated that any options mentioned in view of any of the methods, embodiments, and systems may be used in conjunction with the other methods, embodiments, and systems, and vice versa. It will be appreciated that any of the options may be combined. It will be appreciated that any of the aspects may be combined. It will be appreciated that any of the embodiments and/or variations may be combined with the methods, embodiments, and systems described herein.
Several figures (
Comparison of the uncorrected image (
Attempting to hold the laser speckle imaging device still in one hand is expected to yield somewhat more subject-device motion compared to holding still with two hands. Comparison of the uncorrected image (
The three handheld device scenarios with slow device motion introduced are intended to yield more subj ect-device motion compared to the other conditions, in order to assess how a motion correction method with use of a patch may perform in the presence of such device operator hand motions. Comparison of the uncorrected image (
While the present disclosure has been illustrated and described in connection with various embodiments shown and described in detail, it is not intended to be limited to the details shown, since various modifications and structural changes may be made without departing in any way from the scope of the present disclosure. Various modifications of form, arrangement of components, steps, details and order of operations of the embodiments illustrated, as well as other embodiments of the disclosure may be made without departing in any way from the scope of the present disclosure, and will be apparent to a person of skill in the art upon reference to this description. It is therefore contemplated that the appended claims will cover such modifications and embodiments as they fall within the true scope of the disclosure. For the purpose of clarity and a concise description, features are described herein as part of the same or separate embodiments; however, it will be appreciated that the scope of the disclosure includes embodiments having combinations of all or some of the features described. For the terms “for example” and “such as,” and grammatical equivalences thereof, the phrase “and without limitation” is understood to follow unless explicitly stated otherwise. As used herein, the singular forms “a”, “an”, and “the” include plural referents unless the context clearly dictates otherwise.
This application claims the benefit of U.S. Provisional Application No. 62/955,304, filed Dec. 30, 2019, the entire contents of which are hereby incorporated by reference herein.
Number | Date | Country | |
---|---|---|---|
62955304 | Dec 2019 | US |