Ultrasound imaging system having near-infrared/infrared detection

Information

  • Patent Grant
  • 12287403
  • Patent Number
    12,287,403
  • Date Filed
    Friday, April 15, 2022
    3 years ago
  • Date Issued
    Tuesday, April 29, 2025
    3 days ago
Abstract
Disclosed herein is an ultrasound imaging system including an ultrasound probe and a blood vessel visualization device. The ultrasound probe includes an ultrasound generation device and is configured to detect one or more blood vessels. The blood vessel visualization device is configured to project a depiction of the blood vessel topography within a target area. The blood vessel visualization device can include one or more near-infrared/infrared emitters configured to generate infrared/near-infrared waves within the target area, one or more near-infrared/infrared sensors configured to detect the difference in reflective properties of tissue and blood vessels within the target area, and one or more visual light projectors configured to project a blood vessel visualization depiction of the blood vessel topography onto the target area.
Description
BACKGROUND

Readily accessing a blood vessel can require multiple sticks to access the correct blood vessel. Clinicians may use vein visualizers, medical devices that detect subcutaneous blood vessels and project a depiction of the vein topography on the skin surface, to readily access the correct blood vessel. Vein visualizers must be held some distance away from the skin surface where hands or other medical devices may disturb the depiction of the vein topography, adding difficulty to accessing the correct blood vessel. It would be beneficial to clinicians and patients to have a vein visualizer whose depiction would not be disturbed, allowing clinicians ready access to the correct blood vessel. Disclosed herein are an ultrasound imaging system and method of use that address the foregoing.


SUMMARY

Disclosed herein is an ultrasound imaging system including an ultrasound probe and a blood vessel visualization device. The ultrasound probe includes an ultrasound generation device and is configured to detect one or more blood vessels. The blood vessel visualization device is configured to project a depiction of the blood vessel topography within a target area, the blood vessel visualization device includes one or more near-infrared/infrared emitters configured to generate infrared/near-infrared waves within the target area, one or more near-infrared/infrared sensors configured to detect the difference in reflective properties of tissue and blood vessels within the target area, and one or more visual light projectors configured to project a blood vessel visualization depiction of the blood vessel topography onto the target area.


In some embodiments, the one or more visual light projectors project the blood vessel visualization depiction to one side of the ultrasound probe onto the target area.


In some embodiments, each of the one or more near-infrared sensors, the one or more near-infrared/infrared emitters, and the one or more visual light projectors are coupled to the ultrasound probe.


In some embodiments, the one or more near-infrared/infrared emitters and the one or more near-infrared/infrared sensors are located oblique to an ultrasound acoustic stack of the ultrasound probe.


In some embodiments, the blood vessel visualization device includes a console having non-transitory computer readable medium, an energy source and a plurality of logic modules.


In some embodiments, the console is in communication with each of the ultrasound probe, the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors and the one or more visual light projectors.


In some embodiments, the energy source is in communication with each of the one or more infrared/near-infrared emitters, the one or more infrared/near-infrared sensors, and the one or more visual light projectors.


In some embodiments, the energy source is untethered to one or more of the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors, or the one or more visible light projectors.


In some embodiments, the energy source includes an induction coupling system configured to wirelessly provide energy to the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors or the one or more visible light projectors.


In some embodiments, the energy source is tethered to one or more of the infrared/near-infrared emitters, the infrared/near-infrared sensors or the visible light projectors.


In some embodiments, the plurality of logic modules, when executed by the processor, are configured to perform operations including receiving ultrasound data from the ultrasound probe, correlating a detected location of one or more blood vessels with a starting location for a blood vessel visualization depiction within the target area, activating each of the near-infrared/infrared emitters, the near-infrared/infrared sensors, and the visual light projectors, receiving detected near-infrared/infrared data from the near-infrared/infrared sensors, and generating and projecting to one side of the ultrasound probe, the blood vessel visualization depiction within the target area.


In some embodiments, the ultrasound generation device includes a microelectromechanical systems based device.


In some embodiments, the ultrasound probe includes the ultrasound generation device, the one or more near-infrared/infrared sensors and the one or more visual light projectors, and a vascular access device includes the one or more near-infrared/infrared emitters, the vascular access device configured to be inserted into one of the detected blood vessels within the target area.


In some embodiments, the vascular access device includes a visual indicator configured to be activated based upon a vascular access device trajectory relative to a target blood vessel or an identified blood vessel type including an artery or a vein.


Also disclosed herein is a method of detecting one or more blood vessels in a target area and generating and projecting a blood vessel visualization depiction over the target area. The method includes detecting, by ultrasound and near-infrared/infrared electromagnetic waves, blood vessel topography in the target area, generating a blood vessel visualization depiction, and projecting, to one side of an ultrasound probe, the blood vessel visualization depiction on the target area.


In some embodiments, detecting by near-infrared/infrared electromagnetic waves includes one or more near-infrared/infrared sensors detecting a difference in reflective properties of tissue and blood vessels within the target area.


In some embodiments, detecting by near-infrared/infrared electromagnetic waves includes the electromagnetic waves being emitted from one or more near-infrared/infrared emitters coupled to the ultrasound probe and being detected by the one or more near-infrared/infrared sensors coupled to the ultrasound probe.


In some embodiments, detecting by near-infrared/infrared electromagnetic waves includes the electromagnetic waves being emitted from one or more near-infrared/infrared emitters coupled to a vascular access device and being detected by one or more near-infrared/infrared sensors coupled to the ultrasound probe.


In some embodiments, generating a blood vessel visualization depiction includes a console in communication with each of the ultrasound probe, the one or more near-infrared/infrared emitters and the near-infrared/infrared sensors generating the blood vessel visualization depiction using ultrasound data received from the ultrasound probe and electromagnetic wave data received from the one or more near-infrared/infrared sensors.


In some embodiments, projecting, to one side of an ultrasound probe, the blood vessel visualization depiction on the target area includes one or more visual light projectors coupled to the ultrasound probe projecting the blood vessel visualization depiction onto the target area.


These and other features of the concepts provided herein will become more apparent to those of skill in the art in view of the accompanying drawings and following description, which describe particular embodiments of such concepts in greater detail.





DRAWINGS

A more particular description of the present disclosure will be rendered by reference to specific embodiments thereof that are illustrated in the appended drawings. It is appreciated that these drawings depict only typical embodiments of the invention and are therefore not to be considered limiting of its scope. Example embodiments of the invention will be described and explained with additional specificity and detail through the use of the accompanying drawings in which:



FIG. 1 illustrates a cross sectional view of an ultrasound imaging system including an ultrasound probe having a blood vessel visualization device coupled thereto, in accordance with some embodiments.



FIG. 2 illustrates a block diagram of some components of the ultrasound imaging system including the ultrasound probe having a console, in accordance with some embodiments.



FIG. 3 illustrates a cross sectional view of the ultrasound imaging system including the ultrasound probe and a vascular access device, in accordance with some embodiments.



FIGS. 4A-4C illustrate a side cross-sectional view of an exemplary method of generating and projecting a blood vessel visualization depiction over a target area, in accordance with some embodiments.



FIG. 5 illustrates a flow chart of an exemplary method of detecting one or more blood vessels in a target are, generating and projecting a blood vessel visualization depiction over the target area, in accordance with some embodiments.





DESCRIPTION

Before some particular embodiments are disclosed in greater detail, it should be understood that the particular embodiments disclosed herein do not limit the scope of the concepts provided herein. It should also be understood that a particular embodiment disclosed herein can have features that can be readily separated from the particular embodiment and optionally combined with or substituted for features of any of a number of other embodiments disclosed herein.


Regarding terms used herein, it should also be understood the terms are for the purpose of describing some particular embodiments, and the terms do not limit the scope of the concepts provided herein. Ordinal numbers (e.g., first, second, third, etc.) are generally used to distinguish or identify different features or steps in a group of features or steps, and do not supply a serial or numerical limitation. For example, “first,” “second,” and “third” features or steps need not necessarily appear in that order, and the particular embodiments including such features or steps need not necessarily be limited to the three features or steps. Labels such as “left,” “right,” “top,” “bottom,” “front,” “back,” and the like are used for convenience and are not intended to imply, for example, any particular fixed location, orientation, or direction. Instead, such labels are used to reflect, for example, relative location, orientation, or directions. Singular forms of “a,” “an,” and “the” include plural references unless the context clearly dictates otherwise.


The term “logic” may be representative of hardware, firmware or software that is configured to perform one or more functions. As hardware, the term logic may refer to or include circuitry having data processing and/or storage functionality. Examples of such circuitry may include, but are not limited or restricted to a hardware processor (e.g., microprocessor, one or more processor cores, a digital signal processor, a programmable gate array, a microcontroller, an application specific integrated circuit “ASIC”, etc.), a semiconductor memory, or combinatorial elements.


Additionally, or in the alternative, the term logic may refer to or include software such as one or more processes, one or more instances, Application Programming Interface(s) (API), subroutine(s), function(s), applet(s), servlet(s), routine(s), source code, object code, shared library/dynamic link library (dll), or even one or more instructions. This software may be stored in any type of a suitable non-transitory storage medium, or transitory storage medium (e.g., electrical, optical, acoustical or other form of propagated signals such as carrier waves, infrared signals, or digital signals). Examples of a non-transitory storage medium may include, but are not limited or restricted to a programmable circuit; non-persistent storage such as volatile memory (e.g., any type of random access memory “RAM”); or persistent storage such as non-volatile memory (e.g., read-only memory “ROM”, power-backed RAM, flash memory, phase-change memory, etc.), a solid-state drive, hard disk drive, an optical disc drive, or a portable memory device. As firmware, the logic may be stored in persistent storage.


Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by those of ordinary skill in the art.



FIG. 1 illustrates a perspective view of an ultrasound imaging system (“system”) 100 including an ultrasound probe 110 having a blood vessel visualization device 120 coupled thereto, in accordance with some embodiments. In some embodiments, the ultrasound probe 110 may be configured to detect one or more blood vessels within a target area 140 by contacting a skin surface within the target area 140. In some embodiments, the ultrasound probe 110 may include an ultrasound generation device 112 including an ultrasound acoustic stack or other various modalities of ultrasound generation (e.g., microelectromechanical systems (MEMS) based, etc.). In some embodiments, the target area 140 may include the one or more blood vessels 190 to be accessed by a vascular access device. The blood vessel visualization device 120 may be configured to sense subcutaneous blood vessel topography and project, to one side of the ultrasound probe 110, a depiction of the sensed blood vessel topography on the skin surface of the target area 140,allowing a clinician to access a blood vessel without disturbing the depiction.


In some embodiments, the blood vessel visualization device 120 includes one or more near-infrared/infrared (“NIR/IR”) sensors 122, one or more NIR/IR emitters 124, and one or more visual light projectors 126. The NIR/IR sensors 122, the NIR/IR emitters 124 and the visual light projectors 126 may all be in communication with a console 150. In some embodiments, the NIR/IR sensors 122, the NIR/IR emitters 124 and the visual light projectors 126 may be in wireless communication with the console 150. In some embodiments, the console 150 may be in communication with the ultrasound probe 110 including in communication with the ultrasound generation device 112. In some embodiments, the console 150 may be coupled to the ultrasound probe 110 or integrated into the ultrasound probe 110. In some embodiments, the NIR/IR emitters 124 may be configured to emit electromagnetic waves within the near-infrared/infrared spectrum within the target area 140. The electromagnetic waves reflect off of subcutaneous structures such as blood vessels. In some embodiments, the NIR/IR sensors 122 may be configured to detect the reflected NIR/IR electromagnetic waves and transmit the detected NIR/IR data to the console 150. In some embodiments, the NIR/IR sensors 122 may be configured to detect the difference in reflective properties of tissue and blood vessels within the target area 140. The console 150 may be configured to receive the detected NIR/IR data, correlate the difference in electromagnetic reflectance between blood vessels and surrounding tissues to generate a blood vessel visualization depiction. The one or more visual light projectors 126 may be configured to project the blood vessel visualization depiction to one side of the ultrasound probe 110 on the skin surface within the target area 140. In some embodiments, the NIR/IR emitters 124 and the NIR/IR sensors 122 may be located oblique to the ultrasound acoustic stack 112.



FIG. 2 illustrates a block diagram of some components of the ultrasound imaging system 100 including the console 150, in accordance with some embodiments. In some embodiments, the console 150 may be coupled to or integrated into the ultrasound probe 110. In some embodiments, the console 150 may be in communication with the ultrasound probe 110 and be configured to receive the ultrasound data to generate a portion of the blood vessel visualization depiction within the target area 140. The blood vessel visualization device 120 may include the console 150 in communication with each of: the one or more blood vessel sensors 122, the one or more blood vessel emitters 124, the visible light projectors 126 and the ultrasound probe 110. In some embodiments, the console 150 may be in wireless communication with each of the blood vessel sensors 122, the blood vessel emitters 124, the visible light projectors 126 and the ultrasound probe 110. Exemplary wireless communication modalities can include WiFi, Bluetooth, Near Field Communications (NFC), cellular Global System for Mobile Communication (“GSM”), electromagnetic (EM), radio frequency (RF), combinations thereof, or the like.


In some embodiments, the console 150 may include one or more processors 152, an energy source 154, non-transitory computer readable medium (“memory”) 156 and a plurality of logic modules. The plurality of logic modules may include one or more of: a ultrasound probe receiving logic 158, a ultrasound probe correlation logic 160, a NIR/IR sensor activation logic 162, a NIR/IR emitter activation logic 164, a NIR/IR sensor receiving logic 166, a blood vessel visualization generation logic 168, a blood vessel visualization projection logic 170 and an energy source transmitting logic 172. In some embodiments, the ultrasound probe receiving logic 158 may be configured to receive the ultrasound wave data from the ultrasound probe 110 including from the ultrasound generation device 112. In some embodiments, the ultrasound probe correlation logic 160 may be configured to correlate the detected location of the one or more blood vessels 190 with a starting location for the blood vessel visualization depiction within the target area 140. In some embodiments, the NIR/IR emitter activation logic 164 may be configured to activate the NIR/IR emitters 124 to emit electromagnetic waves within the target area 140. In some embodiments, the NIR/IR sensor activation logic 162 may be configured to activate the NIR/IR sensors 122. In some embodiments, the NIR/IR sensor receiving logic 166 may be configured to detect the reflected electromagnetic waves within the target area 140. In some embodiments, the NIR/IR sensor receiving logic 166 may be configured to detect the difference in the reflective properties of tissue and blood vessels 190 within the target area 140. In some embodiments, the blood vessel visualization generation logic 168 may be configured to generate the blood vessel visualization depiction. The blood vessel visualization generation logic 168 may be configured to receive the detected electromagnetic wave data and the detected ultrasound wave data to generate the blood vessel visualization depiction of the blood vessels 190 within the target area 140. In some embodiments, the blood vessel visualization depiction of the blood vessels 190 within the target area 140 may include a topographical map depicting the blood vessels 190 and surrounding tissues. In some embodiments, the blood vessel visualization projection logic 170 may be configured to use the one or more visible light proj ectors 126 to project the blood vessel visualization depiction on the target area 140.


In some embodiments, the energy source 154 may be configured to provide power to the one or more NIR/IR emitters 124, the one or more NIR/IR sensors 122 and the one or more visible light projectors 126. In some embodiments, the NIR/IR emitters 124, the NIR/IR sensors 122, or the visible light projectors 126 may be tethered to the energy source 154. In some embodiments, the energy source 154 may be in communication with one or more of the NIR/IR sensors 122, the NIR/IR emitters 124 and the visible light projectors 126. In some embodiments, the energy source 154 may include a battery coupled to one or more of the NIR/IR emitters 124, the NIR/IR sensors 122 and the visible light projectors 126. In some embodiments, the NIR/IR sensors 122 or the NIR/IR emitters 124 may be untethered to the ultrasound probe 110 [e.g., coupled to a vascular access device 180 (see FIG. 3)]. The energy source 154 may then be configured to wirelessly provide power to either the NIR/IR sensors 122 or the NIR/IR emitters 124. The energy source 154 may be configured to wirelessly provide power to the NIR/IR sensors 122 through an induction coupling system. In this embodiment, the plurality of logic modules may be configured to include an energy source transmitting logic 172 configured to regulate the amount of power that is wirelessly provided to the NIR/IR sensors 122 or NIR/IR emitters 124.



FIG. 3 illustrates a cross sectional view of the ultrasound imaging system 100 wherein the NIR/IR emitters 124 are coupled to a vascular access device 180 and the NIR/IR sensors 122 and visual light projectors 126 are coupled to the ultrasound probe 110. As used herein a “vascular access device” can include a catheter, peripherally inserted central catheter (“PICC”), peripheral intravenous line (“PIV”), central venous catheter (“CVC”), midline catheter, a needle, or the like. In some embodiments, the NIR/IR emitters 124 and the NIR/IR sensors 122 may be separated between the ultrasound probe 110 and the vascular access device 180. In some embodiments, the NIR/IR emitters 124 may be coupled to the vascular access device 180 and the NIR/IR sensors 122 may be coupled to the ultrasound probe 110. In some embodiments, the NIR/IR emitters 124 may be coupled to the ultrasound probe 110 and the NIR/IR sensors 122 may be coupled to the vascular access device 180. In some embodiments, the NIR/IR emitters 124 may be integrated into the vascular access device 180. In this embodiment, wherein either the NIR/IR emitters 124 or the NIR/IR sensors 122 are coupled to the vascular access device 180, the NIR/IR emitters or NIR/IR sensors 122 may be powered through induction coupling. In an embodiment, the vascular access device 180 is enabled with a visual indicator, wherein the visual indicator is activated based upon the trajectory of the vascular access device 180 relative to a target blood vessel or the visual indicator is activated based upon the identified blood vessel type. The console 150 functions as described above.



FIGS. 4A-4C illustrates a side cross-sectional view of an exemplary method of detecting blood vessel topography and projecting a blood vessel visualization depiction over the target area 140, in accordance with some embodiments. The ultrasound probe 110 includes the NIR/IR emitters 124, the NIR/IR sensors 122 and the visual light projectors 126, as described above. As illustrated in FIG. 4A, in some embodiments, the ultrasound imaging system 100 including the ultrasound probe 110 may be configured to detect one or more blood vessels 190 within the target area 140. In some embodiments, the ultrasound probe 110 may be configured to detect the one or more blood vessels 190 within target area 140 by ultrasound. The NIR/IR emitters 124 may emit electromagnetic waves subcutaneously within the target area 140. As illustrated in FIG. 4A, in some embodiments, the NIR/IR emitters 124, the NIR/IR sensors 122, and the visual light projectors 126 may be arranged in various configurations on the ultrasound probe 110. In some embodiments, the NIR/IR emitters 124 may be located above the NIR/IR sensors 122, and the NIR/IR sensors 122 may be located above the visual light projectors 126. In some embodiments, the NIR/IR emitters 124 may be located adjacent to the NIR/IR sensors 122. In some embodiments, the visual light projectors 126 may be located adjacent to the NIR/IR sensors 122 or the NIR/IR emitters 124.


As illustrated in FIG. 4B, the NIR/IR sensors 122 may be configured to detect the reflected electromagnetic waves, including the difference in reflective properties of the tissue and blood vessels within the target area 140. The NIR/IR sensors 122 may transmit the detected electromagnetic waves to the console 150 where the console 150 may use the detected electromagnetic wave data, along with the detected blood vessel data within the target area 140 to generate a blood vessel visualization depiction. As illustrated in FIG. 4C, the visual light projectors 126 may be configured to receive the blood vessel visualization depiction and project the blood vessel visualization depiction onto the target area 140 including onto a skin surface. A user may use the blood vessel visualization depiction to accurately access the blood vessel 190 within the target area 140.



FIG. 5 illustrates a flow chart of the method 200 of detecting one or more blood vessels 190 in a target area 140 and generating and projecting a blood vessel visualization depiction over the target area 140, in accordance with some embodiments. In some embodiments, the method 200 includes detecting blood vessel topography in a target area 140 (block 202). In some embodiments, detecting blood vessel topography in the target area 140 includes detecting, by near-infrared/infrared electromagnetic waves, blood vessel topography in the target area 140. In some embodiments, detecting blood vessel topography in the target area 140 includes detecting, by ultrasound, blood vessel topography in the target area 140. In some embodiments, detecting blood vessel topography in the target area 140 includes detecting by ultrasound and near-infrared/infrared electromagnetic waves, blood vessel topography in the target area 140. In some embodiments, detecting blood vessel topography includes using the NIR/IR emitters 124 configured to emit near-infrared/infrared electromagnetic waves and NIR/IR sensors 122 coupled to the ultrasound probe 110 configured to detect the near-infrared/infrared electromagnetic waves to detect the blood vessel topography. In some embodiments, detecting blood vessel topography includes using the NIR/IR emitters 124 couped to the vascular access device 180 configured to emit near-infrared/infrared electromagnetic waves and the NIR/IR sensors 122 coupled to the ultrasound probe 110 configured to detect the near-infrared/infrared electromagnetic waves to detect the blood vessel topography. In some embodiments, detecting blood vessel topography in the target area 140 includes the NIR/IR sensors 122 detecting the difference in reflective properties of tissue and blood vessels within the target area 140. The method 200 further includes generating a blood vessel visual depiction of the detected blood vessels 190 within the target area 140 (block 204). In some embodiments, generating the blood vessel visual depiction includes the console 150 generating the blood vessel visual depiction. In some embodiments, the console 150 may use blood vessel topography data received from the NIR/IR sensors 122 to generate the blood vessel visual depiction. In some embodiments, the console 150 may also use ultrasound data received from the ultrasound probe 110 to generate the blood vessel visual depiction. In some embodiments, the method 200 includes projecting, to one side of the ultrasound probe 110, the blood vessel visual depiction of the detected blood vessel on the target area 140 (block 206). In some embodiments projecting the blood vessel visual depiction of the detected blood vessels on the target area 140 includes the blood vessel visual depiction having distinct colors indicating different blood vessel type (e.g., artery versus vein).


While some particular embodiments have been disclosed herein, and while the particular embodiments have been disclosed in some detail, it is not the intention for the particular embodiments to limit the scope of the concepts provided herein. Additional adaptations and/or modifications can appear to those of ordinary skill in the art, and, in broader aspects, these adaptations and/or modifications are encompassed as well. Accordingly, departures may be made from the particular embodiments disclosed herein without departing from the scope of the concepts provided herein.

Claims
  • 1. An ultrasound imaging system, comprising: an ultrasound probe configured to detect one or more blood vessels, the ultrasound probe including an ultrasound generation device defining a bottom side of the ultrasound probe; anda blood vessel visualization device, the blood vessel visualization device configured to project a depiction of blood vessel topography within a target area, the blood vessel visualization device including: one or more near-infrared/infrared emitters configured to generate infrared/near-infrared waves within the target area;one or more near-infrared/infrared sensors configured to detect a difference in reflective properties of tissue and blood vessels within the target area; andone or more visual light projectors configured to project a blood vessel visualization depiction of the blood vessel topography onto the target area,wherein each of the one or more near-infrared/infrared sensors, the one or more near-infrared/infrared emitters, and the one or more visual light projectors are: directly coupled to the ultrasound probe, andarranged on a front side of the ultrasound probe, the front side extending upward away from the bottom side in a substantially perpendicular direction with respect to the bottom side.
  • 2. The ultrasound imaging system according to claim 1, wherein the one or more visual light projectors project the blood vessel visualization depiction to one side of the ultrasound probe onto the target area.
  • 3. The ultrasound imaging system according to claim 1, wherein the one or more near-infrared/infrared emitters and the one or more near-infrared/infrared sensors are located oblique to an ultrasound acoustic stack of the ultrasound probe.
  • 4. The ultrasound imaging system according to claim 1, wherein the blood vessel visualization device includes a console having non-transitory computer readable medium, an energy source and a plurality of logic modules.
  • 5. The ultrasound imaging system according to claim 4, wherein the console is in communication with each of the ultrasound probe, the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors and the one or more visual light projectors.
  • 6. The ultrasound imaging system according to claim 5, wherein the energy source is in communication with each of the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors, and the one or more visual light projectors.
  • 7. The ultrasound imaging system according to claim 6, wherein the energy source is untethered to one or more of the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors, or the one or more visible light projectors.
  • 8. The ultrasound imaging system according to claim 6, wherein the energy source includes an induction coupling system configured to wirelessly provide energy to the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors or the one or more visible light projectors.
  • 9. The ultrasound imaging system according to claim 6, wherein the energy source is tethered to the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors or the one or more visible light projectors.
  • 10. The ultrasound imaging system according to claim 4, wherein the plurality of logic modules, when executed by one or more processors, are configured to perform operations: receiving ultrasound data from the ultrasound probe;correlating a detected location of the one or more blood vessels with a starting location for the blood vessel visualization depiction within the target area;activating each of the one or more near-infrared/infrared emitters, the one or more near-infrared/infrared sensors, and the one or more visual light projectors;receiving detected near-infrared/infrared data from the one or more near-infrared/infrared sensors; andgenerating and projecting to one side of the ultrasound probe, the blood vessel visualization depiction within the target area.
  • 11. The ultrasound imaging system according to claim 1, wherein the ultrasound generation device includes a microelectromechanical systems based device.
  • 12. The ultrasound imaging system according to claim 1, wherein a vascular access device includes a visual indicator configured to be activated based upon a vascular access device trajectory relative to a target blood vessel or an identified blood vessel type including an artery or a vein.
  • 13. A method of detecting one or more blood vessels in a target area and generating and projecting a blood vessel visualization depiction over the target area, comprising: detecting, by ultrasound and near-infrared/infrared electromagnetic waves, blood vessel topography in the target area, wherein detecting by near-infrared/infrared electromagnetic waves includes the electromagnetic waves being emitted from one or more near-infrared/infrared emitters directly coupled to an ultrasound probe and being detected by one or more near-infrared/infrared sensors directly coupled to the ultrasound probe;generating the blood vessel visualization depiction; andprojecting to one side of an ultrasound probe, the blood vessel visualization depiction on the target area,wherein: the ultrasound probe includes an ultrasound generation device defining a bottom side of the ultrasound probe, andeach of the one or more near-infrared/infrared sensors and the one or more near-infrared/infrared emitters are arranged on a front side of the ultrasound probe, the front side extending upward away from the bottom side in a substantially perpendicular direction with respect to the bottom side.
  • 14. The method according to claim 13, wherein detecting by near-infrared/infrared electromagnetic waves includes the one or more near-infrared/infrared sensors detecting a difference in reflective properties of tissue and blood vessels within the target area.
  • 15. The method according to claim 14, wherein generating the blood vessel visualization depiction includes a console in communication with each of the ultrasound probe, the one or more near-infrared/infrared emitters and the one or more near-infrared/infrared sensors generating the blood vessel visualization depiction using ultrasound data received from the ultrasound probe and electromagnetic wave data received from the one or more near-infrared/infrared sensors.
  • 16. The method according to claim 15, wherein projecting to one side of the ultrasound probe, the blood vessel visualization depiction on the target area includes one or more visual light projectors coupled to the ultrasound probe projecting the blood vessel visualization depiction onto the target area.
PRIORITY

This application claims the benefit of priority to U.S. Provisional Application No. 63/175,475, filed Apr. 15, 2021, which is incorporated by reference in its entirety into this application.

US Referenced Citations (394)
Number Name Date Kind
3697917 Orth et al. Oct 1972 A
5148809 Biegeleisen-Knight et al. Sep 1992 A
5181513 Touboul et al. Jan 1993 A
5325293 Dorne Jun 1994 A
5349865 Kavli et al. Sep 1994 A
5441052 Miyajima Aug 1995 A
5549554 Miraki Aug 1996 A
5573529 Haak et al. Nov 1996 A
5758650 Miller et al. Jun 1998 A
5775322 Silverstein et al. Jul 1998 A
5879297 Haynor et al. Mar 1999 A
5897503 Lyon et al. Apr 1999 A
5908387 LeFree et al. Jun 1999 A
5967984 Chu et al. Oct 1999 A
5970119 Hofmann Oct 1999 A
6004270 Urbano et al. Dec 1999 A
6019724 Gronningsaeter et al. Feb 2000 A
6068599 Saito et al. May 2000 A
6074367 Hubbell Jun 2000 A
6129668 Haynor et al. Oct 2000 A
6132379 Patacsil et al. Oct 2000 A
6216028 Haynor et al. Apr 2001 B1
6233476 Strommer et al. May 2001 B1
6245018 Lee Jun 2001 B1
6263230 Haynor et al. Jul 2001 B1
6375615 Flaherty et al. Apr 2002 B1
6436043 Bonnefous Aug 2002 B2
6498942 Esenaliev et al. Dec 2002 B1
6503205 Manor et al. Jan 2003 B2
6508769 Bonnefous Jan 2003 B2
6511458 Milo et al. Jan 2003 B2
6524249 Moehring et al. Feb 2003 B2
6543642 Milliorn Apr 2003 B1
6554771 Buil et al. Apr 2003 B1
6592520 Peszynski et al. Jul 2003 B1
6592565 Twardowski Jul 2003 B2
6601705 Molina et al. Aug 2003 B2
6612992 Hossack et al. Sep 2003 B1
6613002 Clark et al. Sep 2003 B1
6623431 Sakuma et al. Sep 2003 B1
6641538 Nakaya et al. Nov 2003 B2
6647135 Bonnefous Nov 2003 B2
6687386 Ito et al. Feb 2004 B1
6733458 Steins et al. May 2004 B1
6749569 Pellegretti Jun 2004 B1
6754608 Svanerudh et al. Jun 2004 B2
6755789 Stringer et al. Jun 2004 B2
6840379 Franks-Farah et al. Jan 2005 B2
6857196 Dalrymple Feb 2005 B2
6979294 Selzer et al. Dec 2005 B1
7074187 Selzer et al. Jul 2006 B2
7244234 Ridley et al. Jul 2007 B2
7359554 Klingensmith et al. Apr 2008 B2
7534209 Abend et al. May 2009 B2
7599730 Hunter et al. Oct 2009 B2
7637870 Flaherty et al. Dec 2009 B2
7681579 Schwartz Mar 2010 B2
7691061 Hirota Apr 2010 B2
7699779 Sasaki et al. Apr 2010 B2
7720520 Willis May 2010 B2
7727153 Fritz et al. Jun 2010 B2
7734326 Pedain et al. Jun 2010 B2
7831449 Ying et al. Nov 2010 B2
7905837 Suzuki Mar 2011 B2
7925327 Weese Apr 2011 B2
7927278 Selzer et al. Apr 2011 B2
8014848 Birkenbach et al. Sep 2011 B2
8038619 Steinbacher Oct 2011 B2
8060181 Rodriguez Ponce et al. Nov 2011 B2
8075488 Burton Dec 2011 B2
8090427 Eck et al. Jan 2012 B2
8105239 Specht Jan 2012 B2
8172754 Watanabe et al. May 2012 B2
8175368 Sathyanarayana May 2012 B2
8200313 Rambod et al. Jun 2012 B1
8211023 Swan et al. Jul 2012 B2
8228347 Beasley et al. Jul 2012 B2
8298147 Huennekens et al. Oct 2012 B2
8303505 Webler et al. Nov 2012 B2
8323202 Roschak et al. Dec 2012 B2
8328727 Miele et al. Dec 2012 B2
8336536 Wood-Putnam et al. Dec 2012 B1
8388541 Messerly et al. Mar 2013 B2
8409103 Grunwald et al. Apr 2013 B2
8449465 Nair et al. May 2013 B2
8553954 Saikia Oct 2013 B2
8556815 Pelissier et al. Oct 2013 B2
8585600 Liu et al. Nov 2013 B2
8622913 Dentinger et al. Jan 2014 B2
8706457 Hart et al. Apr 2014 B2
8727988 Flaherty et al. May 2014 B2
8734357 Taylor May 2014 B2
8744211 Owen Jun 2014 B2
8754865 Merritt et al. Jun 2014 B2
8764663 Smok et al. Jul 2014 B2
8781194 Malek et al. Jul 2014 B2
8781555 Burnside et al. Jul 2014 B2
8790263 Randall et al. Jul 2014 B2
8849382 Cox et al. Sep 2014 B2
8939908 Suzuki et al. Jan 2015 B2
8961420 Zhang Feb 2015 B2
9022940 Meier May 2015 B2
9087147 Fonte Jul 2015 B1
9138290 Hadjicostis Sep 2015 B2
9199082 Yared et al. Dec 2015 B1
9204858 Pelissier et al. Dec 2015 B2
9220477 Urabe et al. Dec 2015 B2
9295447 Shah Mar 2016 B2
9320493 Visveshwara Apr 2016 B2
9357980 Toji et al. Jun 2016 B2
9364171 Harris et al. Jun 2016 B2
9427207 Sheldon et al. Aug 2016 B2
9445780 Hossack et al. Sep 2016 B2
9456766 Cox et al. Oct 2016 B2
9456804 Tamada Oct 2016 B2
9468413 Hall et al. Oct 2016 B2
9492097 Wilkes et al. Nov 2016 B2
9521961 Silverstein et al. Dec 2016 B2
9554716 Burnside et al. Jan 2017 B2
9582876 Specht Feb 2017 B2
9610061 Ebbini et al. Apr 2017 B2
9636031 Cox May 2017 B2
9649037 Lowe et al. May 2017 B2
9649048 Cox et al. May 2017 B2
9702969 Hope Simpson et al. Jul 2017 B2
9715757 Ng et al. Jul 2017 B2
9717415 Cohen et al. Aug 2017 B2
9731066 Liu et al. Aug 2017 B2
9814433 Benishti et al. Nov 2017 B2
9814531 Yagi et al. Nov 2017 B2
9861337 Patwardhan et al. Jan 2018 B2
9895138 Sasaki Feb 2018 B2
9913605 Harris et al. Mar 2018 B2
9949720 Southard et al. Apr 2018 B2
10043272 Forzoni et al. Aug 2018 B2
10449330 Newman et al. Oct 2019 B2
10524691 Newman et al. Jan 2020 B2
10751509 Misener Aug 2020 B2
11564861 Gaines Jan 2023 B1
11900593 Dhatt et al. Feb 2024 B2
20020038088 Imran et al. Mar 2002 A1
20030047126 Tomaschko Mar 2003 A1
20030106825 Molina et al. Jun 2003 A1
20030120154 Sauer et al. Jun 2003 A1
20030125629 Ustuner Jul 2003 A1
20030135115 Burdette et al. Jul 2003 A1
20030149366 Stringer et al. Aug 2003 A1
20040015080 Kelly et al. Jan 2004 A1
20040055925 Franks-Farah et al. Mar 2004 A1
20040197267 Black et al. Oct 2004 A1
20050000975 Carco et al. Jan 2005 A1
20050049504 Lo et al. Mar 2005 A1
20050075597 Vournakis et al. Apr 2005 A1
20050165299 Kressy et al. Jul 2005 A1
20050251030 Azar et al. Nov 2005 A1
20050267365 Sokulin et al. Dec 2005 A1
20060004290 Smith et al. Jan 2006 A1
20060013523 Childlers et al. Jan 2006 A1
20060015039 Cassidy et al. Jan 2006 A1
20060020204 Serra et al. Jan 2006 A1
20060047617 Bacioiu et al. Mar 2006 A1
20060079781 Germond-Rouet et al. Apr 2006 A1
20060184029 Haim et al. Aug 2006 A1
20060210130 Germond-Rouet et al. Sep 2006 A1
20060241463 Shau et al. Oct 2006 A1
20070043341 Anderson et al. Feb 2007 A1
20070049822 Bunce et al. Mar 2007 A1
20070073155 Park et al. Mar 2007 A1
20070167738 Timinger et al. Jul 2007 A1
20070199848 Ellswood et al. Aug 2007 A1
20070239005 Ogasawara Oct 2007 A1
20070239120 Brock et al. Oct 2007 A1
20070249911 Simon Oct 2007 A1
20070287886 Saadat Dec 2007 A1
20080021322 Stone et al. Jan 2008 A1
20080033293 Beasley et al. Feb 2008 A1
20080033759 Finlay Feb 2008 A1
20080051657 Rold Feb 2008 A1
20080108930 Weitzel et al. May 2008 A1
20080125651 Watanabe et al. May 2008 A1
20080146915 McMorrow Jun 2008 A1
20080177186 Slater et al. Jul 2008 A1
20080221425 Olson et al. Sep 2008 A1
20080269605 Nakaya Oct 2008 A1
20080294037 Richter Nov 2008 A1
20080300491 Bonde et al. Dec 2008 A1
20090012399 Sunagawa et al. Jan 2009 A1
20090012401 Steinbacher Jan 2009 A1
20090074280 Lu et al. Mar 2009 A1
20090124903 Osaka May 2009 A1
20090137887 Shariati et al. May 2009 A1
20090137907 Takimoto et al. May 2009 A1
20090143672 Harms et al. Jun 2009 A1
20090143684 Cermak et al. Jun 2009 A1
20090156926 Messerly et al. Jun 2009 A1
20090281413 Boyden et al. Nov 2009 A1
20090306509 Pedersen et al. Dec 2009 A1
20100010348 Halmann Jan 2010 A1
20100168576 Poland et al. Jul 2010 A1
20100211026 Sheetz et al. Aug 2010 A2
20100249598 Smith et al. Sep 2010 A1
20100286515 Gravenstein et al. Nov 2010 A1
20100312121 Guan Dec 2010 A1
20100324423 El-Aklouk et al. Dec 2010 A1
20110002518 Ziv-Ari et al. Jan 2011 A1
20110026796 Hyun et al. Feb 2011 A1
20110071404 Schmitt et al. Mar 2011 A1
20110074244 Osawa Mar 2011 A1
20110087107 Lindekugel et al. Apr 2011 A1
20110166451 Blaivas et al. Jul 2011 A1
20110282188 Burnside et al. Nov 2011 A1
20110295108 Cox et al. Dec 2011 A1
20110313293 Lindekugel et al. Dec 2011 A1
20120136242 Qi et al. May 2012 A1
20120136256 Nozaki et al. May 2012 A1
20120165679 Orome et al. Jun 2012 A1
20120179038 Meurer et al. Jul 2012 A1
20120179042 Fukumoto et al. Jul 2012 A1
20120179044 Chiang et al. Jul 2012 A1
20120197132 O'Connor Aug 2012 A1
20120220865 Brown et al. Aug 2012 A1
20120277576 Lui Nov 2012 A1
20130041250 Pelissier et al. Feb 2013 A1
20130102889 Southard et al. Apr 2013 A1
20130131499 Chan et al. May 2013 A1
20130131502 Blaivas et al. May 2013 A1
20130144166 Specht et al. Jun 2013 A1
20130150724 Blaivas et al. Jun 2013 A1
20130188832 Ma et al. Jul 2013 A1
20130197367 Smok Aug 2013 A1
20130218024 Boctor et al. Aug 2013 A1
20130323700 Samosky et al. Dec 2013 A1
20130338503 Cohen et al. Dec 2013 A1
20130338508 Nakamura et al. Dec 2013 A1
20130345566 Weitzel et al. Dec 2013 A1
20140005530 Liu et al. Jan 2014 A1
20140031694 Solek Jan 2014 A1
20140066779 Nakanishi Mar 2014 A1
20140073976 Fonte et al. Mar 2014 A1
20140100440 Cheline et al. Apr 2014 A1
20140114194 Kanayama et al. Apr 2014 A1
20140170620 Savitsky et al. Jun 2014 A1
20140180098 Flaherty et al. Jun 2014 A1
20140180116 Lindekugel et al. Jun 2014 A1
20140188133 Misener Jul 2014 A1
20140188440 Donhowe et al. Jul 2014 A1
20140276059 Sheehan Sep 2014 A1
20140276069 Amble et al. Sep 2014 A1
20140276081 Tegels Sep 2014 A1
20140276085 Miller Sep 2014 A1
20140276690 Grace Sep 2014 A1
20140296694 Jaworski Oct 2014 A1
20140343431 Vajinepalli et al. Nov 2014 A1
20140357994 Jin et al. Dec 2014 A1
20150005738 Blacker Jan 2015 A1
20150011887 Ahn Jan 2015 A1
20150065916 Maguire et al. Mar 2015 A1
20150073279 Cai et al. Mar 2015 A1
20150112200 Oberg et al. Apr 2015 A1
20150141821 Yoshikawa et al. May 2015 A1
20150190111 Fry Jul 2015 A1
20150209003 Halmann et al. Jul 2015 A1
20150209113 Burkholz et al. Jul 2015 A1
20150209510 Burkholz et al. Jul 2015 A1
20150209526 Matsubara et al. Jul 2015 A1
20150245820 Tamada Sep 2015 A1
20150257735 Ball Sep 2015 A1
20150272448 Fonte et al. Oct 2015 A1
20150282890 Cohen et al. Oct 2015 A1
20150294497 Ng et al. Oct 2015 A1
20150297097 Matsubara et al. Oct 2015 A1
20150342572 Tahmasebi Maraghoosh et al. Dec 2015 A1
20150359520 Shan et al. Dec 2015 A1
20150359991 Dunbar et al. Dec 2015 A1
20160000367 Lyon Jan 2016 A1
20160000399 Halmann et al. Jan 2016 A1
20160026894 Nagase Jan 2016 A1
20160029995 Navratil et al. Feb 2016 A1
20160038119 Desjardins Feb 2016 A1
20160081674 Bagwan et al. Mar 2016 A1
20160113517 Lee et al. Apr 2016 A1
20160113699 Sverdlik et al. Apr 2016 A1
20160120607 Sorotzkin et al. May 2016 A1
20160125639 Park et al. May 2016 A1
20160157831 Kang et al. Jun 2016 A1
20160166232 Merritt Jun 2016 A1
20160202053 Walker et al. Jul 2016 A1
20160211045 Jeon et al. Jul 2016 A1
20160213398 Liu Jul 2016 A1
20160220124 Grady et al. Aug 2016 A1
20160259992 Knodt et al. Sep 2016 A1
20160278869 Grunwald Sep 2016 A1
20160287214 Ralovich et al. Oct 2016 A1
20160296208 Sethuraman et al. Oct 2016 A1
20160374644 Mauldin, Jr. et al. Dec 2016 A1
20170014105 Chono Jan 2017 A1
20170020561 Cox et al. Jan 2017 A1
20170079548 Silverstein et al. Mar 2017 A1
20170086785 Bjaerum Mar 2017 A1
20170103534 Park et al. Apr 2017 A1
20170143312 Hedlund et al. May 2017 A1
20170164923 Matsumoto Jun 2017 A1
20170172666 Govari et al. Jun 2017 A1
20170215842 Ryu et al. Aug 2017 A1
20170252004 Broad et al. Sep 2017 A1
20170258522 Goshgarian et al. Sep 2017 A1
20170328751 Lemke Nov 2017 A1
20170367678 Sirtori et al. Dec 2017 A1
20180015256 Southard et al. Jan 2018 A1
20180116723 Hettrick et al. May 2018 A1
20180125450 Blackbourne et al. May 2018 A1
20180161502 Nanan et al. Jun 2018 A1
20180199914 Ramachandran et al. Jul 2018 A1
20180214119 Mehrmohammadi et al. Aug 2018 A1
20180228465 Southard Aug 2018 A1
20180235649 Elkadi Aug 2018 A1
20180235709 Donhowe et al. Aug 2018 A1
20180289927 Messerly Oct 2018 A1
20180296185 Cox et al. Oct 2018 A1
20180310955 Lindekugel et al. Nov 2018 A1
20180333135 Kim et al. Nov 2018 A1
20180344293 Raju et al. Dec 2018 A1
20190029636 Lee et al. Jan 2019 A1
20190060001 Kohli et al. Feb 2019 A1
20190060014 Hazelton et al. Feb 2019 A1
20190090855 Kobayashi et al. Mar 2019 A1
20190125210 Govari et al. May 2019 A1
20190200951 Meier Jul 2019 A1
20190239848 Bedi et al. Aug 2019 A1
20190239850 Dalvin et al. Aug 2019 A1
20190307419 Durfee Oct 2019 A1
20190307515 Naito et al. Oct 2019 A1
20190307516 Schotzko et al. Oct 2019 A1
20190365347 Abe Dec 2019 A1
20190365348 Toume et al. Dec 2019 A1
20190365354 Du Dec 2019 A1
20200069929 Mason et al. Mar 2020 A1
20200113540 Gijsbers et al. Apr 2020 A1
20200163654 Satir et al. May 2020 A1
20200200900 Asami et al. Jun 2020 A1
20200229795 Tadross et al. Jul 2020 A1
20200230391 Burkholz et al. Jul 2020 A1
20200237403 Southard et al. Jul 2020 A1
20200281563 Muller et al. Sep 2020 A1
20200359990 Poland et al. Nov 2020 A1
20200390416 Swan et al. Dec 2020 A1
20210059639 Howell Mar 2021 A1
20210077058 Mashood et al. Mar 2021 A1
20210137492 Imai May 2021 A1
20210146167 Barthe et al. May 2021 A1
20210161510 Sasaki et al. Jun 2021 A1
20210186467 Urabe et al. Jun 2021 A1
20210212668 Li et al. Jul 2021 A1
20210267569 Yamamoto Sep 2021 A1
20210267570 Ulman et al. Sep 2021 A1
20210295048 Buras et al. Sep 2021 A1
20210315538 Brandl et al. Oct 2021 A1
20210373602 Min Dec 2021 A1
20210378627 Yarmush et al. Dec 2021 A1
20220019313 He et al. Jan 2022 A1
20220022969 Misener Jan 2022 A1
20220039777 Durfee Feb 2022 A1
20220039829 Zijlstra et al. Feb 2022 A1
20220071593 Tran Mar 2022 A1
20220096053 Sethuraman et al. Mar 2022 A1
20220096797 Prince Mar 2022 A1
20220104791 Matsumoto Apr 2022 A1
20220104886 Blanchard et al. Apr 2022 A1
20220117582 McLaughlin et al. Apr 2022 A1
20220160434 Messerly et al. May 2022 A1
20220168050 Sowards et al. Jun 2022 A1
20220172354 Misener et al. Jun 2022 A1
20220225963 Sutton et al. Jul 2022 A1
20220296303 McLeod et al. Sep 2022 A1
20220304652 Peterson et al. Sep 2022 A1
20220330922 Sowards et al. Oct 2022 A1
20220361840 Matsumoto et al. Nov 2022 A1
20230048327 Lampe et al. Feb 2023 A1
20230107629 Sowards et al. Apr 2023 A1
20230132148 Sowards et al. Apr 2023 A1
20230135562 Misener et al. May 2023 A1
20230135757 Bauer et al. May 2023 A1
20230138970 Sowards et al. May 2023 A1
20230148872 Sowards et al. May 2023 A1
20230201539 Howell Jun 2023 A1
20230277153 Sowards et al. Sep 2023 A1
20230277154 Sowards et al. Sep 2023 A1
20230293143 Sowards et al. Sep 2023 A1
20230338010 Sturm Oct 2023 A1
20230371928 Rajguru et al. Nov 2023 A1
20230397900 Prince Dec 2023 A1
20240065673 Sowards et al. Feb 2024 A1
20240307024 Sowards et al. Sep 2024 A1
20250017559 Denny et al. Jan 2025 A1
Foreign Referenced Citations (52)
Number Date Country
102871645 Jan 2013 CN
105107067 Dec 2015 CN
105107067 May 2018 CN
0933063 Aug 1999 EP
1504713 Feb 2005 EP
1591074 May 2008 EP
2823766 Jan 2015 EP
3181083 Jun 2017 EP
3870059 Sep 2021 EP
2000271136 Oct 2000 JP
2007222291 Sep 2007 JP
2014150928 Aug 2014 JP
2018175547 Nov 2018 JP
20180070878 Jun 2018 KR
102176196 Nov 2020 KR
2004082749 Sep 2004 WO
2007115174 Oct 2007 WO
2010029521 Mar 2010 WO
2010076808 Jul 2010 WO
2013059714 Apr 2013 WO
2014115150 Jul 2014 WO
2015017270 Feb 2015 WO
2016081023 May 2016 WO
2017096487 Jun 2017 WO
2017214428 Dec 2017 WO
2018026878 Feb 2018 WO
2018134726 Jul 2018 WO
2019232451 Dec 2019 WO
2020002620 Jan 2020 WO
2020016018 Jan 2020 WO
2019232454 Feb 2020 WO
2020044769 Mar 2020 WO
2020083660 Apr 2020 WO
WO-2020067897 Apr 2020 WO
2020186198 Sep 2020 WO
2021123905 Jun 2021 WO
2021198226 Oct 2021 WO
2022072727 Apr 2022 WO
2022081904 Apr 2022 WO
2022119853 Jun 2022 WO
2022115479 Jun 2022 WO
2022119856 Jun 2022 WO
2022221703 Oct 2022 WO
2022221714 Oct 2022 WO
2023059512 Apr 2023 WO
2023076268 May 2023 WO
2023081220 May 2023 WO
2023081223 May 2023 WO
2023091424 May 2023 WO
2023167866 Sep 2023 WO
2023177718 Sep 2023 WO
2024044277 Feb 2024 WO
Non-Patent Literature Citations (88)
Entry
PCT/US2022/049983 filed Nov. 15, 2022 International Search Report and Written Opinion dated Mar. 29, 2023.
PCT/US2023/014143 filed Feb. 28, 2023 International Search Report and Written Opinion dated Jun. 12, 2023.
PCT/US2023/015266 filed Mar. 15, 2023 International Search Report and Written Opinion dated May 25, 2023.
Saxena Ashish et al Thermographic venous blood flow characterization with external cooling stimulation Infrared Physics and Technology Elsevier Science GB vol. 90 Feb. 9, 2018 Feb. 9, 2018 pp. 8-19 XP085378852.
U.S. Appl. No. 17/020,476, filed Sep. 14, 2020 Notice of Allowance dated Apr. 28, 2022.
U.S. Appl. No. 17/468,318, filed Sep. 7, 2021 Non-Final Office Action dated Apr. 12, 2023.
U.S. 17/684,180, filed Mar. 1, 2022 Restriction Requirement dated May 19, 2023.
PCT/US2022/025082 filed Apr. 15, 2022 International Search Report and Written Opinion dated Jul. 11, 2022.
PCT/US2022/025097 filed Apr. 15, 2022 International Search Report and Written Opinion dated Jul. 8, 2022.
U.S. Appl. No. 17/020,476, filed Sep. 14, 2020 Advisory Action dated Aug. 19, 2022.
U.S. Appl. No. 17/471,015, filed Sep. 9, 2021 Non-Final Office Action dated Aug. 16, 2022.
PCT/US2022/048716 filed Nov. 2, 2022 International Search Report and Written Opinion dated Feb. 24, 2023.
PCT/US2022/048722 filed Nov. 2, 2022 International Search Report and Written Opinion dated Feb. 24, 2023.
PCT/US2022047727 filed Oct. 25, 2022 International Search Report and Written Opinion dated Jan. 25, 2023.
U.S. Appl. No. 17/020,476, filed Sep. 14, 2020 Final Office Action dated Jan. 5, 2023.
U.S. Appl. No. 17/020,476, filed Sep. 14, 2020 Non-Final Office Action dated Sep. 23, 2022.
U.S. Appl. No. 17/471,015, filed Sep. 9, 2021 Non-Final Office Action dated Mar. 30, 2023.
U.S. Appl. No. 17/534,099, filed Nov. 23, 2021 Non-Final Office Action dated Mar. 31, 2023.
U.S. Appl. No. 17/538,911, filed Nov. 30, 2021 Non-Final Office Action dated Mar. 2, 2023.
EP 20866520.8 filed Apr. 5, 2022 Extended European Search Report dated Aug. 22, 2023.
U.S. Appl. No. 17/468,318, filed Sep. 7, 2021 Final Office Action dated Sep. 8, 2023.
U.S. Appl. No. 17/538,911, filed Nov. 30, 2021 Final Office Action dated Sep. 13, 2023.
U.S. Appl. No. 17/684,180, filed Mar. 1, 2022 Non-Final Office Action dated Jul. 28, 2023.
U.S. Appl. No. 17/722,151, filed Apr. 15, 2022 Non-Final Office Action dated Sep. 7, 2023.
Lu Zhenyu et al “Recent advances in 5 robot-assisted echography combining perception control and cognition.” Cognitive Computation and Systems the Institution of Engineering and Technology, Michael Faraday House, Six Hills Way, Stevenage Herts. SG1 2AY UK vol. 2 No. 3 Sep. 2, 2020 (Sep. 2, 2020).
Pagoulatos, N. et al. “New spatial localizer based on fiber optics with applications in 3D ultrasound imaging” Proceeding of Spie, vol. 3976 (Apr. 18, 2000; Apr. 18, 2000).
PCT/US2021/049294 filed Sep. 7, 2021 International Search Report and Written Opinion dated Dec. 8, 2021.
PCT/US2021/049712 filed Sep. 9, 2021 International Search Report and Written Opinion dated Dec. 14, 2021.
PCT/US2021/060622 filed Nov. 23, 2021 International Search Report and Written Opinion dated Mar. 3, 2022.
PCT/US2021/061267 filed Nov. 30, 2021 International Search Report and Written Opinion dated Mar. 9, 2022.
PCT/US2021/061276 filed Nov. 30, 2021 International Search Report and Written Opinion dated Mar. 9, 2022.
Sebastian Vogt: “Real-Time Augmented Reality for Image-Guided Interventions”, Oct. 5, 2009, XPO55354720, Retrieved from the Internet: URL: https://opus4.kobv.de/opus4-fau/frontdoor/deliver/index/docld/1235/file/SebastianVogtDissertation.pdf.
U.S. Appl. No. 15/650,474, filed Jul. 14, 2017 Final Office Action dated Jun. 2, 2020.
U.S. Appl. No. 15/650,474, filed Jul. 14, 2017 Non-Final Office Action dated Dec. 16, 2019.
U.S. Appl. No. 15/650,474, filed Jul. 14, 2017 Notice of Allowance dated Dec. 11, 2020.
U.S. Appl. No. 15/650,474, filed Jul. 14, 2017 Notice of Allowance dated Mar. 1, 2021.
U.S. Appl. No. 17/020,476, filed Sep. 14, 2020 Final Office Action dated Jun. 9, 2022.
U.S. Appl. No. 17/020,476, filed Sep. 14, 2020 Non-Final Office Action dated Feb. 9, 2022.
William F Garrett et al: “Real-time incremental visualization of dynamic ultrasound volumes using parallel BSP trees”, Visualization '96. Proceedings, IEEE, NE, Oct. 27, 1996, pp. 235-ff, XPO58399771, ISBN: 978-0-89791-864-0 abstract, figures 1-7, pp. 236-240.
M. Ikhsan, K. K. Tan, AS. Putra, C. F. Kong, et al., “Automatic identification of blood vessel cross-section for central venous catheter placement using a cascading classifier,” 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). pp. 1489-1492 (Year: 2017).
PCT/US2022/025097 filed Apr. 15, 2021 International Preliminary Report on Patentability dated Oct. 26, 2023.
PCT/US2023/030970 filed Aug. 23, 2023 International Search Report and Written Opinion dated Oct. 30, 2023.
U.S. Appl. No. 17/468,318, filed Sep. 7, 2021 Advisory Action dated Nov. 6, 2023.
U.S. Appl. No. 17/468,318, filed Sep. 7, 2021 Notice of Allowance dated Jan. 18, 2024.
U.S. Appl. No. 17/471,015, filed Sep. 9, 2021 Advisory Action dated Feb. 2, 2024.
U.S. Appl. No. 17/471,015, filed Sep. 9, 2021 Final Office Action dated Oct. 12, 2023.
U.S. Appl. No. 17/471,015, filed Sep. 9, 2021 Non-Final Office Action dated Mar. 28, 2024.
U.S. Appl. No. 17/534,099, filed Nov. 23, 2021 Advisory Action dated Dec. 8, 2023.
U.S. Appl. No. 17/534,099, filed Nov. 23, 2021 Final Office Action dated Sep. 29, 2023.
U.S. Appl. No. 17/534,099, filed Nov. 23, 2021 Non-Final Office Action dated Mar. 14, 2024.
U.S. Appl. No. 17/538,911, filed Nov. 30, 2021 Advisory Action dated Nov. 22, 2023.
U.S. Appl. No. 17/538,911, filed Nov. 30, 2021 Notice of Allowance dated Mar. 14, 2024.
U.S. Appl. No. 17/538,943, filed Nov. 30, 2021 Non-Final Office Action dated Jan. 30, 2024.
U.S. Appl. No. 17/684,180, filed Mar. 1, 2022 Advisory Action dated Apr. 4, 2024.
U.S. Appl. No. 17/684,180, filed Mar. 1, 2022 Final Office Action dated Jan. 18, 2024.
U.S. Appl. No. 17/722,151, filed Apr. 15, 2022 Advisory Action dated Jan. 2, 2024.
U.S. Appl. No. 17/722,151, filed Apr. 15, 2022 Final Office Action dated Nov. 6, 2023.
U.S. Appl. No. 17/722,151, filed Apr. 15, 2022 Non-Final Office Action dated Mar. 25, 2024.
U.S. Appl. No. 17/894,460, filed Aug. 24, 2022 Advisory Action dated Apr. 4, 2024.
U.S. Appl. No. 17/894,460, filed Aug. 24, 2022 Final Office Action dated Jan. 31, 2024.
U.S. Appl. No. 17/894,460, filed Aug. 24, 2022 Non-Final Office Action dated Nov. 6, 2023.
U.S. Appl. No. 18/238,281, filed Aug. 25, 2023 Non-Final Office Action dated Mar. 22, 2024.
Thermographic venous blood flow characterization with external cooling stimulation (Year: 2018).
U.S. Appl. No. 17/534,099, filed Nov. 23, 2021 Notice of Allowance dated Sep. 18, 2024.
U.S. Appl. No. 17/538,943, filed Nov. 30, 2021 Notice of Allowance dated Aug. 14, 2024.
U.S. Appl. No. 17/684,180, filed Mar. 1, 2022 Final Office Action dated Sep. 23, 2024.
U.S. Appl. No. 17/722,151, filed Apr. 15, 2022 Final Office Action dated Sep. 20, 2024.
U.S. Appl. No. 17/894,460, filed Aug. 24, 2022 Non-Final Office Action dated Sep. 25, 2024.
U.S. Appl. No. 17/979,564, filed Nov. 2, 2022 Final Office Action dated Oct. 18, 2024.
U.S. Appl. No. 17/979,601, filed Nov. 2, 2022 Non-Final Office Action dated Aug. 20, 2024.
U.S. Appl. No. 17/987,698, filed Nov. 15, 2022 Non-Final Office Action dated Sep. 20, 2024.
U.S. Appl. No. 17/684,180, filed Mar. 1, 2022 Non-Final Office Action dated May 8, 2024.
U.S. Appl. No. 17/979,564, filed Nov. 2, 2022 Non-Final Office Action dated Jun. 5, 2024.
U.S. Appl. No. 18/238,281, filed Aug. 25, 2023 Notice of Allowance dated Jul. 16, 2024.
PCT/US2022/045372 filed Sep. 30, 2022 International Search Report and Written Opinion dated Jan. 14, 2023.
PCT/US2024/037647 filed Jul. 11, 2024 International Search Report and Written Opinion dated Oct. 16, 2024.
U.S. Appl. No. 17/471,015, filed Sep. 9, 2021 Notice of Allowance dated Oct. 29, 2024.
U.S. Appl. No. 17/957,562, filed Sep. 30, 2022 Final Office Action dated Nov. 27, 2024.
U.S. Appl. No. 17/957,562, filed Sep. 30, 2022 Non-Final Office Action dated Jun. 20, 2024.
U.S. Appl. No. 17/979,601, filed Nov. 2, 2022 Final Office Action dated Dec. 5, 2024.
U.S. Appl. No. 18/113,003, filed Feb. 22, 2023 Non-Final Office Action dated Nov. 27, 2024.
US 17/684, 180 filed Mar. 1, 2022 Advisory Action dated Dec. 27, 2024.
U.S. Appl. No. 17/722,151 filed Apr. 15, 2022 Advisory Action dated Dec. 27, 2024.
U.S. Appl. No. 17/894,460, filed Aug. 24, 2022 Final Office Action dated Dec. 31, 2024.
U.S. Appl. No. 17/973,171 filed Oct. 25, 2022 Non-Final Office Action dated Dec. 6, 2024.
U.S. Appl. No. 17/987,698, filed Nov. 15, 2022 Final Office Action dated Dec. 13, 2024.
U.S. Appl. No. 18/121,802, filed Mar. 15, 2023 Non-Final Office Action dated Dec. 16, 2024.
U.S. Appl. No. 18/674,601, filed May 24, 2024 Non-Final Office Action dated Jan. 7, 2025.
Related Publications (1)
Number Date Country
20220334251 A1 Oct 2022 US
Provisional Applications (1)
Number Date Country
63175475 Apr 2021 US