This application is also related to U.S. application Ser. No. 11/865,501, filed Oct. 1, 2007, now U.S. Pat. No. 8,007,439, titled “Method and Apparatus to Produce Ultrasonic Images Using Multiple Apertures”, and to U.S. application Ser. No. 11/532,013, filed Sep. 14, 2006, now U.S. Pat. No. 8,105,239, titled “Method and Apparatus to Visualize the Coronary Arteries Using Ultrasound”; all of which are herein incorporated by reference in their entirety.
All publications, including patents and patent applications, mentioned in this specification are herein incorporated by reference in their entirety to the same extent as if each individual publication was specifically and individually indicated to be incorporated by reference.
In conventional ultrasonic imaging, a focused beam of ultrasound energy is transmitted into body tissues to be examined and the returned echoes are detected and plotted to form an image. The basic principles of conventional ultrasonic imaging are well described in the first chapter of “Echocardiography,” by Harvey Feigenbaum (Lippincott Williams & Wilkins, 5th ed., Philadelphia, 1993).
In order to insonify body tissues, an ultrasound beam is typically formed and focused either by a phased array or a shaped transducer. Phased array ultrasound is a commonly used method of steering and focusing a narrow ultrasound beam for forming images in medical ultrasonography. A phased array probe has many small ultrasonic transducer elements, each of which can be pulsed individually. By varying the timing of ultrasound pulses (e.g., by pulsing elements one by one in sequence along a row), a pattern of constructive interference is set up that results in a beam directed at a chosen angle. This is known as beam steering. Such a steered ultrasound beam may then be swept through the tissue or object being examined. Data from multiple beams are then combined to make a visual image showing a slice through the object.
Traditionally, the same transducer or array used for transmitting an ultrasound beam is used to detect the returning echoes. This design configuration lies at the heart of one of the most significant limitations in the use of ultrasonic imaging for medical purposes: poor lateral resolution. Theoretically, the lateral resolution could be improved by increasing the width of the aperture of an ultrasonic probe, but practical problems involved with aperture size increase have kept apertures small. Unquestionably, ultrasonic imaging has been very useful even with this limitation, but it could be more effective with better resolution.
In the practice of cardiology, for example, the limitation on single aperture size is dictated by the space between the ribs (the intercostal spaces). Such intercostal apertures are typically limited to no more than about one to two centimeters. For scanners intended for abdominal and other use, the limitation on aperture size is less a matter of physical constraints, and more a matter of difficulties in image processing. The problem is that it is difficult to keep the elements of a large aperture array in phase because the speed of ultrasound transmission varies with the type of tissue between the probe and the area of interest. According to the book by Wells (cited above), the speed varies up to plus or minus 10% within the soft tissues. When the aperture is kept small (e.g., less than about 2 cm), the intervening tissue is, to a first order of approximation, all the same and any variation is ignored. When the size of the aperture is increased to improve the lateral resolution, the additional elements of a phased array may be out of phase and may actually degrade the image rather than improving it.
US Patent Publication 2008/0103393, now U.S. Pat. No. 8,007,439, to Specht teaches embodiments of ultrasound imaging systems utilizing multiple apertures which may be separated by greater distances, thereby producing significant improvements in lateral resolution of ultrasound images.
One embodiment of a method describes a method of constructing an ultrasound image, comprising transmitting an omni-directional unfocused ultrasound waveform approximating a first point source within a transmit aperture on a first array through a target region, receiving ultrasound echoes from the target region with first and second receiving elements disposed on a first receive aperture on a second array, the first array being physically separated from the second array, determining a first time for the waveform to propagate from the first point source to a first pixel location in the target region to the first receiving element, and determining a second time for the waveform to propagate from the first point source to the first pixel location in the target region to the second receiving element, and forming a first ultrasound image of the first pixel by combining the echo received by the first receiving element at the first time with the echo received by the second receiving element at the second time.
In some embodiments, the method further comprises repeating the determining and forming steps for additional pixel locations in the target region. In one embodiment, additional pixel locations are located on a grid without scan-conversion.
In one embodiment, determining the first time and the second time comprises assuming a uniform speed of sound.
In another embodiment, the method further comprises transmitting a second omni-directional unfocused ultrasound waveform approximating a second point source within the transmit aperture through the target region, receiving ultrasound echoes from the target region with first and second receiving elements disposed on the first receive aperture, determining a third time for the second waveform to propagate from the second point source to the first pixel location in the target region to the first receiving element, and determining a fourth time for the second waveform to propagate from the second point source to the first pixel location in the target region to the second receiving element, and forming a second ultrasound image of the first pixel by combining the echo received by the first receiving element at the third time with the echo received by the second receiving element at the fourth time.
In some embodiments, the method further comprises combining the first ultrasound image with the second ultrasound image. The combining step can comprise coherent addition. In another embodiment, the combining step can comprise incoherent addition. In yet another embodiment, the combining step can comprise a combination of coherent addition and incoherent addition.
In some embodiments, the method can further comprise receiving ultrasound echoes from the target region with third and fourth receiving elements disposed on a second receive aperture on a third array, the third array being physically separated from the first and second arrays, determining a third time for the waveform to propagate from the first point source to the first pixel location in the target region to the third receiving element, and determining a fourth time for the waveform to propagate from the first point source to the first pixel location in the target region to the fourth receiving element, and forming a second ultrasound image of the first pixel by combining the echo received by the third receiving element at the third time with the echo received by the fourth receiving element at the fourth time.
In some embodiments, the method further comprises repeating the determining and forming steps for additional pixel locations in the target region. In some embodiments, the additional pixel locations are located on a grid without scan-conversion.
In one embodiment, the method further comprises transmitting a second omni-directional unfocused ultrasound waveform approximating a second point source within the transmit aperture through the target region, receiving ultrasound echoes from the target region with first and second receiving elements disposed on the first receive aperture and with the third and fourth receiving elements disposed on the second receive aperture, determining a fifth time for the second waveform to propagate from the second point source to the first pixel location in the target region to the first receiving element, determining a sixth time for the second waveform to propagate from the second point source to the first pixel location in the target region to the second receiving element, determining a seventh time for the second waveform to propagate from the second point source to the first pixel location in the target region to the third receiving element, determining an eighth time for the second waveform to propagate from the second point source to the first pixel location in the target region to the fourth receiving element, and forming a third ultrasound image of the first pixel by combining the echo received by the first receiving element at the fifth time with the echo received by the second receiving element at the sixth time, and forming a fourth ultrasound image of the first pixel by combining the echo received by the third receiving element at the seventh time with the echo received by the fourth receiving element at the eighth time.
In some embodiments, the method further comprises combining the first, second, third, and fourth ultrasound images. In some embodiments, the combining step comprises coherent addition. In other embodiments, the combining step comprises incoherent addition. In additional embodiments, the combining step comprises a combination of coherent addition and incoherent addition.
In some embodiments, the method comprises combining the first ultrasound image with the second ultrasound image. The combining step can comprise coherent addition. In another embodiment, the combining step can comprise incoherent addition. In yet another embodiment, the combining step can comprise a combination of coherent addition and incoherent addition.
In some embodiments, the method further comprises comparing the first ultrasound image to the second, third, and fourth ultrasound images to determine displacements of the second, third, and fourth ultrasound images relative to the first ultrasound image.
In another embodiment, the method further comprises correcting the displacements of the second, third, and fourth ultrasound images relative to the first ultrasound image and then combining the first, second, third and fourth ultrasound images.
In an additional embodiment, the method comprises adjusting the third, fourth, fifth, sixth, seventh, and eighth times to correct the displacements of the second, third, and fourth ultrasound images relative to the first ultrasound image.
In some embodiments, the method further comprises comparing the first ultrasound image to the second ultrasound image to determine a displacement of the second ultrasound image relative to the first ultrasound image.
The method can further comprise correcting the displacement of the second ultrasound image relative to the first ultrasound image and then combining the first and second ultrasound images.
In another embodiment, the method comprises adjusting the third time and the fourth time to correct the displacement of the second ultrasound image relative to the first ultrasound image.
In some embodiments, the first pixel is disposed outside a plane defined by the point source, the first receiving element, and the second receiving element. In other embodiments, the first pixel is disposed inside a plane defined by the point source, the first receiving element, and the second receiving element.
Various embodiments of a multi-aperture ultrasound imaging system are also provided, comprising a transmit aperture on a first array configured to transmit an omni-directional unfocused ultrasound waveform approximating a first point source through a target region, a first receive aperture on a second array having first and second receiving elements, the second array being physically separated from the first array, wherein the first and second receiving elements are configured to receive ultrasound echoes from the target region, and a control system coupled to the transmit aperture and the first receive aperture, the control system configured to determine a first time for the waveform to propagate from the first point source to a first pixel location in the target region to the first receiving element, and is configured to determine a second time for the waveform to propagate from the first point source to the first pixel location in the target region to the second receiving element, the control system also being configured to form a first ultrasound image of the first pixel by combining the echo received by the first receiving element at the first time with the echo received by the second receiving element at the second time.
In some embodiments of the system, there are no transducer elements disposed between the physical separation of the transmit aperture and the first receive aperture.
In one embodiment of the system, the transmit aperture and the first receive aperture are separated by at least twice a minimum wavelength of transmission from the transmit aperture. In another embodiment, the transmit aperture and the receive aperture comprise a total aperture ranging from 2 cm to 10 cm.
In some embodiments, the ultrasound system further comprises a second receive aperture on a third array having third and fourth receiving elements, the third array being physically separated from the first and second arrays, wherein the third and fourth receiving elements are configured to receive ultrasound echoes from the target region.
In another embodiment of the multi-aperture ultrasound imaging system, the control system can be coupled to the transmit aperture and the first and second receive apertures, wherein the control system is configured to determine a third time for the waveform to propagate from the first point source to a first pixel location in the target region to the third receiving element, and is configured to determine a fourth time for the waveform to propagate from the first point source to the first pixel location in the target region to the fourth receiving element, the control system also being configured to form a second ultrasound image of the first pixel by combining the echo received by the third receiving element at the third time with the echo received by the fourth receiving element at the fourth time.
In some embodiments, the control system is configured to correct a displacement of the second ultrasound image relative to the first ultrasound image due to speed of sound variation.
In other embodiments of the multi-aperture ultrasound imaging system, the transmit aperture, the first receive aperture, and the second receive aperture are not all in a single scan plane.
Greatly improved lateral resolution in ultrasound imaging can be achieved by using multiple separate apertures for transmit and receive functions. Systems and methods herein may provide for both transmit functions from point sources and for compensation for variations in the speed-of-sound of ultrasound pulses traveling through potentially diverse tissue types along a path between a transmit aperture and one or more receive apertures. Such speed-of-sound compensation may be performed by a combination of image comparison techniques (e.g., cross-correlation), and the coherent and/or incoherent averaging of a plurality of received image frames.
As used herein the terms “ultrasound transducer” and “transducer” may carry their ordinary meanings as understood by those skilled in the art of ultrasound imaging technologies, and may refer without limitation to any single component capable of converting an electrical signal into an ultrasonic signal and/or vice versa. For example, in some embodiments, an ultrasound transducer may comprise a piezoelectric device. In some alternative embodiments, ultrasound transducers may comprise capacitive micromachined ultrasound transducers (CMUT). Transducers are often configured in arrays of multiple elements. An element of a transducer array may be the smallest discrete component of an array. For example, in the case of an array of piezoelectric transducer elements, each element may be a single piezoelectric crystal.
As used herein, the terms “transmit element” and “receive element” may carry their ordinary meanings as understood by those skilled in the art of ultrasound imaging technologies. The term “transmit element” may refer without limitation to an ultrasound transducer element which at least momentarily performs a transmit function in which an electrical signal is converted into an ultrasound signal. Similarly, the term “receive element” may refer without limitation to an ultrasound transducer element which at least momentarily performs a receive function in which an ultrasound signal impinging on the element is converted into an electrical signal. Transmission of ultrasound into a medium may also be referred to herein as “insonifying.” An object or structure which reflects ultrasound waves may be referred to as a “reflector” or a “scatterer.”
As used herein the term “aperture” refers without limitation to one or more ultrasound transducer elements collectively performing a common function at a given instant of time. For example, in some embodiments, the term aperture may refer to a group of transducer elements performing a transmit function. In alternative embodiments, the term aperture may refer to a plurality of transducer elements performing a receive function. In some embodiments, group of transducer elements forming an aperture may be redefined at different points in time.
As used herein the term “point source transmission” may refer to an introduction of transmitted ultrasound energy into a medium from single spatial location. This may be accomplished using a single ultrasound transducer element or combination of adjacent transducer elements transmitting together. A single transmission from said element(s) approximates a uniform spherical wave front, or in the case of imaging a 2D slice it creates a uniform circular wave front within the 2D slice. This point source transmission differs in its spatial characteristics from a “phased array transmission” which focuses energy in a particular direction from the transducer element array. Phased array transmission manipulates the phase of a group of transducer elements in sequence so as to strengthen or steer an insonifying wave to a specific region of interest. A short duration point source transmission is referred to herein as a “point source pulse.” Likewise, a short duration phased array transmission is referred to herein as a “phased array pulse.”
As used herein, the terms “receive aperture,” “insonifying aperture,” and/or “transmit aperture” can carry their ordinary meanings as understood by those skilled in the art of ultrasound imaging, and may refer to an individual element, a group of elements within an array, or even entire arrays within a common housing, that perform the desired transmit or receive function from a desired physical viewpoint or aperture at a given time. In some embodiments, these various apertures may be created as physically separate components with dedicated functionality. In alternative embodiments, the functionality may be electronically designated and changed as needed. In still further embodiments, aperture functionality may involve a combination of both fixed and variable elements.
In some embodiments, an aperture is an array of ultrasound transducers which is separated from other transducer arrays. Such multiple aperture ultrasound imaging systems provide greatly increased lateral resolution. According to some embodiments, a multi-aperture imaging method comprises the steps of insonifying a target object with an ultrasound pulse from a first aperture, detecting returned echoes with a second aperture positioned at a distance from the first aperture, determining the relative positions of the second aperture with respect to the first aperture, and processing returned echo data to combine images while correcting for variations in speed-of-sound through the target object.
In some embodiments, a distance and orientation between adjacent apertures may be fixed relative to one another, such as by use of a rigid housing. In alternative embodiments, distances and orientations of apertures relative to one another may be variable, such as with a movable linkage. In further alternative embodiments, apertures may be defined as groups of elements on a single large transducer array where the groups are separated by at least a specified distance. For example, some embodiments of such a system are shown and described in U.S. Provisional Application No. 61/392,896, filed Oct. 13, 2010, titled “Multiple Aperture Medical Ultrasound Transducers”. In some embodiments of a multi-aperture ultrasound imaging system, a distance between adjacent apertures may be at least a width of one transducer element. In alternative embodiments, a distance between apertures may be as large as possible within the constraints of a particular application and probe design.
A multi-aperture ultrasound imaging system with a large effective aperture (the total aperture of the several sub apertures) can be made viable by compensation for the variation of speed-of-sound in the target tissue. This may be accomplished in one of several ways to enable the increased aperture to be effective rather than destructive, as described below.
Ultrasound elements and arrays described herein may also be multi-function. That is, the designation of transducer elements or arrays as transmitters in one instance does not preclude their immediate re-designation as receivers in the next instance. Moreover, embodiments of the control system described herein include the capabilities for making such designations electronically based on user inputs or pre-set scan or resolution criteria.
Another embodiment of a multi-aperture ultrasound imaging system 200 is shown in
Multi-aperture ultrasound imaging systems described herein may be configured to utilize transducers of any desired construction. For example, 1D, 1.5D, 2D, CMUT or any other transducer arrays may be utilized in multi-aperture configurations to improve overall resolution and field of view.
Point Source Transmission
In some embodiments, acoustic energy may be transmitted to as wide a two-dimensional slice as possible by using point source transmission. For example, in some embodiments, a transmit aperture, such as transmit apertures 102 or 202 in
Each echo detected at a receive aperture, such as receive apertures 104 or 204/206 in
Although several point source pulses are typically used in order to produce a high-quality image, fewer point source pulses are required than if each pulse were focused on a particular scan line. Since the number of pulses that can be transmitted in a given time is strictly limited by the speed of ultrasound in tissue, this yields the practical advantage that more frames can be produced per second by utilizing a point source pulse. This is very important when imaging moving organs, and in particular, the heart.
In some embodiments, a spread spectrum waveform may be imposed on a transmit aperture made up of one or more ultrasound transducer elements. A spread spectrum waveform may be a sequence of frequencies such as a chirp (e.g., frequencies progressing from low to high, or vice versa), random frequency sequence (also referred to as frequency hop), or a signal generated by a pseudo random waveform (PN sequence). These techniques can be collectively referred to as pulse compression. Pulse compression provides longer pulses for greater depth penetration without loss of depth resolution. In fact, the depth resolution may be greatly improved in the process. Spread spectrum processing typically involves much more signal processing in the form of matched filtering of each of the received signals before the delay and summation steps. The above examples of transmit pulse forms are provided for illustration only. The techniques taught herein may apply regardless of the form of the transmit pulse.
Basic Image Rendering
In one embodiment, an ultrasound image may be produced by insonifying the entire region to be imaged, such as internal tissue or target object T, (e.g., a plane through the heart, organ, tumor, or other portion of the body) with a transmitting element (e.g., transmit element T1 of aperture 102), and then receiving echoes from the entire imaged plane on a receive element (e.g., receive element R1 of aperture 104). In some embodiments, receive functions may be performed by all elements in the receive probe (e.g., R1 through Rm). In alternative embodiments, echoes are received on only one or a select few elements of the receive aperture. The method proceeds by using each of the elements on the transmitting aperture 102 (e.g., T2, . . . Tn) and insonifying the entire region to be imaged with each of the transmitting elements in turn, and receiving echoes on the receive aperture after each insonifying pulse. Transmit elements may be operated in any desired sequential order, and need not follow a prescribed pattern. Individually, the images obtained after insonification by each transmitting element may not be sufficient to provide a high resolution image, but the combination of all the images may provide a high resolution image of the entire region to be imaged. For a scanning point represented by coordinates (i,j) as shown in
When the speed of ultrasound in tissue is assumed to be uniform throughout the tissue, it is possible to calculate the time delay from the onset of the transmit pulse to the time that an echo is received at the receive element. (Non uniform speed-of-sound in tissue is discussed below.) This one fact means that a scatterer (i.e., a reflective point within the target object) is a point in the medium for which a+b=the given time delay. The same method can be used to calculate delays for any point in the desired tissue to be imaged, creating a locus of points.
The fact that all points on the ellipse 180 are returned with the same time delay presents a display challenge, since distinguishing points along the ellipse from one another within a single image is not possible. However, by combining images obtained from multiple receive points, the points may be more easily distinguished, since the equal-time-delay ellipses defined by the multiple receive apertures will be slightly different.
A method of rendering the location of all of the scatterers in the target object, and thus forming a two dimensional cross section of the target object, will now be described with reference to multi-aperture ultrasound imaging system 300 of
In one embodiment, the following pseudo code may be used to accumulate all of the information to be gathered from a transmit pulse from one transmit element (e.g., one element of T1 . . . Tn from aperture 302), and the consequent echoes received by one receive element (e.g., one element of R1 . . . Rm from aperture 304) in the arrangement of
The fixed delay is primarily the time from the transmit pulse until the first echoes are received. As will be discussed later, an increment can be added or subtracted to compensate for varying fat layers.
A complete two dimensional image may be formed by repeating this process for every receive element in aperture 304 (e.g., R1 . . . Rm). In some embodiments, it is possible to implement this code in parallel hardware resulting in real time image formation.
Combining similar images resulting from pulses from other transmit elements will improve the quality (e.g., in terms of signal-to-noise ratio) of the image. In some embodiments, the combination of images may be performed by a simple summation of the single point source pulse images (e.g., coherent addition). Alternatively, the combination may involve taking the absolute value of each element of the single point source pulse images first before summation (e.g., incoherent addition). In some embodiments, the first technique (coherent addition) may be best used for improving lateral resolution, and the second technique (incoherent addition) may be best applied for the reduction of speckle noise. In addition, the incoherent technique may be used with less precision required in the measurement of the relative positions of the transmit and receive apertures. A combination of both techniques may be used to provide an optimum balance of improved lateral resolution and reduced speckle noise. Finally, in the case of coherent addition, the final sum should be replaced by the absolute value of each element, and in both cases, some form of compression of the dynamic range may be used so that both prominent features and more-subtle features appear on the same display. In some embodiments, additional pixel locations are located on a grid without scan-conversion.
In some embodiments, compression schemes may include taking the logarithm (e.g., 20 log10 or “dB”) of each element before display, or taking the nth root (e.g., 4th root) of each element before display. Other compression schemes may also be employed.
Referring still to
The speed-of-sound in various soft tissues throughout the body can vary by +/−10%. Using typical ultrasound techniques, it is commonly assumed that the speed-of-sound is constant in the path between the transducer and the organ of interest. This assumption is valid for narrow transducer arrays in systems using one transducer array for both transmit and receive. However, the constant speed-of-sound assumption breaks down as the transducer's aperture becomes wider because the ultrasound pulses pass through more tissue and possibly diverse types of tissue, such as fat, muscle, blood vessels, etc. Tissue diversity under the width of the transducer array affects both the transmit and the receive functions.
When a scatterer is insonified by a point source pulse from a single transmit element, it reflects back an echo to all of the elements of the receiver group. Coherent addition of images collected by elements in this receive aperture can be effective if the speed-of-sound variations in the paths from scatterer (i,j) to each of the receiver elements do not exceed +−180 degrees phase shift relative to one path chosen as reference. Referring to
When insonifying with unfocused point source pulses, the aperture size of the transmit group is not highly critical since variation in the path time from transmitter elements to a scatterer such as scatterer (i,j) will change only the displayed position of the point. For example, a variation resulting in a phase shift of 180 degrees in the receive paths results in complete phase cancellation when using coherent addition, whereas the same variation on the transmit paths results in a displayed position error of only a half wavelength (typically about 0.2 mm), a distortion that would not be noticed.
Thus, in a multi-aperture imaging system with one aperture used only for transmit and the other used only for receive during a single transmit/receive cycle, as is illustrated in
Substantial improvement in lateral resolution is achieved with a receive aperture of the same width as a conventional single array 1D, 1.5D or 2D ultrasound probe used for both transmit and receive, because received energy when imaging adjacent cells (i.e., regions of the target object) to that which represents a scatterer is dependent on the time difference between when an echo is expected to arrive and the time that it actually arrives. When the transmit pulse originates from the same array used for receive, the time difference is small. However, when the transmit pulse originates from a second array at some distance from the receive array, the time difference is larger and therefore more out of phase with the signal for the correct cell. The result is that fewer adjacent cells will have signals sufficiently in phase to falsely represent the true scatterer.
Referring to
Now consider moving the transmitting aperture 402 away from the receive aperture 404 by an angle theta (“θ”). For convenience in comparison, let the distance “b” from aperture 402 to scatterer “S” be equal to the distance “a” from aperture 404 to scatterer “S”. The time for the ultrasound to propagate from the transmit aperture 402 to “S” and be returned to the receive aperture 404 would still be (a+b)/C=2a/C (with a=b), but the expected time for the signal to propagate to the adjacent cell “S′” would be (d+sqrt(a2+c2))/C=(sqrt((a sin θ−c)2+(a cos θ)2)+sqrt(a2+c2))/C. The difference between the expected time of arrival and actual would then be Diff=(sqrt((a sin θ−c)2+(a cos θ)2)+sqrt(a2+c2)−2a)/C.
To put some numbers in this equation, suppose that the separation of aperture 402 and aperture 404 is only 5 degrees, distance a=400 cells, and distance c=1 cell. Then the ratio of the difference in time-of-arrival for θ=5 degrees to that for θ=0 degrees is 33.8. That is, the drop off of display amplitude to adjacent cells is 33 times faster with θ=5 degrees. The larger difference in time-of-arrival greatly simplifies the ability to uniquely distinguish echo information from adjacent cells. Therefore, with high theta angles, the display of a point will be less visible as noise in adjacent cells and the result will be higher resolution of the real image. With multiple aperture transmitters and receivers, we can make the angle as high as needed to improve resolution.
Simulation for a realistic ultrasound system with multiple reflectors in multiple cells shows that the effect is still significant, but not as dramatic as above. For a system comprising a receive aperture of 63 elements, a θ of 10 degrees, and a transmit pulse from a point-source transmit aperture that extends for 5 cycles with cosine modulation, the lateral spread of the point spread function was improved by a factor of 2.3.
Explicit Compensation for Speed-of-Sound Variation
A single image may be formed by coherent averaging of all of the signals arriving at the receiver elements as a result of a single point source pulse for insonification. Summation of these images resulting from multiple point source pulses can be accomplished either by coherent addition, incoherent addition, or a combination of coherent addition by groups and incoherent addition of the images from the groups. Coherent addition (retaining the phase information before addition) maximizes resolution whereas incoherent addition (using the magnitude of the signals and not the phase) minimizes the effects of registration errors and averages out speckle noise. Some combination of the two modes may be preferred. Coherent addition can be used to average point source pulse images resulting from transmit elements that are close together and therefore producing pulses transmitted through very similar tissue layers. Incoherent addition can then be used where phase cancellation would be a problem. In the extreme case of transmission time variation due to speed-of-sound variations, 2D image correlation can be used to align images prior to addition.
When an ultrasound imaging system includes a second aperture, using the second aperture for receiving as well as transmitting produces much better resolution. In combining the images from two or more receive arrays; it is possible and beneficial to use explicit compensation for the speed-of-sound variation.
Consider the tissue layer model for the three-aperture ultrasound imaging system 500 as shown in
Image comparison techniques may be used to determine the amount of displacement needed to align image frames from left and right apertures (e.g., apertures 506 and 504, respectively). In one embodiment, the image comparison technique can be cross-correlation. Cross-correlation involves evaluating the similarity of images or image sections to identify areas with a high degree of similarity. Areas with at least a threshold value of similarity may be assumed to be the same. Thus, by identifying areas within images with high degrees of similarity, one image (or a section thereof) may be shifted such that areas with substantial similarity overlap and enhance overall image quality.
Further, these image comparison techniques can also be used by applying sub-image analysis, which can be used to determine displacement of sub-images and accommodate for localized variation in speed-of-sound in the underlying tissue. In other words, by breaking down the images into smaller segments (e.g., halves, thirds, quarters, etc.), small portions of a first image may be compared to the corresponding small portion of a second image. The two images may then be combined by warping to assure alignment. Warping is a technique understood by those skilled in the art, and is described, for example in U.S. Pat. No. 7,269,299 to Schroeder.
The same technique of incoherent addition of images from multiple receive transducer arrays may be applied to any number of apertures. The same idea may be applied even to a single element array which is too wide to be used for coherent addition all at once. An ultrasound imaging system with a single wide array of elements may be divided into sections (apertures) each of which is small enough for coherent addition, and then the images resulting from these sections may be combined incoherently (with displacement if necessary).
Even a slight distortion of the image may be compensated for with sufficient computational power. Image renderings may be computed for one receive array using varying amounts of delay in the rendering algorithm (echo[i][j]=echo[i][j]+stored receive echo at the computed time+delay). Then the best matched of these (by cross-correlation or some other measure of acuity) may be incoherently added to the image from the other receive array(s). A faster technique includes calculating the cross correlation network for the uncorrected pair of images, and feeding this into a neural network trained to pick the correction delay.
Because multiple aperture ultrasound systems that can correct for speed of sound incongruences allow for significantly larger apertures, some embodiments of the multi-aperture ultrasound systems described herein can have apertures located 10 cm apart from one another. Since resolution is proportional to 2λ/D, this larger aperture leads to higher resolution of tissues located well below the surface of the skin. For instance, the renal arteries are frequently located 10 cm to 15 cm below the skin and are 4 mm to 6 mm in size near the abdominal aorta. Phased array, linear array and synthetic aperture ultrasound systems usually cannot detect this physiology in most patients; specifically because the aperture size is not large enough to have adequate lateral resolution. Typically, phased array systems have aperture sizes of approximately 2 cm. Increasing the aperture size from larger than 2 cm to approximately 10 cm in a multi-aperture ultrasound system can increase the resolution by up to 5×.
3D Imaging
In some embodiments, three-dimensional information may be obtained by moving a two-dimensional imaging system and acquiring 2D slices at a number of positions or angles. From this information and using interpolation techniques, a 3D image at any position or angle may be reconstructed. Alternatively, a 2D projection of all of the data in the 3D volume may be produced. A third alternative is to use the information in a direct 3D display.
Because multi-aperture ultrasound imaging systems may result in wider probe devices, the easiest way to use them to obtain 3D data is to not move them on the patient's skin but merely rock them so that the 2D slices span the 3D volume to be imaged. In some embodiments, a mechanical rotator mechanism which records position data may be used to assist in the collection the 2D slices. In other embodiments, a freely operated ultrasound probe with precision position sensors (such as gyroscopic sensors) located in the head of the probe may be used instead. Such an apparatus allows for complete freedom of movement while collecting 2D slices. Finally, intravenous and intracavity probes may also be manufactured to accommodate wide apertures. Such probes may be manipulated in similar ways in order to collect 2D slices.
This combination is particularly desirable for 3D cardiac imaging using a multi-aperture ultrasound imaging system. Most patients have good acoustic windows in two intercostal spaces next to the sternum. A multi-aperture imaging system is ideal in this case since the intervening rib would render a flat probe useless, while a probe with at least two widely spaced apertures can be positioned such that a send aperture and a receive aperture align with separate intercostal spaces. Once a probe with multiple apertures is in place, it cannot be rotated, but it can be rocked to obtain the 3D information. A multi-aperture probe may also be used in the same intercostal space but across the sternum.
3D information may also be obtained directly with multi-aperture imaging systems having apertures that are not all in the same scan plane. In this case the elements making up the transmit aperture preferably propagate spherical waveforms (rather than circular waveforms confined to one plane of scan). The elements making up the receive apertures may likewise be sensitive to energy arriving from all directions (rather than being sensitive only to ultrasonic energy in a single plane of scan). The reconstruction pseudo code provided above may then be extended to three dimensions.
As for additional details pertinent to the present invention, materials and manufacturing techniques may be employed as within the level of those with skill in the relevant art. The same may hold true with respect to method-based aspects of the invention in terms of additional acts commonly or logically employed. Also, it is contemplated that any optional feature of the inventive variations described may be set forth and claimed independently, or in combination with any one or more of the features described herein. Likewise, reference to a singular item, includes the possibility that there are plural of the same items present. More specifically, as used herein and in the appended claims, the singular forms “a,” “and,” “said,” and “the” include plural referents unless the context clearly dictates otherwise. It is further noted that the claims may be drafted to exclude any optional element. As such, this statement is intended to serve as antecedent basis for use of such exclusive terminology as “solely,” “only” and the like in connection with the recitation of claim elements, or use of a “negative” limitation. Unless defined otherwise herein, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. The breadth of the present invention is not to be limited by the subject specification, but rather only by the plain meaning of the claim terms employed.
This application is a continuation of U.S. application Ser. No. 16/506,570, filed Jul. 9, 2019, which is a continuation of U.S. application Ser. No. 16/000,507, filed Jun. 5, 2018, now U.S. Pat. No. 10,342,518, which is a continuation of U.S. application Ser. No. 15/364,075, filed Nov. 29, 2016, now U.S. Pat. No. 9,986,975, which is a continuation of U.S. application Ser. No. 14/846,374, filed Sep. 4, 2015, now U.S. Pat. No. 9,526,475, which is a continuation of U.S. application Ser. No. 13/029,907, filed Feb. 17, 2011, now U.S. Pat. No. 9,146,313, which application claims the benefit of U.S. Provisional Application No. 61/305,784, filed Feb. 18, 2010, entitled “Alternative Method for Medical Multi-Aperture Ultrasound Imaging”.
Number | Name | Date | Kind |
---|---|---|---|
3174286 | Erickson | Mar 1965 | A |
3895381 | Kock | Jul 1975 | A |
3974692 | Hassler | Aug 1976 | A |
4055988 | Dutton | Nov 1977 | A |
4072922 | Taner et al. | Feb 1978 | A |
4097835 | Green | Jun 1978 | A |
4105018 | Greenleaf et al. | Aug 1978 | A |
4180792 | Lederman et al. | Dec 1979 | A |
4205394 | Pickens | May 1980 | A |
4229798 | Rosie | Oct 1980 | A |
4259733 | Taner et al. | Mar 1981 | A |
4265126 | Papadofrangakis et al. | May 1981 | A |
4271842 | Specht et al. | Jun 1981 | A |
4325257 | Kino et al. | Apr 1982 | A |
4327738 | Green et al. | May 1982 | A |
4328569 | Trott et al. | May 1982 | A |
4333474 | Nigam | Jun 1982 | A |
4339952 | Foster | Jul 1982 | A |
4452084 | Taenzer | Jun 1984 | A |
4501279 | Seo | Feb 1985 | A |
4511998 | Kanda et al. | Apr 1985 | A |
4539847 | Paap | Sep 1985 | A |
4566459 | Umemura et al. | Jan 1986 | A |
4567768 | Satoh et al. | Feb 1986 | A |
4604697 | Luthra et al. | Aug 1986 | A |
4662222 | Johnson | May 1987 | A |
4669482 | Ophir | Jun 1987 | A |
4682497 | Sasaki | Jul 1987 | A |
4694434 | Von Ramm et al. | Sep 1987 | A |
4781199 | Hirama et al. | Nov 1988 | A |
4817434 | Anderson | Apr 1989 | A |
4831601 | Breimesser et al. | May 1989 | A |
4893284 | Magrane | Jan 1990 | A |
4893628 | Angelsen | Jan 1990 | A |
4990462 | Sliwa, Jr. | Feb 1991 | A |
5027658 | Anderson | Jul 1991 | A |
5050588 | Grey et al. | Sep 1991 | A |
5060205 | Phelan | Oct 1991 | A |
5062295 | Shakkottai et al. | Nov 1991 | A |
5141738 | Rasor et al. | Aug 1992 | A |
5161536 | Vilkomerson et al. | Nov 1992 | A |
5197475 | Antich et al. | Mar 1993 | A |
5226019 | Bahorich | Jul 1993 | A |
5230339 | Charlebois | Jul 1993 | A |
5269309 | Fort et al. | Dec 1993 | A |
5278757 | Hoctor et al. | Jan 1994 | A |
5293871 | Reinstein et al. | Mar 1994 | A |
5299576 | Shiba | Apr 1994 | A |
5301674 | Erikson et al. | Apr 1994 | A |
5305756 | Entrekin et al. | Apr 1994 | A |
5339282 | Kuhn et al. | Aug 1994 | A |
5340510 | Bowen | Aug 1994 | A |
5345426 | Lipschutz | Sep 1994 | A |
5349960 | Gondo | Sep 1994 | A |
5355888 | Kendall | Oct 1994 | A |
5381794 | Tel et al. | Jan 1995 | A |
5398216 | Hall et al. | Mar 1995 | A |
5409010 | Beach et al. | Apr 1995 | A |
5442462 | Guissin | Aug 1995 | A |
5454372 | Banjanin et al. | Oct 1995 | A |
5477858 | Norris et al. | Dec 1995 | A |
5503152 | Oakley et al. | Apr 1996 | A |
5515853 | Smith et al. | May 1996 | A |
5515856 | Olstad et al. | May 1996 | A |
5522393 | Phillips et al. | Jun 1996 | A |
5526815 | Granz et al. | Jun 1996 | A |
5544659 | Banjanin | Aug 1996 | A |
5558092 | Unger | Sep 1996 | A |
5564423 | Mele et al. | Oct 1996 | A |
5568812 | Murashita et al. | Oct 1996 | A |
5570691 | Wright et al. | Nov 1996 | A |
5581517 | Gee et al. | Dec 1996 | A |
5625149 | Gururaja et al. | Apr 1997 | A |
5628320 | Teo | May 1997 | A |
5666953 | Wilk | Sep 1997 | A |
5673697 | Bryan et al. | Oct 1997 | A |
5675550 | Ekhaus | Oct 1997 | A |
5720291 | Schwartz | Feb 1998 | A |
5720708 | Lu et al. | Feb 1998 | A |
5744898 | Smith et al. | Apr 1998 | A |
5769079 | Hossack | Jun 1998 | A |
5784334 | Sena et al. | Jul 1998 | A |
5785654 | Iinuma et al. | Jul 1998 | A |
5795297 | Daigle | Aug 1998 | A |
5797845 | Barabash et al. | Aug 1998 | A |
5798459 | Ohba et al. | Aug 1998 | A |
5817023 | Daft | Oct 1998 | A |
5820561 | Olstad et al. | Oct 1998 | A |
5838564 | Bahorich et al. | Nov 1998 | A |
5850622 | Vassiliou et al. | Dec 1998 | A |
5862100 | VerWest | Jan 1999 | A |
5870691 | Partyka et al. | Feb 1999 | A |
5871446 | Wilk | Feb 1999 | A |
5876342 | Chen et al. | Mar 1999 | A |
5891038 | Seyed-Bolorforosh et al. | Apr 1999 | A |
5892732 | Gersztenkorn | Apr 1999 | A |
5916169 | Hanafy et al. | Jun 1999 | A |
5919139 | Lin | Jul 1999 | A |
5920285 | Benjamin | Jul 1999 | A |
5930730 | Marfurt et al. | Jul 1999 | A |
5938612 | Kline-Schoder | Aug 1999 | A |
5940778 | Marfurt et al. | Aug 1999 | A |
5951479 | Holm et al. | Sep 1999 | A |
5964707 | Fenster et al. | Oct 1999 | A |
5969661 | Benjamin | Oct 1999 | A |
5999836 | Nelson et al. | Dec 1999 | A |
6007499 | Martin et al. | Dec 1999 | A |
6013032 | Savord | Jan 2000 | A |
6014473 | Hossack et al. | Jan 2000 | A |
6048315 | Chiao et al. | Apr 2000 | A |
6049509 | Sonneland et al. | Apr 2000 | A |
6050943 | Slayton et al. | Apr 2000 | A |
6056693 | Haider | May 2000 | A |
6058074 | Swan et al. | May 2000 | A |
6077224 | Lang et al. | Jun 2000 | A |
6092026 | Bahorich et al. | Jul 2000 | A |
6122538 | Sliwa, Jr. et al. | Sep 2000 | A |
6123670 | Mo | Sep 2000 | A |
6129672 | Seward et al. | Oct 2000 | A |
6135960 | Holmberg | Oct 2000 | A |
6138075 | Yost | Oct 2000 | A |
6148095 | Prause et al. | Nov 2000 | A |
6162175 | Marian, Jr. et al. | Dec 2000 | A |
6166384 | Dentinger et al. | Dec 2000 | A |
6166853 | Sapia et al. | Dec 2000 | A |
6190318 | Bab et al. | Feb 2001 | B1 |
6193665 | Hall et al. | Feb 2001 | B1 |
6196739 | Silverbrook | Mar 2001 | B1 |
6200266 | Shokrollahi et al. | Mar 2001 | B1 |
6210335 | Miller | Apr 2001 | B1 |
6213958 | Winder | Apr 2001 | B1 |
6221019 | Kantorovich | Apr 2001 | B1 |
6224556 | Schwartz et al. | May 2001 | B1 |
6231511 | Bae | May 2001 | B1 |
6238342 | Feleppa et al. | May 2001 | B1 |
6246901 | Benaron | Jun 2001 | B1 |
6251073 | Imran et al. | Jun 2001 | B1 |
6264609 | Herrington et al. | Jul 2001 | B1 |
6266551 | Osadchy et al. | Jul 2001 | B1 |
6278949 | Alam | Aug 2001 | B1 |
6279399 | Holm | Aug 2001 | B1 |
6289230 | Chaiken et al. | Sep 2001 | B1 |
6299580 | Asafusa | Oct 2001 | B1 |
6304684 | Niczyporuk et al. | Oct 2001 | B1 |
6309356 | Ustuner et al. | Oct 2001 | B1 |
6324453 | Breed et al. | Nov 2001 | B1 |
6345539 | Rawes et al. | Feb 2002 | B1 |
6361500 | Masters | Mar 2002 | B1 |
6363033 | Cole et al. | Mar 2002 | B1 |
6370480 | Gupta et al. | Apr 2002 | B1 |
6373984 | Gouge et al. | Apr 2002 | B1 |
6374185 | Taner et al. | Apr 2002 | B1 |
6394955 | Perlitz | May 2002 | B1 |
6423002 | Hossack | Jul 2002 | B1 |
6431175 | Penner et al. | Aug 2002 | B1 |
6436046 | Napolitano et al. | Aug 2002 | B1 |
6449821 | Sudol et al. | Sep 2002 | B1 |
6450965 | Williams et al. | Sep 2002 | B2 |
6464637 | Criton et al. | Oct 2002 | B1 |
6468216 | Powers et al. | Oct 2002 | B1 |
6471650 | Powers et al. | Oct 2002 | B2 |
6475150 | Haddad | Nov 2002 | B2 |
6480790 | Calvert et al. | Nov 2002 | B1 |
6487502 | Taner | Nov 2002 | B1 |
6490477 | Zylka et al. | Dec 2002 | B1 |
6499536 | Ellingsen | Dec 2002 | B1 |
6503204 | Sumanaweera et al. | Jan 2003 | B1 |
6508768 | Hall et al. | Jan 2003 | B1 |
6508770 | Cai | Jan 2003 | B1 |
6514205 | Lee et al. | Feb 2003 | B1 |
6517484 | Wilk et al. | Feb 2003 | B1 |
6526163 | Halmann et al. | Feb 2003 | B1 |
6543272 | Vitek | Apr 2003 | B1 |
6547732 | Jago | Apr 2003 | B2 |
6551246 | Ustuner et al. | Apr 2003 | B1 |
6565510 | Haider | May 2003 | B1 |
6582367 | Robinson et al. | Jun 2003 | B1 |
6585647 | Winder | Jul 2003 | B1 |
6597171 | Hurlimann et al. | Jul 2003 | B2 |
6604421 | Li | Aug 2003 | B1 |
6614560 | Silverbrook | Sep 2003 | B1 |
6620101 | Azzam et al. | Sep 2003 | B2 |
6629929 | Jago et al. | Oct 2003 | B1 |
6645147 | Jackosn et al. | Nov 2003 | B1 |
6652461 | Levkovitz | Nov 2003 | B1 |
6668654 | Dubois et al. | Dec 2003 | B2 |
6672165 | Rather et al. | Jan 2004 | B2 |
6681185 | Young et al. | Jan 2004 | B1 |
6690816 | Aylward et al. | Feb 2004 | B2 |
6692450 | Coleman | Feb 2004 | B1 |
6695778 | Golland et al. | Feb 2004 | B2 |
6702745 | Smythe | Mar 2004 | B1 |
6704692 | Banerjee et al. | Mar 2004 | B1 |
6719693 | Richard | Apr 2004 | B2 |
6728567 | Rather et al. | Apr 2004 | B2 |
6752762 | DeJong et al. | Jun 2004 | B1 |
6755787 | Hossack et al. | Jun 2004 | B2 |
6780152 | Ustuner et al. | Aug 2004 | B2 |
6790182 | Eck et al. | Sep 2004 | B2 |
6835178 | Wilson et al. | Dec 2004 | B1 |
6837853 | Marian | Jan 2005 | B2 |
6843770 | Sumanaweera | Jan 2005 | B2 |
6847737 | Kouri et al. | Jan 2005 | B1 |
6854332 | Alleyne | Feb 2005 | B2 |
6865140 | Thomenius et al. | Mar 2005 | B2 |
6932767 | Landry et al. | Aug 2005 | B2 |
7033320 | Von Behren et al. | Apr 2006 | B2 |
7087023 | Daft et al. | Aug 2006 | B2 |
7104956 | Christopher | Sep 2006 | B1 |
7217243 | Takeuchi | May 2007 | B2 |
7221867 | Silverbrook | May 2007 | B2 |
7231072 | Yamano et al. | Jun 2007 | B2 |
7269299 | Schroeder | Sep 2007 | B2 |
7283652 | Mendonca et al. | Oct 2007 | B2 |
7285094 | Nohara et al. | Oct 2007 | B2 |
7293462 | Lee et al. | Nov 2007 | B2 |
7313053 | Wodnicki | Dec 2007 | B2 |
7366704 | Reading et al. | Apr 2008 | B2 |
7402136 | Hossack et al. | Jul 2008 | B2 |
7410469 | Talish et al. | Aug 2008 | B1 |
7415880 | Renzel | Aug 2008 | B2 |
7443765 | Thomenius et al. | Oct 2008 | B2 |
7444875 | Wu et al. | Nov 2008 | B1 |
7447535 | Lavi | Nov 2008 | B2 |
7448998 | Robinson | Nov 2008 | B2 |
7466848 | Metaxas et al. | Dec 2008 | B2 |
7469096 | Silverbrook | Dec 2008 | B2 |
7474778 | Shinomura et al. | Jan 2009 | B2 |
7481577 | Ramamurthy et al. | Jan 2009 | B2 |
7491171 | Barthe et al. | Feb 2009 | B2 |
7497828 | Wilk et al. | Mar 2009 | B1 |
7497830 | Li | Mar 2009 | B2 |
7510529 | Chou et al. | Mar 2009 | B2 |
7514851 | Wilser et al. | Apr 2009 | B2 |
7549962 | Dreschel et al. | Jun 2009 | B2 |
7574026 | Rasche et al. | Aug 2009 | B2 |
7625343 | Cao et al. | Dec 2009 | B2 |
7637869 | Sudol | Dec 2009 | B2 |
7668583 | Fegert et al. | Feb 2010 | B2 |
7674228 | Williams et al. | Mar 2010 | B2 |
7682311 | Simopoulos et al. | Mar 2010 | B2 |
7699776 | Walker et al. | Apr 2010 | B2 |
7722541 | Cai | May 2010 | B2 |
7744532 | Ustuner et al. | Jun 2010 | B2 |
7750311 | Daghighian | Jul 2010 | B2 |
7764984 | Desmedt et al. | Jul 2010 | B2 |
7785260 | Umemura et al. | Aug 2010 | B2 |
7787680 | Ahn et al. | Aug 2010 | B2 |
7806828 | Stringer | Oct 2010 | B2 |
7819810 | Stringer et al. | Oct 2010 | B2 |
7822250 | Yao et al. | Oct 2010 | B2 |
7824337 | Abe et al. | Nov 2010 | B2 |
7833163 | Cai | Nov 2010 | B2 |
7837624 | Hossack et al. | Nov 2010 | B1 |
7846097 | Jones et al. | Dec 2010 | B2 |
7850613 | Stribling | Dec 2010 | B2 |
7862508 | Davies et al. | Jan 2011 | B2 |
7876945 | Lötjönen | Jan 2011 | B2 |
7880154 | Otto | Feb 2011 | B2 |
7887486 | Ustuner et al. | Feb 2011 | B2 |
7901358 | Mehl et al. | Mar 2011 | B2 |
7914451 | Davies | Mar 2011 | B2 |
7919906 | Cerofolini | Apr 2011 | B2 |
7926350 | Kröning et al. | Apr 2011 | B2 |
7927280 | Davidsen | Apr 2011 | B2 |
7972271 | Johnson et al. | Jul 2011 | B2 |
7984637 | Ao et al. | Jul 2011 | B2 |
7984651 | Randall et al. | Jul 2011 | B2 |
8002705 | Napolitano et al. | Aug 2011 | B1 |
8007439 | Specht | Aug 2011 | B2 |
8057392 | Hossack et al. | Nov 2011 | B2 |
8057393 | Yao et al. | Nov 2011 | B2 |
8079263 | Randall et al. | Dec 2011 | B2 |
8079956 | Azuma et al. | Dec 2011 | B2 |
8088067 | Vortman et al. | Jan 2012 | B2 |
8088068 | Yao et al. | Jan 2012 | B2 |
8088071 | Hwang et al. | Jan 2012 | B2 |
8105239 | Specht | Jan 2012 | B2 |
8135190 | Bae et al. | Mar 2012 | B2 |
8157737 | Zhang et al. | Apr 2012 | B2 |
8182427 | Wu et al. | May 2012 | B2 |
8202219 | Luo et al. | Jun 2012 | B2 |
8265175 | Barsoum et al. | Sep 2012 | B2 |
8277383 | Specht | Oct 2012 | B2 |
8279705 | Choi et al. | Oct 2012 | B2 |
8412307 | Willis et al. | Apr 2013 | B2 |
8414564 | Goldshleger et al. | Apr 2013 | B2 |
8419642 | Sandrin et al. | Apr 2013 | B2 |
8473239 | Specht et al. | Jun 2013 | B2 |
8478382 | Burnside et al. | Jul 2013 | B2 |
8483804 | Hsieh et al. | Jul 2013 | B2 |
8532951 | Roy et al. | Sep 2013 | B2 |
8539838 | Yoo et al. | Sep 2013 | B2 |
8582848 | Funka-Lea et al. | Nov 2013 | B2 |
8602993 | Specht et al. | Dec 2013 | B2 |
8627724 | Papadopoulos et al. | Jan 2014 | B2 |
8634615 | Brabec | Jan 2014 | B2 |
8672846 | Napolitano et al. | Mar 2014 | B2 |
8684936 | Specht | Apr 2014 | B2 |
9036887 | Fouras et al. | May 2015 | B2 |
9072495 | Specht | Jul 2015 | B2 |
9146313 | Specht et al. | Sep 2015 | B2 |
9192355 | Smith et al. | Nov 2015 | B2 |
9220478 | Smith et al. | Dec 2015 | B2 |
9247926 | Smith et al. | Feb 2016 | B2 |
9265484 | Brewer et al. | Feb 2016 | B2 |
9282945 | Smith et al. | Mar 2016 | B2 |
9339256 | Specht et al. | May 2016 | B2 |
9392986 | Ning et al. | Jul 2016 | B2 |
9420994 | Specht | Aug 2016 | B2 |
9510806 | Smith et al. | Dec 2016 | B2 |
9526475 | Specht et al. | Dec 2016 | B2 |
9572549 | Belevich et al. | Feb 2017 | B2 |
9582876 | Specht | Feb 2017 | B2 |
9606206 | Boernert et al. | Mar 2017 | B2 |
9668714 | Call et al. | Jun 2017 | B2 |
9788813 | Adam et al. | Oct 2017 | B2 |
9883848 | Specht et al. | Feb 2018 | B2 |
9986969 | Call et al. | Jun 2018 | B2 |
9986975 | Specht et al. | Jun 2018 | B2 |
10064605 | Belevich et al. | Sep 2018 | B2 |
10130333 | Specht | Nov 2018 | B2 |
10206662 | Smith et al. | Feb 2019 | B2 |
10226234 | Specht et al. | Mar 2019 | B2 |
10267913 | Smith et al. | Apr 2019 | B2 |
10342518 | Specht et al. | Jul 2019 | B2 |
10380399 | Call et al. | Aug 2019 | B2 |
10401493 | Call et al. | Sep 2019 | B2 |
10617384 | Brewer et al. | Apr 2020 | B2 |
10653392 | Specht et al. | May 2020 | B2 |
10675000 | Specht et al. | Jun 2020 | B2 |
10695027 | Call et al. | Jun 2020 | B2 |
10835208 | Smith et al. | Nov 2020 | B2 |
10856846 | Davis et al. | Dec 2020 | B2 |
10925577 | Adam et al. | Feb 2021 | B2 |
11016191 | Call et al. | May 2021 | B2 |
11051791 | Smith et al. | Jul 2021 | B2 |
11068689 | Call et al. | Jul 2021 | B2 |
11096662 | Specht | Aug 2021 | B2 |
11172911 | Call et al. | Nov 2021 | B2 |
11253233 | Belevich et al. | Feb 2022 | B2 |
11464492 | Specht et al. | Oct 2022 | B2 |
20020035864 | Paltieli et al. | Mar 2002 | A1 |
20020073781 | Hashimoto et al. | Jun 2002 | A1 |
20020087071 | Schmitz et al. | Jul 2002 | A1 |
20020111568 | Bukshpan | Aug 2002 | A1 |
20020138003 | Bukshpan | Sep 2002 | A1 |
20020161299 | Prater et al. | Oct 2002 | A1 |
20030007598 | Wang et al. | Jan 2003 | A1 |
20030013962 | Bjaerum et al. | Jan 2003 | A1 |
20030028111 | Vaezy et al. | Feb 2003 | A1 |
20030040669 | Grass et al. | Feb 2003 | A1 |
20030163271 | Chell et al. | Sep 2003 | A1 |
20030181806 | Medan et al. | Sep 2003 | A1 |
20030220554 | Grenon et al. | Nov 2003 | A1 |
20030228053 | Li et al. | Dec 2003 | A1 |
20040015079 | Berger et al. | Jan 2004 | A1 |
20040054283 | Corey et al. | Mar 2004 | A1 |
20040068184 | Trahey et al. | Apr 2004 | A1 |
20040100163 | Baumgartner et al. | May 2004 | A1 |
20040111028 | Abe et al. | Jun 2004 | A1 |
20040122313 | Moore et al. | Jun 2004 | A1 |
20040122322 | Moore et al. | Jun 2004 | A1 |
20040127793 | Mendlein et al. | Jul 2004 | A1 |
20040138565 | Trucco | Jul 2004 | A1 |
20040144176 | Yoden | Jul 2004 | A1 |
20040215075 | Zagzebski et al. | Oct 2004 | A1 |
20040236217 | Cerwin et al. | Nov 2004 | A1 |
20040236223 | Barnes et al. | Nov 2004 | A1 |
20040258127 | Ramamurthy et al. | Dec 2004 | A1 |
20040267132 | Podany | Dec 2004 | A1 |
20050004449 | Mitschke et al. | Jan 2005 | A1 |
20050053305 | Li et al. | Mar 2005 | A1 |
20050054910 | Tremblay et al. | Mar 2005 | A1 |
20050061536 | Proulx | Mar 2005 | A1 |
20050090743 | Kawashima et al. | Apr 2005 | A1 |
20050090745 | Steen | Apr 2005 | A1 |
20050111846 | Steinbacher et al. | May 2005 | A1 |
20050113689 | Gritzky | May 2005 | A1 |
20050113694 | Haugen et al. | May 2005 | A1 |
20050124883 | Hunt | Jun 2005 | A1 |
20050131300 | Bakircioglu et al. | Jun 2005 | A1 |
20050147297 | McLaughlin et al. | Jul 2005 | A1 |
20050165312 | Knowles et al. | Jul 2005 | A1 |
20050203404 | Freiburger | Sep 2005 | A1 |
20050215883 | Hundley et al. | Sep 2005 | A1 |
20050240125 | Makin et al. | Oct 2005 | A1 |
20050252295 | Fink et al. | Nov 2005 | A1 |
20050281447 | Moreau-Gobard et al. | Dec 2005 | A1 |
20050288588 | Weber et al. | Dec 2005 | A1 |
20060036170 | Lachaine et al. | Feb 2006 | A1 |
20060058664 | Barthe et al. | Mar 2006 | A1 |
20060062447 | Rinck et al. | Mar 2006 | A1 |
20060074313 | Slayton et al. | Apr 2006 | A1 |
20060074315 | Liang et al. | Apr 2006 | A1 |
20060074320 | Yoo et al. | Apr 2006 | A1 |
20060079759 | Vaillant et al. | Apr 2006 | A1 |
20060079778 | Mo et al. | Apr 2006 | A1 |
20060079782 | Beach et al. | Apr 2006 | A1 |
20060094962 | Clark | May 2006 | A1 |
20060111634 | Wu | May 2006 | A1 |
20060122506 | Davies et al. | Jun 2006 | A1 |
20060173327 | Kim | Aug 2006 | A1 |
20060256231 | Sasaki et al. | Nov 2006 | A1 |
20060262961 | Holsing et al. | Nov 2006 | A1 |
20060270934 | Savord et al. | Nov 2006 | A1 |
20070016022 | Blalock et al. | Jan 2007 | A1 |
20070016044 | Blalock et al. | Jan 2007 | A1 |
20070036414 | Georgescu et al. | Feb 2007 | A1 |
20070043290 | Goepp et al. | Feb 2007 | A1 |
20070055155 | Owen et al. | Mar 2007 | A1 |
20070073781 | Adkins et al. | Mar 2007 | A1 |
20070078345 | Mo et al. | Apr 2007 | A1 |
20070083109 | Ustuner et al. | Apr 2007 | A1 |
20070088213 | Poland | Apr 2007 | A1 |
20070138157 | Dane et al. | Jun 2007 | A1 |
20070161898 | Hao et al. | Jul 2007 | A1 |
20070161904 | Urbano | Jul 2007 | A1 |
20070167752 | Proulx et al. | Jul 2007 | A1 |
20070167824 | Lee et al. | Jul 2007 | A1 |
20070232914 | Chen et al. | Oct 2007 | A1 |
20070238985 | Smith et al. | Oct 2007 | A1 |
20070242567 | Daft et al. | Oct 2007 | A1 |
20080044572 | Loeffler et al. | Feb 2008 | A1 |
20080110261 | Randall et al. | May 2008 | A1 |
20080110263 | Klessel et al. | May 2008 | A1 |
20080112265 | Urbano et al. | May 2008 | A1 |
20080114241 | Randall et al. | May 2008 | A1 |
20080114245 | Randall et al. | May 2008 | A1 |
20080114246 | Randall et al. | May 2008 | A1 |
20080114247 | Urbano et al. | May 2008 | A1 |
20080114248 | Urbano et al. | May 2008 | A1 |
20080114249 | Randall et al. | May 2008 | A1 |
20080114250 | Urbano et al. | May 2008 | A1 |
20080114251 | Weymer et al. | May 2008 | A1 |
20080114252 | Randall et al. | May 2008 | A1 |
20080114253 | Randall et al. | May 2008 | A1 |
20080114255 | Schwartz et al. | May 2008 | A1 |
20080125659 | Wilser et al. | May 2008 | A1 |
20080132790 | Burton | Jun 2008 | A1 |
20080181479 | Yang et al. | Jul 2008 | A1 |
20080183075 | Govari et al. | Jul 2008 | A1 |
20080188747 | Randall et al. | Aug 2008 | A1 |
20080188750 | Randall et al. | Aug 2008 | A1 |
20080194957 | Hoctor et al. | Aug 2008 | A1 |
20080194958 | Lee et al. | Aug 2008 | A1 |
20080194959 | Wang et al. | Aug 2008 | A1 |
20080208061 | Halmann | Aug 2008 | A1 |
20080242996 | Hall et al. | Oct 2008 | A1 |
20080249408 | Palmeri et al. | Oct 2008 | A1 |
20080255452 | Entrekin | Oct 2008 | A1 |
20080262357 | Wodnicki | Oct 2008 | A1 |
20080269604 | Boctor et al. | Oct 2008 | A1 |
20080269613 | Summers et al. | Oct 2008 | A1 |
20080275344 | Glide-Hurst et al. | Nov 2008 | A1 |
20080285819 | Konofagou et al. | Nov 2008 | A1 |
20080287787 | Sauer et al. | Nov 2008 | A1 |
20080294045 | Ellington et al. | Nov 2008 | A1 |
20080294050 | Shinomura et al. | Nov 2008 | A1 |
20080294052 | Wilser et al. | Nov 2008 | A1 |
20080306382 | Guracar et al. | Dec 2008 | A1 |
20080306386 | Baba et al. | Dec 2008 | A1 |
20080319317 | Kamiyama et al. | Dec 2008 | A1 |
20080319318 | Johnson et al. | Dec 2008 | A1 |
20090005679 | Dala-Krishna | Jan 2009 | A1 |
20090010459 | Garbini et al. | Jan 2009 | A1 |
20090012393 | Choi | Jan 2009 | A1 |
20090015665 | Willsie | Jan 2009 | A1 |
20090016163 | Freeman et al. | Jan 2009 | A1 |
20090018445 | Schers et al. | Jan 2009 | A1 |
20090024039 | Wang et al. | Jan 2009 | A1 |
20090036780 | Abraham | Feb 2009 | A1 |
20090043206 | Towfiq et al. | Feb 2009 | A1 |
20090048519 | Hossack et al. | Feb 2009 | A1 |
20090069681 | Lundberg et al. | Mar 2009 | A1 |
20090069686 | Daft et al. | Mar 2009 | A1 |
20090069692 | Cooley et al. | Mar 2009 | A1 |
20090079299 | Bradley et al. | Mar 2009 | A1 |
20090099483 | Rybyanets | Apr 2009 | A1 |
20090112095 | Daigle | Apr 2009 | A1 |
20090131797 | Jeong et al. | May 2009 | A1 |
20090143680 | Yao et al. | Jun 2009 | A1 |
20090148012 | Altmann et al. | Jun 2009 | A1 |
20090150094 | Van Velsor et al. | Jun 2009 | A1 |
20090182233 | Wodnicki | Jul 2009 | A1 |
20090182237 | Angelsen et al. | Jul 2009 | A1 |
20090198134 | Hashimoto et al. | Aug 2009 | A1 |
20090203997 | Ustuner | Aug 2009 | A1 |
20090208080 | Grau et al. | Aug 2009 | A1 |
20090259128 | Stribling | Oct 2009 | A1 |
20090264760 | Lazebnik et al. | Oct 2009 | A1 |
20090306510 | Hashiba et al. | Dec 2009 | A1 |
20090326379 | Daigle et al. | Dec 2009 | A1 |
20100010354 | Skert et al. | Jan 2010 | A1 |
20100016725 | Thiele | Jan 2010 | A1 |
20100036258 | Dietz et al. | Feb 2010 | A1 |
20100063397 | Wagner | Mar 2010 | A1 |
20100063399 | Walker et al. | Mar 2010 | A1 |
20100069751 | Hazard et al. | Mar 2010 | A1 |
20100069756 | Ogasawara et al. | Mar 2010 | A1 |
20100085383 | Cohen et al. | Apr 2010 | A1 |
20100106431 | Baba et al. | Apr 2010 | A1 |
20100109481 | Buccafusca | May 2010 | A1 |
20100121193 | Fukukita et al. | May 2010 | A1 |
20100121196 | Hwang et al. | May 2010 | A1 |
20100130855 | Lundberg et al. | May 2010 | A1 |
20100145195 | Hyun | Jun 2010 | A1 |
20100168566 | Bercoff et al. | Jul 2010 | A1 |
20100168578 | Garson, Jr. et al. | Jul 2010 | A1 |
20100174194 | Chiang et al. | Jul 2010 | A1 |
20100174198 | Young et al. | Jul 2010 | A1 |
20100191110 | Insana et al. | Jul 2010 | A1 |
20100217124 | Cooley | Aug 2010 | A1 |
20100217125 | Kadokura et al. | Aug 2010 | A1 |
20100228126 | Emery et al. | Sep 2010 | A1 |
20100240994 | Zheng | Sep 2010 | A1 |
20100249570 | Carson et al. | Sep 2010 | A1 |
20100249596 | Magee | Sep 2010 | A1 |
20100256488 | Kim et al. | Oct 2010 | A1 |
20100262013 | Smith et al. | Oct 2010 | A1 |
20100266176 | Masumoto et al. | Oct 2010 | A1 |
20100286525 | Osumi | Nov 2010 | A1 |
20100286527 | Cannon et al. | Nov 2010 | A1 |
20100298712 | Pelissier et al. | Nov 2010 | A1 |
20100310143 | Rao et al. | Dec 2010 | A1 |
20100317971 | Fan et al. | Dec 2010 | A1 |
20100324418 | El-Aklouk et al. | Dec 2010 | A1 |
20100324423 | El-Aklouk et al. | Dec 2010 | A1 |
20100329521 | Beymer et al. | Dec 2010 | A1 |
20110005322 | Ustuner | Jan 2011 | A1 |
20110016977 | Guracar | Jan 2011 | A1 |
20110021920 | Shafir et al. | Jan 2011 | A1 |
20110021923 | Daft et al. | Jan 2011 | A1 |
20110033098 | Richter et al. | Feb 2011 | A1 |
20110044133 | Tokita | Feb 2011 | A1 |
20110066030 | Yao | Mar 2011 | A1 |
20110098565 | Masuzawa | Apr 2011 | A1 |
20110112400 | Emery et al. | May 2011 | A1 |
20110112404 | Gourevitch | May 2011 | A1 |
20110125017 | Ramamurthy et al. | May 2011 | A1 |
20110178441 | Tyler | Jul 2011 | A1 |
20110213244 | Frinking et al. | Sep 2011 | A1 |
20110270088 | Shiina | Nov 2011 | A1 |
20110301470 | Sato et al. | Dec 2011 | A1 |
20110306886 | Daft et al. | Dec 2011 | A1 |
20110319764 | Okada et al. | Dec 2011 | A1 |
20120004545 | Ziv-Ari et al. | Jan 2012 | A1 |
20120035482 | Kim et al. | Feb 2012 | A1 |
20120036934 | Kröning et al. | Feb 2012 | A1 |
20120085173 | Papadopoulos et al. | Apr 2012 | A1 |
20120101378 | Lee | Apr 2012 | A1 |
20120114210 | Kim et al. | May 2012 | A1 |
20120121150 | Murashita | May 2012 | A1 |
20120137778 | Kitazawa et al. | Jun 2012 | A1 |
20120140595 | Amemiya | Jun 2012 | A1 |
20120141002 | Urbano et al. | Jun 2012 | A1 |
20120165670 | Shi et al. | Jun 2012 | A1 |
20120179044 | Chiang et al. | Jul 2012 | A1 |
20120209150 | Zeng et al. | Aug 2012 | A1 |
20120226201 | Clark et al. | Sep 2012 | A1 |
20120243763 | Wen et al. | Sep 2012 | A1 |
20130030296 | Miyaki | Jan 2013 | A1 |
20130070062 | Fouras et al. | Mar 2013 | A1 |
20130131516 | Katsuyama | May 2013 | A1 |
20130144165 | Ebbini et al. | Jun 2013 | A1 |
20130258805 | Hansen et al. | Oct 2013 | A1 |
20130261463 | Chiang et al. | Oct 2013 | A1 |
20140073921 | Specht et al. | Mar 2014 | A1 |
20140147013 | Shandas et al. | May 2014 | A1 |
20150045668 | Smith et al. | Feb 2015 | A1 |
20180153511 | Specht et al. | Jun 2018 | A1 |
20190200961 | Specht et al. | Jul 2019 | A1 |
20200275910 | Specht et al. | Sep 2020 | A1 |
20200297320 | Specht et al. | Sep 2020 | A1 |
20200323513 | Call et al. | Oct 2020 | A1 |
20210068787 | Smith et al. | Mar 2021 | A1 |
20210085292 | Davis et al. | Mar 2021 | A1 |
20210278531 | Call et al. | Sep 2021 | A1 |
20210350101 | Call et al. | Nov 2021 | A1 |
20210378633 | Specht et al. | Dec 2021 | A1 |
20220071601 | Call et al. | Mar 2022 | A1 |
20220167949 | Belevich et al. | Jun 2022 | A1 |
20230277158 | Brewer et al. | Sep 2023 | A1 |
20230380805 | Specht et al. | Nov 2023 | A1 |
Number | Date | Country |
---|---|---|
1535243 | Oct 2004 | CN |
1636150 | Jul 2005 | CN |
1781460 | Jun 2006 | CN |
101103927 | Jan 2008 | CN |
101116622 | Feb 2008 | CN |
101190134 | Jun 2008 | CN |
101453955 | Jun 2009 | CN |
100545650 | Sep 2009 | CN |
101609150 | Dec 2009 | CN |
101852773 | Jun 2010 | CN |
101785684 | Jul 2010 | CN |
101843501 | Sep 2010 | CN |
101912278 | Dec 2010 | CN |
102018533 | Apr 2011 | CN |
102112047 | Jun 2011 | CN |
102123668 | Jul 2011 | CN |
1346689 | Sep 2003 | EP |
1949856 | Jul 2008 | EP |
2058796 | May 2009 | EP |
2101191 | Sep 2009 | EP |
2182352 | May 2010 | EP |
2187813 | May 2010 | EP |
2198785 | Jun 2010 | EP |
1757955 | Nov 2010 | EP |
2319417 | May 2011 | EP |
2325672 | May 2011 | EP |
1462819 | Jul 2011 | EP |
2356941 | Aug 2011 | EP |
1979739 | Oct 2011 | EP |
2385391 | Nov 2011 | EP |
2294400 | Feb 2012 | EP |
2453256 | May 2012 | EP |
1840594 | Jun 2012 | EP |
1850743 | Dec 2012 | EP |
1594404 | Sep 2013 | EP |
2026280 | Oct 2013 | EP |
2851662 | Aug 2004 | FR |
49-11189 | Jan 1974 | JP |
54-44375 | Apr 1979 | JP |
55-103839 | Aug 1980 | JP |
57-31848 | Feb 1982 | JP |
58-223059 | Dec 1983 | JP |
59-101143 | Jun 1984 | JP |
59-174151 | Oct 1984 | JP |
60-13109 | Jan 1985 | JP |
60-68836 | Apr 1985 | JP |
01164354 | Jun 1989 | JP |
02501431 | May 1990 | JP |
03015455 | Jan 1991 | JP |
03126443 | May 1991 | JP |
04017842 | Jan 1992 | JP |
04067856 | Mar 1992 | JP |
05042138 | Feb 1993 | JP |
H05146437 | Jun 1993 | JP |
06125908 | May 1994 | JP |
06254092 | Sep 1994 | JP |
07051266 | Feb 1995 | JP |
07204201 | Aug 1995 | JP |
H07204202 | Aug 1995 | JP |
08154930 | Jun 1996 | JP |
08252253 | Oct 1996 | JP |
H0315455 | Jan 1997 | JP |
09103429 | Apr 1997 | JP |
09201361 | Aug 1997 | JP |
2777197 | May 1998 | JP |
10216128 | Aug 1998 | JP |
11089833 | Apr 1999 | JP |
11239578 | Sep 1999 | JP |
2001507794 | Jun 2001 | JP |
2001245884 | Sep 2001 | JP |
2002209894 | Jul 2002 | JP |
2002253548 | Sep 2002 | JP |
2002253549 | Sep 2002 | JP |
2003235839 | Aug 2003 | JP |
2003290224 | Oct 2003 | JP |
2004167092 | Jun 2004 | JP |
2004215987 | Aug 2004 | JP |
2004337457 | Dec 2004 | JP |
2004340809 | Dec 2004 | JP |
2004351214 | Dec 2004 | JP |
2005046192 | Feb 2005 | JP |
2005046193 | Feb 2005 | JP |
2005152187 | Jun 2005 | JP |
2005523792 | Aug 2005 | JP |
2005526539 | Sep 2005 | JP |
2006051356 | Feb 2006 | JP |
2006061203 | Mar 2006 | JP |
2006122657 | May 2006 | JP |
2006130313 | May 2006 | JP |
2006204923 | Aug 2006 | JP |
2007325937 | Dec 2007 | JP |
2008122209 | May 2008 | JP |
2008513763 | May 2008 | JP |
2008515557 | May 2008 | JP |
2008132342 | Jun 2008 | JP |
2008522642 | Jul 2008 | JP |
2008259541 | Oct 2008 | JP |
2008279274 | Nov 2008 | JP |
2008307087 | Dec 2008 | JP |
2009178448 | Aug 2009 | JP |
2009240667 | Oct 2009 | JP |
2010005375 | Jan 2010 | JP |
2010124842 | Jun 2010 | JP |
2010526626 | Aug 2010 | JP |
2010227503 | Oct 2010 | JP |
2011529362 | Dec 2011 | JP |
100715132 | Apr 2007 | KR |
1020080044737 | May 2008 | KR |
1020090009258 | Jan 2009 | KR |
1020090103408 | Oct 2009 | KR |
1020100051108 | May 2010 | KR |
WO9218054 | Oct 1992 | WO |
WO9800719 | Jan 1998 | WO |
WO0164109 | Sep 2001 | WO |
WO02084594 | Oct 2002 | WO |
WO2005009245 | Feb 2005 | WO |
WO2006113445 | Oct 2006 | WO |
WO2006114735 | Nov 2006 | WO |
WO2007127147 | Nov 2007 | WO |
WO2008097479 | Aug 2008 | WO |
WO2008137030 | Nov 2008 | WO |
WO2009060182 | May 2009 | WO |
WO2010095094 | Aug 2010 | WO |
WO2010137453 | Dec 2010 | WO |
WO2010139519 | Dec 2010 | WO |
WO2011004661 | Jan 2011 | WO |
WO2011057252 | May 2011 | WO |
WO2011064688 | Jun 2011 | WO |
WO2011094585 | Aug 2011 | WO |
WO2011100697 | Aug 2011 | WO |
WO2011123529 | Oct 2011 | WO |
WO2012028896 | Mar 2012 | WO |
WO2012049124 | Apr 2012 | WO |
WO2012049612 | Apr 2012 | WO |
WO2012078639 | Jun 2012 | WO |
WO2012091280 | Jul 2012 | WO |
WO2012112540 | Aug 2012 | WO |
Entry |
---|
Abeysekera et al.; Alignment and calibration of dual ultrasound transducers using a wedge phantom; Ultrasound in Medicine and Biology; 37(2); pp. 271-279; Feb. 2011. |
Arigovindan et al.; Full motion and flow field recovery from echo doppler data; IEEE Transactions on Medical Imaging; 26(1); pp. 31-45; Jan. 2007. |
Cai et al.; Off-axis directional acoustic wave beaming control by an asymmetric rubber heterostructures film deposited on steel plate in water; IEEE Intl.; 2009 Ultrasonics Symposium (IUS); pp. 1552-1554; Rome; Sep. 2009. |
Capineri et al.; A doppler system for dynamic vector velocity maps; Ultrasound in Medicine & Biology; 28(2); pp. 237-248; Feb. 28, 2002. |
Carson et al.; Measurement of photoacoustic transducer position by robotic source placement and nonlinear parameter estimation; Biomedical Optics (BIOS); International Society for Optics and Photonics (9th Conf. on Biomedical Thermoacoustics, Optoacoustics, and Acousto-optics; vol. 6656; 9 pages; Feb. 28, 2008. |
Chen et al.; Maximum-likelihood source localization and unknown sensor location estimation for wideband signals in the near-field; IEEE Transactions On Signal Processing; 50(8); pp. 1843-1854; Aug. 2002. |
Chen et al.; Source localization and tracking of a wideband source using a randomly distributed beamforming sensor array; International Journal of High Performance Computing Applications: 16(3); pp. 259-272; Fall 2002. |
Cristianini et al.; An Introduction to Support Vector Machines; Cambridge University Press; pp. 93-111; Mar. 2000. |
Dunmire et al.; A brief history of vector doppler; Medical Imaging 2001; International Society for Optics and Photonics; pp. 200-214; May 30, 2001. |
Dunmire et al.; Cross-beam vector Doppler ultrasound for angle-independent velocity measurements; Ultrasound in medicine & biology; 26(8); pp. 1213-1235; Oct. 2000. |
Du et al.; User parameter free approaches to multistatic adaptive ultrasound imaging; 5th IEEE International Symposium; pp. 1287-1290, May 2008. |
Feigenbaum, Harvey, M.D.; Echocardiography; Lippincott Williams & Wilkins; Philadelphia; 5th Ed.; pp. 482, 484; Feb. 1994. |
Fernandez et al.; High resolution ultrasound beamforming using synthetic and adaptive imaging techniques; Proceedings IEEE International Symposium on Biomedical Imaging; Washington, D.C.; pp. 433-436; Jul. 7-10, 2002. |
Gazor et al.; Wideband multi-source beamforming with array location calibration and direction finding; Conference on Acoustics, Speech and Signal Processing ICASSP-95; Detroit, MI; vol. 3 IEEE; pp. 1904-1907; May 9-12, 1995. |
Gran et al.; Directional velocity estimation using a spatio-temporal encoding technique based on frequency division for synthetic transmit aperture ultrasound; IEEE transactions on ultrasonics, ferroelectrics, and frequency control; 53 (7); pp. 1289-1299, Aug. 2006. |
Haun et al.; Efficient three-dimensional imaging from a small cylindrical aperture; IEEE Trans. on Ultrasonics, Ferroelectrics, and Frequency Control; 49(7); pp. 861-870; Jul. 2002. |
Haykin, Simon; Neural Networks: A Comprehensive Foundation (2nd Ed.); Prentice Hall; pp. 156-187; Jul. 16, 1998. |
Heikkila et al.; A four-step camera calibration procedure with implicit image correction; Proceedings IEEE Computer Scociety Conference on Computer Vision and Pattern Recognition; San Juan; pp. 1106-1112; Jun. 17-19, 1997. |
Hendee et al.; Medical Imaging Physics; Wiley-Liss, Inc. 4th Edition; Chap. 19-22; pp. 303-353; (year of pub. sufficiently earlier than effective US filing date and any foreign priority date) © 2002. |
Hsu et al.; Real-time freehand 3D ultrasound calibration; CUED/F-INFENG/TR 565; Department of Engineering, University of Cambridge, United Kingdom; 14 pages; Sep. 2006. |
Jeffs; Beamforming: a brief introduction; Brigham Young University; 14 pages; retrieved from the internet (http://ens.ewi.tudelft.nl/Education/courses/et4235/Beamforming.pdf); Oct. 2004. |
Khamene et al.; A novel phantom-less spatial and temporal ultrasound calibration method; Medical Image Computing and Computer-Assisted Intervention—MICCAI (Proceedings 8th Int. Conf.); Springer Berlin Heidelberg; Palm Springs, CA; pp. 65-72; Oct. 26-29, 2005. |
Kramb et al,.; Considerations for using phased array ultrasonics in a fully automated inspection system. Review of Quantitative Nondestructive Evaluation, 2004 Edition, ed. D. O. Thompson and D. E. Chimenti, American Inst. of Physics, pp. 817-825, Mar. 2004. |
Ledesma-Carbayo et al.; Spatio-temporal nonrigid registration for ultrasound cardiac motion estimation; IEEE Trans. On Medical Imaging; vol. 24; No. 9; Sep. 2005. |
Leotta et al.; Quantitative three-dimensional echocardiography by rapid imaging . . . ; J American Society of Echocardiography; vol. 10; No. 8; pp. 830-839; Oct. 1997. |
Li et al.; An efficient speckle tracking algorithm for ultrasonic imaging; 24; pp. 215-228; Oct. 1, 2002. |
Montaldo et al.; Building three-diminsional images using a time-reversal chaotic cavity; IEEE Trans. on Ultrasonics, Ferroelectrics, and Frequency Control; 52(9); pp. 1489-1497; Sep. 2005. |
Morrison et al.; A probabilistic neural network based image segmentation network for magnetic resonance images; Proc. Conf. Neural Networks; Baltimore, MD; vol. 3; pp. 60-65; Jun. 1992. |
Nadkarni et al.; Cardiac motion synchronization for 3D cardiac ultrasound imaging; Ph.D. Dissertation, University of Western Ontario; Jun. 2002. |
Press et al.; Cubic spline interpolation; §3.3 in “Numerical Recipes in FORTRAN: The Art of Scientific Computing”, 2nd Ed.; Cambridge, England; Cambridge University Press; pp. 107-110; Sep. 1992. |
Saad et al.; Computer vision approach for ultrasound doppler angle estimation; Journal of Digital Imaging; 22(6); pp. 681-688; Dec. 1, 2009. |
Sakas et al.; Preprocessing and volume rendering of 3D ultrasonic data; IEEE Computer Graphics and Applications; pp. 47-54, Jul. 1995. |
Sapia et al.; Deconvolution of ultrasonic waveforms using an adaptive wiener filter; Review of Progress in Quantitative Nondestructive Evaluation; vol. 13A; Plenum Press; pp. 855-862; Jan. 1994. |
Sapia et al.; Ultrasound image deconvolution using adaptive inverse filtering; 12 IEEE Symposium on Computer-Based Medical Systems, CBMS, pp. 248-253; Jun. 1999. |
Sapia, Mark Angelo; Multi-dimensional deconvolution of optical microscope and ultrasound imaging using adaptive least-mean-square (LMS) inverse filtering; Ph.D. Dissertation; University of Connecticut; Jan. 2000. |
Scabia et al.; A real-time two-dimensional pulsed-wave Doppler system; Ultrasound in medicine & biology: 26(1); pp. 121-131; Jan. 1, 2000. |
Slavine et al.; Construction, calibration and evaluation of a tissue phantom with reproducible optical properties for investigations in light emission tomography; Engineering in Medicine and Biology Workshop; Dallas, TX; IEEE pp. 122-125; Nov. 11-12, 2007. |
Smith et al.; High-speed ultrasound volumetric imaging system. 1. Transducer design and beam steering; IEEE Trans. Ultrason., Ferroelect., Freq. Contr.; vol. 38; pp. 100-108; Mar. 1991. |
Specht et al.; Deconvolution techniques for digital longitudinal tomography; SPIE; vol. 454; presented at Application of Optical Instrumentation in Medicine XII; pp. 319-325; Jun. 1984. |
Specht et al.; Experience with adaptive PNN and adaptive GRNN; Proc. IEEE International Joint Conf. on Neural Networks; vol. 2; pp. 1203-1208; Orlando, FL; Jun. 1994. |
Specht, D.F.; A general regression neural network; IEEE Trans. On Neural Networks; vol. 2.; No. 6; Nov. 1991. |
Specht, D.F.; Blind deconvolution of motion blur using LMS inverse filtering; Lockheed Independent Research (unpublished); Jun. 23, 1975. |
Specht, D.F.; Enhancements to probabilistic neural networks; Proc. IEEE International Joint Conf. on Neural Networks; Baltimore, MD; Jun. 1992. |
Specht, D.F.; GRNN with double clustering; Proc. IEEE International Joint Conf. Neural Networks; Vancouver, Canada; Jul. 16-21, 2006. |
Specht, D.F.; Probabilistic neural networks; Pergamon Press; Neural Networks; vol. 3; pp. 109-118; Feb. 1990. |
Stern; The basic concepts of diagnostic ultrasound. Yale-New Haven Teachers Institute; Apr. 2005. |
UCLA Academic Technology; SPSS learning module: How can I analyze a subset of my data; 6 pages; retrieved from the internet (http://www.ats.ucla.edu/stat/spss/modules/subset_analyze.htm) Nov. 26, 2001. |
Urban et al.; Implementation of vibro-acoustography on a clinical ultrasound system; IEEE Ultrasonics Symposium (IUS); pp. 326-329; Oct. 14, 2010. |
Von Ramm et al.; High-speed ultrasound volumetric imaging—System. 2. Parallel processing and image display; IEEE Trans. Ultrason., Ferroelect., Freq. Contr.; vol. 38; pp. 109-115; Mar. 1991. |
Wang et al.; Photoacoustic tomography of biological tissues with high cross-section resolution: reconstruction and experiment; Medical Physics; 29(12); pp. 2799-2805; Dec. 2002. |
Wells, P.N.T.; Biomedical ultrasonics; Academic Press; London, New York, San Francisco; pp. 124-125; Mar. 1977. |
Widrow et al.; Adaptive signal processing; Prentice-Hall; Englewood Cliffs, NJ; pp. 99-116; Mar. 1985. |
Wikipedia; Point cloud; 2 pages; retrieved Nov. 24, 2014 from the internet (https://en.wikipedia.org/w/index.php?title=Point_cloud&oldid=472583138). |
Wikipedia; Curve fitting; 5 pages; retrieved from the internet (http:en.wikipedia.org/wiki/Curve_fitting) Dec. 19, 2010. |
Wikipedia; Speed of sound; 17 pages; retrieved from the internet (http:en.wikipedia.org/wiki/Speed_of_sound) Feb. 15, 2011. |
Yang et al.; Time-of-arrival calibration for improving the microwave breast cancer imaging; 2011 IEEE Topical Conf. on Biomedical Wireless Technologies, Networks, and sensing Systems (BioWireleSS); Phoenix, AZ; pp. 67-70; Jan. 16-19, 2011. |
Zhang et al.; A high-frequency high frame rate duplex ultrasound linear array imaging system for small animal imaging; IEEE transactions on ultrasound, ferroelectrics, and frequency control; 57(7); pp. 1548-1567; Jul. 2010. |
Bajikar et al.; U.S. Appl. No. 18/165,276 entitled “Multiple aperture ultrasound imaging systems and methods,” filed Feb. 6, 2023. |
Jensen et al.; Synthetic aperture ultrasound imaging; Ultrasonics; vol. 44, pp. e5-e15; Dec. 22, 2006. |
Call et al.; U.S. Appl. No. 18/330,699 entitled “Network-based ultrasound imaging system,” filed Jun. 7, 2023. |
Atmeh et al.; U.S. Appl. No. 18/251,421 entitled “Systems and methods for improving ultrasound image quality,” filed May 2, 2023. |
Smith et al.; U.S. Appl. No. 18/344,278 entitled “Concave ultrasound transducers and 3d arrays,” filed Jun. 29, 2023. |
Call et al.; U.S. Appl. No. 18/344,479 entitled “Ultrasound imaging systems and methods for detecting object motion,” filed Jun. 29, 2023. |
Lockwood et al.; Real-time 3-D ultrasound imaging using sparse synthetic aperture beamforming; IEEE transactions on ultrasonics, ferroelectrics, and frequency control; 45(4); pp. 980-988, Jul. 1998. |
Number | Date | Country | |
---|---|---|---|
20230270416 A1 | Aug 2023 | US |
Number | Date | Country | |
---|---|---|---|
61305784 | Feb 2010 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16506570 | Jul 2019 | US |
Child | 17936420 | US | |
Parent | 16000507 | Jun 2018 | US |
Child | 16506570 | US | |
Parent | 15364075 | Nov 2016 | US |
Child | 16000507 | US | |
Parent | 14846374 | Sep 2015 | US |
Child | 15364075 | US | |
Parent | 13029907 | Feb 2011 | US |
Child | 14846374 | US |