There are various modalities for performing medical imaging of patients. For example, magnetic resonance imaging (MRI) is a medical imaging technique used in radiology to visualize internal structures of the body in detail. An MRI scanner is a device in which the patient or a portion of the patient's body is positioned within a powerful magnet where a magnetic field is used to align the magnetization of some atomic nuclei (usually hydrogen nuclei-protons) and radio frequency magnetic fields are applied to systematically alter the alignment of this magnetization. This causes the nuclei to produce a rotating magnetic field detectable by the scanner and this information is recorded to construct an image of the scanned region of the body. These scans typically take several minutes (up to about 40 minutes in some scanners) and in prior art devices any significant movement can ruin the images and require the scan to be repeated.
U.S. Pat. No. 8,121,361 issued Feb. 21, 2012, entitled Motion Tracking System for Real Time Adaptive Imaging and Spectroscopy describes a system that adaptively compensates for subject motion. and the disclosure therein is hereby incorporated herein by reference.
An accurate and reliable method of determining the dynamic position and orientation of a patient's head or other body portion during MRI scanning is a requirement in any attempt to compensate for subject motion during the procedure. Toward this end, disclosed herein are systems and methods that include practical optical head tracking capability using at least a first sensor, e.g., a first camera, and a second sensor, e.g., a second camera, such as a pair of cameras, for example ordinary CCD cameras, ordinary white light or LED illumination, and a marker target, such as a compact, inexpensive target which mounts comfortably and rigidly to the subject's skeletal frame The camera can be configured to detect any desired wavelength or range of wavelengths of energy, including one or more of the infrared, near-infrared, visible, or ultraviolet spectra for example. Some preferred embodiments can track head and other body motion having up to and including six degrees of freedom (sometimes referred to as 6-DOF).
In an embodiment, the system is configured for tracking a moving target having up to six degrees of freedom and rapidly determining positions of the target, said system includes an easy to locate precision optical target fixed to the target. The system can also include at least two cameras positioned so as to view the optical camera from different directions with each of the at least two cameras being adapted to record two dimensional images of the precision optical target defining a precise target point. In an embodiment, a computer processor is programmed to determine the target movement in Cartesian coordinates of x, y and z and pitch, roll and yaw utilizing an algorithm adapted to identify a set of precise target points on the precision optical target and the x, y and z displacement and the pitch, roll and yaw rotation of the precise target points based on optical images collected by the at least two cameras. The system can utilize an iteration procedure whereby an approximate first-order solution is proposed and tested against the identified precise target point projections on the cameras to determine residual errors which are then divided by the local derivatives with respect to each component of rotation and translation, to determine an iterative correction. The system can be configured to repeat the above actions until residual error becomes smaller than desired accuracy. Using this process the system can be configured to determine the position of the target at rates of at least 100 times per second with translations accuracies of about or no more than about 0.1 mm and angle accuracies of about or no more than about 0.1 degrees. With repetition rates in the range of 100 times per second, the full 6-DOF movement determination can be performed for each repetition. In these embodiments the results of each movement determination is used for the initial first order solution during the next iteration.
The six degrees of freedom movements are over orthogonal directions x, y, and z and roll, pitch and yaw angles. Direction x is along the spinal axis. Direction y perpendicular to x is along the shoulder to shoulder direction and direction z is perpendicular to both x and y and in the floor-to-ceiling direction assuming the patient is lying on his back parallel to the floor. The roll angle is about the x-axis; the angle made by a shaking head “No”. The pitch angle is about the y-axis; the angle made by shaking head “Yes” and the Yaw angle is about the z-axis, the angle made by leaning head toward a shoulder.
In an embodiment, the desired accuracy is about 0.1 mm for each of the directions and about 0.1 degrees for each of the angles. Movements are measured relative to a pivot point in the patient's neck. In an embodiment the pivot point is located at the base of the patient's neck where the head swivels for nod turn and lean motions. The offset of the precision optical target from this pivot point position is Δy=0, Δx−4.5″, Δz=5.5″. The precision of these offsets is not critical since all motions of interest are relative motions. The six measurements are x, y, and z distances and roll, pitch and yaw angles. In some embodiments, the measurements are up-dated at a rate of about 100 solutions per second with a latency of about 10 milliseconds. The system can be configured to report to MRI systems the exact position or the approximate position of the head with accuracies of about or better than about 0.1 mm in distances and about 0.1 degree in angles.
One possible coordinate system for reporting 6-DOF motions to the MRI field compensation system is a Cartesian system aligned with the symmetry axis of the head coil. The head coil coordinate system is coincident with body coordinates in the nominal (“square”) head position. Target displacements and rotations can be reported to the coil field compensation system using this system of coordinates.
Optical Target Fixed to an Anatomical Location, e.g., the Head
To overcome the challenge of the elastic nature of skin, in an embodiment, an optical tracking target can be coupled to the upper teeth of the patient. One accessible feature which is rigid to the skull is the upper teeth. Unlike the teeth on the lower jawbone, the upper teeth are rigidly affixed to the skull of the patient. In an embodiment, a compact and reliable optical tracking target can be attached to one or more of the upper teeth with a clip-on or other coupling device. Such attachment devices can be configured to be extremely comfortable. In an embodiment, a printed precision optical target is attached to the top front teeth of a patient.
An optical target can be configured to be easy to locate with a high degree of accuracy regardless of orientation in a sensor field of view. A circle or series of concentric circles or ellipses can be potentially advantageous in this regard. Furthermore, to accommodate the fastest composite 2D data processing methods, a number (at least 3) of centroid positions can be discernible at every instant in time. The target can be, in some embodiments, composed of three sets of concentric circles or ellipses located at the vertices of an equilateral triangle. Compactness is desired for practical reasons, but the minimum size and spacing of the targets is dictated to large extent by characteristics of the sensors and the available non-occluded optical lines of sight through the MRI field compensation coil. A tradeoff arises, for instance, between the minimum size of the target and the cost of the imaging cameras used to sense the head motion—the smaller the edge dimension of the target triangle, the more pixels required in the camera sensor, and the faster the readout and processing electronics required.
As a reasonable compromise, in some embodiments an equilateral triangle side length of 0.5 inches can be adopted. The printed target pattern includes a solid central elliptical dot of 1/16″ minor diameter at each triangle vertex, and each dot is surrounded by a first concentric ellipse of 3/16″ minor diameter and 1/32″ line width, and a second concentric ellipse of 5/16″ minor diameter and 1/32″ line width (ellipses scaled to look circular from camera nominal 45° look angle). In this embodiment, the entire target measures about 1 inch wide by about 0.93 inches high. Other dimensions are possible.
A camera viewing this target is able to determine the centroid of each ellipse on the target pattern using a simple brightness moment calculation, independent of orientation of the target. The target itself subtends only a small portion of the camera field of view, but is recognizable by its high contrast and lack of gray scale. In embodiments the computer processor is programmed to track each of the three sub-targets by enclosing each of the three sub-targets within a sub-pixel array of 48×48 pixels and to calculate centroids of each sub-target by dividing (a) the sum of the product of pixel darkness and pixel position by (b) the sum of the pixel darkness of all of the pixels in the 48×48 sub-pixel array. The processor is also programmed to move each of the 48×48 pixel arrays so that its target is always located fully within the sub-pixel array. With sufficient camera spatial and brightness resolution and target illumination and contrast, centroid positional accuracy of about 0.1 pixels in row and/or column or less is achievable using this target.
Latency
Latency in the measurement of head motion using optical tracking techniques is comprised of the camera sensor integration and readout time, the target centroid determination time and the 6-DOF decomposition time. In order to reliably track head motions as fast as 2 cm/second and head rotations as fast as 10 degrees per second, a camera frame rate of about 100 Hz is desired, with electronic shuttering to freeze motion at rates up to 10 times this speed for sharp resolution of the optical target without blurring. A significant field of view is required to accommodate large motions, so fast camera readout without expensive mechanical tracking capabilities will require either a low pixel density or a camera with a larger focal plane but the ability to window a smaller region of interest for readout. Centroid and 6-DOF decomposition algorithms running in composite 2D, rather than full 3D space, and utilizing rapidly converging solution methods can be capable of returning solutions to the compensating head coil electronics at 100 solutions per second, with about 10 ms of latency. In some embodiments, the system can be configured to operate with a latency that enables it to update the scanner in between each image acquisition
Cameras
For a subject wearing or coupled with the optical head tracking target, the target size and subject rotation angles and translation position determine the physical location of the three target centroids precisely in three dimensions. With precise knowledge of these angles and the optical sensor (camera and lens) parameters—pixel pitch, lens focal length and radial distortion, camera location and orientation relative to nominal target position—the location of the target centroid projections on the focal plane sensor can be predicted to any level of accuracy even prior to measurement.
In principle, the inverse problem should be equally simple as long as the 3D position of the target centroids can be ascertained optically. Using two cameras, a stereo view of the centroid projections can be used to determine the 3D location in space of each target centroid, and the 6-DOF displacement vector can then be determined through a simple matrix inversion. In practice, however, this approach leads to expensive and complicated requirements on camera pixel density, pixel count, camera alignment and camera calibration.
An alternate unfolding approach dispenses with stereo ranging but uses separate 2D projections from two cameras without attempting to correlate absolute target positions on the two cameras. This approach eliminates the strict requirements on camera alignment and magnification matching characteristic of the stereo vision approach, and also relaxes the pixel density and count requirements needed to obtain the required positional accuracy (about 0.1 mm in translation and about 0.1 degrees in rotation) by about a factor of 20, resulting in significant savings in cost and processing speed.
Even for this 2D measurement approach some basic steps can be taken to calibrate camera parameters once the cameras are integrated with the head coil; these can be performed at the manufacturing facility. These include measuring the projected pixel location of a single reference point on both cameras, as well as the camera magnification factors for pixel displacement per degree of rotation in pitch, yaw and roll, and per mm of translation along x, y and z. However, as stated before, it is not necessary that the cameras be exactly aligned in space (e.g. perfectly normal) or that their magnifications (lens focal length and distance to reference point) be identical, as is easily verified by simulation.
Stereo Versus Composite 2D Vision Requirements
With a single camera viewing the target from 45 degrees off of vertical in the target plane, the camera sees very little centroid displacement when the target moves in the direction of the camera (e.g. upward vertical translation equal to horizontal translation in the camera direction, with no rotation). Assuming a 7 micron pixel pitch, a 25 mm lens, and a working distance of 14 inches, target displacement in the camera direction may be at least 0.6 mm before the target can be detected as a 0.1-pixel increase in target centroid separation. However, as shown in
Camera Depth of Field
To accommodate head roll of +/−15 degrees plus the 0.85-inch target width at a working distance of 14 inches, the lens can be configured to provide sharp focus for distances between 13″ and 15.5″. At f/22, assuming a circle of confusion slightly smaller than a camera pixel (7 microns), a 25 mm focal-length lens provides this necessary depth of field a nominal 14-inch focus. At this working distance, the optical path can be folded with a turning mirror (
In some embodiments, one possible camera that can be utilized or modified for use with systems and methods as disclosed herein, is produced by Allied Vision Technologies and designated the Prosilica GE-680 Monochrome CCD Camera. This camera features a Kodak KAI-0340 ⅓″ 640×480 VGA focal plane sensor with 7.4 μm square pixels and a fast Gigabit Ethernet output delivering up to 205 frames per second at 12-bit pixel depth. An inexpensive possible lens for use is an Edmund Optics TechSpec 25 mm high-resolution fixed focal length lens.
For this camera and lens, at 14 inches from the target at 45° incidence, the 5/16″ diameter target circles project to ellipses on the camera, with the minor diameter of the largest ellipses at about 28 pixels and the major diameter at about 40 pixels. With sufficient S/N ratio (target illumination) and lens MTF (sharpness), this pattern should allow accurate centroiding to about 0.1 pixels in row and/or column or less. The entire projected target subtends about 128 H×168 V pixels, and allowing for head roll of +/−11.5 degrees, a camera with 640 horizontal pixels (pixel columns) can accommodate the entire field of interest without mechanical tracking provisions.
Six Degree-of-Freedom Measurement and Reporting Algorithm
In some embodiments, the MRI Head Tracker takes real-time input from two 2D imaging sensors and analyzes these data to determine and report motions in six degrees of freedom with minimal latency. This task can be performed by detecting and measuring the three centroid positions on the target and utilizing those positions with a reporting algorithm to determine the position of the patient's head.
Six-Degree-of-Freedom Coordinate System
In an embodiment, the system is configured to use a coordinate system for reporting 6-DOF motions to the MRI field compensation system that is a Cartesian system aligned with the symmetry axis of the head coil as shown in
Coordinate definitions are adopted by the same conventions used in defining aircraft motion, except that the rotation directions are taken to be right-handed (positive for counter-clockwise rotation about the basis direction vectors):
x is the longitudinal (chin-to-crown) direction, with values increasing toward the top of the head
y is the transverse (left-to-right) direction, with increasing values toward the patient's right ear
z is the up-down direction, with increasing values toward the ceiling
is the yaw angle or right-handed rotation about the z-axis (head lean toward shoulder while facing forward, zero at normal “square” position, positive values for patient leaning toward patient's right shoulder)
is the pitch angle or right-handed rotation about the y-axis (nodding “yes,” zero at normal “square” position, positive values for patient looking “upward”)
is the roll angle or right-handed rotation about the x-axis (shaking the head “no,” zero at normal “square” position, positive values for patient looking toward patient's left side).
The origin of coordinates and angle zero references are arbitrary, as only relative motions are reported, however two convenient reference origin positions exist: 1) at the center of the target in its normal (“square”) head position, and 2) at the base of the neck at the point where the head swivels for nod, turn and lean motions. The latter is adopted here (as shown in
Target Displacement Equations
The full 6-DOF translation is composed of a 3-D displacement as well as a 3-axis rotation. To first order we assume that the skull moves as a rigid body about a single rotation point somewhere in the neck. From this point the translation becomes separable from the rotation, so this is chosen as the coordinate origin. The rotations are separated into roll, pitch and yaw as described above, and the translated position through rotation follows the Euler rotation matrix formulation as follows (using right-handed angle conventions). The x, y, and z displacement coordinates then follow the independent translations:
Decomposing the six independent translations from the absolute and relative displacements of the measured target centroids is the subject of this effort. The 2D inverse problem is somewhat more difficult than the 3D problem, in that after the target centroid projections in focal plane row and column are determined, significant degeneracies remain in the unfolding matrices for each camera. Combining the data from both cameras removes these degeneracies through a series of interrelated, nonlinear equations. The fastest procedure for solving this inverse problem is obtained by the Newton-Raphson method or a variant thereof, whereby an approximate first-order solution is proposed and tested against the known (measured) centroid locations on the two camera focal planes. The residual error is divided by the local derivatives with respect to each component of rotation and translation, to determine an iterative correction. The first-order solution is chosen by considering the features of the projected target pattern which are most strongly affected by a single rotation angle or displacement, and linearizing the inversion problem along these feature axes.
A 6-DOF motion simulation and decomposition algorithm was developed and tested to allow simulation of arbitrary motions and then verify the ability of a pair of orthogonal cameras to decompose centroid measurements at the 0.1-pixel level into distinct x, y, z, roll, pitch and yaw components at the requisite level of accuracy.
Six-Degree-of-Freedom Motion Determination Algorithm
General subject motion is a superposition of translation along x, y, and z as well as rotation about the x, y and z axes (designated roll, pitch and yaw respectively). Displacements along each of these degrees of freedom are not sensitive to coordinate system origin; however it is convenient (as explained above) for modeling purposes to place an origin near the region of the spine about which the head rotates and swivels, and a secondary reference point at the center of the optical tracking target in the nominal (“correct”) head position and orientation. This secondary reference is typically offset from the spinal origin by ˜10 cm in x and ˜10 cm in z.
The target shown in
Yaw
Rotation about the z-axis is designated as yaw; a positive or “right handed” rotation about this axis (head leaning to subject's right shoulder) results in a counterclockwise rotation of the target. Because this rotation usually occurs about a point lower in the neck, it is typically accompanied by a translation to the subject's right side (camera left), as seen in
The median of the centered target triangle (as shown at the right in
Pitch
Rotation about the y-axis is designated as pitch; a positive or “right-handed” rotation about this axis (head tipped back) results in motion of the target upward off the gantry (+z) and toward the top of the head (+x). For a single camera this projection is not easily distinguishable from a simultaneous target displacement in x and y (see
Roll
Rotation about the x-axis is designated as roll; a positive or “right-handed” rotation about this axis (head pointing toward subject's left side) results in a motion of the target toward the subject's left (−y). For a single camera this motion is not easily distinguishable from a displacement in y (see
As shown in
X-Axis Translation
Translation along the x-axis (positive toward top of head) results in a motion of the target along the vertical direction of the camera focal plane (see
Y-Axis Translation
Translation along the y-axis (positive toward subject's right side) results in a motion of the target along the horizontal axis of the camera focal plane (see
Z-Axis Translation
Translation along the z-axis (positive toward the ceiling) results in apparent motion of the target along the horizontal axis of the camera focal plane. Unlike for y translation, however, the direction of the horizontal displacement is opposite between the left-side and right-side cameras (see
Pitch Versus (X+Z) Translation Degeneracy
Pitch is nearly degenerate with simultaneous x and z translation, except for a small tilt in the triangle vertical which results from the tilt of the target plane about the y axis. This tilt creates an apparent clockwise rotation of the triangle from the left-side view and an apparent counterclockwise rotation from the right side view, as shown in
Roll Versus (Y+Z) Translation Degeneracy
Roll is nearly degenerate with simultaneous y and z translation, except for larger camera-to-camera differences in apparent target size encountered with roll, resulting from tilt of the target's plane about the x-axis. A significant difference in the apparent length of the target triangle base is a reliable distinguishing characteristic of roll motion rather than simple translation.
For a roll of 0.1 degrees and translations in y and z of −0.244 mm and 0.0002 mm respectively, the lower centroid is unchanged in both camera views. In this case, the left-side camera sees the target triangle base shrink by 0.15 pixels while the right-side camera sees the triangle base grow by 0.15 pixels. Assuming the centroiding routine can locate the target centroids to an accuracy of 0.1 pixels, shifts of 0.14 pixels should be discernible, so a pitch as small as 0.1 degrees is distinguishable from a simple translation by comparison of the length of the target triangle base.
Complementary Projections Versus Stereo Imaging
The target size, rotation angles and translation vector determine the relative displacement of the three target centroids precisely in three dimensions. Precise knowledge of camera and lens parameters (e.g., pixel pitch, lens focal length and radial distortion, camera location and orientation relative to nominal target position), are then sufficient to predict the location of the target centroid projections to better than 0.1 pixels in row and column for each camera. In principle, the inverse problem should be equally simple; the stereo view of the centroid projections determine the 3D location in space of each target centroid, and the 6-DOF displacement vector can then be determined through a simple matrix inversion. In practice, however, this approach leads to expensive and complicated requirements on camera pixel density, pixel count, camera alignment and camera calibration. An alternate unfolding approach dispenses with stereo ranging and uses the two camera projections separately without strict requirements on precise matching of camera alignment and magnification, to determine the 6-DOF displacement vector to within 0.1 degrees in each rotation angle and 0.1 mm along each translation axis. This approach relaxes the pixel density and count requirements by about a factor of 20 relative to the stereo approach, resulting in significant savings in cost and processing speed.
Even for this 2D approach some basic measurements can be made to calibrate camera parameters once the cameras are integrated with the head coil; these can be easily performed at the manufacturing facility. These measurements include the projected pixel location of a single reference point on both cameras, as well as the camera magnification factors for pixel displacement per degree of rotation in pitch, yaw and roll, and per mm of translation along x, y and z. However, as stated before, it is not necessary that the cameras be exactly aligned in space (e.g. perfectly normal) or that their magnifications (lens focal length and distance to reference point) be identical, as has been easily verified by simulation.
Inversion Equations
The 2D inversion problem is somewhat more difficult than the 3D problem, in that after the target centroid projections in focal plane row and column are determined, significant degeneracies remain in the unfolding matrices for each camera. Combining the data from both cameras removes these degeneracies through a series of interrelated, nonlinear equations. The fastest procedure for solving this inverse problem is obtained by a variant of the Newton-Raphson method, whereby an approximate first-order solution is proposed and tested against the known (measured) centroid locations on the two camera focal planes. The residual error is divided by the local derivatives with respect to each component of rotation and translation, to determine an iterative correction. The first-order solution is chosen by considering the features of the projected target pattern which are most strongly affected by a single rotation angle or displacement, and linearizing the inversion problem along these feature axes.
6-DOF Extraction Algorithm
The method for extracting the 6 degree of freedom displacement matrix from the observed target location on two focal plane cameras is described.
Step 1: Characterizing the Target Images
The optical target consists of elliptical targets shown in
a) ΣHD—the sum of the horizontal displacements (in pixels) of the target center on camera 1 and camera 2; the formula used is Σi=13Σj=12(Xi,j−X0i,j), where X0i,j is the initial (zero displacement) horizontal camera coordinate of each centroid projection.
b) ΔHD—the difference between the horizontal displacements (in pixels) of the target center for camera 1 and camera 2; the formula used is Σi=13(Xi,1−X0i,1)−(Xi,2−X0i,2).
c) ΣVD—the sum of the vertical displacements (in pixels) of the target center for camera 1 and camera 2; the formula used is Σi=13Σj=12(Yi,j−Y0i,j), where Y0i,j is the initial (zero displacement) vertical camera coordinate of each centroid projection.
d) ΔBL—the difference in the apparent base length of the target triangle (in pixels) for camera 1 and camera 2; the formula used is {√{square root over ((X3,1−X1,1)2+(Y3,1−Y1,1)2)}−√{square root over ((X3,2−X1,2)2+(Y3,2−Y1,2)2)}}.
e) ΣMT—the sum of the apparent median tilt of the target triangle (offset in horizontal pixels between center-of-base and apex) for camera 1 and camera 2; the formula used is
f) ΔMT—the difference between the apparent median tilt of the target triangle (in pixels) for camera 1 and camera 2; the formula used is
Step 2: Characterizing Global Variation in Principal Quantities with 6-DOF Motions
Partial derivatives relative to subject displacements and rotations (φ, θ, ψ, Δx, Δy, Δz), of the principal quantities described above, about the initial (non-displaced) position, are computed numerically. Here:
Roll φ is right-handed rotation about the x-axis
Pitch θ is right-handed rotation about the y-axis
Yaw ψ is right-handed rotation about the z-axis
Δx is toe-to-head direction
Δy is left-to-right direction
Δz is down-to-up direction
Starting from an initial target position in 3-D world space, defined as (φ, θ, ψ, Δx, Δy, Δz)=(0, 0, 0, 0, 0, 0), the initial target vertex world coordinates (x0i, y0i, z0i) are determined for vertex index i=1 to 3, based on the geometric size and shape of the target triangle and definition of a convenient coordinate origin.
Local partial derivatives of each of the principal quantities, with respect to each of the 6 degrees of freedom (roll, pitch, yaw, dx, dy, dz), are performed numerically by evaluating changes in these quantities for small increments in each degree of freedom. Changes in the target vertex positions for specified motions along the six degrees of freedom are computed using the Euler rotation matrices and translation vector:
Subsequently, the camera projections of these new target vertex positions are determined using a geometric projection calculation. Given accurate knowledge of camera positions, pixel size and lens focal length, the horizontal and vertical pixel numbers on each camera focal plane (camera index j equal to 1 or 2) that these new 3-D positions in space should project onto is as follows:
Here Xi,j and Yi,j are the horizontal and vertical pixel numbers for translated target vertex i projected onto the camera j sensor, X0,j and Y0,j are the horizontal and vertical number of the pixel column and row intersected by the optical axis of that camera (typically at or very near the camera center), f.l. and spix are the lens focal length and camera pixel pitch, and the angles αi,j and βi,j are the polar and azimuth angles locating target vertex i, relative to the camera j focal axis. These angles are calculated from the vertex world coordinates as follows:
where the point (x⊥,j, y⊥,j, z⊥,j) is the point of intersection between the camera optical axis and the plane perpendicular to the optical axis which includes the translated target vertex (xi, yi, zi):
x⊥,j=x0+κ(xcj−x0);y⊥,j=y0+κ(ycj−y0);z⊥,j=z0+κ(zcj−z0), [5]
with (xcj, ycj, zcj) defining the 3-D position of camera j, (x0, y0, z0) defining the nominal boresight position of both cameras at the un-displaced target center and the constant κ based on geometric projection and given by:
In equation [4], the inverse cosine function is taken to range from 0 to π, and the appropriate sign for βi,j is given by:
sign[βi,j]=sign[(zcj−zi){(xcj−x0)(x⊥,j−xi)+(ycj−y0)(y⊥,j−yi)}−(z⊥,j−zi){(xcj−x0)2+(ycj−y0)2}]
During this determination of the camera projection of the 3-D target vertices, a compensation function may be applied for large values of the polar angle αi,j to account for barrel distortion in the lens, based on prior lens calibration measurements. The geometric value for αi,j is first computed based on equation [3] and then adjusted for lens distortion by way of a pre-determined look-up table or measured fit function, and this new compensated value for αi,j is then used in the calculation of Xi,j and Yi,j through equation [2].
To numerically evaluate the partial derivatives of the principal quantities about the initialized target position, the un-displaced 3-D target vertex coordinates (x0i, y0i, z0i) are first projected to camera coordinates using equations [2] through [6] above, and initial values are computed for each of the principal quantities described in Step 1 (most should be zero or near-zero at the starting position). Then small increments of roll, pitch, yaw, x-, y- and z-axis displacements are introduced one at a time; for each increment the new world coordinates and the new camera projections of the target vertices are computed and the principal quantities are re-calculated. The change in each principal quantity is divided by the small angular or displacement increment to determine the partial derivative.
For instance, to determine the partial derivatives with respect to roll, the displacement vector (φ, θ, ω, Δx, Δy, Δz)=(βφ, 0, 0, 0, 0, 0) is introduced to the general displacement equation [1] to determine the translated target vertex positions (xi, yi, zi). The conversion to camera coordinates (X1,j, Yi,j) is then performed using equations [2] through [6], and the principal quantities are calculated as outlined in Step 1. The difference between each principal quantity and the corresponding value of that quantity for the un-displaced calculation is divided by the small increment in roll, to give the partial derivative of each quantity with respect to roll. To determine partial derivatives with respect to pitch, the displacement vector (φ, θ, ψ, Δx, Δy, Δz)=(0, βθ, 0, 0, 0, 0) is used to initiate the calculations, and so on for all six degrees of freedom.
Each of these six repetitions produces one column of the global partial derivative matrix:
Step 3: Determining First-Order Displacement Vector
A first-order approximation to the displacement matrix is determined by multiplying the matrix of measured principal quantities, as determined in Step 1, by the inverse of the partial derivative matrix computed in Step 2:
Step 4: Characterizing Local Variation in Principal Quantities with 6-DOF Motions
First order values for (φ, θ, ψ, Δx, Δy, Δz) determined in Step 3 are entered into the translation equation [1] to determine the corresponding translated 3-D target position (xi, yi, z1) for each of the three target vertices. These world coordinates are projected to camera coordinates (Xi,j, Yi,j) using equations [2] through [6], and the principal quantities are re-calculated. These six quantities are compared against the measured values of these quantities determined in Step 1, to create a residual error matrix:
(σΣ
Local partial derivatives of the principal quantities are calculated by introducing small increments in roll, pitch, yaw, x-, y- and z-axis displacements one at a time as before, but this time the increments are relative to the first-order displacement vector. For each increment, the new world coordinates and the new camera projections of the target vertices are re-computed and the principal quantities are re-calculated. The change in each principal quantity is divided by the small angular or displacement increment to determine a local partial derivative. For instance, to calculate partial derivatives with respect to roll, the first-order displacement vector {φ0, θ0, ψ0, (Δx)0, (Δy)0, (Δz)0} is replaced by {φ0+δφ, θ0, ω0, (Δx)0, (Δy)0, (Δz)0} and resulting changes to each of the principal quantities is divided by δφ to determine the local derivative with respect to roll. This is repeated for each of the six degrees of freedom.
Each of these six repetitions produces one column of the new local partial derivative matrix:
Step 5: Determining Coarse Correction to First-Order Displacement Vector
A coarse correction is computed to improve the first-order displacement vector and reduce residual error, by multiplying the residual error matrix determined in Step 4 by the inverse of the local partial derivative matrix, also determined in Step 4:
The first-order displacement vector is incremented by the coarse correction matrix to create a better approximation to the displacement vector:
{φ0+Δφ),θ0+Δθ,ψ0+Δψ,(Δx)0+Δ(Δx),(Δy)0+Δ(Δy),(Δz)0+Δ(Δz)}.
Step 6: Performing Fine Correction to Determine Final 6DOF Displacement Vector
Steps 4 and 5 are repeated, starting with the coarse-corrected displacement vector, to determine a final fine correction to the displacement vector. After this iteration, the resultant fine correction increments are added to the coarse-corrected vector to create the final 6-DOF displacement vector. Empirical results from a general simulation indicate that this fine correction is sufficient in all cases to reduce residual errors to well below the stated 0.1-degree, 0.1-mm tolerances.
Algorithm Numerical Simulation to Verify Absolute Convergence
As cameras, targets and rotation stages are being procured and assembled, the 6 DOF decomposition algorithm can be coded and tested for a test set of rotations. It is clear that the routine will converge for small translations and rotations, but it can be potentially advantageous to determine whether there are limitations on its convergence for extreme displacements in all six degrees of freedom. To this end, we imagine an extreme target displacement, calculate the 3D position of the displaced target, calculate the centroid positions that will be seen on each of the two cameras, and run the decomposition algorithm to determine speed of convergence.
In some embodiments, to demonstrate absolute convergence of the iterative 6DOF unfolding algorithm, the simulation is started with a test set of very large rotations and displacements, as listed in Table 1 below.
The simulation begins by determining the locations of the displaced centroids that will be seen by each camera, allowing for some degree of mispointing and misalignment of each camera. The original (nominal) target location is rotated and displaced by the Euler rotation formalism presented in Section 2.5.2.2, to determine the three displaced target centroid locations in three-dimensional space. Next these “world coordinates” are translated to 2-D “camera coordinates” for each of the two cameras independently, as described in the same Section.
Assuming the target is imaged into these camera coordinates, but that the operator has no prior knowledge of the displacement matrix giving rise to this target position, we use the algorithm as described in Section 2.5.2 from end to end to recreate the displacement matrix. By the end of Step 3 (Section 2.5.2.3), the algorithm returns an initial estimate of the 6DOF displacement vector, as shown in Table 2 below.
As expected, residual errors at this stage are atypically large, due to the extreme magnitudes of the translations and rotations chosen for this simulation along and about each axis; this situation creates a good test for absolute convergence of the Newton Raphson algorithm methodology. Assuming this estimate to be correct, the algorithm in Step 4 (Section 2.5.2.4) again calculates the displaced position of the target, the resulting centroid positions seen by each camera, and the principal quantities (vertical tip sum and difference, base length difference, vertical displacement sum, and horizontal displacement sum and difference) which would result, for comparison with the actual observed values. The residual errors, in pixels, and the local derivatives of each of the principal values for small changes (pixels per 0.1 degrees) in yaw, pitch, and roll, and for small changes (pixels per 0.1 mm) in dx, dy and dz are calculated as described in Section 2.5.2.4, and tabulated as shown in Table 3 below.
The matrix of derivatives at the left of Table 3 is inverted and multiplied by the residual error vector at the right, to yield first-order corrections to the initial estimate of the displacement vector, as described in Section 2.5.2.5, and as shown at the left of Table 4 below. These are added to the initial estimates, to produce the more refined estimate of the 6 DOF displacement vector, shown at the right of Table 4.
This process is repeated for a second and final time as described in Section 2.5.2.6, assuming again that the (now refined) 6 DOF displacement vector is accurate, and calculating first the 3D target centroid positions and then the locations of the target centroids as projected onto each of the two camera focal planes. Again the six principal quantities are computed and compared with the actual observations to produce a vector of residual errors. Again the local derivatives are computed, this time at the location of the first-order displacement vector. The results are tabulated as shown in Table 5 below.
The matrix of derivatives at the left of Table 5 is inverted and multiplied by the residual error vector at the right, to yield final corrections to the first-order estimate of the displacement vector, as shown at the left of Table 6 below. These corrections are added to the first-order estimates, to produce the final second-order estimate of the 6 DOF displacement vector, shown at the right of Table 6.
Even for the extreme rotations and displacements used in this model, the algorithm is shown to converge to within 0.003 degrees and 0.006 mm in only two iterations. Given the number of floating-point operations needed to perform the initial estimate and two successive iterations of the Newton method, the algorithm can produce a solution on a typical laptop computer in less than 5 milliseconds.
Quaternion Representation
The head coil ICD specifies the rotation vector in terms of the quaternion, for which (still using right-handed Euler angle rotation conventions):
The translation vector is unchanged from the form calculated here.
Centroid Determination Algorithm
The centroid location on the focal plane is given by:
This calculation is performed for three subregions on the target as shown in
Centroid Determination
In some embodiments, a test target can be printed and mounted in the view field of a monochrome camera at an angle of approximately 45 degrees. At this angle the elliptical target projected to an approximately round target on the camera focal plane.
The calculated target centroid is displayed as a red dot at the center of the LabView image in
Camera Calibration
As with any camera lens, the lens used for the head tracker could have some level of distortion as a function of distance from imaging axis. Azimuthal distortion should be negligible, but radial distortion can be measured after lens installation and fit to a polynomial curve to allow rapid compensation of centroid positions near the edges of the camera field of view. The 6DOF unfolding algorithm can be constructed to accommodate typical levels of radial distortion as a second-order compensation during the application of the Newton Raphson iteration method.
Radial distortion can be determined using a printed reference target with concentric circles of diameter ⅓″, ⅔″, 1″, and so on up to a maximum diameter of 4 inches, as shown in
In one embodiment, the measured radial distortion measured for the TechSpec High Resolution Fixed Focus 25 mm lens follows camera polar angle θc=(1+0.0053144θ−0.0016804θ2+0.0002483θ3−0.0000138θ4)θ, with laboratory polar angle θ in degrees. At the extreme corner of the viewing field, where θ˜6.75°, camera aberration results in a radial growth in camera angle of about 0.7% relative to true angle, or about 2.8 pixels in radius.
Full 6-DOF Tracking
The full 6-DOF tracking algorithm was coded in LabView with the Graphical User Interface (GUI). The upper left side of the GUI screen gives centroid information for target circles in the current frame, and the lower left side gives the same information for the prior frame. For each, one nested target circle from the set of three is displayed in negative (white on black) along with a histogram of its pixel brightness within a 48-by-48 pixel box centered on the centroid location of the previous frame. This histogram is split into two sections to display (at left) the peak from background pixels at one end of the brightness scale, and (at right) the peak from the pixels of the target itself, at the other end of the brightness scale. A long continuum of pixels in between represents pixels at dark-light boundaries in the target frame. From analysis of the two histograms, the target field is clipped at the lower-brightness shoulder on the bright side, and the upper brightness shoulder on the dark side, to create a binary target field that is not sensitive to variations in illumination across the target. Although displayed in real time for only one target circle, all three target circles are processed in this way.
Next to the target histograms, the x-y camera centroid locations are displayed to two-decimal precision for each of the three nested circle targets; again at the upper half of the screen for the current data and at the lower half of the screen for the prior frame.
The right side of the screen displays the processed 6-DOF data, after analysis using the approach described in Section 2.5. An analog meter-style display shows the acquisition and processing time per frame, which is limited at its low end to the camera frame integration and readout time of about 8 milliseconds. Using a single iteration of the Newton-Raphson routine described in Section 2.5, the algorithm runs during the integration period for the successive frame, so the processing time is approximately 8 milliseconds, corresponding to a 120 Hz camera readout rate. The 6-DOF data can be displayed in either analog or digital format, but the digital format can be read to precision of 0.01 mm and 0.01 degree for comparison with the 0.1 mm, 0.1 degree accuracy requirements.
Laboratory Mechanical Layout for Head Tracking Simulation
The laboratory setup was designed to mimic head rotation and displacement using a six-degree-of-freedom optical rotation mount. This mount included three ganged translation stages along the x-, y-, and z-axes of the optical table, and three ganged rotation stages corresponding to yaw, roll and pitch respectively. The two monochrome cameras and turning mirrors were mounted in the appropriate geometry for use with an existing 12-channel head coil. The two monochrome cameras are in foreground, mounted at ±45° relative to horizontal to accommodate rotation by the turning mirrors. The turning mirrors are mounted 10 inches behind cameras (slightly obscured by the cameras in the picture). The target is partially visible in the reflection of each mirror. The 6-DOF rotation stage is at center in foreground, with the y-axis stage at bottom, x-axis stage next, and z-axis stage above that, followed by the yaw rotation stage, the roll stage, and finally the pitch stage with target at the top (the pitch rotation handle is obscured by the stage). A near-IR illumination LED is at the center in background; light from this stage is within the camera spectral range, but hardly visible to the human eye.
X-Axis Translation
The second translation stage from the bottom in the 6-DOF displacement assembly controls x-axis displacement (aligned with the patient's spine). The x-axis translation stage control knob is turned four full rotations (corresponding to −2.54 mm), and the absolute position change is calculated from the resulting motion of the centroid camera coordinates. Results are: the displacement determined by the unfolding algorithm is −2.56 mm in x, less than 0.1 mm in y and z, and less than 0.1° in roll, pitch and yaw. The target displacement by dx=−2.54 mm, with zoom on lower right display section of GUI showed calculated dx=−2.56 mm, dy=0.08 mm, dz=0.02 mm, dϕ=0.05°, dθ=−0.03°, and dψ=−0.01°.
Y-Axis Translation
The bottom translation stage in the 6-DOF displacement assembly controls y-axis displacement (patient's left-to-right). The y-axis translation stage control knob is turned four full rotations (corresponding to −2.54 mm), and the absolute position change is calculated from the resulting motion of the centroid camera coordinates. This resulted in a target displacement by dy=−2.54 mm, with zoom on lower right display section of GUI showing dx=0.00 mm, dy=−2.47 mm, dz=−0.01 mm, dϕ=0.64°, dθ=−0.04°, and dψ=−0.03°.
Z-Axis Translation
The top translation stage in the 6-DOF displacement assembly controls z-axis displacement (patient's down to up, with the patient lying on his back). The z-axis translation stage control knob is turned four full rotations (corresponding to −2.54 cm), and the absolute position change is calculated from the resulting motion of the centroid camera coordinates. The displacement determined by the unfolding algorithm was −2.54 mm in z, less than 0.1 mm in x and y, and less than 0.1° in roll, pitch and yaw. The results were a target displacement by dz=−2.54 mm, with zoom on lower right display section of GUI showing dx=0.01 mm, dy=−0.01 mm, dz=−2.59 mm, dϕ=−0.02°, dθ=−0.06° and dφ=0.01°.
Yaw Rotation
The bottom rotation stage in the 6-DOF displacement assembly controls yaw rotation (patient's left shoulder-to-right shoulder lean direction). The yaw rotation stage control knob is turned by +4° degrees (heading 315° to heading 311° on stage, corresponds to movement toward right shoulder), and the absolute position change is calculated from the resulting motion of the centroid camera coordinates. The displacement determined by the unfolding algorithm is less than 0.1 mm in dx, dy and dz, 0.1° in roll and less than 0.1° in pitch, and 3.94° in yaw. The results were a target rotation by dψ=+4.00°, with zoom on lower right display section of GUI showing dx=0.07 mm, dy=−0.05 mm, dz=0.02 mm, dϕ=0.10°, dθ=−0.01°, and dψ=3.94°.
Roll Rotation
The middle rotation stage in the 6-DOF displacement assembly controls roll rotation (patient's right shoulder-to-left shoulder “head shaking” direction). The roll goniometer control knob is turned by +5° degrees, and the absolute position change is calculated from the resulting motion of the centroid camera coordinates. The displacement determined by the unfolding algorithm is less than 0.1 mm in dx, and dz, 1.78 mm in dy, 4.97° in roll and less than 0.1° in pitch and yaw. Displacement in y is expected due to the fact that the center of rotation for the Thorlabs GNL18 goniometer stage is 44.5 mm above the mount surface, while the target is only 22 mm above the stage. For the resulting −20.5 mm lever arm, the y-displacement due to a 5° roll rotation is −(−20.5 mm)*)sin(5°)=+1.79 mm, in good agreement with the measured data.
The results were a target rotation by dϕ=+5.00°, with zoom on lower right display section of GUI showing dx=0.07 mm, dy=1.78 mm, dz=−0.01 mm, dϕ=4.97°, dθ=−0.03°, and dψ=0.08°.
Pitch Rotation
The top rotation stage in the 6-DOF displacement assembly controls pitch rotation (patient's “nodding” direction). The pitch goniometer control knob is turned by +5° degrees, and the absolute position change is calculated from the resulting motion of the centroid camera coordinates. The calculated pitch is 4.95°, with less than 0.1° in yaw. The center of rotation for the Thorlabs GNL10 goniometer stage is 25.4 mm above the mount surface, while the target is only 6.4 mm above the stage. For the resulting −19 mm lever arm, the x-displacement due to a 5° rotation is −19 mm*)sin(5°)=−1.66 mm, the y-displacement is 0.00 mm, and the z-displacement is −19 mm*)[1−cos(5°)]=0.07 mm. These displacements are all within 0.1 mm of measured data.
The results were a target pitch rotation by dθ=+5.00°, with zoom on lower right display section of GUI showing dx=−1.63 mm, dy=0.09 mm, dz=0.17 mm, dϕ=0.21°, dθ=4.95°, and dψ=−0.07°.
Specific embodiments have been described in detail above with emphasis on medical application and in particular MRI examination of a patient's head. However, the teachings of the present invention can be utilized for other MRI examinations of other body parts where movements of up to six degrees of freedom are possible. In addition medical procedures involving imaging devices other than MRI equipment (e.g., CT, PET, ultrasound, plain radiography, and others) may benefit from the teaching of the present invention. The teachings of the present invention may be useful in many non-medical applications where tracking of a target having several degrees of freedom are possible. Some of these applications could be military applications. Furthermore, while particular algorithms are disclosed, variations, combinations, and subcombinations are also possible.
Computing System
In some embodiments, the computer clients and/or servers described above take the form of a computing system 1500 illustrated in
Motion Correction Control Systems
In an embodiment, the system 700 comprises a motion correction control system module 1514 that carries out the functions described herein with reference to motion correction mechanism, including any one of the motion correction methods described above. The motion correction control system module 1514 may be executed on the computing system 1500 by a central processing unit 1504 discussed further below.
In general, the word “module,” as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, COBOL, CICS, Java, Lua, C or C++ or Objective C. A software module may be compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language such as, for example, BASIC, Perl, or Python. It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts. Software instructions may be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors. The modules described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage.
Computing System Components
In an embodiment, the computing system 1500 also comprises a workstation or other computing devices suitable for controlling and/or communicating with large databases, performing transaction processing, and generating reports from large databases. The computing system 1500 also comprises a central processing unit (“CPU”) 1504, which may comprise a conventional microprocessor. The computing system 1500 further comprises a memory 1508, such as random access memory (“RAM”) for temporary storage of information and/or a read only memory (“ROM”) for permanent storage of information, and a mass storage device 1502, such as a hard drive, diskette, or optical media storage device. Typically, the modules of the computing system 1500 are connected to the computer using a standards based bus system. In different embodiments, the standards based bus system could be Peripheral Component Interconnect (PCI), Microchannel, SCSI, Industrial Standard Architecture (ISA) and Extended ISA (EISA) architectures, for example.
The computing system 1500 comprises one or more commonly available input/output (I/O) devices and interfaces 1512, such as a keyboard, mouse, touchpad, and printer. In one embodiment, the I/O devices and interfaces 1512 comprise one or more display devices, such as a monitor, that allows the visual presentation of data to a user. More particularly, a display device provides for the presentation of GUIs, application software data, and multimedia presentations, for example. In the embodiment of
Computing System Device/Operating System
The computing system 1500 may run on a variety of computing devices, such as, for example, a mobile device or a server or a desktop or a workstation, a Windows server, an Structure Query Language server, a Unix server, a personal computer, a mainframe computer, a laptop computer, a cell phone, a personal digital assistant, a kiosk, an audio player, a smartphone, a tablet computing device, and so forth. The computing system 1500 is generally controlled and coordinated by operating system software, such as iOS, z/OS, Windows 95, Windows 98, Windows NT, Windows 2000, Windows XP, Windows Vista, Windows 7, Linux, BSD, SunOS, Solaris, or other compatible operating systems. In Macintosh systems, the operating system may be any available operating system, such as MAC OS X. In other embodiments, the computing system 1500 may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, and I/O services, and provide a user interface, such as a graphical user interface (“GUI”), among other things.
Network
In the embodiment of
Access to the motion correction control system module 1514 of the computer system 1500 by computing systems 1520 and/or by data sources 1522 may be through a web-enabled user access point such as the computing systems' 1520 or data source's 1522 personal computer, cellular phone, laptop, or other device capable of connecting to the network 1518. Such a device may have a browser module is implemented as a module that uses text, graphics, audio, video, and other media to present data and to allow interaction with data via the network 1518.
The browser module may be implemented as a combination of an all points addressable display such as a cathode-ray tube (CRT), a liquid crystal display (LCD), a plasma display, touch screen display or other types and/or combinations of displays. In addition, the browser module may be implemented to communicate with input devices 1512 and may also comprise software with the appropriate interfaces which allow a user to access data through the use of stylized screen elements such as, for example, menus, windows, dialog boxes, toolbars, and controls (for example, radio buttons, check boxes, sliding scales, and so forth). Furthermore, the browser module may communicate with a set of input and output devices to receive signals from the user.
The input device(s) may comprise a keyboard, roller ball, pen and stylus, mouse, trackball, voice recognition system, or pre-designated switches or buttons. The output device(s) may comprise a speaker, a display screen, a printer, or a voice synthesizer. In addition a touch screen may act as a hybrid input/output device. In another embodiment, a user may interact with the system more directly such as through a system terminal connected to the score generator without communications over the Internet, a WAN, or LAN, or similar network.
In some embodiments, the system 1500 may comprise a physical or logical connection established between a remote microprocessor and a mainframe host computer for the express purpose of uploading, downloading, or viewing interactive data and databases on-line in real time. The remote microprocessor may be operated by an entity operating the computer system 1500, including the client server systems or the main server system, an/or may be operated by one or more of the data sources 1522 and/or one or more of the computing systems. In some embodiments, terminal emulation software may be used on the microprocessor for participating in the micro-mainframe link.
In some embodiments, computing systems 1520 that are internal to an entity operating the computer system 1500 may access the motion correction control system module 1514 internally as an application or process run by the CPU 1504.
User Access Point
In an embodiment, the computing system 1500 comprises a computing system, a smartphone, a tablet computing device, a mobile device, a personal computer, a laptop computer, a portable computing device, a server, a computer workstation, a local area network of individual computers, an interactive kiosk, a personal digital assistant, an interactive wireless communications device, a handheld computer, an embedded computing device, or the like.
Other Systems
In addition to the systems that are illustrated in
Conditional language, such as, among others, “can,” “could,” “might,” or “may,” unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment. The headings used herein are for the convenience of the reader only and are not meant to limit the scope of the inventions or claims.
Although this invention has been disclosed in the context of certain preferred embodiments and examples, it will be understood by those skilled in the art that the present invention extends beyond the specifically disclosed embodiments to other alternative embodiments and/or uses of the invention and obvious modifications and equivalents thereof. Additionally, the skilled artisan will recognize that any of the above-described methods can be carried out using any appropriate apparatus. Further, the disclosure herein of any particular feature, aspect, method, property, characteristic, quality, attribute, element, or the like in connection with an embodiment can be used in all other embodiments set forth herein. For all of the embodiments described herein the steps of the methods need not be performed sequentially. Thus, it is intended that the scope of the present invention herein disclosed should not be limited by the particular disclosed embodiments described above.
This application is a continuation of U.S. patent application Ser. No. 15/428,495, entitled SYSTEMS, DEVICES, AND METHODS FOR TRACKING MOVING TARGETS, filed Feb. 9, 2017, which is a continuation of U.S. patent application Ser. No. 14/837,554, entitled SYSTEMS, DEVICES, AND METHODS FOR TRACKING MOVING TARGETS, filed Aug. 27, 2015, which is a continuation of U.S. patent application Ser. No. 13/831,115, entitled SYSTEMS, DEVICES, AND METHODS FOR TRACKING MOVING TARGETS, filed Mar. 14, 2013, which claims the benefit as a nonprovisional application of U.S. Provisional Patent Application No. 61/849,338, entitled SIX DEGREES OF FREEDOM OPTICAL TRACKER, filed Jan. 24, 2013. Each of the foregoing applications is hereby incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
3811213 | Eaves | May 1974 | A |
4689999 | Shkedi | Sep 1987 | A |
4724386 | Haacke et al. | Feb 1988 | A |
4894129 | Leiponen et al. | Jan 1990 | A |
4923295 | Sireul et al. | May 1990 | A |
4953554 | Zerhouni et al. | Sep 1990 | A |
4988886 | Palum et al. | Jan 1991 | A |
5075562 | Greivenkamp et al. | Dec 1991 | A |
5318026 | Pelc | Jun 1994 | A |
5515711 | Hinkle | May 1996 | A |
5545993 | Taguchi et al. | Aug 1996 | A |
5615677 | Pelc et al. | Apr 1997 | A |
5687725 | Wendt | Nov 1997 | A |
5728935 | Czompo | Mar 1998 | A |
5802202 | Yamada et al. | Sep 1998 | A |
5808376 | Gordon et al. | Sep 1998 | A |
5835223 | Zawemer et al. | Nov 1998 | A |
5877732 | Ziarati | Mar 1999 | A |
5886257 | Gustafson et al. | Mar 1999 | A |
5889505 | Toyama | Mar 1999 | A |
5891060 | McGregor | Apr 1999 | A |
5936722 | Armstrong et al. | Aug 1999 | A |
5936723 | Schmidt et al. | Aug 1999 | A |
5947900 | Derbyshire et al. | Sep 1999 | A |
5987349 | Schulz | Nov 1999 | A |
6016439 | Acker | Jan 2000 | A |
6031888 | Ivan et al. | Feb 2000 | A |
6044308 | Huissoon | Mar 2000 | A |
6057680 | Foo et al. | May 2000 | A |
6061644 | Leis | May 2000 | A |
6088482 | He | Jul 2000 | A |
6144875 | Schweikard et al. | Nov 2000 | A |
6175756 | Ferre | Jan 2001 | B1 |
6236737 | Gregson et al. | May 2001 | B1 |
6246900 | Cosman et al. | Jun 2001 | B1 |
6279579 | Riaziat et al. | Aug 2001 | B1 |
6285902 | Kienzle, III et al. | Sep 2001 | B1 |
6289235 | Webber | Sep 2001 | B1 |
6292683 | Gupta et al. | Sep 2001 | B1 |
6298262 | Franck et al. | Oct 2001 | B1 |
6381485 | Hunter et al. | Apr 2002 | B1 |
6384908 | Schmidt et al. | May 2002 | B1 |
6390982 | Bova et al. | May 2002 | B1 |
6402762 | Hunter et al. | Jun 2002 | B2 |
6405072 | Cosman | Jun 2002 | B1 |
6421551 | Kuth et al. | Jul 2002 | B1 |
6467905 | Stahl et al. | Oct 2002 | B1 |
6474159 | Foxlin et al. | Nov 2002 | B1 |
6484131 | Amoral-Moriya et al. | Nov 2002 | B1 |
6490475 | Seeley et al. | Dec 2002 | B1 |
6501981 | Schweikard et al. | Dec 2002 | B1 |
6587707 | Nehrke et al. | Jul 2003 | B2 |
6621889 | Mostafavi | Sep 2003 | B1 |
6650920 | Schaldach et al. | Nov 2003 | B2 |
6662036 | Cosman | Dec 2003 | B2 |
6687528 | Gupta et al. | Feb 2004 | B2 |
6690965 | Riaziat et al. | Feb 2004 | B1 |
6711431 | Sarin et al. | Mar 2004 | B2 |
6731970 | Schlossbauer et al. | May 2004 | B2 |
6758218 | Anthony | Jul 2004 | B2 |
6771997 | Schaffer | Aug 2004 | B2 |
6794869 | Brittain | Sep 2004 | B2 |
6856827 | Seeley et al. | Feb 2005 | B2 |
6856828 | Cossette et al. | Feb 2005 | B2 |
6876198 | Watanabe et al. | Apr 2005 | B2 |
6888924 | Claus et al. | May 2005 | B2 |
6891374 | Brittain | May 2005 | B2 |
6892089 | Prince et al. | May 2005 | B1 |
6897655 | Brittain et al. | May 2005 | B2 |
6913603 | Knopp et al. | Jul 2005 | B2 |
6937696 | Mostafavi | Aug 2005 | B1 |
6959266 | Mostafavi | Oct 2005 | B1 |
6973202 | Mostafavi | Dec 2005 | B2 |
6980679 | Jeung et al. | Dec 2005 | B2 |
7007699 | Martinelli et al. | Mar 2006 | B2 |
7107091 | Jutras et al. | Sep 2006 | B2 |
7110805 | Machida | Sep 2006 | B2 |
7123758 | Jeung et al. | Oct 2006 | B2 |
7171257 | Thomson | Jan 2007 | B2 |
7173426 | Bulumulla et al. | Feb 2007 | B1 |
7176440 | Cofer et al. | Feb 2007 | B2 |
7191100 | Mostafavi | Mar 2007 | B2 |
7204254 | Riaziat et al. | Apr 2007 | B2 |
7209777 | Saranathan et al. | Apr 2007 | B2 |
7209977 | Acharya et al. | Apr 2007 | B2 |
7260253 | Rahn et al. | Aug 2007 | B2 |
7260426 | Schweikard et al. | Aug 2007 | B2 |
7295007 | Dold | Nov 2007 | B2 |
7313430 | Urquhart et al. | Dec 2007 | B2 |
7327865 | Fu et al. | Feb 2008 | B2 |
7348776 | Aksoy et al. | Mar 2008 | B1 |
7403638 | Jeung et al. | Jul 2008 | B2 |
7494277 | Setala | Feb 2009 | B2 |
7498811 | Macfarlane et al. | Mar 2009 | B2 |
7502413 | Guillaume | Mar 2009 | B2 |
7505805 | Kuroda | Mar 2009 | B2 |
7535411 | Falco | May 2009 | B2 |
7551089 | Sawyer | Jun 2009 | B2 |
7561909 | Pai et al. | Jul 2009 | B1 |
7567697 | Mostafavi | Jul 2009 | B2 |
7573269 | Yao | Aug 2009 | B2 |
7602301 | Stirling et al. | Oct 2009 | B1 |
7603155 | Jensen | Oct 2009 | B2 |
7623623 | Raanes et al. | Nov 2009 | B2 |
7657300 | Hunter et al. | Feb 2010 | B2 |
7657301 | Mate et al. | Feb 2010 | B2 |
7659521 | Pedroni | Feb 2010 | B2 |
7660623 | Hunter et al. | Feb 2010 | B2 |
7668288 | Conwell et al. | Feb 2010 | B2 |
7689263 | Fung et al. | Mar 2010 | B1 |
7702380 | Dean | Apr 2010 | B1 |
7715604 | Sun et al. | May 2010 | B2 |
7742077 | Sablak et al. | Jun 2010 | B2 |
7742621 | Hammoud et al. | Jun 2010 | B2 |
7742804 | Faul et al. | Jun 2010 | B2 |
7744528 | Wallace et al. | Jun 2010 | B2 |
7760908 | Curtner et al. | Jul 2010 | B2 |
7766837 | Pedrizzetti et al. | Aug 2010 | B2 |
7769430 | Mostafavi | Aug 2010 | B2 |
7772569 | Bewersdorf et al. | Aug 2010 | B2 |
7787011 | Zhou et al. | Aug 2010 | B2 |
7787935 | Dumoulin et al. | Aug 2010 | B2 |
7791808 | French et al. | Sep 2010 | B2 |
7792249 | Gertner et al. | Sep 2010 | B2 |
7796154 | Senior et al. | Sep 2010 | B2 |
7798730 | Westerweck | Sep 2010 | B2 |
7801330 | Zhang et al. | Sep 2010 | B2 |
7805987 | Smith | Oct 2010 | B1 |
7806604 | Bazakos et al. | Oct 2010 | B2 |
7817046 | Coveley et al. | Oct 2010 | B2 |
7817824 | Liang et al. | Oct 2010 | B2 |
7819818 | Ghajar | Oct 2010 | B2 |
7833221 | Voegele | Nov 2010 | B2 |
7834846 | Bell | Nov 2010 | B1 |
7835783 | Aletras | Nov 2010 | B1 |
7839551 | Lee et al. | Nov 2010 | B2 |
7840253 | Tremblay et al. | Nov 2010 | B2 |
7844094 | Jeung et al. | Nov 2010 | B2 |
7844320 | Shahidi | Nov 2010 | B2 |
7850526 | Zalewski et al. | Dec 2010 | B2 |
7860301 | Se et al. | Dec 2010 | B2 |
7866818 | Schroeder et al. | Jan 2011 | B2 |
7868282 | Lee et al. | Jan 2011 | B2 |
7878652 | Chen et al. | Feb 2011 | B2 |
7883415 | Larsen et al. | Feb 2011 | B2 |
7889907 | Engelbart et al. | Feb 2011 | B2 |
7894877 | Lewin et al. | Feb 2011 | B2 |
7902825 | Bammer et al. | Mar 2011 | B2 |
7907987 | Dempsey | Mar 2011 | B2 |
7908060 | Basson et al. | Mar 2011 | B2 |
7908233 | Angell et al. | Mar 2011 | B2 |
7911207 | Macfarlane et al. | Mar 2011 | B2 |
7912532 | Schmidt et al. | Mar 2011 | B2 |
7920250 | Robert et al. | Apr 2011 | B2 |
7920911 | Hoshino et al. | Apr 2011 | B2 |
7925066 | Ruohonen et al. | Apr 2011 | B2 |
7925549 | Looney et al. | Apr 2011 | B2 |
7931370 | Prat Bartomeu | Apr 2011 | B2 |
7944354 | Kangas et al. | May 2011 | B2 |
7944454 | Zhou et al. | May 2011 | B2 |
7945304 | Feinberg | May 2011 | B2 |
7946921 | Ofek et al. | May 2011 | B2 |
7962197 | Rioux et al. | Jun 2011 | B2 |
7971999 | Zinser | Jul 2011 | B2 |
7977942 | White | Jul 2011 | B2 |
7978925 | Souchard | Jul 2011 | B1 |
7988288 | Donaldson | Aug 2011 | B2 |
7990365 | Marvit et al. | Aug 2011 | B2 |
8005571 | Sutherland et al. | Aug 2011 | B2 |
8009198 | Alhadef | Aug 2011 | B2 |
8019170 | Wang et al. | Sep 2011 | B2 |
8021231 | Walker et al. | Sep 2011 | B2 |
8022982 | Thorn | Sep 2011 | B2 |
8024026 | Groszmann | Sep 2011 | B2 |
8031909 | Se et al. | Oct 2011 | B2 |
8031933 | Se et al. | Oct 2011 | B2 |
8036425 | Hou | Oct 2011 | B2 |
8041077 | Bell | Oct 2011 | B2 |
8041412 | Glossop et al. | Oct 2011 | B2 |
8048002 | Ghajar | Nov 2011 | B2 |
8049867 | Bridges et al. | Nov 2011 | B2 |
8055020 | Meuter et al. | Nov 2011 | B2 |
8055049 | Stayman et al. | Nov 2011 | B2 |
8060185 | Hunter et al. | Nov 2011 | B2 |
8063929 | Kurtz et al. | Nov 2011 | B2 |
8073197 | Xu et al. | Dec 2011 | B2 |
8077914 | Kaplan | Dec 2011 | B1 |
8085302 | Zhang et al. | Dec 2011 | B2 |
8086026 | Schulz | Dec 2011 | B2 |
8086299 | Adler et al. | Dec 2011 | B2 |
RE43147 | Aviv | Jan 2012 | E |
8094193 | Peterson | Jan 2012 | B2 |
8095203 | Wright et al. | Jan 2012 | B2 |
8095209 | Flaherty | Jan 2012 | B2 |
8098889 | Zhu et al. | Jan 2012 | B2 |
8113991 | Kutliroff | Feb 2012 | B2 |
8116527 | Sabol | Feb 2012 | B2 |
8121356 | Friedman | Feb 2012 | B2 |
8121361 | Ernst et al. | Feb 2012 | B2 |
8134597 | Thorn | Mar 2012 | B2 |
8135201 | Smith et al. | Mar 2012 | B2 |
8139029 | Boillot | Mar 2012 | B2 |
8139896 | Ahiska | Mar 2012 | B1 |
8144118 | Hildreth | Mar 2012 | B2 |
8144148 | El Dokor | Mar 2012 | B2 |
8150063 | Chen | Apr 2012 | B2 |
8150498 | Gielen et al. | Apr 2012 | B2 |
8160304 | Rhoads | Apr 2012 | B2 |
8165844 | Luinge et al. | Apr 2012 | B2 |
8167802 | Baba et al. | May 2012 | B2 |
8172573 | Sonenfeld et al. | May 2012 | B2 |
8175332 | Herrington | May 2012 | B2 |
8179604 | Prada Gomez et al. | May 2012 | B1 |
8180428 | Kaiser et al. | May 2012 | B2 |
8180432 | Sayeh | May 2012 | B2 |
8187097 | Zhang | May 2012 | B1 |
8189869 | Bell | May 2012 | B2 |
8189889 | Pearlstein et al. | May 2012 | B2 |
8189926 | Sharma | May 2012 | B2 |
8190233 | Dempsey | May 2012 | B2 |
8191359 | White et al. | Jun 2012 | B2 |
8194134 | Furukawa | Jun 2012 | B2 |
8195084 | Xiao | Jun 2012 | B2 |
8199983 | Qureshi | Jun 2012 | B2 |
8206219 | Shum | Jun 2012 | B2 |
8207967 | El Dokor | Jun 2012 | B1 |
8208758 | Wang | Jun 2012 | B2 |
8213693 | Li | Jul 2012 | B1 |
8214012 | Zuccolotto et al. | Jul 2012 | B2 |
8214016 | Lavallee et al. | Jul 2012 | B2 |
8216016 | Yamagishi et al. | Jul 2012 | B2 |
8218818 | Cobb | Jul 2012 | B2 |
8218819 | Cobb | Jul 2012 | B2 |
8218825 | Gordon | Jul 2012 | B2 |
8221399 | Amano | Jul 2012 | B2 |
8223147 | El Dokor | Jul 2012 | B1 |
8224423 | Faul | Jul 2012 | B2 |
8226574 | Whillock | Jul 2012 | B2 |
8229163 | Coleman | Jul 2012 | B2 |
8229166 | Teng | Jul 2012 | B2 |
8229184 | Benkley | Jul 2012 | B2 |
8232872 | Zeng | Jul 2012 | B2 |
8235529 | Raffle | Aug 2012 | B1 |
8235530 | Maad | Aug 2012 | B2 |
8241125 | Huges | Aug 2012 | B2 |
8243136 | Aota | Aug 2012 | B2 |
8243269 | Matousek | Aug 2012 | B2 |
8243996 | Steinberg | Aug 2012 | B2 |
8248372 | Saila | Aug 2012 | B2 |
8249691 | Chase et al. | Aug 2012 | B2 |
8253770 | Kurtz | Aug 2012 | B2 |
8253774 | Huitema | Aug 2012 | B2 |
8253778 | Atsushi | Aug 2012 | B2 |
8259109 | El Dokor | Sep 2012 | B2 |
8260036 | Hamza et al. | Sep 2012 | B2 |
8279288 | Son | Oct 2012 | B2 |
8284157 | Markovic | Oct 2012 | B2 |
8284847 | Adermann | Oct 2012 | B2 |
8287373 | Marks et al. | Oct 2012 | B2 |
8289390 | Aggarwal | Oct 2012 | B2 |
8289392 | Senior et al. | Oct 2012 | B2 |
8290208 | Kurtz | Oct 2012 | B2 |
8290229 | Qureshi | Oct 2012 | B2 |
8295573 | Bredno et al. | Oct 2012 | B2 |
8301226 | Csavoy et al. | Oct 2012 | B2 |
8306260 | Zhu | Nov 2012 | B2 |
8306267 | Gossweiler, III | Nov 2012 | B1 |
8306274 | Grycewicz | Nov 2012 | B2 |
8306663 | Wickham | Nov 2012 | B2 |
8310656 | Zalewski | Nov 2012 | B2 |
8310662 | Mehr | Nov 2012 | B2 |
8311611 | Csavoy et al. | Nov 2012 | B2 |
8314854 | Yoon | Nov 2012 | B2 |
8315691 | Sumanaweera et al. | Nov 2012 | B2 |
8316324 | Boillot | Nov 2012 | B2 |
8320621 | McEldowney | Nov 2012 | B2 |
8320709 | Arartani et al. | Nov 2012 | B2 |
8323106 | Zalewski | Dec 2012 | B2 |
8325228 | Mariadoss | Dec 2012 | B2 |
8330811 | Maguire, Jr. | Dec 2012 | B2 |
8330812 | Maguire, Jr. | Dec 2012 | B2 |
8331019 | Cheong | Dec 2012 | B2 |
8334900 | Qu et al. | Dec 2012 | B2 |
8339282 | Noble | Dec 2012 | B2 |
8351651 | Lee | Jan 2013 | B2 |
8368586 | Mohamadi | Feb 2013 | B2 |
8369574 | Hu | Feb 2013 | B2 |
8374393 | Cobb | Feb 2013 | B2 |
8374411 | Ernst et al. | Feb 2013 | B2 |
8374674 | Gertner | Feb 2013 | B2 |
8376226 | Dennard | Feb 2013 | B2 |
8376827 | Cammegh | Feb 2013 | B2 |
8379927 | Taylor | Feb 2013 | B2 |
8380284 | Saranathan et al. | Feb 2013 | B2 |
8386011 | Wieczorek | Feb 2013 | B2 |
8390291 | Macfarlane et al. | Mar 2013 | B2 |
8390729 | Long | Mar 2013 | B2 |
8395620 | El Dokor | Mar 2013 | B2 |
8396654 | Simmons et al. | Mar 2013 | B1 |
8400398 | Schoen | Mar 2013 | B2 |
8400490 | Apostolopoulos | Mar 2013 | B2 |
8405491 | Fong | Mar 2013 | B2 |
8405656 | El Dokor | Mar 2013 | B2 |
8405717 | Kim | Mar 2013 | B2 |
8406845 | Komistek et al. | Mar 2013 | B2 |
8411931 | Zhou | Apr 2013 | B2 |
8427538 | Ahiska | Apr 2013 | B2 |
8428319 | Tsin et al. | Apr 2013 | B2 |
8571293 | Ernst et al. | Oct 2013 | B2 |
8600213 | Mestha et al. | Dec 2013 | B2 |
8615127 | Fitzpatrick | Dec 2013 | B2 |
8617081 | Mestha et al. | Dec 2013 | B2 |
8744154 | Van Den Brink | Jun 2014 | B2 |
8747382 | D'Souza | Jun 2014 | B2 |
8768438 | Mestha et al. | Jul 2014 | B2 |
8790269 | Xu et al. | Jul 2014 | B2 |
8792969 | Bernal et al. | Jul 2014 | B2 |
8805019 | Jeanne et al. | Aug 2014 | B2 |
8848977 | Bammer et al. | Sep 2014 | B2 |
8855384 | Kyal et al. | Oct 2014 | B2 |
8862420 | Ferran et al. | Oct 2014 | B2 |
8873812 | Larlus-Larrondo et al. | Oct 2014 | B2 |
8953847 | Moden | Feb 2015 | B2 |
8971985 | Bernal et al. | Mar 2015 | B2 |
8977347 | Mestha et al. | Mar 2015 | B2 |
8995754 | Wu et al. | Mar 2015 | B2 |
8996094 | Schouenborg et al. | Mar 2015 | B2 |
9020185 | Mestha et al. | Apr 2015 | B2 |
9036877 | Kyal et al. | May 2015 | B2 |
9076212 | Ernst et al. | Jul 2015 | B2 |
9082177 | Sebok | Jul 2015 | B2 |
9084629 | Rosa | Jul 2015 | B1 |
9103897 | Herbst et al. | Aug 2015 | B2 |
9138175 | Ernst et al. | Sep 2015 | B2 |
9173715 | Baumgartner | Nov 2015 | B2 |
9176932 | Baggen et al. | Nov 2015 | B2 |
9194929 | Siegert et al. | Nov 2015 | B2 |
9226691 | Bernal et al. | Jan 2016 | B2 |
9305365 | Lovberg et al. | Apr 2016 | B2 |
9318012 | Johnson | Apr 2016 | B2 |
9336594 | Kyal et al. | May 2016 | B2 |
9395386 | Corder et al. | Jul 2016 | B2 |
9433386 | Mestha et al. | Sep 2016 | B2 |
9436277 | Furst et al. | Sep 2016 | B2 |
9443289 | Xu et al. | Sep 2016 | B2 |
9451926 | Kinahan et al. | Sep 2016 | B2 |
9453898 | Nielsen et al. | Sep 2016 | B2 |
9504426 | Kyal et al. | Nov 2016 | B2 |
9606209 | Ernst et al. | Mar 2017 | B2 |
9607377 | Lovberg et al. | Mar 2017 | B2 |
9629595 | Walker | Apr 2017 | B2 |
9693710 | Mestha et al. | Jul 2017 | B2 |
9717461 | Yu et al. | Aug 2017 | B2 |
9734589 | Yu et al. | Aug 2017 | B2 |
9779502 | Lovberg et al. | Oct 2017 | B1 |
9782141 | Yu et al. | Oct 2017 | B2 |
20020082496 | Kuth | Jun 2002 | A1 |
20020087101 | Barrick et al. | Jul 2002 | A1 |
20020091422 | Greenberg et al. | Jul 2002 | A1 |
20020115931 | Strauss et al. | Aug 2002 | A1 |
20020118373 | Eviatar et al. | Aug 2002 | A1 |
20020180436 | Dale et al. | Dec 2002 | A1 |
20020188194 | Cosman | Dec 2002 | A1 |
20030063292 | Mostafavi | Apr 2003 | A1 |
20030088177 | Totterman et al. | May 2003 | A1 |
20030116166 | Anthony | Jun 2003 | A1 |
20030130574 | Stoyle | Jul 2003 | A1 |
20030195526 | Vilsmeir | Oct 2003 | A1 |
20040071324 | Norris et al. | Apr 2004 | A1 |
20040116804 | Mostafavi | Jun 2004 | A1 |
20040140804 | Polzin et al. | Jul 2004 | A1 |
20040171927 | Lowen et al. | Sep 2004 | A1 |
20050027194 | Adler et al. | Feb 2005 | A1 |
20050054910 | Tremblay et al. | Mar 2005 | A1 |
20050070784 | Komura et al. | Mar 2005 | A1 |
20050105772 | Voronka et al. | May 2005 | A1 |
20050107685 | Seeber | May 2005 | A1 |
20050137475 | Dold et al. | Jun 2005 | A1 |
20050148845 | Dean et al. | Jul 2005 | A1 |
20050148854 | Ito et al. | Jul 2005 | A1 |
20050283068 | Zuccoloto et al. | Dec 2005 | A1 |
20060004281 | Saracen | Jan 2006 | A1 |
20060045310 | Tu et al. | Mar 2006 | A1 |
20060074292 | Thomson et al. | Apr 2006 | A1 |
20060241405 | Leitner et al. | Oct 2006 | A1 |
20070049794 | Glassenberg et al. | Mar 2007 | A1 |
20070093709 | Abernathie | Apr 2007 | A1 |
20070206836 | Yoon | Sep 2007 | A1 |
20070239169 | Plaskos et al. | Oct 2007 | A1 |
20070280508 | Ernst et al. | Dec 2007 | A1 |
20080039713 | Thomson et al. | Feb 2008 | A1 |
20080129290 | Yao | Jun 2008 | A1 |
20080181358 | Van Kampen et al. | Jul 2008 | A1 |
20080183074 | Carls et al. | Jul 2008 | A1 |
20080212835 | Tavor | Sep 2008 | A1 |
20080221442 | Tolowsky et al. | Sep 2008 | A1 |
20080273754 | Hick et al. | Nov 2008 | A1 |
20080287728 | Mostafavi | Nov 2008 | A1 |
20080287780 | Chase et al. | Nov 2008 | A1 |
20080317313 | Goddard et al. | Dec 2008 | A1 |
20090028411 | Pfeuffer | Jan 2009 | A1 |
20090052760 | Smith et al. | Feb 2009 | A1 |
20090185663 | Gaines, Jr. et al. | Jul 2009 | A1 |
20090187112 | Meir et al. | Jul 2009 | A1 |
20090209846 | Bammer | Aug 2009 | A1 |
20090253985 | Shachar et al. | Oct 2009 | A1 |
20090304297 | Adabala et al. | Dec 2009 | A1 |
20090306499 | Van Vorhis et al. | Dec 2009 | A1 |
20100054579 | Okutomi | Mar 2010 | A1 |
20100057059 | Makino | Mar 2010 | A1 |
20100059679 | Albrecht | Mar 2010 | A1 |
20100069742 | Partain et al. | Mar 2010 | A1 |
20100091089 | Cromwell et al. | Apr 2010 | A1 |
20100099981 | Fishel | Apr 2010 | A1 |
20100125191 | Sahin | May 2010 | A1 |
20100137709 | Gardner et al. | Jun 2010 | A1 |
20100148774 | Kamata | Jun 2010 | A1 |
20100149099 | Elias | Jun 2010 | A1 |
20100149315 | Qu | Jun 2010 | A1 |
20100160775 | Pankratov | Jun 2010 | A1 |
20100164862 | Sullivan | Jul 2010 | A1 |
20100165293 | Tanassi et al. | Jul 2010 | A1 |
20100167246 | Ghajar | Jul 2010 | A1 |
20100172567 | Prokoski | Jul 2010 | A1 |
20100177929 | Kurtz | Jul 2010 | A1 |
20100178966 | Suydoux | Jul 2010 | A1 |
20100179390 | Davis | Jul 2010 | A1 |
20100179413 | Kadour et al. | Jul 2010 | A1 |
20100183196 | Fu et al. | Jul 2010 | A1 |
20100191631 | Weidmann | Jul 2010 | A1 |
20100194879 | Pasveer | Aug 2010 | A1 |
20100198067 | Mahfouz | Aug 2010 | A1 |
20100198101 | Song | Aug 2010 | A1 |
20100198112 | Maad | Aug 2010 | A1 |
20100199232 | Mistry | Aug 2010 | A1 |
20100210350 | Walker | Aug 2010 | A9 |
20100214267 | Radivojevic | Aug 2010 | A1 |
20100231511 | Henty | Sep 2010 | A1 |
20100231692 | Perlman | Sep 2010 | A1 |
20100245536 | Huitema | Sep 2010 | A1 |
20100245593 | Kim | Sep 2010 | A1 |
20100251924 | Taylor | Oct 2010 | A1 |
20100253762 | Cheong | Oct 2010 | A1 |
20100268072 | Hall et al. | Oct 2010 | A1 |
20100277571 | Xu | Nov 2010 | A1 |
20100282902 | Rajasingham | Nov 2010 | A1 |
20100283833 | Yeh | Nov 2010 | A1 |
20100284119 | Coakley | Nov 2010 | A1 |
20100289899 | Hendron | Nov 2010 | A1 |
20100290668 | Friedman | Nov 2010 | A1 |
20100292841 | Wickham | Nov 2010 | A1 |
20100295718 | Mohamadi | Nov 2010 | A1 |
20100296701 | Hu | Nov 2010 | A1 |
20100302142 | French | Dec 2010 | A1 |
20100303289 | Polzin | Dec 2010 | A1 |
20100311512 | Lock | Dec 2010 | A1 |
20100321505 | Kokubun | Dec 2010 | A1 |
20100328055 | Fong | Dec 2010 | A1 |
20100328201 | Marbit | Dec 2010 | A1 |
20100328267 | Chen | Dec 2010 | A1 |
20100330912 | Saila | Dec 2010 | A1 |
20110001699 | Jacobsen | Jan 2011 | A1 |
20110006991 | Elias | Jan 2011 | A1 |
20110007939 | Teng | Jan 2011 | A1 |
20110007946 | Liang | Jan 2011 | A1 |
20110008759 | Usui | Jan 2011 | A1 |
20110015521 | Faul | Jan 2011 | A1 |
20110019001 | Rhoads | Jan 2011 | A1 |
20110025853 | Richardson | Feb 2011 | A1 |
20110038520 | Yui | Feb 2011 | A1 |
20110043631 | Marman | Feb 2011 | A1 |
20110043759 | Bushinsky | Feb 2011 | A1 |
20110050562 | Schoen | Mar 2011 | A1 |
20110050569 | Marvit | Mar 2011 | A1 |
20110050947 | Marman | Mar 2011 | A1 |
20110052002 | Cobb | Mar 2011 | A1 |
20110052003 | Cobb | Mar 2011 | A1 |
20110052015 | Saund | Mar 2011 | A1 |
20110054870 | Dariush | Mar 2011 | A1 |
20110057816 | Noble | Mar 2011 | A1 |
20110058020 | Dieckmann | Mar 2011 | A1 |
20110064290 | Punithakaumar | Mar 2011 | A1 |
20110069207 | Steinberg | Mar 2011 | A1 |
20110074675 | Shiming | Mar 2011 | A1 |
20110081000 | Gertner | Apr 2011 | A1 |
20110081043 | Sabol | Apr 2011 | A1 |
20110085704 | Han | Apr 2011 | A1 |
20110087091 | Olson | Apr 2011 | A1 |
20110092781 | Gertner | Apr 2011 | A1 |
20110102549 | Takahashi | May 2011 | A1 |
20110105883 | Lake et al. | May 2011 | A1 |
20110105893 | Akins et al. | May 2011 | A1 |
20110115793 | Grycewicz | May 2011 | A1 |
20110115892 | Fan | May 2011 | A1 |
20110116683 | Kramer et al. | May 2011 | A1 |
20110117528 | Marciello et al. | May 2011 | A1 |
20110118032 | Zalewski | May 2011 | A1 |
20110133917 | Zeng | Jun 2011 | A1 |
20110142411 | Camp | Jun 2011 | A1 |
20110150271 | Lee | Jun 2011 | A1 |
20110157168 | Bennett | Jun 2011 | A1 |
20110157358 | Bell | Jun 2011 | A1 |
20110157370 | Livesey | Jun 2011 | A1 |
20110160569 | Cohen et al. | Jun 2011 | A1 |
20110172060 | Morales | Jul 2011 | A1 |
20110172521 | Zdeblick et al. | Jul 2011 | A1 |
20110175801 | Markovic | Jul 2011 | A1 |
20110175809 | Markovic | Jul 2011 | A1 |
20110175810 | Markovic | Jul 2011 | A1 |
20110176723 | Ali et al. | Jul 2011 | A1 |
20110180695 | Li | Jul 2011 | A1 |
20110181893 | MacFarlane | Jul 2011 | A1 |
20110182472 | Hansen | Jul 2011 | A1 |
20110187640 | Jacobsen | Aug 2011 | A1 |
20110193939 | Vassigh | Aug 2011 | A1 |
20110199461 | Horio | Aug 2011 | A1 |
20110201916 | Duyn et al. | Aug 2011 | A1 |
20110201939 | Hubschman et al. | Aug 2011 | A1 |
20110202306 | Eng | Aug 2011 | A1 |
20110205358 | Aota | Aug 2011 | A1 |
20110207089 | Lagettie | Aug 2011 | A1 |
20110208437 | Teicher | Aug 2011 | A1 |
20110216002 | Weising | Sep 2011 | A1 |
20110216180 | Pasini | Sep 2011 | A1 |
20110221770 | Kruglick | Sep 2011 | A1 |
20110229862 | Parikh | Sep 2011 | A1 |
20110230755 | MacFarlane et al. | Sep 2011 | A1 |
20110234807 | Jones | Sep 2011 | A1 |
20110234834 | Sugimoto | Sep 2011 | A1 |
20110235855 | Smith | Sep 2011 | A1 |
20110237933 | Cohen | Sep 2011 | A1 |
20110242134 | Miller | Oct 2011 | A1 |
20110244939 | Cammegh | Oct 2011 | A1 |
20110250929 | Lin | Oct 2011 | A1 |
20110251478 | Wieczorek | Oct 2011 | A1 |
20110255845 | Kikuchi | Oct 2011 | A1 |
20110257566 | Burdea | Oct 2011 | A1 |
20110260965 | Kim | Oct 2011 | A1 |
20110262002 | Lee | Oct 2011 | A1 |
20110267427 | Goh | Nov 2011 | A1 |
20110267456 | Adermann | Nov 2011 | A1 |
20110275957 | Bhandari | Nov 2011 | A1 |
20110276396 | Rathod | Nov 2011 | A1 |
20110279663 | Fan | Nov 2011 | A1 |
20110285622 | Marti | Nov 2011 | A1 |
20110286010 | Kusik et al. | Nov 2011 | A1 |
20110291925 | Isarel | Dec 2011 | A1 |
20110293143 | Narayanan et al. | Dec 2011 | A1 |
20110293146 | Grycewicz | Dec 2011 | A1 |
20110298708 | Hsu | Dec 2011 | A1 |
20110298824 | Lee | Dec 2011 | A1 |
20110300994 | Verkaaik | Dec 2011 | A1 |
20110301449 | Maurer, Jr. | Dec 2011 | A1 |
20110301934 | Tardis | Dec 2011 | A1 |
20110303214 | Welle | Dec 2011 | A1 |
20110304541 | Dalal | Dec 2011 | A1 |
20110304650 | Canpillo | Dec 2011 | A1 |
20110304706 | Border et al. | Dec 2011 | A1 |
20110306867 | Gopinathan | Dec 2011 | A1 |
20110310220 | McEldowney | Dec 2011 | A1 |
20110310226 | McEldowney | Dec 2011 | A1 |
20110316994 | Lemchen | Dec 2011 | A1 |
20110317877 | Bell | Dec 2011 | A1 |
20120002112 | Huang | Jan 2012 | A1 |
20120004791 | Buelthoff | Jan 2012 | A1 |
20120007839 | Tsao et al. | Jan 2012 | A1 |
20120019645 | Maltz | Jan 2012 | A1 |
20120020524 | Ishikawa | Jan 2012 | A1 |
20120021806 | Maltz | Jan 2012 | A1 |
20120027226 | Desenberg | Feb 2012 | A1 |
20120029345 | Mahfouz et al. | Feb 2012 | A1 |
20120032882 | Schlachta | Feb 2012 | A1 |
20120033083 | Horvinger | Feb 2012 | A1 |
20120035462 | Maurer, Jr. et al. | Feb 2012 | A1 |
20120039505 | Bastide et al. | Feb 2012 | A1 |
20120044363 | Lu | Feb 2012 | A1 |
20120045091 | Kaganovich | Feb 2012 | A1 |
20120049453 | Morichau-Beauchant et al. | Mar 2012 | A1 |
20120051588 | McEldowney | Mar 2012 | A1 |
20120051664 | Gopalakrishnan et al. | Mar 2012 | A1 |
20120052949 | Weitzner | Mar 2012 | A1 |
20120056982 | Katz | Mar 2012 | A1 |
20120057640 | Shi | Mar 2012 | A1 |
20120065492 | Gertner et al. | Mar 2012 | A1 |
20120065494 | Gertner et al. | Mar 2012 | A1 |
20120072041 | Miller | Mar 2012 | A1 |
20120075166 | Marti | Mar 2012 | A1 |
20120075177 | Jacobsen | Mar 2012 | A1 |
20120076369 | Abramovich | Mar 2012 | A1 |
20120081504 | Ng | Apr 2012 | A1 |
20120083314 | Ng | Apr 2012 | A1 |
20120083960 | Zhu | Apr 2012 | A1 |
20120086778 | Lee | Apr 2012 | A1 |
20120086809 | Lee | Apr 2012 | A1 |
20120092445 | McDowell | Apr 2012 | A1 |
20120092502 | Knasel | Apr 2012 | A1 |
20120093481 | McDowell | Apr 2012 | A1 |
20120098938 | Jin | Apr 2012 | A1 |
20120101388 | Tripathi | Apr 2012 | A1 |
20120105573 | Apostolopoulos | May 2012 | A1 |
20120106814 | Gleason et al. | May 2012 | A1 |
20120108909 | Slobounov et al. | May 2012 | A1 |
20120113140 | Hilliges | May 2012 | A1 |
20120113223 | Hilliges | May 2012 | A1 |
20120116202 | Bangera | May 2012 | A1 |
20120119999 | Harris | May 2012 | A1 |
20120120072 | Se | May 2012 | A1 |
20120120237 | Trepess | May 2012 | A1 |
20120120243 | Chien | May 2012 | A1 |
20120120277 | Tsai | May 2012 | A1 |
20120121124 | Bammer | May 2012 | A1 |
20120124604 | Small | May 2012 | A1 |
20120127319 | Rao | May 2012 | A1 |
20120133616 | Nishihara | May 2012 | A1 |
20120133889 | Bergt | May 2012 | A1 |
20120143029 | Silverstein | Jun 2012 | A1 |
20120143212 | Madhani | Jun 2012 | A1 |
20120147167 | Mason | Jun 2012 | A1 |
20120154272 | Hildreth | Jun 2012 | A1 |
20120154511 | Hsu | Jun 2012 | A1 |
20120154536 | Stoker | Jun 2012 | A1 |
20120154579 | Hanpapur | Jun 2012 | A1 |
20120156661 | Smith | Jun 2012 | A1 |
20120158197 | Hinman | Jun 2012 | A1 |
20120162378 | El Dokor et al. | Jun 2012 | A1 |
20120165964 | Flaks | Jun 2012 | A1 |
20120167143 | Longet | Jun 2012 | A1 |
20120169841 | Chemali | Jul 2012 | A1 |
20120176314 | Jeon | Jul 2012 | A1 |
20120184371 | Shum | Jul 2012 | A1 |
20120188237 | Han | Jul 2012 | A1 |
20120188371 | Chen | Jul 2012 | A1 |
20120194422 | El Dokor | Aug 2012 | A1 |
20120194517 | Izadi et al. | Aug 2012 | A1 |
20120194561 | Grossinger | Aug 2012 | A1 |
20120195466 | Teng | Aug 2012 | A1 |
20120196660 | El Dokor et al. | Aug 2012 | A1 |
20120197135 | Slatkine | Aug 2012 | A1 |
20120200676 | Huitema | Aug 2012 | A1 |
20120201428 | Joshi et al. | Aug 2012 | A1 |
20120206604 | Jones | Aug 2012 | A1 |
20120212594 | Barns | Aug 2012 | A1 |
20120218407 | Chien | Aug 2012 | A1 |
20120218421 | Chien | Aug 2012 | A1 |
20120220233 | Teague | Aug 2012 | A1 |
20120224666 | Speller | Sep 2012 | A1 |
20120224743 | Rodriguez | Sep 2012 | A1 |
20120225718 | Zhang | Sep 2012 | A1 |
20120229643 | Chidanand | Sep 2012 | A1 |
20120229651 | Takizawa | Sep 2012 | A1 |
20120230561 | Qureshi | Sep 2012 | A1 |
20120235896 | Jacobsen | Sep 2012 | A1 |
20120238337 | French | Sep 2012 | A1 |
20120242816 | Cruz | Sep 2012 | A1 |
20120249741 | Maciocci | Oct 2012 | A1 |
20120253201 | Reinhold | Oct 2012 | A1 |
20120253241 | Levital et al. | Oct 2012 | A1 |
20120262540 | Rondinelli | Oct 2012 | A1 |
20120262558 | Boger | Oct 2012 | A1 |
20120262583 | Bernal | Oct 2012 | A1 |
20120268124 | Herbst et al. | Oct 2012 | A1 |
20120275649 | Cobb | Nov 2012 | A1 |
20120276995 | Lansdale | Nov 2012 | A1 |
20120277001 | Lansdale | Nov 2012 | A1 |
20120281093 | Fong | Nov 2012 | A1 |
20120281873 | Brown | Nov 2012 | A1 |
20120288142 | Gossweiler, III | Nov 2012 | A1 |
20120288852 | Willson | Nov 2012 | A1 |
20120289334 | Mikhailov | Nov 2012 | A9 |
20120289822 | Shachar et al. | Nov 2012 | A1 |
20120293412 | El Dokor | Nov 2012 | A1 |
20120293506 | Vertucci | Nov 2012 | A1 |
20120293663 | Liu | Nov 2012 | A1 |
20120294511 | Datta | Nov 2012 | A1 |
20120300961 | Moeller | Nov 2012 | A1 |
20120303839 | Jackson | Nov 2012 | A1 |
20120304126 | Lavigne | Nov 2012 | A1 |
20120307075 | Margalit | Dec 2012 | A1 |
20120307207 | Abraham | Dec 2012 | A1 |
20120314066 | Lee | Dec 2012 | A1 |
20120315016 | Fung | Dec 2012 | A1 |
20120319946 | El Dokor | Dec 2012 | A1 |
20120319989 | Argiro | Dec 2012 | A1 |
20120320219 | David | Dec 2012 | A1 |
20120326966 | Rauber | Dec 2012 | A1 |
20120326976 | Markovic | Dec 2012 | A1 |
20120326979 | Geisert | Dec 2012 | A1 |
20120327241 | Howe | Dec 2012 | A1 |
20120327246 | Senior et al. | Dec 2012 | A1 |
20130002866 | Hanpapur | Jan 2013 | A1 |
20130002879 | Weber | Jan 2013 | A1 |
20130002900 | Gossweiler, III | Jan 2013 | A1 |
20130009865 | Valik | Jan 2013 | A1 |
20130010071 | Valik | Jan 2013 | A1 |
20130013452 | Dennard | Jan 2013 | A1 |
20130016009 | Godfrey | Jan 2013 | A1 |
20130016876 | Wooley | Jan 2013 | A1 |
20130021434 | Ahiska | Jan 2013 | A1 |
20130021578 | Chen | Jan 2013 | A1 |
20130024819 | Rieffel | Jan 2013 | A1 |
20130030283 | Vortman et al. | Jan 2013 | A1 |
20130033640 | Lee | Feb 2013 | A1 |
20130033700 | Hallil | Feb 2013 | A1 |
20130035590 | Ma et al. | Feb 2013 | A1 |
20130035612 | Mason | Feb 2013 | A1 |
20130040720 | Cammegh | Feb 2013 | A1 |
20130041368 | Cunninghan | Feb 2013 | A1 |
20130053683 | Hwang et al. | Feb 2013 | A1 |
20130057702 | Chavan | Mar 2013 | A1 |
20130064426 | Watkins, Jr. | Mar 2013 | A1 |
20130064427 | Picard | Mar 2013 | A1 |
20130065517 | Svensson | Mar 2013 | A1 |
20130066448 | Alonso | Mar 2013 | A1 |
20130066526 | Mondragon | Mar 2013 | A1 |
20130069773 | Li | Mar 2013 | A1 |
20130070201 | Shahidi | Mar 2013 | A1 |
20130070257 | Wong | Mar 2013 | A1 |
20130072787 | Wallace et al. | Mar 2013 | A1 |
20130076863 | Rappel | Mar 2013 | A1 |
20130076944 | Kosaka | Mar 2013 | A1 |
20130077823 | Mestha | Mar 2013 | A1 |
20130079033 | Gupta | Mar 2013 | A1 |
20130084980 | Hammontree | Apr 2013 | A1 |
20130088584 | Malhas | Apr 2013 | A1 |
20130093866 | Ohlhues et al. | Apr 2013 | A1 |
20130096439 | Lee | Apr 2013 | A1 |
20130102879 | MacLaren et al. | Apr 2013 | A1 |
20130102893 | Vollmer | Apr 2013 | A1 |
20130108979 | Daon | May 2013 | A1 |
20130113791 | Isaacs et al. | May 2013 | A1 |
20130211421 | Abovitz et al. | Aug 2013 | A1 |
20130281818 | Vija et al. | Oct 2013 | A1 |
20140055563 | Jessop | Feb 2014 | A1 |
20140073908 | Biber | Mar 2014 | A1 |
20140088410 | Wu | Mar 2014 | A1 |
20140148685 | Liu et al. | May 2014 | A1 |
20140159721 | Grodzki | Jun 2014 | A1 |
20140171784 | Ooi et al. | Jun 2014 | A1 |
20140378816 | Oh et al. | Dec 2014 | A1 |
20150085072 | Yan | Mar 2015 | A1 |
20150094597 | Mestha et al. | Apr 2015 | A1 |
20150094606 | Mestha et al. | Apr 2015 | A1 |
20150212182 | Nielsen et al. | Jul 2015 | A1 |
20150245787 | Kyal et al. | Sep 2015 | A1 |
20150257661 | Mestha et al. | Sep 2015 | A1 |
20150265187 | Bernal et al. | Sep 2015 | A1 |
20150265220 | Ernst et al. | Sep 2015 | A1 |
20150297120 | Son et al. | Oct 2015 | A1 |
20150297314 | Fowler | Oct 2015 | A1 |
20150316635 | Stehning et al. | Nov 2015 | A1 |
20150323637 | Beck et al. | Nov 2015 | A1 |
20150331078 | Speck et al. | Nov 2015 | A1 |
20150359464 | Oleson | Dec 2015 | A1 |
20160000383 | Lee et al. | Jan 2016 | A1 |
20160000411 | Raju et al. | Jan 2016 | A1 |
20160045112 | Weissler et al. | Feb 2016 | A1 |
20160091592 | Beall et al. | Mar 2016 | A1 |
20160166205 | Ernst et al. | Jun 2016 | A1 |
20160198965 | Mestha et al. | Jul 2016 | A1 |
20160228005 | Bammer et al. | Aug 2016 | A1 |
20160249984 | Janssen | Sep 2016 | A1 |
20160256713 | Saunders et al. | Sep 2016 | A1 |
20160262663 | MacLaren et al. | Sep 2016 | A1 |
20160287080 | Olesen et al. | Oct 2016 | A1 |
20160310229 | Bammer et al. | Oct 2016 | A1 |
20160313432 | Feiweier et al. | Oct 2016 | A1 |
20170032538 | Ernst | Feb 2017 | A1 |
20170038449 | Voigt et al. | Feb 2017 | A1 |
20170143271 | Gustafsson et al. | May 2017 | A1 |
20170303859 | Robertson et al. | Oct 2017 | A1 |
20170319143 | Yu et al. | Nov 2017 | A1 |
20170345145 | Nempont et al. | Nov 2017 | A1 |
Number | Date | Country |
---|---|---|
100563551 | Dec 2009 | CN |
105392423 | Mar 2016 | CN |
106572810 | Apr 2017 | CN |
106714681 | May 2017 | CN |
29519078 | Mar 1996 | DE |
102004024470 | Dec 2005 | DE |
0904733 | Mar 1991 | EP |
1319368 | Jun 2003 | EP |
1354564 | Oct 2003 | EP |
1524626 | Apr 2005 | EP |
2023812 | Feb 2009 | EP |
2515139 | Oct 2012 | EP |
2747641 | Jul 2014 | EP |
2948056 | Dec 2015 | EP |
2950714 | Dec 2015 | EP |
3157422 | Apr 2017 | EP |
3188660 | Jul 2017 | EP |
03023838 | May 1991 | JP |
WO 9617258 | Jun 1996 | WO |
WO 9938449 | Aug 1999 | WO |
WO 0072039 | Nov 2000 | WO |
WO 03003796 | Jan 2003 | WO |
WO 2004023783 | Mar 2004 | WO |
WO 2005077293 | Aug 2005 | WO |
WO 2007025301 | Mar 2007 | WO |
WO 2007085241 | Aug 2007 | WO |
WO 2007136745 | Nov 2007 | WO |
WO 2009101566 | Aug 2009 | WO |
WO 2009129457 | Oct 2009 | WO |
WO 2010066824 | Jun 2010 | WO |
WO 2011047467 | Apr 2011 | WO |
WO 2011113441 | Sep 2011 | WO |
WO 2012046202 | Apr 2012 | WO |
WO 2013032933 | Mar 2013 | WO |
WO 2014005178 | Jan 2014 | WO |
WO 2014116868 | Jul 2014 | WO |
WO 2014120734 | Aug 2014 | WO |
WO 2015022684 | Feb 2015 | WO |
WO 2015042138 | Mar 2015 | WO |
WO 2015092593 | Jun 2015 | WO |
WO 2015148391 | Oct 2015 | WO |
WO 2016014718 | Jan 2016 | WO |
WO2017091479 | Jun 2017 | WO |
WO2017189427 | Nov 2017 | WO |
Entry |
---|
Ashouri, H., L. et al., Unobtrusive Estimation of Cardiac Contractility and Stroke Volume Changes Using Ballistocardiogram Measurements on a High Bandwidth Force Plate, Sensors 2016, 16, 787; doi:10.3390/s16060787. |
Communication pursuant to Article 94(3) EPC for application No. 14743670.3, which is an EP application related to the present application, dated Feb. 6, 2018. |
Extended Europen Search Report for application No. 14743670.3 which is a EP application related to the present application, dated Aug. 17, 2017. |
Extended Europen Search Report for application No. 15769296.3 which is a EP application related to the present application, dated Dec. 22, 2017. |
Extended European Search Report for application No. 15824707.2 which is a EP application related to the present appliation, dated Apr. 16, 2018. |
Gordon, J. W. Certain molar movements of the human body produced by the circulation of the blood. J. Anat. Physiol. 11, 533-536 (1877). |
Herbst et al., “Reproduction of Motion Artifacts for Performance Analysis of Prospective Motion Correction in MRI”, Magnetic Resonance in Medicine., vol. 71, No. 1, p. 182-190 (Feb. 25, 2013). |
Kim, Chang-Sei et al. “Ballistocardiogram: Mechanism and Potential for Unobtrusive Cardiovascular Health Monitoring”, Scientific Reports, Aug. 9, 2016. |
MacLaren et al., “Prospective Motion Correction in Brain Imaging: A Review” Online Magnetic Resonance in Medicine, vol. 69, No. 3, pp. 621-636 (Mar. 1, 2013. |
Tarvainen, M.P. et al., “An advanced de-trending method with application to HRV analysis,” IEEE Trans. Biomed. Eng., vol. 49, No. 2, pp. 172-175, Feb. 2002. |
Wang, Ching-Cheng, “Extrinsic Calibration of a Vision Sensor Mounted on a Robot”, IEEE Transactions on Robotics and Automation, 8(2):161-175 (Apr. 1992). |
Aksoy et al., “Hybrind Prospective and Retrospective Head Motion Correction to Mitigate Cross-Calibration Errors”, NIH Publication, Nov. 2012. |
Aksoy et al., “Real-Time Optical Motion Correction for Diffusion Tensor Imaging, Magnetic Resonance in Medicine” (Mar. 22, 2011) 66 366-378. |
Alusi et al., “A study of tremor in multiple sclerosis”, Brain, vol. 124:720-730 (2001). |
Andrews et al., “Prospective Motion Correction for Magnetic Resonance Spectroscopy Using Single Camera Retro-Grate Reflector Optical Tracking, Journal of Magnetic Resonance Imaging” (Feb. 2011) 33(2): 498-504. |
Angeles et al., “The Online Solution of the Hand-Eye Problem”, IEEE Transactions on Robotics and Automation, 16(6): 720-731 (Dec. 2000). |
Anishenko et al., “A Motion Correction System for Brain Tomography Based on Biologically Motivated Models.” 7th IEEE International Conference on Cybernetic Intelligent Systems, dated Sep. 9, 2008, in 9 pages. |
Armstrong et al., RGR-6D: Low-cost, high-accuracy measurement of 6-DOF Pose from a Single Image. Publication date unknown. |
Armstrong et al., “RGR-3D: Simple, cheap detection of 6-DOF pose for tele-operation, and robot programming and calibration”, In Proc. 2002 Int. Conf. on Robotics and Automation, IEEE, Washington (May 2002). |
Bandettini, Peter A., et al., “Processing Strategies for Time-Course Data Sets in Functional MRI of the Human Breain”, Magnetic Resonance in Medicine 30: 161-173 (1993). |
Barmet et al, Spatiotemporal Magnetic Field Monitoring for MR, Magnetic Resonance in Medicine (Feb. 1, 2008) 60: 187-197. |
Bartels, LW, et al., “Endovascular interventional magnetic resonance imaging”, Physics in Medicine and Biology 48: R37-R64 (2003). |
Benchoff, Brian, “Extremely Precise Positional Tracking”, https://hackaday.com/2013/10/10/extremely-precise-positional-tracking/, printed on Sep. 16, 2017, in 7 pages. |
Carranza-Herrezuelo et al, “Motion estimation of tagged cardiac magnetric resonance images using variational techniques” Elsevier, Computerized Medical Imaging and Graphics 34 (2010), pp. 514-522. |
Chou, Jack C. K., et al., “Finding the Position and Orientation of a Sensor on a Robot Manipulator Using Quaternions”, The International Journal of Robotics Research, 10(3): 240-254 (Jun. 1991). |
Cofaru et al “Improved Newton-Raphson digital image correlation method for full-field displacement and strain calculation,” Department of Materials Science and Engineering, Ghent University ST-Pietersnieuwstraat, Nov. 20, 2010. |
Ernst et al., “A Novel Phase and Frequency Navigator for Proton Magnetic Resonance Spectroscopy Using Water-Suppression Cycling, Magnetic Resonance in Medicine” (Jan. 2011) 65(1): 13-7. |
Eviatar et al., “Real time head motion correction for functional MRI”, In: Proceedings of the International Society for Magnetic Resonance in Medicine (1999) 269. |
Forbes, Kristen P. N., et al., “Propeller MRI: Clinical Testing of a Novel Technique for Quantification and Compensation of Head Motion”, Journal of Magnetic Resonance Imaging 14: 215-222 (2001). |
Fulton et al., “Correction for Head Movements in Positron Emission Tomography Using an Optical Motion-Tracking System”, IEEE Transactions on Nuclear Science, vol. 49(1):116-123 (Feb. 2002). |
Glover, Gary H., et al., “Self-Navigated Spiral fMRI: Interleaved versus Single-shot”, Magnetic Resonance in Medicine 39: 361-368 (1998). |
Gumus et al., “Elimination of DWI signal dropouts using blipped gradients for dynamic restoration of gradient moment”, ISMRM 20th Annual Meeting & Exhibition, May 7, 2012. |
Herbst et al., “Preventing Signal Dropouts in DWI Using Continous Prospective Motion Correction”, Proc. Intl. Soc. Mag. Reson. Med. 19 (May 2011) 170. |
Herbst et al., “Prospective Motion Correction With Continuous Gradient Updates in Diffusion Weighted Imaging, Magnetic Resonance in Medicine” (2012) 67:326-338. |
Hoff et al., “Analysis of Head Pose Accuracy in Augmented Reality”, IEEE Transactions on Visualization and Computer Graphics 6, No. 4 (Oct.-Dec. 2000): 319-334. |
Horn, Berthold K. P., “Closed-form solution of absolute orientation using unit quaternions”, Journal of the Optical Society of America, vol. 4, p. 629-642 (Apr. 1987). |
International Preliminary Report on Patentability for Application No. PCT/US2015/022041, dated Oct. 6, 2016, in 8 pages. |
International Preliminary Report on Patentability for Application No. PCT/US2007/011899, dated Jun. 8, 2008, in 13 pages. |
International Search Report and Written Opinion for Application No. PCT/US2007/011899, dated Nov. 14, 2007. |
International Search Report and Written Opinion for Application No. PCT/US2014/012806, dated May 15, 2014, in 15 pages. |
International Search Report and Written Opinion for Application No. PCT/US2015/041615, dated Oct. 29, 2015, in 13 pages. |
International Preliminary Report on Patentability for Application No. PCT/US2014/013546, dated Aug. 4, 2015, in 9 pages. |
International Search Report and Written Opinion for Application No. PCT/US2015/022041, dated Jun. 29, 2015, in 9 pages. |
Jochen Triesch, et al.“Democratic Integration: Self-Organized Integration of Adaptive Cues”, Neural Computation., vol. 13, No. 9, dated Sep. 1, 2001, pp. 2049-2074. |
Josefsson et al. “A flexible high-precision video system for digital recording of motor acts through lightweight reflect markers”, Computer Methods and Programs in Biomedicine, vol. 49:111-129 (1996). |
Katsuki, et al., “Design of an Artificial Mark to Determine 3D Pose by Monocular Vision”, 2003 IEEE International Conference on Robotics and Automation (Cat. No. 03CH37422), Sep. 14-19, 2003, pp. 995-1000 vol. 1. |
Kiebel et al., “MRI and PET coregistration—a cross validation of statistical parametric mapping and automated image registration”, Neuroimage 5(4):271-279 (1997). |
Kiruluta et al., “Predictive Head Movement Tracking Using a Kalman Filter”, IEEE Trans. on Systems, Man, and Cybernetics—Part B: Cybernetics, 27(2):326-331 (Apr. 1997). |
Lerner, “Motion correction in fmri images”, Technion-Israel Institute of Technology, Faculty of Computer Science ( Feb. 2006). |
Maclaren et al., “Combined Prospective and Retrospective Motion Correction to Relax Navigator Requirements”, Magnetic Resonance in Medicine (Feb. 11, 2011) 65:1724-1732. |
MacLaren et al., “Navigator Accuracy Requirements for Prospective Motion Correction”, Magnetic Resonance in Medicine (Jan. 2010) 63(1): 162-70. |
MacLaren, “Prospective Motion Correction in MRI Using Optical Tracking Tape”, Book of Abstracts, ESMRMB (2009). |
Maclaren et al., “Measurement and correction of microscopic head motion during magnetic resonance imaging of the brain”, PLOS One, vol. 7(11):1-9 (2012). |
McVeigh et al., “Real-time, Interactive MRI for Cardiovascular Interventions”, Academic Radiology, 12(9): 1121-1127 (2005). |
Nehrke et al., “Prospective Correction of Affine Motion for Arbitrary MR Sequences on a Clinical Scanner”, Magnetic Resonance in Medicine (Jun. 28, 2005) 54:1130-1138. |
Norris et al., “Online motion correction for diffusion-weighted imaging using navigator echoes: application to RARE imaging without sensitivity loss”, Magnetic Resonance in Medicine, vol. 45:729-733 (2001). |
Olesen et al., “Structured Light 3D Tracking System for Measuring Motions in PET Brain Imaging”, Proceedings of SPIE, the International Society for Optical Engineering (ISSN: 0277-786X), vol. 7625:76250X (2010). |
Olesen et al., “Motion Tracking in Narrow Spaces: A Structured Light Approach”, Lecture Notes in Computer Science (ISSN: 0302-9743)vol. 6363:253-260 (2010). |
Olesen et al., “Motion Tracking for Medical Imaging: A Nonvisible Structured Light Tracking Approach”, IEEE Transactions on Medical Imaging, vol. 31(1), Jan. 2012. |
Ooi et al., “Prospective Real-Time Correction for Arbitrary Head Motion Using Active Markers”, Magnetic Resonance in Medicine (Apr. 15, 2009) 62(4): 943-54. |
Orchard et al., “MRI Reconstruction using real-time motion tracking: A simulation study”, Signals, Systems and Computers, 42nd Annual Conference IEEE, Piscataway, NJ, USA (Oct. 26, 2008). |
Park, Frank C. and Martin, Bryan J., “Robot Sensor Calibration: Solving AX-XB on the Euclidean Group”, IEEE Transaction on Robotics and Automation, 10(5): 717-721 (Oct. 1994). |
PCT Search Report from the International Searching Authority, dated Feb. 28, 2013, in 16 pages, regarding International Application No. PCT/US2012/052349. |
Qin et al., “Prospective Head-Movement Correction for High-Resolution MRI Using an In-Bore Optical Tracking System”, Magnetic Resonance in Medicine (Apr. 13, 2009) 62: 924-934. |
Schulz et al., “First Embedded In-Bore System for Fast Optical Prospective Head Motion-Correction in MRI”, Proceedings of the 28th Annual Scientific Meeting of the ESMRMB (Oct. 8, 2011) 369. |
Shiu et al., “Calibration of Wrist-Mounted Robotic Sensors by Solving Homogeneous Transform Equations of the Form AX=XB”, IEEE Transactions on Robotics and Automation, 5(1): 16-29 (Feb. 1989). |
Speck, et al., “Prospective real-time slice-by-slice Motion Correction for fMRI in Freely Moving Subjects”, Magnetic Resonance Materials in Physics, Biology and Medicine., 19(2), 55-61, published May 9, 2006. |
Tremblay et al., “Retrospective Coregistration of Functional Magnetic Resonance Imaging Data using External monitoring”, Magnetic Resonance in Medicine 53:141-149 (2005). |
Tsai et al., “A New Technique for Fully Autonomous and Efficient 3D Robotics Hand/Eye Calibration”, IEEE Transaction on Robotics and Automation, 5(3): 345-358 (Jun. 1989). |
Ward et al., “Prospective Multiaxial Motion Correction for fMRI”, Magnetic Resonance in Medicine 43:459-469 (2000). |
Welch at al., “Spherical Navigator Echoes for Full 3D Rigid Body Motion Measurement in MRI”, Magnetic Resonance in Medicine 47:32-41 (2002). |
Wilm et al., “Accurate and Simple Calibration of DLP Projector Systems”, Proceedings of SPIE, the International Society for Optical Engineering (ISSN: 0277-786X), vol. 8979 (2014). |
Wilm et al., “Correction of Motion Artifacts for Real-Time Structured Light”, R.R. Paulsen and K.S. Pedersen (Eds.): SCIA 2015, LNCS 9127, pp. 142-151 (2015). |
Yeo, et al. Motion correction in fMRI by mapping slice-to-volume with concurrent field-inhomogeneity correction:, International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 752-760 (2004). |
Zaitsev, M., et al., “Prospective Real-Time Slice-by-Slice 3D Motion Correction for EPI Using an External Optical Motion Tracking System”, Proc.Intl.Soc.Mag.Reson.Med.11:517(2004). |
Zeitsev et al., “Magnetic resonance imaging of freely moving objects: Prospective real-time motion correction using an external optical motion tracking system”, NeuroImage 31 (Jan. 29, 2006) 1038-1050. |
Gaul, Scott, Quiet Mind Cafe, https://www.youtube.com/watch?v=7wFX9Wn70eM. |
https://www.innoveremedical.com/. |
Ming-Zhere Pho, D.J. McDuff, and R.W. Picard, “Advancements in Noncontact, Multiparameter Physiological Measurements Using a Webcam”, IEEE Transactions on Biomedical Engineering, vol. 58, No. 1, Jan. 2011. |
Rostaminia, A. Mayberry, D. Ganesan, B. Marlin, and J. Gummeson, “Low-power Sensing of Fatigue and Drowsiness Measures on a Computational Eyeglass”, Proc ACM Interact Mob Wearable Ubiquitous Technol.; 1(2): 23; doi: 10.1145/3090088, Jun. 2017. |
Number | Date | Country | |
---|---|---|---|
20180122069 A1 | May 2018 | US |
Number | Date | Country | |
---|---|---|---|
61849338 | Jan 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15428495 | Feb 2017 | US |
Child | 15680860 | US | |
Parent | 14837554 | Aug 2015 | US |
Child | 15428495 | US | |
Parent | 13831115 | Mar 2013 | US |
Child | 14837554 | US |