System and methods for the reduction and elimination of image artifacts in the calibration of x-ray imagers

Information

  • Patent Grant
  • 6370224
  • Patent Number
    6,370,224
  • Date Filed
    Monday, June 12, 2000
    24 years ago
  • Date Issued
    Tuesday, April 9, 2002
    22 years ago
Abstract
Image processing operations are used to improve images that include visual artifacts generated by calibration markers used in intrinsic calibration of an x-ray image. Artifacts introduced by opaque or semi-transparent calibration markers may be completely or partially removed from the image. More particularly, artifacts caused by opaque calibration markers are removed by changing the pixels corresponding to the projections of the calibration markers to blend in with pixels surrounding the calibration markers. Artifacts may also be generated with semi-transparent calibration markers. These artifacts may be eliminated from the image, while leaving intact the underlying image, by subtracting a constant offset from each marker projection.
Description




FIELD OF THE INVENTION




This invention relates generally to x-ray imaging systems, and more specifically, to the calibration of x-ray imaging systems.




BACKGROUND OF THE INVENTION




Modem diagnostic medicine has benefitted significantly from radiology, which is the use of radiation, such as x-rays, to generate images of internal body structures. In general, to create an x-ray image, x-ray beams are passed through the body and absorbed, in varying amounts, by tissues in the body. An x-ray image is created based on the relative differences in the transmitted x-ray intensities.





FIG. 1A

is a diagram illustrating a fluoroscopic C-arm x-ray imaging device. Imaging device


100


includes C-arm


103


attached to mobile base


102


. X-ray source


105


is located at one end of C-arm


103


and x-ray receiving section


106


is located at the other end of C-arm


103


. Receiving section


106


generates an image representing the intensities of received x-rays. Typically, receiving section


106


comprises an image intensifier that converts the x-rays to visible light and a charge coupled device (CCD) video camera that converts the visible light to digital images.




Images taken at the mobile base


102


are transmitted to control unit


120


for analysis. In particular, control unit


120


typically provides facilities for displaying, saving, digitally manipulating, or printing a hard copy of the received images. Control unit


120


additionally includes controls for controlling base unit


102


.




In operation, the patient is positioned in area


110


, between the x-ray source


105


and the x-ray receiving section


106


. In response to an operator's command input at control unit


120


, x-rays emanating from source


105


pass through patient area


110


and into receiving section


106


, which generates a two-dimensional image of the patient.




Although each individual image taken by base unit


102


is a two-dimensional image, techniques are known in the art through which multiple two-dimensional images taken from multiple perspectives can be used to infer the three-dimensional location of an anatomical projection. To change image perspective, C-arm


103


rotates as shown, for example, in FIG.


1


B. By taking multiple two-dimensional images of point


124


, but from different perspectives, the three-dimensional position of point


124


may be determined.




Raw images generated by receiving section


106


tend to suffer from undesirable distortion caused by a number of factors, including inherent image distortion in the image intensifier and external electromagnetic fields. An example of a true and a distorted image is shown in FIG.


2


. Checkerboard


202


represents the true image of a checkerboard shaped object placed in image taking area


110


. The image taken by receiving section


106


, however, suffers significant distortion, as illustrated by distorted image


204


.




Intrinsic calibration, which is the process of correcting image distortion in a received image and learning the projective geometry of the imager, involves placing “calibration markers” in the path of the x-ray, where a calibration marker is an object opaque to x-rays. The calibration markers are rigidly arranged in predetermined patterns in one or more planes in the path of the x-rays and are visible in the recorded images.




Because the true relative position of the calibration markers in the recorded images is known, control unit


120


is able to calculate an amount of distortion at each pixel in the image (where a pixel is a single point in the image). Accordingly, control unit


120


can digitally compensate for the distortion in the image and generate a distortion-free, or at least a distortion improved image. A more detailed explanation of a method for performing intrinsic calibration is described in U.S. Pat. No. 5,442,674 to Picard et al, the contents of which are incorporated by reference herein.




A notable disadvantage in the conventional method of compensating for image distortion, as described above, is that although there is significantly less distortion in the image, projections of the calibration markers are present in the image. This is undesirable, as the projections of the markers may occlude important portions of the patient's anatomy and/or act as a visual distraction that prevents the clinician from concentrating on important features of the image.




There is, therefore, a need in the art to improve the intrinsic calibration process.




SUMMARY OF THE INVENTION




Objects and advantages of the invention will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention. The objects and advantages of the invention will be realized and attained by means of the elements and combinations particularly pointed out in the appended claims.




To achieve the objects and in accordance with the purpose of the invention, as embodied and broadly described herein, a first aspect consistent with the present invention includes a method for causing a computer processor to perform the steps of: storing a digital image representing anatomy of a patient, the digital image including representations of calibration markers that at least partially occlude portions of the patient anatomy; and performing image processing operations on the digital image to de-emphasize the representations of the calibration markers.




Additional aspects of the present invention, related to the first aspect, are directed to a computer readable medium and a computer system.




A second aspect of the present invention is directed to a medical imaging system comprising a combination of elements, including: an x-ray source for generating x-rays; semi-transparent calibration markers positioned in a path of the x-rays; and an x-ray receiving device for receiving the generated x-rays and deriving a digital image representing objects through which the generated x-rays have passed, the digital image including representations of the calibration markers. A processor is coupled to the x-ray receiving device and performs image processing operations on the digital image, the digital processing operations removing distortion from the image by performing intrinsic calibration on the image based on projections of the semi-transparent calibration markers in the image.




A third aspect of the present invention is directed to a method of creating an image of an object. The method comprises the steps of: transmitting x-rays in a path including a target object and calibration markers arranged in a predetermined pattern; receiving the transmitted x-rays; deriving a digital image representing the object and the calibration markers; and de-emphasizing the representations of the calibration markers in the digital image.




Additional aspects of the present invention, related to the third aspect, are directed to a computer readable medium and a computer system.











BRIEF DESCRIPTION OF THE DRAWINGS




The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate several embodiments consistent with this invention and, together with the description, help explain the principles of the invention. In the drawings,





FIGS. 1A and 1B

are diagrams illustrating a fluoroscopic C-arm x-ray imaging device;





FIG. 2

is a diagram illustrating a true and a distorted image taken with a fluoroscopic C-arm x-ray imaging device;





FIG. 3

is a block diagram illustrating a control unit of an imaging device;





FIG. 4

is an image illustrating two-dimensional circular artifacts projected from spherical calibration markers;





FIG. 5

is a flow chart of image processing methods consistent with the present invention for reducing the artifacts caused by calibration markers;





FIG. 6

is an image of an expanded view of a calibration marker projection;





FIG. 7A

is an image illustrating two-dimensional circular artifacts projected from spherical calibration markers;





FIGS. 7B and 7C

are versions of the image shown in

FIG. 7A

after application of methods consistent with the present invention;





FIG. 8

is an image of a calibration marker projection divided into four regions;





FIG. 9

is a flow chart of image processing methods consistent with the present invention for eliminating artifacts caused by semi-transparent calibration markers; and





FIG. 10

is a flow chart of image processing methods consistent with a second aspect of the present invention for eliminating artifacts caused by semi-transparent calibration markers.











DETAILED DESCRIPTION




As described herein, image processing operations are used to improve images that include visual artifacts generated by calibration markers used in intrinsic calibration of the image. Artifacts introduced by opaque or semi-transparent calibration markers may be completely or partially removed from the image.




Referring to the accompanying drawings, detailed description of embodiments consistent with the present invention will now be described.




System Overview




Methods consistent with the present invention may be implemented on images taken with an x-ray imaging device in which intrinsic image calibration is implemented. One such imaging device is the “Series9600 Mobile Digital Imaging System,” from OEC Medical Systems, Inc., of Salt Lake City, Utah. The “Series9600 Mobile Digital Imaging System” is structurally similar to imaging system


100


. Alternatively, methods consistent with the present invention may be implemented on images at a computer system not associated with the imaging device.





FIG. 3

is a block diagram illustrating control unit


120


in more detail. Communications between base unit


102


and control unit


120


are performed via transmission medium


302


, which may be, for example, a radio or cable link. Digital images may be received from base unit


102


and commands transmitted to base unit


102


. Control unit


120


may include an additional external connection, such as network connection


315


. Through network connection


315


, data, such as images stored in memory


304


, may be transmitted to additional computing resources, such as computer


305


.




Control unit


120


further comprises a computer processor


303


and a memory


304


coupled to processor


303


through a bus


306


. Processor


303


fetches computer instructions from memory


304


and executes those instructions. Processor


303


also (1) reads data from and writes data to memory


304


, (2) sends data and control signals through bus


306


to one or more peripheral output devices


312


and


313


; and (3) receives data and control signals through bus


306


from input device(s)


314


.




Memory


304


can include any type of computer memory, including, without limitation, random access memory (RAM), read-only memory (ROM), and storage devices that include storage media such as magnetic and/or optical disks. Memory


304


includes a computer process


310


that processor


303


executes. A computer process in this description is a collection of computer instructions and data that collectively define a task performed by control unit


120


.




Input device


314


is used by an operator to enter commands to control unit


120


. The commands may be executed directly by control unit


120


or transmitted to base unit


102


. Input device


314


may be, for example, a keyboard, a pointing device such as a mouse, or a combination thereof Output devices


3




12


and


313


are preferably a display and a printer, respectively. Display


312


is typically used to exhibit images taken by base unit


102


and printer


3




13


is used to create hard copies of the images.




In operation, images stored in memory


304


may be processed by processor


303


to perform various image processing operations. For example, processor


303


may perform intrinsic calibration on an image or generate the location of a three-dimensional point from a series of two-dimensional images. Consistent with the present invention, processing section


303


also removes artifacts caused by calibration markers used in the intrinsic calibration process. Computer


305


, instead of processing section


303


, may alternatively perform image processing operations consistent with the present invention.




The above-described architecture of control unit


120


is exemplary only. One of ordinary skill in the art will recognize that many modifications could be made to the described architecture and still achieve the described functionality.




Intrinsic Calibration




As previously discussed, intrinsic calibration uses calibration markers placed at fixed, predetermined positions in the x-ray imaging path to either obtain an image transformation that removes distortion from the original image generated by receiving section


106


or to learn the projective geometry of the imager (i.e., to discern how a pixel in the image projects into three-dimensional space). Typically, each calibration marker is a three-dimensional shape that appears in the image as a two-dimensional object, although calibration markers can also be constructed using thin films that are essentially two-dimensional in nature. Many possible shapes, such as spheres and cylindrical rods can be used to implement the calibration markers. Spheres appear in the two-dimensional image as circles and cylindrical rods appear as lines. Throughout this disclosure, spherical calibration markers are illustrated, although one of ordinary skill in the art will recognize that calibration markers of any shape could be used.




A typical C-arm calibration target contains a large set of calibration markers (e.g., 25+) with the markers positioned over one or more depth planes.




Artifact Reduction




Consistent with a first aspect of the present invention, artifacts introduced into an x-ray image by radio-opaque markers are reduced.





FIG. 4

is an image having two-dimensional circular artifacts projected from spherical calibration markers. Two different calibration marker patterns were used to generate image


400


. Large circles


402


represent a first spherical pattern of the calibration markers and smaller circles


404


represent a second spherical pattern of the calibration markers. Preferably, each spherical pattern is rigidly fixed in a separate plane traversed by the x-rays. As shown, markers


402


and


404


were opaque to the x-rays used to take the image, thus the two-dimensional projection of the markers appears as solid black circles.





FIG. 5

is a flow chart of image processing methods consistent with the present invention for reducing the artifacts caused by calibration markers, such as artifacts


402


and


404


of image


400


. The methods illustrated in

FIG. 5

may be performed after a received image has been intrinsically calibrated to reduce image distortion.




For each digitized image that is to be processed, processor


303


begins by identifying the calibration marker projections in the image (step


502


). As the shape, general pixel intensity, and relative position of the markers are known a priori, detection of the marker projections is a straightforward image processing operation well within the skill of one of ordinary skill in the art, and therefore will not be described further. Identification of the marker projections classifies the image pixels into those corresponding to the marker projections and those corresponding to anatomy or other non-marker objects.




For each marker projection in the image, processor


303


identifies pixels surrounding the identified marker artifacts, (steps


503


and


504


), and reads the values (intensities) of the surrounding pixels (step


506


). Finally, the processor changes the pixel values of the marker projections to values based on that of the pixels surrounding the marker (step


508


). The modified pixels of the marker projections tend to blend in more smoothly with the actual image, thereby reducing the visual distraction caused by the marker artifacts.




Because the new marker projection values are only estimates of the intensities of the true underlying image data, it is possible that the new marker projection values will not accurately reflect the true image and will mislead the clinician. Accordingly, processor


303


may modify the marker pixels so that they are visible to the clinician but yet are still visibly less distracting than the original marker projections (optional step


509


). Preferably, this step is achieved by supplementing the new marker projection values with a small constant offset (e.g., 5% of the maximum pixel value), thus causing the new marker projections to be visibly distinct but not visually distracting.





FIG. 6

is an image of an expanded view of one of calibration marker projections


404


. Small squares


601


highlight pixels defined as surrounding pixels in step


504


. As shown, the “surrounding pixels” are not necessarily limited to just those pixels that immediately border marker projection


404


, but may include additional neighboring pixels. For example, the surrounding pixels may include all the pixels with a certain radius of the outer border of the marker projection (e.g., a radius of five pixels) or all the non-marker pixels within a square aligned with the center of the marker projection.




There are many possible approaches to appropriately modifying the pixel values within the marker projections as performed in step


508


. The best approach used by processor


303


in any particular situation may vary depending on the circumstances, and may be selectable by the user or selected automatically by processor


303


. Exemplary ones of these approaches will now be discussed.




In a first method, processor


303


simply calculates the average intensity value of surrounding pixels


601


(i.e., the sum of the surrounding pixel values divided by the number of surrounding pixels in the sample) and sets each of the pixels in marker projection


604


to that intensity value.

FIG. 7A

is an image, similar to image


400


, having two-dimensional circular artifacts projected from spherical calibration markers.

FIG. 7B

is an image after application of the averaging method applied to the image of FIG.


7


A.

FIG. 7C

is the image shown in

FIG. 7B

after application of the averaging method and the addition of a small constant offset chosen to make the marker projection visibly distinct but not visibly distracting.




In a second method, processor


303


divides marker projection


604


into multiple regions and separately calculate average intensity values of surrounding pixels for each region. An example of a marker projection divided into four regions (quadrants) is shown in FIG.


8


. Marker projection


804


is surrounded by pixels


801


. Processor


303


separately calculates the average value of the surround pixels in each of quadrants


810


-


813


and then sets the marker projection pixels in that quadrant to the calculated value.




Other methods, in addition to the average and multiple region averaging methods discussed above, may also be used to calculate new marker projection pixel values. In particular, a second general class of approaches for determining underlying marker projection intensity values uses estimators that optimize a criterion function in order to derive the pixel intensities. This class of methods involves maximum likelihood estimators such as the Expectation Maximization (EM) algorithm, neural networks, fuzzy systems, and other methods which estimate a set of parameters (i.e., the new marker projection intensity values) by maximizing a criterion function. For example, an EM algorithm could estimate underlying pixel intensities in a statistically optimal sense given the measured image and the current marker location. Any of these approaches may incorporate statistical models of the image that mathematically describe the expected image structure (e.g., measures of image texture or image variation, measures of feature orientation, etc.).




Artifact Elimination




Consistent with a second aspect of the present invention, artifacts introduced into an x-ray image by semi-transparent markers may be substantially eliminated while preserving much of the true underlying image.




The semi-transparent calibration markers should be opaque enough so that they are visible enough to be automatically identified in the x-ray images, and transparent enough so that the features underlying the markers (i.e., the features along the x-ray projection path passing through the markers) will also influence the image intensity. When these conditions are satisfied, the marker projections may be completely eliminated while preserving the underlying image features by subtracting offset values from the detected marker projections.




The semi-transparent calibration markers may be made from a material such as a thin layer of copper (e.g., 0.5-2 mm thick) or a solid ceramic layer.





FIG. 9

is a flow chart of image processing methods consistent with the present invention for substantially eliminating artifacts caused by semi-transparent calibration markers.




Essentially, artifact elimination is performed by subtracting a pre-measured offset from each pixel in the marker projections. The appropriate offset value to subtract is initially determined by processor


303


by acquiring an intensity image of a calibration marker projection in which no anatomy or other material is visible (step


901


). That is, a preselected calibration marker is placed in an x-ray imaging path in which the x-rays pass only through the calibration marker. If all the pixels corresponding to the preselected calibration marker are of the same intensity, then the offset is simply that intensity value. If the intensity values of the pixels corresponding to the preselected calibration marker projection vary, whether by design or because of consistent variance in the calibration marker's material composition, then a separate offset value may be saved for each pixel.




Once the offset for a particular image has been determined, processor


303


proceeds with eliminating the artifacts by identifying the calibration marker projections, (step


902


), and, for each identified projection, (step


903


), subtracting the acquired offset(s) from the pixels of the projection (step


904


). Ideally, steps


901


-


904


will completely eliminate the artifacts from the image while leaving the true underlying image (e.g., the patient anatomy). Practically, image noise may prevent a perfect result. In these situations, processor


303


refines the result by applying an estimator function, such as the EM algorithm described above, to further improve the result (optional step


905


). The input to the EM algorithm is the output of step


904


, while the output is a refined estimate of the true underlying pixel intensities.





FIG. 10

is a flow chart of image processing methods consistent with a second aspect of the present invention for substantially eliminating artifacts caused by semi-transparent calibration markers. The process illustrated in

FIG. 10

is similar to that illustrated in

FIG. 9

, except that instead of subtracting offset intensities from the pixels of the marker projections, an estimator optimizing a criterion function, such as the EM function, is used to modify the marker projections. More specifically, processor


303


eliminates, or substantially eliminates, the artifacts by identifying the calibration marker projections, (step


1002


), and, for each identified projection, (step


1003


), applies the estimator function (step


1005


).




As described in this disclosure, artifacts present in x-ray images are de-emphasized. More particularly, artifacts may either be reduced in prominence (artifact reduction) or eliminated all together (artifact elimination), thereby improving the image presented to the clinician.




While there has been illustrated and, described what are at present considered to be preferred embodiments and methods of the present invention, it will be understood by those skilled in the art that various changes and modifications may be made, and equivalents may be substituted for elements thereof without departing from the true scope of the invention. For example, although described in the context of a medical imaging system using x-rays, methods consistent with the present invention can be performed on any digitized input image.




In addition, many modifications may be made to adapt a particular element, technique or implementation to the teachings of the present invention without departing from the central scope of the invention. Therefore, it is intended that this invention not be limited to the particular embodiments and methods disclosed herein, but that the invention include all embodiments falling within the scope of the appended claims.



Claims
  • 1. A medical imaging system comprising:an x-ray source for generating x-rays; calibration markers positioned in a path of the x-rays; an x-ray receiving device for receiving the generated x-rays and deriving a digital image representing objects through which the generated x-rays have passed, the digital image including representations of the calibration markers; and a processor coupled to the x-ray receiving device for performing image processing operations on the digital image, the image processing operations modifying the visual presence of the representations of the calibration markers, based on characteristics of an image area surrounding the calibration markers.
  • 2. The system of claim 1, wherein the x-ray receiving device further comprises an image intensifier and a charge coupled device (CCD) array for deriving the digital image from the x-rays.
  • 3. The system of claim 1, wherein the processor modifies the visual presence of the representations of the calibration markers using a maximum likelihood estimator that maximizes a criterion function.
  • 4. The system of claim 3, wherein the maximum likelihood estimator is an Expectation Maximizing algorithm.
  • 5. The system of claim 1, wherein the x-ray receiving device provides a representation of calibration markers that are opaque or semi-transparent to the x-rays.
  • 6. The system of claim 5, wherein the processor further includes means for identifying pixels in the digital image that surround the calibration markers and means for modifying the representations of the calibration markers based on the values of the pixels that surround the representations of the calibration markers.
  • 7. The system of claim 6, wherein modifying the representation of the calibration markers further includes using an average value of the pixels that surround the representations of the calibration markers.
  • 8. The system of claim 6, wherein modifying the representations of the calibration markers further includes separately calculating average intensity values for separate regions of the digital image surrounding the representation of the calibration markers.
  • 9. The system of claim 6, wherein the modifying means includes an estimator using the value of the pixels surrounding the representation of the calibration markers.
  • 10. The system of claim 6, wherein identifying the representations of the calibration markers further includes identifying the calibration markers by classifying a plurality of image pixels into image pixels corresponding to the calibration markers and image pixels corresponding to anatomy or other objects.
  • 11. The system of claim 6, wherein identifying the pixels in the digital image includes identifying pixels that are not directly adjacent to the calibration markers.
  • 12. The system of claim 6, wherein the processor further includes means for adding a constant offset to the modified representations of the calibration markers to enhance the visibility of the representations.
  • 13. The system of claim 6, wherein modifying the representations of the calibration markers further includes adding an offset to the representations of the calibration markers that have been modified to enhance the visibility of the representations.
  • 14. The system of claim 13, wherein the processor further includes means for acquiring an offset value corresponding to an intensity of one of the calibration marker representations in which the x-rays have not traversed the objects.
  • 15. The system of claim 14, wherein the processor further includes means for subtracting the offset value from each of the calibration marker representations.
  • 16. The system of claim 15, wherein the processor further includes means for refining the subtracted versions of the calibration marker representations by using an estimator algorithm that optimizes a criterion function.
  • 17. A method of creating an image of a target comprising:transmitting x-rays at the target and calibration markers; collecting the x-rays; producing a digital image representing the target and calibration markers; and processing the digital image to modify the representation of the calibration markers, based on characteristics of an image area surrounding the calibration markers.
  • 18. The method of claim 17, wherein the identifying of the representations of the calibration markers includes identifying the calibration markers by classifying a plurality of image pixels into image pixels corresponding to the calibration markers and image pixels corresponding to anatomy or other objects.
  • 19. The method of claim 17, wherein the transmitting further includes transmitting the x-rays through semi-transparent calibration markers arranged in a predetermined pattern.
  • 20. The method of claim 19, wherein processing includes eliminating the representations of the calibration markers by subtracting an offset from pixels in the digital image that comprise the calibration marker representations.
  • 21. The method of claim 17, wherein the transmitting further includes transmitting the x-rays at calibration markers that are opaque or semi-transparent to the x-rays.
  • 22. The method of claim 21, wherein the processing includes identifying pixels in the digital image that surround the calibration markers and modifying the representations of the calibration markers based on the values of the pixels that surround the representations of the calibration markers.
  • 23. The method of claim 22, wherein the identifying further includes identifying pixels that are not directly adjacent to the calibration markers.
  • 24. The method of claim 22, wherein the modifying of the representation of the calibration markers further includes using an average value of the pixels that surround the representation of the calibration markers.
  • 25. The method of claim 22, further comprising adding an offset to the modified representations of the calibration markers to enhance the visibility of the representations.
  • 26. The method of claim 22, wherein modifying of the representation of the calibration markers includes separately calculating average intensity values for separate regions of the digital image surrounding the representation of the calibration markers.
  • 27. The method of claim 22, wherein the modifying of the representation of the calibration markers further includes using an estimator utilizing the value of the pixels surrounding the representation of the calibration markers.
  • 28. A computer readable medium containing computer instructions for causing a processor to perform processing comprising:storing a digital image representing anatomy of a patient, the digital image including representations of calibration markers that at least partially occlude portions of the patient anatomy; and performing image processing operations on the digital image to modify the representations of the calibration markers, based on characteristics of an image area surrounding the calibration markers.
  • 29. The computer readable medium of claim 28, wherein the instructions further include instructions for eliminating the representations of the calibration markers by subtracting an offset from pixels in the digital image that comprise the calibration marker representations.
  • 30. The computer readable medium of claim 28, wherein the instructions further include instructions for modifying the visual presence of the representations of the calibration markers using a maximum likelihood estimator that maximizes a criterion function.
  • 31. The computer readable medium of claim 30 wherein the maximum likelihood estimator is an Expectation Maximizing algorithm.
  • 32. The computer readable medium of claim 28, wherein the instructions further include instructions for identifying pixels in the digital image that surround the calibration markers and modifying the representations of the calibration markers based on the values of the pixels that surround the representations of the calibration markers.
  • 33. The computer readable medium of claim 32, wherein the instructions further include instructions for identifying pixels in the digital image that are not directly adjacent to the calibration markers.
  • 34. The computer readable medium of claim 32, wherein the instructions for modifying further include using an average value of the pixels that surround the representation of the calibration markers.
  • 35. The computer readable medium of claim 32, wherein the instructions further include separately calculating average intensity values for separate regions of the digital image surrounding the representation of the calibration markers.
  • 36. The computer readable medium of claim 32, wherein the instructions further include using an estimator to calculate the value of the pixels surrounding the representation of the calibration markers.
  • 37. The computer readable medium of claim 32, wherein the instructions include identifying the calibration markers by classifying a plurality of image pixels into image pixels corresponding to the calibration markers and image pixels corresponding to anatomy or other objects.
  • 38. The computer readable medium of claim 32, further including instructions for adding an offset to the modified representations of the calibration markers to enhance the visibility of the representations.
  • 39. A computer system comprising:a first computer memory storing a digital image representing anatomy of a patient, the digital image including representations of calibration markers that at least partially occlude portions of the patient anatomy; a second memory storing instruction for performing image processing operations on the digital image to modify the representations of the calibration markers based on characteristics of an image area surrounding the calibration markers; and a processor coupled to the first and second memory for executing the instructions stored in the second memory.
  • 40. The computer system of claim 39, wherein the processor modifies the visual presence of the representations of the calibration markers using a maximum likelihood estimator that maximizes a criterion function.
  • 41. The computer system of claim 40, wherein the maximum likelihood estimator is an Expectation Maximizing algorithm.
  • 42. The computer system of claim 39, wherein the processor further includes means for identifying pixels in the digital image that surround the calibration markers and means for modifying the representations of the calibration markers based on the values of the pixels that surround the representations of the calibration markers.
  • 43. The computer system of claim 42, wherein the processor further includes means for adding a constant offset to the modified representations of the calibration markers to enhance the visibility of the representations.
  • 44. The computer system of claim 42, wherein the processor further includes means for modifying the representations of the calibration markers by identifying pixels in the digital image that are not directly adjacent to the calibration markers.
  • 45. The computer system of claim 42, wherein the processor further includes means for modifying the representations of the calibration markers by using an average value of the pixels that surround the representation of the calibration markers.
  • 46. The computer system of claim 42, wherein the processor further includes means for modifying the representations of the calibration markers by separately calculating average intensity values for separate regions of the digital image surrounding the representation of the calibration markers.
  • 47. The computer system of claim 42, wherein the processor further includes means for modifying the representations of the calibration markers using an estimator utilizing the value of the pixels surrounding the representation of the calibration markers.
  • 48. The computer system of claim 42, wherein the processor further includes means for modifying the representations of the calibration markers by classifying a plurality of image pixels into image pixels corresponding to the calibration markers and image pixels corresponding to anatomy or other objects.
Parent Case Info

This is a continuation of application Ser. No. 09/106,109, filed Jun. 29, 1998 now U.S. Pat. No. 6,118,845, which is incorporated herein by reference.

US Referenced Citations (193)
Number Name Date Kind
1576781 Philips Mar 1926 A
1735726 Bornhardt Nov 1929 A
2407845 Nemeyer Sep 1946 A
2650588 Drew Sep 1953 A
2697433 Zehnder Dec 1954 A
3016899 Stenvall Jan 1962 A
3017887 Heyer Jan 1962 A
3061936 Dobbeleer Nov 1962 A
3073310 Mocarski Jan 1963 A
3294083 Alderson Dec 1966 A
3367326 Frazier Feb 1968 A
3577160 White May 1971 A
3702935 Carey et al. Nov 1972 A
3704707 Halloran Dec 1972 A
3847157 Caillouette et al. Nov 1974 A
3868565 Kuipers Feb 1975 A
3941127 Froning Mar 1976 A
4037592 Kronner Jul 1977 A
4054881 Raab Oct 1977 A
4068556 Foley Jan 1978 A
4071456 McGee et al. Jan 1978 A
4117337 Staats Sep 1978 A
4202349 Jones May 1980 A
4228779 Anichkov et al. Oct 1980 A
4259725 Andrews et al. Mar 1981 A
4341220 Perry Jul 1982 A
4358856 Stivender et al. Nov 1982 A
4360028 Barbier et al. Nov 1982 A
4403321 DiMarco Sep 1983 A
4418422 Richter et al. Nov 1983 A
4465069 Barbier et al. Aug 1984 A
4485815 Amplatz Dec 1984 A
4506676 Duska Mar 1985 A
4543959 Sepponen Oct 1985 A
4583538 Onik et al. Apr 1986 A
4621628 Bludermann Nov 1986 A
4625718 Olerud et al. Dec 1986 A
4651732 Frederick Mar 1987 A
4653509 Oloff et al. Mar 1987 A
4706665 Gouda Nov 1987 A
4722056 Roberts et al. Jan 1988 A
4722336 Kim et al. Feb 1988 A
4727565 Ericson Feb 1988 A
4737921 Goldwasser et al. Apr 1988 A
4750487 Zanetti Jun 1988 A
4764944 Finlayson Aug 1988 A
4771787 Wurster et al. Sep 1988 A
4791934 Brunnett Dec 1988 A
4797907 Anderton Jan 1989 A
4803976 Frigg et al. Feb 1989 A
4821206 Arora Apr 1989 A
4821213 Cline et al. Apr 1989 A
4821731 Martinelli et al. Apr 1989 A
4829373 Leberl et al. May 1989 A
4923459 Nambu May 1990 A
4945914 Allen Aug 1990 A
4977655 Martinelli Dec 1990 A
4989608 Ratner Feb 1991 A
4991579 Allen Feb 1991 A
5005578 Greer et al. Apr 1991 A
5013317 Cole et al. May 1991 A
5016639 Allen May 1991 A
5027818 Bova et al. Jul 1991 A
5030222 Calandruccio et al. Jul 1991 A
5031203 Trecha Jul 1991 A
5054492 Scribner et al. Oct 1991 A
5059789 Salcudean Oct 1991 A
5070454 Griffith Dec 1991 A
5078142 Siczek et al. Jan 1992 A
5079699 Tuy et al. Jan 1992 A
5086401 Glassman et al. Feb 1992 A
5094241 Allen Mar 1992 A
5097839 Allen Mar 1992 A
5107843 Aarnio et al. Apr 1992 A
5119817 Allen Jun 1992 A
5142930 Allen et al. Sep 1992 A
5154179 Ratner Oct 1992 A
5178164 Allen Jan 1993 A
5178621 Cook et al. Jan 1993 A
5186174 Schlondorff et al. Feb 1993 A
5189690 Samuel Feb 1993 A
5193106 DeSena Mar 1993 A
5197476 Nowacki et al. Mar 1993 A
5198877 Schulz Mar 1993 A
5211164 Allen May 1993 A
5212720 Landi et al. May 1993 A
5219351 Teubner et al. Jun 1993 A
5222499 Allen et al. Jun 1993 A
5229935 Yamagishi et al. Jul 1993 A
5230338 Allen et al. Jul 1993 A
5230623 Guthrie et al. Jul 1993 A
5249581 Horbal et al. Oct 1993 A
5251127 Raab Oct 1993 A
5257629 Kitney et al. Nov 1993 A
5276927 Day Jan 1994 A
5295483 Nowacki et al. Mar 1994 A
5299254 Dancer et al. Mar 1994 A
5299288 Glassman et al. Mar 1994 A
5305203 Raab Apr 1994 A
5309913 Kormos et al. May 1994 A
5315630 Sturm et al. May 1994 A
5320111 Livingston Jun 1994 A
5333168 Fernandes et al. Jul 1994 A
5368030 Zinreich et al. Nov 1994 A
5369678 Chiu et al. Nov 1994 A
5383454 Bucholz Jan 1995 A
5389101 Heilbrun et al. Feb 1995 A
5391199 Ben-Haim Feb 1995 A
5394457 Leibinger et al. Feb 1995 A
5397329 Allen Mar 1995 A
5398684 Hardy Mar 1995 A
5399146 Nowacki et al. Mar 1995 A
5400384 Fernandes et al. Mar 1995 A
5402801 Taylor Apr 1995 A
5408409 Glassman et al. Apr 1995 A
5426683 O'Farrell, Jr. et al. Jun 1995 A
5426687 Goodall et al. Jun 1995 A
5427097 Depp Jun 1995 A
RE35025 Anderton Aug 1995 E
5442674 Picard et al. Aug 1995 A
5446548 Gerig et al. Aug 1995 A
5447154 Cinquin et al. Sep 1995 A
5478341 Cook et al. Dec 1995 A
5478343 Ritter Dec 1995 A
5483961 Kelly et al. Jan 1996 A
5490196 Rudich et al. Feb 1996 A
5494034 Schlöndorff et al. Feb 1996 A
5503416 Aoki et al. Apr 1996 A
5515160 Schulz et al. May 1996 A
5517990 Kalfas et al. May 1996 A
5531227 Schneider Jul 1996 A
5531520 Grimson et al. Jul 1996 A
5551429 Fitzpatrick et al. Sep 1996 A
5551431 Wells, III et al. Sep 1996 A
5558091 Acker et al. Sep 1996 A
5568809 Ben-haim Oct 1996 A
5572999 Funda et al. Nov 1996 A
5583909 Hanover Dec 1996 A
5588430 Bova et al. Dec 1996 A
5590215 Allen Dec 1996 A
5596228 Anderton et al. Jan 1997 A
5603318 Heilbrun et al. Feb 1997 A
5611025 Lorensen et al. Mar 1997 A
5617462 Spratt Apr 1997 A
5617857 Chader et al. Apr 1997 A
5619261 Anderton Apr 1997 A
5622170 Schulz Apr 1997 A
5627873 Hanover et al. May 1997 A
5628315 Vilsmeier et al. May 1997 A
5630431 Taylor May 1997 A
5638819 Manwaring et al. Jun 1997 A
5642395 Anderton et al. Jun 1997 A
5647361 Damadian Jul 1997 A
5651047 Moorman et al. Jul 1997 A
5662111 Cosman Sep 1997 A
5676673 Ferre et al. Oct 1997 A
5682886 Delp et al. Nov 1997 A
5682890 Kormos et al. Nov 1997 A
5690108 Chakeres Nov 1997 A
5695500 Taylor et al. Dec 1997 A
5695501 Carol et al. Dec 1997 A
5711299 Manwaring et al. Jan 1998 A
5727553 Saad Mar 1998 A
5748767 Raab May 1998 A
5749362 Funda et al. May 1998 A
5755725 Druais May 1998 A
RE35816 Schulz Jun 1998 E
5769789 Wang et al. Jun 1998 A
5769861 Vilsmeier Jun 1998 A
5772594 Barrick Jun 1998 A
5795294 Luber et al. Aug 1998 A
5799055 Peshkin et al. Aug 1998 A
5799099 Wang et al. Aug 1998 A
5800535 Howard, III Sep 1998 A
5802719 O'Farrell, Jr. et al. Sep 1998 A
5823958 Truppe Oct 1998 A
5828725 Levinson Oct 1998 A
5833608 Acker Nov 1998 A
5834759 Glossop Nov 1998 A
5836954 Heilbrun et al. Nov 1998 A
5848967 Cosman Dec 1998 A
5851183 Bucholz Dec 1998 A
5868675 Henrion et al. Feb 1999 A
5871445 Bucholz Feb 1999 A
5871487 Warner et al. Feb 1999 A
5891034 Bucholz Apr 1999 A
5891157 Day et al. Apr 1999 A
5904691 Barnett et al. May 1999 A
5907395 Schulz et al. May 1999 A
5920395 Schulz Jul 1999 A
5921992 Costales et al. Jul 1999 A
5987349 Schulz Nov 1999 A
6118845 Simon et al. Sep 2000 A
Foreign Referenced Citations (44)
Number Date Country
964149 Mar 1975 CA
3042343 Jun 1982 DE
3042343 Jun 1982 DE
3508730 Sep 1986 DE
3717871 Dec 1988 DE
3831278 Mar 1989 DE
3838011 Jul 1989 DE
3904595 Apr 1990 DE
3902249 Aug 1990 DE
4225112 Dec 1993 DE
4233978 Apr 1994 DE
4432890 Mar 1996 DE
19829230 Mar 2000 DE
0018166 Oct 1980 EP
0155857 Sep 1985 EP
350996 Jan 1990 EP
0359773 Mar 1990 EP
0427358 May 1991 EP
0456103 Nov 1991 EP
0469966 Feb 1992 EP
0501993 Sep 1992 EP
0908146 Apr 1999 EP
79 04241 Feb 1979 FR
2417970 Sep 1979 FR
2094590 Sep 1982 GB
WO 8809151 Dec 1988 WO
WO9005494 May 1990 WO
WO 9103982 Apr 1991 WO
WO 9104711 Apr 1991 WO
WO 9107726 May 1991 WO
WO9200702 Jan 1992 WO
WO 9206645 Apr 1992 WO
WO9406352 Mar 1994 WO
WO 9423647 Oct 1994 WO
WO 9424933 Nov 1994 WO
WO 9611624 Apr 1996 WO
WO 9838908 Sep 1998 WO
WO9915097 Apr 1999 WO
WO9921498 May 1999 WO
WO9926549 Jun 1999 WO
WO9927839 Jun 1999 WO
WO9929253 Jun 1999 WO
WO9933406 Jul 1999 WO
WO9938449 Aug 1999 WO
Non-Patent Literature Citations (102)
Entry
Adams, L., et al., Aide au Reperage Tridimensionnel pour la Chirurgie de la Base du Crane, Innov. Tech. Biol. Med., vol. 13, No. 4, pp. 409-424 (1992).
Barrick, E. F., Journal of Orthopaedic Trauma: Distal Locking Screw Insertion Using a Cannulated Drill Bit: Technical Note, Raven Press, vol. 7, No. 3, pp. 248-251 (1993).
Batnitzky, S., et al., Three-Dimensional Computer Reconstructions of Brain Lesions from Surface Contours Provided by Computed Tomography: A Prospectus, Neurosurgery, vol. 11, No. 1, Part 1, pp. 73-84 (1982).
Bouazza-Marouf et al., Robotic-Assisted Internal Fixation of Femoral Fractures, IMECHE, pp. 51-58 (1995).
Brack, C. et al., Accurate X-ray Based Navigation in Computer-Assisted Orthopedic Surgery, CAR '98, pp. 716-722.
Brack, C., et al., Towards Accurate X-Ray Camera Calibration in Computer-Assisted Robotic Surgery, CAR '96 Computer-Assisted Radiology, Proceedings of the International Symposium on Computer and Communication Systems for Image Guided Diagnosis and Therapy, Paris, pp. 721-728 (Jun. 1996).
Champleboux, G., et al., Accurate Calibration of Cameras and Range Imaging Sensors: The NPBS Method, Proceedings 1992 IEEE International Conference on Robotics and Automation, pp. 1552-1557 (May 12-14, 1992).
Champleboux, G., Utilisation de Fonctions Splines pour la Mise au Point d'Un Capteur Tridimensionnel sans Contact (Jul. 1991).
Cinquin, P., et al, Computer-Assisted Medical Interventions, pp. 63-65 (Sep. 1989).
Cinquin, P., et al., Computer-Assisted Medical Interventions, IEEE Engineering in Medicine and Biology, pp. 254-263 (May/Jun. 1995).
Cinquin, P., et al., Computer-Assisted Medical Interventions, International Advanced Robotics Programme, pp. 63-65 (1989).
Clarysse, P., et al., A Computer-Assisted System for 3-D Frameless Localization in Stereotaxic MRI, IEEE Transactions on Medical Imaging, vol. 10., No. 4, pp. 523-529 (1991).
Colchester, A. et al., Information Processing in Medical Imaging, 12th International Conference, IPMI, Lecture Notes in Computer Science, pp. 135-141 (1991).
Feldmar, J. et al., 3D-2D Projective Registration of Free-Form Curves and Surfaces, Rapport de recherche (Inria Sophia Antipolis), pp. 1-44 (1994).
Foley, J. D., et al. Fundamentals of Interactive Computer Graphics, Addison-Wesley Systems Programming Series, pp. 245-266 (1982).
Foley, K. T., et al., Image-Guided Intraoperative Spinal Localization, Intraoperative Neuroprotection, Chapter 19, pp. 325-340 (1996).
Foley, K. T., The SteathStation™, Three-Dimensional Image-Interactive Guidance of the Spine Surgeon, Spinal Frontiers, pp. 7-9 (Apr. 1996).
Frederick et al., Prophylactic Intramedullary Fixation of the Tibia for Stress Fracture in a Professional Athlete, Journal of Orthopaedic Trauma, vol. 6, No. 2, pp. 241-244 (1992).
Frederick et al., Technical Difficulties with the Brooker-Wills Nail in Acute Fractures of the Femur, Journal of Orthopaedic Trauma, vol. 6, No. 2, pp. 144-150 (1990).
Gildenberg, P. L., et al., Calculation of Stereotactic Coordinates from the Computed Tomographic Scan, CT Scan Stereotactic Coordinates, pp. 580-586 (May 1982).
Gonzalez, R. C. et al., Digital Image Fundamentals, Digital Image Processing, Second Edition, Addison-Wesley Publishing, pp. 52-54 (1987).
Gottesfeld-Brown, L. M. et al., Registration of Planar Film Radiographs with Computer Tomography, Proceedings of MMBIA, pp. 42-51 (Jun. 1996).
Guéziec , A. P. et al., Registration of Computer Tomography Data to a Surgical Robot Using Fluoroscopy: A Feasibility Study, Computer Science/Mathematics, 6 pages (Sep. 27, 1996).
Hamadeh, A. et al., Automated 3-Dimensional Computer Tomographic and Fluoroscopic Image Registration, Computer Aided Surgery, 3: 11-19 (1998).
Hamadeh, A. et al., Towards Automatic Registration Between CT and X-ray Images: Cooperation Between 3D/2D Registration and 2D Edge Detection, MRCAS'95, pp. 39-46.
Hamadeh, A., et al., Kinematic Study of Lumbar Spine Using Functional Radiographies and 3D/2D Registration, TIMC UMR 5525—IMAG.
Hatch, J. F., Reference-Display System for the Integration of CT Scanning and the Operating Microscope, A Thesis Submitted to the Thayer School of Engineering, Dartmouth College, pp. 1-189 (Oct., 1984).
Henderson, J. M., et al., An Accurate and Ergonomic Method of Registration for Image-Guided Neurosurgery, Computerized Medical Imaging and Graphics, vol. 18, No. 4, pp. 273-277 (1994).
Hoerenz, P., The Operating Microscope I. Optical Principles, Illumination Systems, and Support Systems, Journal of Microsurgery, vol. 1, pp. 364-369 (1980).
Hofstetter, R. et al., Fluoroscopy Based Surgical Navigation—Concept and Clinical Applications, Computer-Assisted Radiology and Surgery, pp. 956-960 (1997).
Hounsfield, G. N., Computerized Transverse Axial Scanning (Tomography): Part I. Description of System, British Journal of Radiology, vol. 46, No. 552, pp. 1016-1022 (Dec. 1973).
Jacques, S., et al., A Computerized Microstereotactic Method to Approach, 3-Dimensionally Reconstruct, Remove and Adjuvantly Treat Small CNS Lesions, Meeting of the Amer. Soc. Stereotactic and Functional Neurosurgery, Houston, Appl. Neurophysiology, 43:176-182 (1980).
Jacques, S., et al., Computerized Three-Dimensional Stereotaxic Removal of Small Central Nervous System Lesions in Patients, J. Neurosurg., 53:816-820 (1980).
Joskowicz, L. et al., Computer-Aided Image-Guided Bone Fracture Surgery: Concept and Implementation, CAR '98, pp. 710-715.
Kelly, P. J., et al., Precision Resection of Intra-Axial CNS Lesions by CT-Based Stereotactic Craniotomy and Computer Monitored CO2 Laser, Acta Neurochirurgica 68, pp. 1-9 (1983).
Lavallée, S., A New System for Computer-Assisted Neurosurgery, IEEE Engineering in Medicine & Biology Society 11th Annual International Conference, pp. 926-927 (1989).
Lavallée, S., et al., Matching of Medical Images for Computed and Robot-Assisted Surgery, IEEE EMBS (1991).
Lavallée, S., et al., Computer-Assisted Interventionist Imaging: The Instance of Stereotactic Brain Surgery, North-Holland MEDINFO 89, pp. 613-617 (1989).
Lavallée, S., et al., Computer-Assisted Driving of a Needle into the Brain, Proceedings of the International Symposium, CAR 89, Computer-Assisted Radiology; pp. 416-420 (1989).
Lavallée, S., et al, Computer-Assisted Spine Surgery: A Technique For Accurate Transpedicular Screw Fixation Using CT Data and a 3-D Optical Localizer, TIMC, Faculte de Medecine de Grenoble.
Lavallée, S., et al. Computer-Assisted Interventionist Imaging: The Instance of Stereotactic Brain Surgery, North-Holland MEDINFO 89, Part 1, pp. 613-617 (1989).
Lavallée, S., et al., Matching 3-D Smooth Surfaces with Their 2-D Projections Using 3-D Distance Maps, SPIE, vol. 1570, Genometric Methods in Computer Vision, pp. 322-336 (1991).
Lavallée, S., VI Adaption de la Methodologie a Quelques Applications Cliniques, Chapitre VI, pp. 133-148.
Lavallée, S., et al., Image Guided Operating Robot: A Clinical Application in Stereotactic Neurosurgery, Proceedings of the 1992 IEEE International Conference on Robotics and Automation, pp. 618-624 (May 1992).
Leksell, L. et al., Stereotaxis and Tomography—A Technical Note, ACTA Neurochirugica, vol. 52, pp. 1-7 (1980).
Lemieux, L. et al., A Patient-to-Computer Tomography Image Registration Method Based on Digitally Reconstructed Radiographs, Med. Phys. 21 (11), pp. 1749-1760 (Nov. 1994).
Levin, D. N., et al., The Brain: Integrated Three-dimensional Display of MR and PET Images, Radiology, pp. 172:783-789 (Sep. 1989).
Mazier, B., et al., Chirurgie de la Colonne Vertebrale Assistee Par Ordinateur: Application au Vissage Pediculaire, Innov. Tech. Biol. Med., vol. 11, No. 5, pp. 559-566 (1990).
Mazier, B., et al., Computer-Assisted Interventionist Imaging: Application to the Vertebral Column Surgery, Annual International Conference of the IEEE Engineering in Medicine and Biology Society, vol. 12, No. 1, pp. 430-431 (1990).
PCT International Search Report, PCT/US99/14565 (Oct. 20, 1999).
Pelizzari, C. A. et al., No. 528—Three-Dimensional Correlation of PET, CT and MRI Images, The Journal of Nuclear Medicine, vol. 28, No. 4, p. 682 (Apr. 1987).
Pelizzari, C. A., et al., Accurate Three-Dimensional Registration of CT, PET, and/or MR Images of the Brain, Journal of Computer-Assisted Tomography, vol. 13, No. 1, pp. 20-26 (Jan./Feb. 1989).
Phillips, R. et al., Image Guided Orthopaedic Surgery Design and Analysis, Trans Inst MC, vol. 17, No. 5, pp. 251-264 (1995).
Potamianos, P., et al., Intra-Operative Imaging Guidance for Keyhole Surgery Methodology and Calibration, First International Symposium on Medical Robotics and Computer-Assisted Surgery, pp. 98-104 (Sep. 22-24, 1994).
Potamianos, P., et al., Intra-Operative Registration for Percutaneous Surgery, Proceedings of the Second International Symposium on Medical Robotics and Computer-Assisted Surgery—Baltimore, MD—(Nov. 1995).
Reinhardt, H. F., et al., CT-Guided “Real Time” Stereotaxy, ACTA Neurochirurgica (1989).
Roberts, D. W., et al., A Frameless Stereotaxic Integration of Computerized Tomographic Imaging and the Operating Microscope, J. Neurosurg., vol. 65, pp. 545-549 (Oct. 1986).
Rosenbaum, A. E., et al., Computerized Tomography Guided Stereotaxis: A New Approach, Meeting of the Amer. Soc. Stereotactic and Functional Neurosurgery, Houston, Appl. Neurophysiol., 43:172-173 (1980).
Rougee, A., et al., Geometrical Calibration of X-Ray Imaging Chains For Three-Dimensional Reconstruction, Computerized Medical Imaging and Graphics, vol. 17, Nos. 4/5, pp. 295-300 (1993).
Sautot, P., “Vissage Pediculaire Assiste par Ordinateur,” (Sep. 20, 1994).
Schreiner, S., et al., Accuracy Assessment of a Clinical Biplane Fluoroscope for Three-Dimensional Measurements and Targeting, Proceedings of SPIE, Image Display, vol. 3031, pp. 160-166 (Feb. 23-25, 1997).
Schueler, B., Correction of Image Intensifier Distortion for Three-Dimensional X-ray Angiography, Proceedings of SPIE, Physics of Medical Imaging, vol. 2432, pp. 272-279 (Feb. 26-27, 1995).
Selvik, G., et al., A Roentgen Stereophotogrammetric System, Acta Radiologica Diagnosis, pp. 343-352 (1983).
Shelden, C. H., et al., Development of a Computerized Microstereotaxic Method for Localization and Removal of Minute CNS Lesions Under Direct 3-D Vision, J. Neurosurg., 52:21-27 (1980).
Simon, D., Fast and Accurate Shape-Based Registration, Carnegie Mellon University (Dec. 12, 1996).
Smith, K. R., et al., Computer Methods for Improved Diagnostic Image Display Applied to Stereotactic Neurosurgery, Automedical, vol. 14, pp. 371-386 (1991).
Viant, W. J. et al., A Computer-Assisted Orthopaedic System for Distal Locking of Intramedullary Nails, Proc. of MediMEC '95, Bristol, pp. 86-91 (1995).
Watanabe, E., et al., Three-Dimensional Digitizer (Neuronavigator): New Equipment for Computed Tomography-Guided Stereotaxic Surgery, Surgical Neurology, vol. 27, No. 6, pp. 543-547 (Jun. 1987).
Watanabe, H., Neuronavigator, Igaku-no-Ayumi, vol. 137, No. 6, pp. 1-4 (May 10, 1986).
Weese, J., et al., An Approach to 2D/3D Registration of a Vertebra in 2D X-ray Fluoroscopies with 3D CT Images, First Joint Conference Computer Vision, Virtual Reality and Robotics in Medicine and Med. Robotics and Computer-Assisted Surgery, Grenoble, France, pp. 119-128 (Mar. 19-22, 1997).
Afshar, F. et al., A Three-Dimensional Reconstruction of the Human Brain Stem, Journal of Neurosurgery, vol. 57, No. 3, pp. 491-495 (Oct. 1982).
Awwad, E. et al., MRI Imaging of Lumber Juxtaarticular Cysts, Journal of Computer Assisted Tomography, pp. 415-417, vol. 14, No. 3 (May 1990).
Bajcsy, et al., Computerized Anatomy Atlas of the Human Brain, NCGA ′81 Conference Proceedings, Second Annual Conference & Exhibition, Baltimore, MD, pp. 435-441 (Jun. 1981).
Benzel, E. et al., Magnetic Source Imaging: A Review of the Magnes System of Biomagnetic Technologies Incorporated, Neurosurgery, vol. 33, No. 2, pp. 252-259 (Aug. 1983).
Birg, W. et al., A Computer Programme System for Stereotactic Neurosurgery, Acta Neurochirurgica, Suppl. 24, pp. 99-108 (1977).
Bo{haeck over (e)}thius, J. et al., Stereotaxic Computerized Tomography With a GE 8800 Scanner, J. Neurosurg., vol. 52, pp. 794-800 (Jun. 1980).
Bo{haeck over (e)}thius J. et al, Stereotactic Biopsies and Computer Tomography in Gliomas, Acta Neurochirurgica, vol. 49, pp. 223-232 (1978).
Brunie, L. et al., Pre-and Intra-Irradiation Multimodal Image Registration: Principles and First Experiments, Radiotherapy and Oncology 29, pp. 244-252 (1993).
Bucholz, R. et al., A Comparison of Sonic Digitizers Versus Light Emitting Diode-Based Localization, Interactive Image-Guided Neurosurgery, Chapter 16, pp. 179-200.
Bucholz, R. et al., Image-Guided Surgical Techniques for Infections and the Trauma of the Central Nervous System, Neurosurgery Clinics of North America, vol. 7, No. 2, pp. 187-200 (Apr. 1996).
Bucholz, R. et al., The Correction of Stereotactic Inaccuracy Caused by Brain Shift Using an Intraoperative Ultrasound Device, CVRMed-MRCAS '97, First Joint Conference, Computer Vision, Virtual Reality and Robotics in Medicine and Medical Robotics and Computer-Assisted Surgery, pp. 459-466 (Mar. 19-22, 1997).
Bucholz, R. et al., Variables Affecting the Accuracy of Sterotactic Localization Using Computerized Tomography, Journal of Neurosurgery, vol. 79, pp. 667-673 (Nov. 1993).
Castleman, K. et al., Stereometric Ranging, Chapter 17: Three-Dimensional Image Processing, Digital Image Processing, pp. 364-369 (1979).
Davatzikos, C. et al., Image Registration Based on Boundary Mapping, Thesis (Johns Hopkins University), pp. 1-30 (1995).
Friston, K. et al., Plastic Transformation of PET Images, Journal of Computer-Assisted Tomography, vol. 15, No. 4, pp. 634-639 (1991).
Gallen, C. et al., Intracranial Neurosurgery Guided by Functional Imaging, Surgical Neurology, vol. 42, pp. 523-530 (Dec. 1994).
Gouda, K. et al., New Frame for Stereotaxic Surgery, Journal of Neurosurgery, vol. 53, pp. 256-259 (Aug. 1980).
Greitz, T. et al., Head Fixation System for Integration of Radiodiagnostic and Therapeutic Procedures, Neuroradiology, vol. 19, pp. 1-6 (1980).
Hamadeh, A. et al., Towards Automatic Registration Between CT and X-Ray Images: Cooperation Between 3D/2D Registration and 2D Edge Detection, TIMC-IMAG Faculté de Medecine de Grenoble, pp. 39-46 (with 2 pages of drawings) (1995).
Hamadeh A., X-Ray Data Acquisition in OR—Test Specifications, Operating Room Test Report Laboratoire TIMC, pp. 1-9 (Jun. 7, 1995).
Heilbrun, M. P. Progressive Technology Applications, Neurosurgery for the Third Millenium, Chapter 15, pp. 191-198 (Oct. 1992).
Heilbrun, M. P. et al., Stereotactic Localization and Guidance Using a Machine Vision Technique, Stereotactic and Functional Neurosurgery, vol. 58, pp. 94-98 (Sep. 1992).
Kelly, P. J. et al., Precision Resection of Intra-Axial CNS Lesions by CT-Based Stereotactic Craniotomy and Computer Monitored CO2 Laser, Acta Neurochirurgica, vol. 68, pp. 1-9 (1983).
Lavallee, S. et al, Computer-Assisted Spine Surgery: A Technique for Accurate Transpedicular Screw Fixation Using CT Data and a 3-D Optical Localizer, pp. 315-322 (1995).
Leavitt, D. et al., Dynamic Field Shaping to Optimize Stereotactic Radiosurgery, International Journal of Radiation Oncology, Biology, Physics, vol. 21, pp. 1247-1255 (Oct. 1990).
Mundinger, F. et al., Computer-Assisted Stereotactic Brain Operations by Means Including Computerized Axial Tomography, Applied Neurophysiology, vol. 41, No. 1-4, Proceedings of the Seventh Meeting of the World Society for Stereotactic and Functional Neurosurgery (1978).
Perry, J. et al., Computed Tomography—Guided Stererotactic Surgery: Conception and Development of a New Stereotactic Methodology, Neurosurgery, vol. 7, No. 4, pp. 376-381 (Oct. 1980).
Potamianos, P. et al., Manipulator Assisted Renal Treatment, Centre for Robotics, Imperial College of Science, Technology & Medicine, London, pp. 214-226 (Jul. 1993).
Sautot P., Computer Assisted Introduction of Screws Into Pedicles, Thesis, pp. 1-163 (Sep. 1994).
Sautot, P., Part C—Methodology for Computer Assisted Introduction of a Screw Into a Pedicle, pp. 1-35 (Sep. 1994).
Smith, K. et al., The Neurostation™—A Highly Accurate, Minimally Invasive Solution to Frameless Stereotactic Neurosurgery, Computerized Medical Imaging and Graphics, vol. 18, No. 4, pp. 247-256 (Jul.-Aug. 1994).
Troccaz, J. et al., Conformal External Radiotherapy of Prosthetic Carcinoma: Requirements and Experimental Results, Radiotherapy and Oncology 29, pp. 176-183 (1993).
Continuations (1)
Number Date Country
Parent 09/106109 Jun 1998 US
Child 09/591512 US