Method and apparatus for positioning a guide relative to a base

Information

  • Patent Grant
  • 8165658
  • Patent Number
    8,165,658
  • Date Filed
    Friday, September 26, 2008
    16 years ago
  • Date Issued
    Tuesday, April 24, 2012
    12 years ago
Abstract
A system for positioning a guide relative to an anatomy is provided. The system can include a base adapted to be coupled to the anatomy, and a guide that can move relative to the base. The system can include at least one tracking device that can be coupled to the base and the guide, and a tracking system that tracks a position of the tracking device. The system can include a navigation system that determines a position of the base and the guide relative to the anatomy, and whether the position of the base and the guide are in a desired position. The system can include a display, which can display at least one icon superimposed on the image of the anatomy that graphically indicates a manipulation of the guide required to move the guide into the desired position.
Description
FIELD

The present disclosure relates to a surgical navigation system, and particularly to a method and apparatus for positioning a guide relative to a base.


BACKGROUND

The statements in this section merely provide background information related to the present disclosure and may not constitute prior art.


Image guided medical and surgical procedures utilize patient images (image data) obtained prior to or during a medical procedure to guide a physician performing the procedure. Recent advances in imaging technology, especially in imaging technologies that produce highly-detailed, two, three, and four dimensional images, such as computed tomography (CT), magnetic resonance imaging (MRI), fluoroscopic imaging (such as with a C-arm device), positron emission tomography (PET), and ultrasound imaging (US) has increased the interest in navigated medical procedures.


Generally, during a navigated procedure, images are acquired by a suitable imaging device for display on a workstation. The navigation system tracks the patient, instruments and other devices in the surgical field or patient space. These tracked devices are then displayed relative to the image data on the workstation in image space. In order to track the patient, instruments and other devices, the patient, instruments and other devices can be equipped with tracking devices.


SUMMARY

During a navigated procedure, a dynamic reference frame (DRF) including a tracking device can be fixed relative to the patient in order to track the patient. In addition, it may be desirable to use a holder, such as a screw, head frame, etc. to couple the tracking device to the patient. In some cases, the holder can comprise one or more parts, such as in the case of a head frame. Generally, the head frame can include a positionable device coupled to a fixed base. Typically, the positionable device can be coupled to one or more instruments to enable the instruments to be guided into the anatomy. As the positionable device can guide the instruments into the anatomy, it may be desirable to track the orientation of the positionable device relative to the anatomy. Further, it may be desirable to notify a user of the navigation system if the positionable device and/or the base is/are not properly aligned with the anatomy.


According to various embodiments, a system for positioning a guide relative to an anatomy is disclosed. The system can include a base adapted to be coupled to the anatomy, and a guide that can move relative to the base. The system can include at least one tracking device that can be coupled to the base and the guide, and a tracking system, which can track a position of the at least one tracking device relative to the anatomy. The system can include a navigation system, which can determine a position of the base and the guide relative to the anatomy based on the position of the at least one tracking device. The navigation system can also determine if the position of the base and the position of the guide are about equal to a desired position of the base and the movable member. The system can further include a display, which can display the desired position of the base and the guide superimposed on an image of the anatomy. The display can also display at least one icon superimposed on the image of the anatomy, which can graphically indicate a desired manipulation of the guide required to move the guide into the desired position for the guide relative to the anatomy.


Further disclosed is a method for positioning a guide relative to an anatomy. The method can include acquiring a surgical plan for a surgical procedure to be performed on the anatomy that includes a trajectory for an instrument. The method can also include tracking at least one first tracking device coupled to a base relative to the anatomy, and tracking at least one second tracking device coupled to a guide relative to the anatomy. The guide can also be movable relative to the base. The method can include determining a position of the base and guide relative to the anatomy. The method can also include comparing the position of the base and the guide relative to the anatomy with a desired position for the base and the guide in the surgical plan that enables the guide to guide the instrument along the trajectory. The method can include displaying at least one icon superimposed onto an image of the anatomy that indicates an amount and direction of a first manipulation required to move the guide into the desired position and at least one icon superimposed onto an image of the anatomy that indicates an amount and direction of a second manipulation required to move the guide relative to the base into the desired position.


Also disclosed is a system for positioning a reference frame relative to an anatomy. The system can include a base adapted to be fixedly coupled to the anatomy, and a guide that moves relative to the base. The guide can be operable to be coupled to a drive system to guide at least one instrument into the anatomy. The system can include at least one first tracking device coupled to the base, and at least one second tracking device coupled to the guide. The system can also include a tracking system, which can track a position of the at least one first tracking device and the at least one second device relative to the anatomy. The system can include a navigation control module, which can receive patient image data and surgical plan data. The surgical plan data can include a desired position for the base and the guide relative to the anatomy. The system can also include a navigation system, which can determine a position of the base and the guide relative to the anatomy based on the position of the at least one first tracking device and the at least one second tracking device. The navigation system can also determine if the position of the base and the position of the guide are about equal to the desired position of the base and the guide. The navigation control module can output image data that can include a first arrow icon superimposed onto the patient image data, which can graphically represent an amount of rotation and a direction of rotation for the guide, a second arrow icon superimposed onto the patient image data, which can graphically represent an amount of translation and a direction of translation for the guide, or combinations thereof.


Further areas of applicability will become apparent from the description provided herein. It should be understood that the description and specific examples are intended for purposes of illustration only and are not intended to limit the scope of the present disclosure.





DRAWINGS

The drawings described herein are for illustration purposes only and are not intended to limit the scope of the present disclosure in any way.



FIG. 1 is an environmental view of a surgical navigation system or computer aided surgical system, according to various embodiments;



FIG. 2 is a detail environmental view of an exemplary dynamic reference frame (DRF) holder, for example, a head frame, for use with the navigation system of FIG. 1 according to various embodiments;



FIG. 3 is a simplified block diagram illustrating the navigation system of FIG. 1;



FIG. 4 is a graphical representation of an exemplary display produced by the navigation system of FIG. 1;



FIG. 5 is a dataflow diagram illustrating a control system performed by a control module associated with the navigation system of FIG. 1; and



FIG. 6 is a flowchart illustrating a control method performed by the control module.





DETAILED DESCRIPTION

The following description is merely exemplary in nature and is not intended to limit the present disclosure, application, or uses. It should be understood that throughout the drawings, corresponding reference numerals indicate like or corresponding parts and features. As indicated above, the present teachings are directed toward providing a system and method for positioning a reference frame for use in a navigated surgical procedure. It should be noted, however, that the present teachings could be applicable to any appropriate procedure in which it is desirable to assist a user in orientating a device within an operating theater, via a user interface. Further, as used herein, the term “module” can refer to an application specific integrated circuit (ASIC), an electronic circuit, a processor (shared, dedicated, or group) and memory that executes one or more software or firmware programs, a combinational logic circuit, and/or other suitable hardware or software, firmware programs or components that provide the described functionality. Therefore, it will be understood that the following discussions are not intended to limit the scope of the appended claims.


As will be discussed in greater detail herein, the present disclosure is directed toward a system and method for guiding a positioning of a reference frame relative to an anatomy, such as a brain, to facilitate a surgical procedure on the anatomy. In this regard, this system and method can include an exemplary surgical navigation system 10, which can include a drive system 12 to control the insertion and withdrawal of one or more instruments 14 from the anatomy. Various surgical navigation systems can include those described in U.S. patent application Ser. No. 10/651,267 (now U.S. Pat. App. Pub No. 2005/0049486), filed on Aug. 28, 2003, incorporated herein by reference. As will be discussed, the drive system 12 and the instruments 14 can be positioned relative to a dynamic reference frame, which can be coupled to a patient 16.


With reference to FIG. 1, the exemplary surgical navigation system 10 can include an image based system, an imageless system, an atlas or diagram based system, or combinations thereof. One skilled in the art will understand that the surgical navigation system 10 can require the registration of a patient 16, which defines patient space, to an image space, discussed further herein. According to various embodiments, registration can include registration between image space, defined by image data or atlas data, and the patient space.


With continued reference to FIG. 1, the navigation system 10 that can be used for various procedures is illustrated. The navigation system 10 can be used to track the location of an implant, such as a spinal implant or orthopedic implant, or a surgical device, such as an electrode, relative to a patient 16. Also the navigation system 10 can track the position and orientation of various instruments 14. It should further be noted that the navigation system 10 may be used to navigate any type of instrument, implant, or delivery system, including: guide wires, arthroscopic systems, cardiac leads, orthopedic implants, spinal implants, deep-brain stimulator (DBS) probes, microelectrode recorder probes, macroelectrode stimulation probes, etc. Moreover, these instruments may be used to navigate or map any region of the body. The navigation system 10 and the various instruments may be used in any appropriate procedure, such as one that is generally minimally invasive, arthroscopic, percutaneous, stereotactic, or an open procedure.


Although an exemplary navigation system 10 that can include the imaging device 18 is discussed herein, one skilled in the art will understand that the disclosure is merely for clarity of the present discussion and any appropriate imaging system, navigation system, patient specific data, and non-patient specific data can be used. For example, the intraoperative imaging system can include an MRI imaging system, such as the PoleStar® MRI or an O-arm™ imaging system sold by Medtronic, Inc. It will be understood that the navigation system 10 can incorporate or be used with any appropriate preoperatively or intraoperatively acquired image data. For example, various imageless systems can be used or images from atlas models can be used to produce patient images, such as those disclosed in U.S. Patent Pub. No. 2005-0085714, filed Oct. 16, 2003, entitled “Method and Apparatus for Surgical Navigation of a Multiple Piece Construct for Implantation,” incorporated herein by reference. The imaging device 18 can be, for example, a fluoroscopic x-ray imaging device that may be configured as an O-arm™ or a C-arm, which can have an x-ray source 20, an x-ray receiving section 22, an optional calibration and tracking target 24 and optional radiation sensors 26.


In operation, the imaging device 18 can generate x-rays from the x-ray source 20 that can propagate through the patient 16 and calibration and/or tracking target 24, into the x-ray receiving section 22. This allows direct visualization of the patient 16 and radio-opaque instruments in the cone of the X-rays. In the example of FIG. 1, a longitudinal axis 16a of the patient 16 is substantially in line with a mechanical axis 28 of the C-arm. This can enable the imaging device 18 to be rotated relative to the patient 16, allowing images of the patient 16 to be taken from multiple directions or about multiple planes. An example of a fluoroscopic C-arm X-ray device that may be used as the optional imaging device 18 is the “Series 9600 Mobile Digital Imaging System,” from GE Healthcare, (formerly OEC Medical Systems, Inc.) of Salt Lake City, Utah. Other exemplary fluoroscopes include bi-plane fluoroscopic systems, ceiling fluoroscopic systems, cath-lab fluoroscopic systems, fixed C-arm fluoroscopic systems, isocentric C-arm fluoroscopic systems, 3D fluoroscopic systems, etc.


When the x-ray source 20 generates the x-rays that propagate to the x-ray receiving section 22, the radiation sensors 26 can sense the presence of radiation, which is forwarded to an imaging device controller 30, to identify whether or not the imaging device 18 is actively imaging. This information can also be transmitted to a coil array controller 32, further discussed herein.


The imaging device controller 30 can capture the x-ray images received at the x-ray receiving section 22 and store the images for later use. Multiple two-dimensional images taken by the imaging device 18 may also be captured and assembled by the imaging device controller 30 to provide a larger view or image of a whole region of the patient 16, as opposed to being directed to only a portion of a region of the patient 16. The imaging device controller 30 may also be separate from the imaging device 18 and/or control the rotation of the imaging device 18. For example, a C-arm can move in the direction of arrow A or rotate about the longitudinal axis 16a of the patient 16, allowing anterior or lateral views of the patient 16 to be imaged. Each of these movements involves rotation about the mechanical axis 28 of the C-arm. The movements of the imaging device 18, such as the C-arm, can be tracked with a tracking device 34.


While the imaging device 18 is shown in FIG. 1 as a C-arm, any other alternative 2D, 3D or 4D imaging modality may also be used. For example, any 2D, 3D or 4D imaging device, such as the O-arm™ imaging device, isocentric fluoroscopy, bi-plane fluoroscopy, ultrasound, computed tomography (CT), multi-slice computed tomography (MSCT), magnetic resonance imaging (MRI), high frequency ultrasound (HFU), positron emission tomography (PET), optical coherence tomography (OCT), intra-vascular ultrasound (IVUS), ultrasound, intra-operative CT or MRI may also be used to acquire 2D, 3D or 4D pre- or post-operative and/or real-time images or patient image data 36 of the patient 16. For example, an intra-operative MRI system, may be used such as the PoleStar® MRI system sold by Medtronic, Inc.


In addition, image datasets from hybrid modalities, such as positron emission tomography (PET) combined with CT, or single photon emission computer tomography (SPECT) combined with CT, could also provide functional image data superimposed onto anatomical data to be used to confidently reach target sites within the patient 16. It should further be noted that the imaging device 18 as shown in FIG. 1, provides a virtual bi-plane image using a single-head C-arm fluoroscope as the imaging device 18 by simply rotating the C-arm about at least two planes, which could be orthogonal planes, to generate two-dimensional images that can be converted to three-dimensional volumetric images. By acquiring images in more than one plane, an icon 38 representing the location of the instrument 14, such as an impacter, stylet, reamer driver, taps, drill, deep-brain stimulator (DBS) probes, cardiac leads or other instrument, or implantable devices introduced and advanced in the patient 16, may be superimposed in more than one view and included in image data 42 displayed on a display 44.


If the imaging device 18 is employed, patient image data 36 can be forwarded from the imaging device controller 30 to a navigation computer and/or processor or workstation 46. It will also be understood that the patient image data 36 is not necessarily first retained in the imaging device controller 30, but may also be directly transmitted to the workstation 46. The workstation 46 can include the display 44, a user input device 48 and a control module 50. The workstation 46 can also include or be connected to an image processor, navigation processor, and memory to hold instruction and data. The workstation 46 can provide facilities for displaying the patient image data 36 as an image on the display 44, saving, digitally manipulating, or printing a hard copy image of the received patient image data 36.


The user input device 48 can comprise any device that can enable a user to interface with the workstation 46, such as a touchpad, touch pen, touch screen, keyboard, mouse, wireless mouse, or a combination thereof. The user input device 48 allows a physician or user 52 to provide inputs to control the imaging device 18, via the imaging device controller 30, adjust the display settings of the display 44, or control a tracking system 54, as further discussed herein. The control module 50 can determine the location of a dynamic reference frame (DRF) 64 (FIG. 2) with respect to the patient space, and can output image data 42 to the display 44 to assist in positioning the DRF 64.


With continuing reference to FIG. 1, the navigation system 10 can further include the electromagnetic navigation or tracking system 54. A representative electromagnetic navigation or tracking system 54 can include the AXIEM™ electromagnetic tracking system sold by Medtronic Navigation, Inc. The tracking system 54 can include a localizer, such as a first coil array 58 and/or second coil array 60, the coil array controller 32, a navigation probe interface 62, the device or instrument 14, a patient tracker or dynamic reference frame (DRF) 64, and one or more tracking devices 56. Other tracking systems can include an optical tracking system 54b, for example the StealthStation® Treon® and the StealthStation® Tria® both sold by Medtronic Navigation, Inc. Further, other tracking systems can be used that include acoustic, radiation, radar, infrared, etc., or hybrid systems, such as a system that includes components of both an electromagnetic and optical tracking system, etc. The drive system 12, the instrument 14 and the DRF 64 can each include tracking device(s) 56.


The tracking device 56 or any appropriate tracking device as discussed herein, can include both a sensor, a transmitter, or combinations thereof and can be indicated by the reference numeral 56. Further, the tracking device 56 can be wired or wireless to provide a signal or emitter or receive a signal from a system. For example, a tracking device 56a can include one or more electromagnetic coils, such as a tri-axial coil, to sense a field produced by the localizing coil array 58 or 60. One will understand that the tracking device(s) 56 can receive a signal, transmit a signal, or combinations thereof to provide information to the navigation system 10, which can be used to determine a location of the tracking device 56. The navigation system 10 can determine a position of the instrument 14 and the DRF 64 based on the location of the tracking device(s) 56 to allow for accurate navigation relative to the patient 16 in the patient space.


With regard to the optical localizer or tracking system 54b, the optical tracking system 54b can transmit and receive an optical signal, or combinations thereof. An optical tracking device 56b can be interconnected with the instrument 14, or other devices such as the DRF 64. As generally known, the optical tracking device 56b can reflect, transmit or receive an optical signal to/from the optical localizer or tracking system 54b that can be used in the navigation system 10 to navigate or track various elements. Therefore, one skilled in the art will understand, that the tracking device(s) 56 can be any appropriate tracking device to work with any one or multiple tracking systems.


The coil arrays 58, 60 can transmit signals that are received by the tracking device(s) 56. The tracking device(s) 56 can then transmit or receive signals based upon the transmitted or received signals from or to the coil arrays 58, 60. The coil arrays 58, 60 are shown attached to an operating table 66. It should be noted, however, that the coil arrays 58, 60 can also be positioned at any other location, as well and can also be positioned in the items being navigated. The coil arrays 58, 60 include a plurality of coils that are each operable to generate distinct electromagnetic fields into the navigation region of the patient 16, which is sometimes referred to as patient space. Representative electromagnetic systems are set forth in U.S. Pat. No. 5,913,820, entitled “Position Location System,” issued Jun. 22, 1999 and U.S. Pat. No. 5,592,939, entitled “Method and System for Navigating a Catheter Probe,” issued Jan. 14, 1997, each of which are hereby incorporated by reference.


The coil arrays 58, 60 can be controlled or driven by the coil array controller 32. The coil array controller 32 can drive each coil in the coil arrays 58, 60 in a time division multiplex or a frequency division multiplex manner. In this regard, each coil can be driven separately at a distinct time or all of the coils can be driven simultaneously with each being driven by a different frequency. Upon driving the coils in the coil arrays 58, 60 with the coil array controller 32, electromagnetic fields are generated within the patient 16 in the area where the medical procedure is being performed, which is again sometimes referred to as patient space. The electromagnetic fields generated in the patient space induce currents in a tracking device(s) 56 positioned on or in the drive system 12, the instrument 14 and the DRF 64. These induced signals from the drive system 12, the instrument 14 and the DRF 64 are delivered to the navigation probe interface 62 and can be subsequently forwarded to the coil array controller 32.


The navigation probe interface 62 may provide the necessary electrical isolation for the navigation system 10. The navigation probe interface 62 can also include amplifiers, filters and buffers to directly interface with the tracking device(s) 56 in the instrument 14 and DRF 64. Alternatively, the tracking device(s) 56, or any other appropriate portion, may employ a wireless communications channel, such as that disclosed in U.S. Pat. No. 6,474,341, entitled “Surgical Communication Power System,” issued Nov. 5, 2002, incorporated herein by reference, as opposed to being coupled directly to the navigation probe interface 62.


The instrument 14 may be any appropriate instrument, such as an instrument for preparing a portion of the patient 16, an instrument for recording activity in a portion of the anatomy or an instrument for positioning an implant. With reference to FIGS. 1 and 2, the DRF 64 may be fixed to the patient 16 adjacent to the region being navigated so that any movement of the patient 16 is detected as relative motion between the coil arrays 58, 60 and the DRF 64. This relative motion is forwarded to the coil array controller 32, which updates registration correlation and maintains accurate navigation, further discussed herein. The DRF 64 may include any appropriate tracking device 56 used by the navigation system 10. Therefore, the DRF 64 can include an electromagnetic tracking device, optical, acoustic, etc. Further, with reference to FIG. 2, the DRF 64 can include a DRF holder or head frame 68 and one or more tracking device(s) 56. Alternatively, the DRF 64 can include tracking device(s) 56 that can be formed integrally with the head frame 68.


Moreover, the DRF 64 can be provided as separate pieces and can be positioned at any appropriate position on the anatomy. For example, the tracking device(s) 56 of the DRF 64 can be fixed to the skin of the patient 16 with an adhesive. Also, the DRF 64 can be positioned near a leg, arm, etc. of the patient 16. Thus, the DRF 64 does not need to be provided with the head frame 68 or require any specific base or holding portion. If the DRF 64 is used with an electromagnetic tracking device 56a, it can be configured as a pair of orthogonally oriented coils, each having the same center or may be configured in any other non-coaxial or co-axial coil configurations, such as a tri-axial coil configuration (not specifically shown).


Briefly, the navigation system 10 operates as follows. The navigation system 10 creates a translation map between all points in the radiological image generated from the imaging device 18 in image space and the corresponding points in the anatomical structure of the patient 16 in patient space. After this map is established, whenever a tracked instrument, such as the instrument 14 is used, the workstation 46 in combination with the coil array controller 32 and the imaging device controller 30 uses the translation map to identify the corresponding point on the pre-acquired image or atlas model, which is displayed on display 44. This identification is known as navigation or localization. The icon 38 representing the localized point or instruments 14 can be shown as image data 42 on the display 44.


To enable navigation, the navigation system 10 must be able to detect both the position of the anatomical structure of the patient 16 and the position of the instrument 14. Knowing the location of these two items allows the navigation system 10 to compute and display the position of the instrument 14 in relation to the patient 16 on the display 44. The tracking system 54 can be employed to track the instrument 14 and the anatomical structure simultaneously.


The tracking system 54, if using an electromagnetic tracking assembly, essentially works by positioning the coil arrays 58, 60 adjacent to the patient space to generate a low-energy electromagnetic field generally referred to as a navigation field. Because every point in the navigation field or patient space is associated with a unique field strength, the tracking system 54 can determine the position of the instrument 14 by measuring the field strength at the tracking device 56 location. The DRF 64 can be fixed to the patient 16 to identify a first location of the patient 16 in the navigation field. The tracking system 54 can continuously recompute the relative position of the DRF 64 and the instrument 14 during localization and relate this spatial information to patient registration data to enable image guidance of the instrument 14 within and/or relative to the patient 16.


Patient registration is the process of determining how to correlate the position of the drive system 12 and/or the instrument 14 relative to the patient 16 to the position on the diagnostic or pre-acquired images. To register the patient 16, a physician or user 52 may use point registration by selecting and storing particular points from the pre-acquired images and then touching the corresponding points on the anatomical structure of the patient 16 with a tracked pointer probe 40. The navigation system 10 analyzes the relationship between the two sets of points that are selected and computes a match, which correlates every point in the patient image data 36 with its corresponding point on the anatomical structure of the patient 16 or the patient space, as discussed herein. The points that are selected to perform registration are fiducial markers, such as anatomical landmarks. Again, the landmarks or fiducial markers are identifiable on the images and identifiable and accessible on the patient 16. The fiducial markers can be artificial markers that are positioned on the patient 16 or anatomical landmarks that can be easily identified in the patient image data 36. The artificial landmarks, such as the fiducial markers, can also form part of the DRF 64, such as those disclosed in U.S. Pat. No. 6,381,485, entitled “Registration of Human Anatomy Integrated for Electromagnetic Localization,” issued Apr. 30, 2002, herein incorporated by reference.


The navigation system 10 may also perform registration using anatomic surface information or path information as is known in the art. The navigation system 10 may also perform 2D to 3D registration by utilizing the acquired 2D images to register 3D volume images by use of contour algorithms, point algorithms or density comparison algorithms, as is known in the art. An exemplary 2D to 3D registration procedure, is set forth in U.S. Ser. No. 10/644,680, entitled “Method and Apparatus for Performing 2D to 3D Registration” filed on Aug. 20, 2003, hereby incorporated by reference.


In order to maintain registration accuracy, the navigation system 10 continuously tracks the position of the patient 16 during registration and navigation. This is because the patient 16, DRF 64 and coil arrays 58, 60 may all move during the procedure, even when this movement is not desired. Therefore, if the navigation system 10 did not track the position of the patient 16 or area of the anatomy, any patient movement after image acquisition would result in inaccurate navigation within that image. Because the DRF 64 can be coupled to the patient 16, any movement of the anatomical structure of the patient 16 or the coil arrays 58, 60 can be detected as the relative motion between the coil arrays 58, 60 and the DRF 64. Both the relative motion of the coil arrays 58, 60 and the DRF 64 can be communicated to the coil array controller 32, via the navigation probe interface 62, which can update the registration correlation to thereby maintain accurate navigation.


With continued reference to FIG. 1 and with additional reference to FIG. 2, the guide or drive system 12 for use with an exemplary head frame 68, such as a stereotactic head frame or a small scale sterotactic head frame 68 is illustrated. The drive system 12 can be used to drive various instruments, such as one or more instruments 14, such as electrodes, into an anatomy, such as a brain 98. For example, a procedure on the brain 98 can include a recorder for detecting electrical activity in the brain 98 with a microelectrode (ME) or macroelectrode. Once a recording of the brain 98 has occurred, a stimulator probe, such as a deep brain stimulator probe or a macroelectrode can be delivered to an area identified with the ME or macroelectrode. Generally, the ME or macroelectrode, after identifying an area of interest in the brain 98, can be removed and the stimulator probe can be driven and guided along a similar or identical trajectory or axis relative to the removed ME or macroelectrode. The stimulator probe can be provided to electrically stimulate the selected region of the anatomy, either short term or long term.


With reference to FIG. 2, the drive system 12 can include any appropriate drive system. For example, the drive system 12 can comprise the microtargeting Drive® system produced by Fred Haer Corp., FHC Inc. of 9 Main Street, Bowdoinham Me. 04008, USA. The drive system 12 can be interconnected with various guide or support portions, such as the stereotactic head frame, the small-scale head frame 68, robotic devices, or guide devices, to drive various instruments into selected portions of the anatomy. For example, the stereotactic head frame can comprise any suitable stereotactic head frame known in the art, such as the Leksell Stereotactic System® provided by Elekta AB, and the small-scale head frame 68 can be any appropriate mechanism, such as the NEXFRAME™ sold by Medtronic Image Guided Neurologics of Minnesota, USA. The drive system 12 can be interconnected with the head frames 68 to position the drive system 12 at any appropriate location to drive various instruments 14 into a cranium 108. The drive system 12 can also include one or more tracking devices 56 (FIG. 1), which can be used with the navigation system 10 to determine a position of the drive system 12 relative to the patient 16.


With continued reference to FIG. 2, the head frame 68 can be positioned about an entry point, such as a burr hole 107, formed in the cranium 108. The head frame 68 can include various portions, but generally includes a base 200, a trajectory guide portion, movable portion, saddle or guide 202 and one or more tracking devices 56, which are generally referred to by reference numeral 56. The base 200 can define an aperture or opening 204 that allows the instruments 14 to pass through the base 200 into the cranium 108. One or more bone screws can be used to fix the base 200 relative to the patient 16. The trajectory guide portion, moveable base or guide 202 can be interconnected to the base 200 and the drive system 12 can be connected to the guide 202.


The guide 202 can move relative to the base 200 via a track or slot 208, 210 defined by the guide 202. The guide 202 can be guided or navigated to a selected location or orientation relative to the opening 204, which in turn, can align the drive system 12 into a selected location or orientation relative to the patient 16. The guide 202 can then be fixed in place via any appropriate mechanism, such as one or more locking screws 212, 214. Various markings can be provided on the guide 202 or the fixed base 200 to assist in obtaining a selected orientation of the guide 202 to the cranium 108.


For example, the guide 202 may be allowed to move through a predefined range of motion relative to the base 200. In a further example, the slots 208, 210 can be angled relative to the base 200, and the guide 202 can move or slide in an arcuate manner relative to the base 200 from about 30 to about 60 degrees relative to the base 200, as indicated by the arrow A in FIG. 2. In one example, the guide 202 can also rotate relative to the base 200, and thus, the anatomy. For example, the guide 202 can rotate about 360 degrees relative to the base 200, as indicated by the arrow B. As the guide 202 can move relative to the base 200, and thus, the patient 16, the guide 202 may limit to the amount of trajectory of the instrument 14 can achieve relative to the patient 16 from any one location of the guide 202. Thus, the position of the guide 202 relative to the patient 16 can ensure an appropriate trajectory can be aligned with the instrument 14.


The tracking devices 56 can be interconnected to the base 200 and the guide 202. For example, the base 200 can include a first tracking device 56d, which can be fixed to or positioned within the base 200. Optionally, the base 200 can include a second tracking device 56e and a third tracking device 56f. The guide 202 can include a fourth tracking device 56g, which can be fixed to or positioned within the guide 202. The tracking devices 56d-g can enable the tracking system 54 to track the position of the base 200 and the guide 202. It will be understood that the number of tracking devices 56 coupled to the base 200 and the guide 202 are merely exemplary, and any suitable number of tracking devices 56 can be coupled to the base 200 and guide 202.


The tracking devices 56d-g can comprise the electromagnetic tracking device 56a for use with the electromagnetic tracking system 54, the optical tracking device 56b for use with the optical tracking system 54a, or a combination thereof. If the tracking devices 56d-g comprise electromagnetic tracking devices 56a, the tracking devices 56d-g can also be used as field generating electromagnetic coils for various reasons, such as guiding the instrument 14. Thus, the tracking devices 56d-g can both generate a field and receive or sense a field generated by other electromagnetic coils, similar to the coils of the coil arrays 58, 60. It will also be understood, that in the case of electromagnetic tracking devices, the first tracking device 56d can either sense a field generated by another coil array 58, 60 or the first tracking device 56d can transmit a field to be sensed by the coil arrays 58, 60. In either case, the position of the first tracking device 56d in the base 200 can be determined and this determination can be used for navigation of the base 200. Similar methods can be applied to any other electromagnetic tracking devices 56, such as those on the guide 202, according to various embodiments. The various tracking devices 56d-g can be used with the navigation system 10, to determine a position of the various components of the head frame 68 or the drive system 12 relative the cranium 108, as further discussed herein. In addition, it will be understood that the use of tracking devices 56d-g is merely exemplary, as any suitable mechanism could be employed to position the head frame 68, such as fluoroscopy, for example.


As the base 200 can be fixed relative to the patient 16, the tracking devices 56 associated with the base 200 can be fixed relative to the patient 16, and thus, can form a dynamic reference frame. Moreover, the tracking devices 56 coupled to the base 200 can generate an electromagnetic field, which can be utilized during a portion of a procedure, such as for guiding the instrument 14. Additional detail regarding the placement and use of tracking devices 56 with the head frame 68 is disclosed in U.S. patent Ser. No. 12/110,666, entitled “Method and Apparatus for Performing a Navigated Procedure,” which is hereby incorporated by reference herein in its entirety.


The head frame 68 can be positioned on the patient 16 according to a plan or at a predetermined location. As discussed above, the patient image data 36 can be acquired of the patient 16 including the cranium 108. The patient image data 36 or any appropriate portion can be used with the workstation 46 to plan or predetermine the location for the head frame 68. Then, the navigation system 10 can be used to navigate and track the position of the head frame 68 relative to the patient 16 and the predetermined location for the head frame 68, as discussed further herein. The position of the head frame 68 can be displayed on the display 44 relative to the image data 42, which can include an image of the cranium 108, and instructions on how to position the head frame 68, as will be discussed.


With reference now to FIG. 3, a simplified block diagram schematically illustrates an exemplary navigation system 10 for implementing the control module 50. The navigation system 10 can include the tracking system 54, the head frame 68, a navigation control module 300 and the display 44. The navigation control module 300 can form a portion of the control module 50. The head frame 68 can include the base 200 and the guide 202. Each of the base 200 and the guide 202 can include the tracking device(s) 56.


The tracking system 54 can comprise the electromagnetic tracking system 54 or the optical tracking system 54b, and will generally be referred to as the tracking system 54. The tracking system 54 can receive start-up data 302 from the navigation control module 300. In the case of an electromagnetic tracking system 54, based on the start-up data 302, the tracking system 54 can set activation signal data 304 that can activate or drive the coil arrays 58, 60 to generate an electromagnetic field to which the tracking device(s) 56 coupled to the head frame 68 can receive. The tracking system 54 can also set tracking data 308 for the navigation control module 300, as will be discussed. The tracking data 308 can include data regarding the location or position of the tracking device(s) 56 coupled to the head frame 68 in the patient space as computed from data received from the tracking device(s) 56 or sensor data 310.


When the tracking device(s) 56 receive the electromagnetic field generated by the coil arrays 58, 60, the tracking device(s) 56 can transmit sensor data 310 indicative of a position of the tracking device 56 in the patient space to the tracking system 54. Based on the sensor data 310 received by the tracking system 54, the tracking system 54 can generate and set the tracking data 308 for the navigation control module 300.


The navigation control module 300 can receive the tracking data 308 from the tracking system 54. The navigation control module 300 can also receive patient image data 36 and plan data 312 as input. The patient image data 36 can comprise images of the anatomical structure of the patient 16 obtained from a pre- or intra-operative imaging device, such as the images obtained by the imaging device 18. The plan data 312 can comprise a trajectory for the instrument 14, which can be entered by the user 52, via the workstation 46, or could be loaded onto the workstation 46 from a portable electronic device, such as a portable data device.


Based on the tracking data 308, the patient image data 36 and the plan data 312, the navigation control module 300 can generate image data 42 for display on the display 44. The image data 42 can comprise the patient image data 36 superimposed with an icon 316 of the head frame 68 and a target 318 identified from the plan data 312. The image data 42 can also include a substantially real-time indication of the position and orientation of the head frame 68 in patient space, as indicated by at least one directional icon 315, shown in FIG. 4. The image data 42 could also comprise a schematic illustration of the instrument 14 within the anatomical structure of the patient 16, as shown in FIG. 1, or the icon 38 of the instrument 14 superimposed on the patient image data 36, etc.


For example, as shown in FIG. 4, the at least one directional icon 315 can comprise one or more arrows 317. For example, a first arrow 317a can denote a desired movement of the base 200 to align the base 200 relative to the anatomy such as the cranium 108, and a second arrow 317b can denote a desired movement of the guide 202 to align the guide 202 relative to the anatomy, in accordance with the surgical plan outlined in the plan data 312. In this regard, as the trajectory of the instrument 14 can be defined by the guide 202, the guide 202 and the base 200 need to be properly aligned in accordance with the plan data 312 to ensure the instrument 14 can reach the target 318, via the trajectory outlined in the surgical plan. By providing the at least one directional icon 315, the user 52 can easily determine how to move the base 200 and/or guide 202 relative to the anatomy and to each other to execute the surgical plan.


In one example, with reference to FIGS. 2 and 4, the user 52 can position the base 200 onto the cranium 108. When the position of the base 200 relative to the anatomy corresponds to the desired position of the base 200 denoted in the surgical plan, one or more bone screws can be used to couple the base 200 to the anatomy (FIG. 2). Next, the arrow 317a can indicate a desired amount and direction of rotation for the guide 202 about a first axis relative to the anatomy. Once the guide 202 has reached the desired position, the arrow 317a can diminish or disappear, thereby indicating that no additional rotation of the guide 202 is necessary. Then, the guide 202 can be coupled to the base 200, such that the slides 208, 210 can move the guide 202 about a second axis relative to the base 200 (FIG. 2). Based on the position of the guide 202 relative to the anatomy, the arrow 317b can indicate a desired movement or translation of the guide 202 relative to the base 200. Once the guide 202 has reached the desired position based on the surgical plan, the arrow 317b can diminish and the locking screws 212, 214 (FIG. 2) can be tightened to couple the guide 202 relative to the base 200 and the anatomy.


In addition, the at least one directional icon 315 can reduce in size as the base 200 and/or guide 202 reach their respective desired position relative to the anatomy and each other, which can further assure the user 52 that the head frame 68 is properly aligned relative to the anatomy. It should be noted, however, that although the at least one direction icon 315 is described and illustrated herein as comprising one or more arrows 317, the at least one directional icon 315 could comprise any suitable means to notify the user 52 of the necessary movement of the head frame 68 relative to the anatomy, such as a dialogue box, pop-up message, text, graphical representation of the desired final alignment, etc. Further, although the at least one directional icon 315 is described herein as reducing in size as the user 52 approaches the desired orientation for the head frame 68, any other suitable notification means could be employed, such as a change in color, pop-up message, audible signal, text, etc.


With reference now to FIG. 5, a dataflow diagram illustrates an exemplary control system that can be embedded within the control module 50. Various embodiments of the control system according to the present disclosure can include any number of sub-modules embedded within the control module 50. The sub-modules shown may be combined and/or further partitioned to similarly determine the orientation or position of the head frame 68 within the patient space based on the signals generated by the tracking device(s) 56. In various embodiments, the control module 50 includes the tracking system 54 that can implement a tracking control module 320, and the workstation 46 that can implement the navigation control module 300. It should be noted, however, that the tracking control module 320 and the navigation control module 300 could be implemented on the workstation 46, if desired.


The tracking control module 320 can receive as input the start-up data 302 from the navigation control module 300 and sensor data 310 from the tracking device(s) 56. Upon receipt of the start-up data 302, the tracking control module 320 can output the activation signal data 304 for the tracking device(s) 56. Upon receipt of the sensor data 310, the tracking control module 320 can set the tracking data 308 for the navigation control module 300. As discussed, the tracking data 308 can include data regarding the positions of the base 200 and the guide 202.


The navigation control module 300 can receive as input the tracking data 308, the patient image data 36 and the plan data 312. Based on the tracking data 308 and the plan data 312, the navigation control module 300 can determine the appropriate patient image data 36 for display on the display 44, and can output the tracking data 308, the plan data 312 and the patient image data 36 as image data 42. Further, the navigation control module 300 can output notification data 314 to the display 44 if the position of the head frame 68 does not correspond with the plan data 312. The notification data 314 can comprise the at least one directional icon 315.


With reference now to FIG. 6, a flowchart diagram illustrates an exemplary method performed by the control module 50. At block 400, the method can acquire the plan data 312. The plan data 312 can be acquired by the workstation 46 from an input from the user 52, but could be acquired from another control module. At block 402, the method can determine a desired position for the head frame 68 based on the plan data 312. At block 404, the method can determine the actual position of the head frame 68 based on the tracking data 308 received from the tracking device(s) 56. At decision block 406, the method can determine if the actual position of the head frame 68 is about equal to the desired position of the head frame 68. If the actual position of the headframe 68 is about equal to the desired position of the headframe 68, then the method can end.


Otherwise, at block 408, the method can output image data 42, which can include icon(s), such as the at least one directional icon 315, that can indicate a movement or manipulation of the headframe 68 required for the headframe 68 to reach the desired position. For example, the at least one directional icon 315 can indicate a desired rotation of the guide 202 or translation of the guide 202. Next, the method can go to block 404.


Therefore, the control module 50 of the present disclosure can provide a user 52, such as a surgeon, with an accurate representation of the position and orientation of the head frame 68, within the patient space to enable the user 52 to orientate the head frame 68 prior to the surgical procedure. In this regard, the use of the tracking device(s) 56 on the head frame 68 can enable an accurate depiction on the display 44 of the position of the head frame 68 relative to the anatomical structure of the patient 16. The use of the tracking device(s) 56 in combination with the at least one directional icon 315, can enable the display 44 to graphically indicate a desired manipulation or movement for the head frame 68 to easily enable the user 52 to align the head frame 68 in the desired position. As the head frame 68 in some instances can be used to guide the instrument(s) 14 into the anatomy, the proper alignment of the head frame 68 relative to the anatomy can ensure the instrument(s) 14 can be guided to reach the desired target within the anatomy.


One skilled in the art will understand that the processes and systems discussed above can be used in a surgical procedure. The processes and systems, however, are understood to not be limited to use during or with a surgical procedure. The systems and processes can be used to acquire information regarding inanimate objects, inform or build a database of information; plan a procedure; formulate teaching aids, etc. Registration of image space to physical space can be performed relative to any object in physical space, including a patient, an inanimate object, etc. Also, the registration can occur for any appropriate reason, which may or may not be a surgical procedure.


While specific examples have been described in the specification and illustrated in the drawings, it will be understood by those of ordinary skill in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from the scope of the present disclosure as defined in the claims. Furthermore, the mixing and matching of features, elements and/or functions between various examples is expressly contemplated herein so that one of ordinary skill in the art would appreciate from this disclosure that features, elements and/or functions of one example may be incorporated into another example as appropriate, unless described otherwise, above. Moreover, many modifications may be made to adapt a particular situation or material to the teachings of the present disclosure without departing from the essential scope thereof. Therefore, it is intended that the present disclosure not be limited to the particular examples illustrated by the drawings and described in the specification as the best mode presently contemplated for carrying out this disclosure, but that the scope of the present disclosure will include any embodiments falling within the foregoing description and the appended claims.


For example, while the navigation system 10 has been described as displaying the at least one directional icon 315, those of skill in the art will appreciate that the present disclosure, in its broadest aspects, may be constructed somewhat differently. In this regard, the navigation system 10 could use one or more audible cues to guide the user 52 in the movement or manipulation of the head frame 68. For example, the audible cues could increase as the user 52 reaches the desired position for the head frame 68.

Claims
  • 1. A system for positioning a guide relative to an anatomy comprising: a base adapted to be moved relative to the anatomy and thereafter immoveably coupled to the anatomy;a guide moveably coupled to the base, wherein the guide is configured to move relative to the base at least when the base is immoveably coupled to the anatomy;a first tracking device coupled to the base and a second tracking device coupled to the guide;a tracking system that tracks a position of the first tracking device and the second tracking device relative to the anatomy;a navigation system configured to independently determine a position of both the base and the guide relative to the anatomy based on the position of the first tracking device and the second tracking device and determine if the determined position of the base and the determined position of the guide are about equal to a desired position of the base and a desired position of the guide; anda display configured to display the desired position of the base and the guide superimposed on an image of the anatomy, a first directional icon superimposed on the image of the anatomy that graphically indicates a desired manipulation of the base required to move the base into the desired position of the base relative to the anatomy and a second directional icon superimposed on the image of the anatomy that graphically indicates a desired manipulation of the guide required to move the guide into the desired position of the guide relative to the anatomy.
  • 2. The system of claim 1, wherein the desired position of the base and the desired position the guide relative to the anatomy are both predetermined from a surgical plan.
  • 3. The system of claim 2, wherein the base and the guide each define a central bore, and the system further comprises: at least one instrument movable through the central bore of the base and the guide to perform a procedure on the anatomy.
  • 4. The system of claim 3, wherein the surgical plan includes a target site within the anatomy for the at least one instrument to perform the procedure, and the desired position of the base and the desired position the guide is aligned with a trajectory for the at least one instrument to travel to reach the target site.
  • 5. The system of claim 1, wherein the image of the anatomy is acquired by an imaging device selected from at least one of a fluoroscopy device, an O-arm device, a bi-plane fluoroscopy device, an ultrasound device, a computed tomography (CT) device, a multi-slice computed tomography (MSCT) device, a magnetic resonance imaging (MRI) device, a high frequency ultrasound (HFU) device, a positron emission tomography (PET) device, an optical coherence tomography (OCT) device, an intra-vascular ultrasound (IVUS) device, an intra-operative CT device, an intra-operative MRI device or combinations thereof.
  • 6. The system of claim 1, wherein the target site is selected from at least one of a surgical site, an anatomical site, a navigation area and combinations thereof.
  • 7. The system of claim 1, wherein the first tracking device comprises at least one optical tracking device to track at least one degree of freedom information.
  • 8. The system of claim 1, wherein the first tracking device and the second tracking device comprises at least one electromagnetic tracking device selected from the group including: an electromagnetic receiver tracking device, an electromagnetic transmitter tracking device or combinations thereof.
  • 9. The system of claim 1, wherein the guide is rotatable relative to the anatomy.
  • 10. The system of claim 9, wherein the first directional icon includes a third arrow that graphically represents an amount of rotation and a direction of rotation for the base, a second arrow that graphically represents an amount of translation and a direction of translation for the base, or combinations thereof; wherein the second directional icon includes a first arrow that graphically represents an amount of rotation and a direction of rotation for the guide, a second arrow that graphically represents an amount of translation and a direction of translation for the guide, or combinations thereof.
  • 11. The system of claim 4, wherein based on the surgical plan, the navigation system displays the second directional icon as a first arrow and a second arrow that indicates an alignment for the guide to allow the instrument to follow a trajectory through the base and the guide.
  • 12. A method for positioning a guide relative to an anatomy comprising: acquiring a surgical plan for a surgical procedure to be performed on the anatomy that includes a trajectory for an instrument;tracking a first tracking device coupled to a base relative to the anatomy;tracking a second tracking device coupled to a guide relative to the anatomy, the guide movable relative to the base;determining a position of the base and guide relative to the anatomy;comparing the position of the base and the guide relative to the anatomy with a desired base position for the base and a desired guide position for the guide in the surgical plan that enables the guide to guide the instrument along the trajectory; anddisplaying at least a first icon superimposed onto an image of the anatomy that indicates an amount and direction of a first manipulation required to move the guide into the desired position and at least a second icon superimposed onto an image of the anatomy that indicates an amount and direction of a second manipulation required to move the guide relative to the base into the desired position.
  • 13. The method of claim 12, further comprising: moving the guide within a track defined on the guide relative to the base.
  • 14. The method of claim 12, further comprising: tracking a position of the base;displaying a third icon superimposed onto an image of the anatomy that indicates an amount and direction of a third manipulation required to move the base into the desired base position;coupling the base to the anatomy with at least one bone screw when the base is in the desired base position; andcoupling the guide to the base with at least one locking screw to immoveably fix the guide relative to the base when the guide is in the desired position.
  • 15. The method of claim 12, further comprising: superimposing an icon representing the base onto the image of the anatomy.
  • 16. The method of claim 12 wherein the displayed at least first icon reduces in size as the guide approaches the desired location during the first manipulation, wherein the displayed at least second icon reduces in size as the guide approaches the desired location during the second manipulation, or combinations thereof.
  • 17. A system for positioning a guide relative to an anatomy comprising: a base adapted to be fixedly coupled at a selected location to the anatomy;a guide that moves relative to the base to guide at least one instrument into the anatomy, the guide operable to be coupled to a drive system;a first tracking device coupled to the base;a second tracking device coupled to the guide;a tracking system that tracks a position of the first tracking device and a position of the second device relative to the anatomy;a navigation control module that receives patient image data and surgical plan data, the surgical plan data including a desired base position for the base and a desired guide position for the guide relative to the anatomy;a navigation system that determines a position of the base and the guide relative to the anatomy based on the position of the first tracking device and the second tracking device, and determines if both the position of the base is about equal to the desired base position and if the position of the guide is about equal to the desired guide position; andwherein the navigation control module outputs image data that includes: a first arrow icon superimposed onto the patient image data that graphically represents an amount of movement required to move the base to the desired base position,a second arrow icon superimposed onto the patient image data that graphically represents an amount of rotation and a direction of rotation for the guide to move the guide to the desired guide position, anda third arrow icon superimposed onto the patient image data that graphically represents an amount of translation and a direction of translation for the guide to move the guide to the desired guide position, or combinations thereof.
  • 18. The system of claim 17, further comprising: a display that displays the image data.
  • 19. The system of claim 17, wherein the surgical plan data includes a target site within the anatomy, and the image data further comprises an icon of the target site superimposed onto the patient image data.
  • 20. The system of claim 17, wherein the first arrow icon reduces in size as the base approaches the desired position; wherein the second arrow icon and the third arrow icon also reduce in size as the guide approaches the desired position.
  • 21. The system of claim 1, wherein at least one of the first directional icon or the second directional icon includes at least one of a change in color, a pop-up message, an audible signal, or a text display.
  • 22. The system of claim 1, further comprising: a drive system coupled to the guide to be moved to a selected location and/or orientation relative to the anatomy by movement of the guide.
  • 23. The system of claim 10, wherein the base is moveable independently of the guide and the first directional icon illustrates a direction of movement of the base to reach the desired base position independent of movement of the guide; wherein the first directional icon and the second directional icon are the same icon and shown sequentially regarding movement of the base or movement of the guide.
  • 24. The system of claim 10, wherein the first directional icon is removed when the base reaches the desired base position and the second directional icon is removed when the guide reaches the desired guide position.
  • 25. The method of claim 12, wherein at least one of the first icon or the second icon diminishes or disappears as the guide moves towards the desired guide position, and wherein the third icon diminishes or disappears as the base moves towards the desired base position.
  • 26. The method of claim 14, wherein the base is moved independently of moving the guide; wherein the base is coupled to the anatomy prior to coupling the guide to the base.
US Referenced Citations (622)
Number Name Date Kind
1576781 Phillips Mar 1926 A
1735726 Bornhardt Nov 1929 A
2407845 Nemeyer Sep 1946 A
2650588 Drew Sep 1953 A
2697433 Sehnder Dec 1954 A
3016899 Stenvall Jan 1962 A
3017887 Heyer Jan 1962 A
3061936 Dobbeleer Nov 1962 A
3073310 Mocarski Jan 1963 A
3109588 Polhemus et al. Nov 1963 A
3294083 Alderson Dec 1966 A
3367326 Frazier Feb 1968 A
3439256 Kahne Apr 1969 A
3577160 White May 1971 A
3614950 Rabey Oct 1971 A
3644825 Davis, Jr. et al. Feb 1972 A
3674014 Tillander Jul 1972 A
3702935 Carey et al. Nov 1972 A
3704707 Halloran Dec 1972 A
3821469 Whetstone et al. Jun 1974 A
3868565 Kuipers Feb 1975 A
3941127 Froning Mar 1976 A
3983474 Kuipers Sep 1976 A
4017858 Kuipers Apr 1977 A
4037592 Kronner Jul 1977 A
4052620 Brunnett Oct 1977 A
4054881 Raab Oct 1977 A
4117337 Staats Sep 1978 A
4173228 Van Steenwyk et al. Nov 1979 A
4182312 Mushabac Jan 1980 A
4202349 Jones May 1980 A
4228799 Anichkov et al. Oct 1980 A
4256112 Kopf et al. Mar 1981 A
4262306 Renner Apr 1981 A
4287809 Egli et al. Sep 1981 A
4298874 Kuipers Nov 1981 A
4314251 Raab Feb 1982 A
4317078 Weed et al. Feb 1982 A
4319136 Jinkins Mar 1982 A
4328548 Crow et al. May 1982 A
4328813 Ray May 1982 A
4339953 Iwasaki Jul 1982 A
4341220 Perry Jul 1982 A
4346384 Raab Aug 1982 A
4358856 Stivender et al. Nov 1982 A
4368536 Pfeiler Jan 1983 A
4396885 Constant Aug 1983 A
4396945 DiMatteo et al. Aug 1983 A
4403321 Kruger Sep 1983 A
4418422 Richter et al. Nov 1983 A
4419012 Stephenson et al. Dec 1983 A
4422041 Lienau Dec 1983 A
4431005 McCormick Feb 1984 A
4485815 Amplatz et al. Dec 1984 A
4506676 Duska Mar 1985 A
4543959 Sepponen Oct 1985 A
4548208 Niemi Oct 1985 A
4571834 Fraser et al. Feb 1986 A
4572198 Codrington Feb 1986 A
4583538 Onik et al. Apr 1986 A
4584577 Temple Apr 1986 A
4608977 Brown Sep 1986 A
4613866 Blood Sep 1986 A
4617925 Laitinen Oct 1986 A
4618978 Cosman Oct 1986 A
4621628 Brudermann Nov 1986 A
4625718 Olerud et al. Dec 1986 A
4638798 Shelden et al. Jan 1987 A
4642786 Hansen Feb 1987 A
4645343 Stockdale et al. Feb 1987 A
4649504 Krouglicof et al. Mar 1987 A
4651732 Frederick Mar 1987 A
4653509 Oloff et al. Mar 1987 A
4659971 Suzuki et al. Apr 1987 A
4660970 Ferrano Apr 1987 A
4673352 Hansen Jun 1987 A
4688037 Krieg Aug 1987 A
4701049 Beckman et al. Oct 1987 A
4705395 Hageniers Nov 1987 A
4705401 Addleman et al. Nov 1987 A
4706665 Gouda Nov 1987 A
4709156 Murphy et al. Nov 1987 A
4710708 Rorden et al. Dec 1987 A
4719419 Dawley Jan 1988 A
4722056 Roberts et al. Jan 1988 A
4722336 Kim et al. Feb 1988 A
4723544 Moore et al. Feb 1988 A
4727565 Ericson Feb 1988 A
RE32619 Damadian Mar 1988 E
4733969 Case et al. Mar 1988 A
4737032 Addleman et al. Apr 1988 A
4737794 Jones Apr 1988 A
4737921 Goldwasser et al. Apr 1988 A
4742356 Kuipers May 1988 A
4742815 Ninan et al. May 1988 A
4743770 Lee May 1988 A
4743771 Sacks et al. May 1988 A
4745290 Frankel et al. May 1988 A
4750487 Zanetti Jun 1988 A
4753528 Hines et al. Jun 1988 A
4761072 Pryor Aug 1988 A
4764016 Johansson Aug 1988 A
4771787 Wurster et al. Sep 1988 A
4779212 Levy Oct 1988 A
4782239 Hirose et al. Nov 1988 A
4788481 Niwa Nov 1988 A
4791934 Brunnett Dec 1988 A
4793355 Crum et al. Dec 1988 A
4794262 Sato et al. Dec 1988 A
4797907 Anderton Jan 1989 A
4803976 Frigg et al. Feb 1989 A
4804261 Kirschen Feb 1989 A
4805615 Carol Feb 1989 A
4809694 Ferrara Mar 1989 A
4821200 Oberg Apr 1989 A
4821206 Arora Apr 1989 A
4821731 Martinelli et al. Apr 1989 A
4822163 Schmidt Apr 1989 A
4825091 Breyer et al. Apr 1989 A
4829373 Leberl et al. May 1989 A
4836778 Baumrind et al. Jun 1989 A
4838265 Cosman et al. Jun 1989 A
4841967 Chang et al. Jun 1989 A
4845771 Wislocki et al. Jul 1989 A
4849692 Blood Jul 1989 A
4860331 Williams et al. Aug 1989 A
4862893 Martinelli Sep 1989 A
4869247 Howard, III et al. Sep 1989 A
4875165 Fencil et al. Oct 1989 A
4875478 Chen Oct 1989 A
4884566 Mountz et al. Dec 1989 A
4889526 Rauscher et al. Dec 1989 A
4896673 Rose et al. Jan 1990 A
4905698 Strohl, Jr. et al. Mar 1990 A
4923459 Nambu May 1990 A
4931056 Ghajar et al. Jun 1990 A
4945305 Blood Jul 1990 A
4945914 Allen Aug 1990 A
4951653 Fry et al. Aug 1990 A
4955891 Carol Sep 1990 A
4961422 Marchosky et al. Oct 1990 A
4977655 Martinelli Dec 1990 A
4989608 Ratner Feb 1991 A
4991579 Allen Feb 1991 A
5002058 Martinelli Mar 1991 A
5005592 Cartmell Apr 1991 A
5013317 Cole et al. May 1991 A
5016639 Allen May 1991 A
5017139 Mushabac May 1991 A
5027818 Bova et al. Jul 1991 A
5030196 Inoue Jul 1991 A
5030222 Calandruccio et al. Jul 1991 A
5031203 Trecha Jul 1991 A
5042486 Pfeiler et al. Aug 1991 A
5047036 Koutrouvelis Sep 1991 A
5050608 Watanabe et al. Sep 1991 A
5054492 Scribner et al. Oct 1991 A
5057095 Fabian Oct 1991 A
5059789 Salcudean Oct 1991 A
5078140 Kwoh Jan 1992 A
5079699 Tuy et al. Jan 1992 A
5086401 Glassman et al. Feb 1992 A
5094241 Allen Mar 1992 A
5097839 Allen Mar 1992 A
5098426 Sklar et al. Mar 1992 A
5099845 Besz et al. Mar 1992 A
5099846 Hardy Mar 1992 A
5105829 Fabian et al. Apr 1992 A
5107839 Houdek et al. Apr 1992 A
5107843 Aarnio et al. Apr 1992 A
5107862 Fabian et al. Apr 1992 A
5109194 Cantaloube Apr 1992 A
5117836 Millar Jun 1992 A
5119817 Allen Jun 1992 A
5142930 Allen et al. Sep 1992 A
5143076 Hardy et al. Sep 1992 A
5152288 Hoenig et al. Oct 1992 A
5160337 Cosman Nov 1992 A
5161536 Vilkomerson et al. Nov 1992 A
5178164 Allen Jan 1993 A
5178621 Cook et al. Jan 1993 A
5186174 Schlondorff et al. Feb 1993 A
5187475 Wagener et al. Feb 1993 A
5188126 Fabian et al. Feb 1993 A
5190059 Fabian et al. Mar 1993 A
5193106 DeSena Mar 1993 A
5197476 Nowacki et al. Mar 1993 A
5197965 Cherry et al. Mar 1993 A
5198768 Keren Mar 1993 A
5198877 Schulz Mar 1993 A
5207688 Carol May 1993 A
5211164 Allen May 1993 A
5211165 Dumoulin et al. May 1993 A
5211176 Ishiguro et al. May 1993 A
5212720 Landi et al. May 1993 A
5214615 Bauer May 1993 A
5219351 Teubner et al. Jun 1993 A
5222499 Allen et al. Jun 1993 A
5224049 Mushabac Jun 1993 A
5228442 Imran Jul 1993 A
5230338 Allen et al. Jul 1993 A
5230623 Guthrie et al. Jul 1993 A
5233990 Barnea Aug 1993 A
5237996 Waldman et al. Aug 1993 A
5249581 Horbal et al. Oct 1993 A
5251127 Raab Oct 1993 A
5251635 Dumoulin et al. Oct 1993 A
5253647 Takahashi et al. Oct 1993 A
5255680 Darrow et al. Oct 1993 A
5257636 White Nov 1993 A
5257998 Ota et al. Nov 1993 A
5261404 Mick et al. Nov 1993 A
5265610 Darrow et al. Nov 1993 A
5265611 Hoenig et al. Nov 1993 A
5269759 Hernandez et al. Dec 1993 A
5271400 Dumoulin et al. Dec 1993 A
5273025 Sakiyama et al. Dec 1993 A
5274551 Corby, Jr. Dec 1993 A
5279309 Taylor et al. Jan 1994 A
5285787 Machida Feb 1994 A
5291199 Overman et al. Mar 1994 A
5291889 Kenet et al. Mar 1994 A
5295483 Nowacki et al. Mar 1994 A
5297549 Beatty et al. Mar 1994 A
5299253 Wessels Mar 1994 A
5299254 Dancer et al. Mar 1994 A
5299288 Glassman et al. Mar 1994 A
5300080 Clayman et al. Apr 1994 A
5305091 Gelbart et al. Apr 1994 A
5305203 Raab Apr 1994 A
5306271 Zinreich et al. Apr 1994 A
5307072 Jones, Jr. Apr 1994 A
5309913 Kormos et al. May 1994 A
5315630 Sturm et al. May 1994 A
5316024 Hirschi et al. May 1994 A
5318025 Dumoulin et al. Jun 1994 A
5320111 Livingston Jun 1994 A
5325728 Zimmerman et al. Jul 1994 A
5325873 Hirschi et al. Jul 1994 A
5329944 Fabian et al. Jul 1994 A
5330485 Clayman et al. Jul 1994 A
5333168 Fernandes et al. Jul 1994 A
5353795 Souza et al. Oct 1994 A
5353800 Pohndorf et al. Oct 1994 A
5353807 DeMarco Oct 1994 A
5359417 Muller et al. Oct 1994 A
5368030 Zinreich et al. Nov 1994 A
5371778 Yanof et al. Dec 1994 A
5375596 Twiss et al. Dec 1994 A
5377678 Dumoulin et al. Jan 1995 A
5383454 Bucholz Jan 1995 A
5385146 Goldreyer Jan 1995 A
5385148 Lesh et al. Jan 1995 A
5386828 Owens et al. Feb 1995 A
5389101 Heilbrun et al. Feb 1995 A
5391199 Ben-Haim Feb 1995 A
5394457 Leibinger et al. Feb 1995 A
5394875 Lewis et al. Mar 1995 A
5397329 Allen Mar 1995 A
5398684 Hardy Mar 1995 A
5399146 Nowacki et al. Mar 1995 A
5400384 Fernandes et al. Mar 1995 A
5402801 Taylor Apr 1995 A
5408409 Glassman et al. Apr 1995 A
5413573 Koivukangas May 1995 A
5417210 Funda et al. May 1995 A
5419325 Dumoulin et al. May 1995 A
5423334 Jordan Jun 1995 A
5425367 Shapiro et al. Jun 1995 A
5425382 Golden et al. Jun 1995 A
5426683 O'Farrell, Jr. et al. Jun 1995 A
5426687 Goodall et al. Jun 1995 A
5427097 Depp Jun 1995 A
5429132 Guy et al. Jul 1995 A
5433198 Desai Jul 1995 A
RE35025 Anderton Aug 1995 E
5437277 Dumoulin et al. Aug 1995 A
5443066 Dumoulin et al. Aug 1995 A
5443489 Ben-Haim Aug 1995 A
5444756 Pai et al. Aug 1995 A
5445144 Wodicka et al. Aug 1995 A
5445150 Dumoulin et al. Aug 1995 A
5445166 Taylor Aug 1995 A
5446548 Gerig et al. Aug 1995 A
5447154 Cinquin et al. Sep 1995 A
5448610 Yamamoto et al. Sep 1995 A
5453686 Anderson Sep 1995 A
5456718 Szymaitis Oct 1995 A
5457641 Zimmer et al. Oct 1995 A
5458718 Venkitachalam Oct 1995 A
5464446 Dreessen et al. Nov 1995 A
5469847 Zinreich et al. Nov 1995 A
5478341 Cook et al. Dec 1995 A
5478343 Ritter Dec 1995 A
5480422 Ben-Haim Jan 1996 A
5480439 Bisek et al. Jan 1996 A
5483961 Kelly et al. Jan 1996 A
5484437 Michelson Jan 1996 A
5485849 Panescu et al. Jan 1996 A
5487391 Panescu Jan 1996 A
5487729 Avellanet et al. Jan 1996 A
5487757 Truckai et al. Jan 1996 A
5490196 Rudich et al. Feb 1996 A
5494034 Schlondorff et al. Feb 1996 A
5503416 Aoki et al. Apr 1996 A
5513637 Twiss et al. May 1996 A
5514146 Lam et al. May 1996 A
5515160 Schulz et al. May 1996 A
5517990 Kalfas et al. May 1996 A
5531227 Schneider Jul 1996 A
5531520 Grimson et al. Jul 1996 A
5531673 Helenowski Jul 1996 A
5542938 Avellanet et al. Aug 1996 A
5543951 Moehrmann Aug 1996 A
5546940 Panescu et al. Aug 1996 A
5546949 Frazin et al. Aug 1996 A
5546951 Ben-Haim Aug 1996 A
5551429 Fitzpatrick et al. Sep 1996 A
5558091 Acker et al. Sep 1996 A
5558638 Evers et al. Sep 1996 A
5566681 Manwaring et al. Oct 1996 A
5568384 Robb et al. Oct 1996 A
5568809 Ben-haim Oct 1996 A
5571109 Bertagnoli et al. Nov 1996 A
5572999 Funda et al. Nov 1996 A
5573533 Strul Nov 1996 A
5575794 Walus et al. Nov 1996 A
5575798 Koutrouvelis Nov 1996 A
5583909 Hanover Dec 1996 A
5588430 Bova et al. Dec 1996 A
5590215 Allen Dec 1996 A
5592939 Martinelli Jan 1997 A
5595193 Walus et al. Jan 1997 A
5596228 Anderton et al. Jan 1997 A
5600330 Blood Feb 1997 A
5603318 Heilbrun et al. Feb 1997 A
5611025 Lorensen et al. Mar 1997 A
5617462 Spratt Apr 1997 A
5617857 Chader et al. Apr 1997 A
5619261 Anderton Apr 1997 A
5622169 Golden et al. Apr 1997 A
5622170 Schulz Apr 1997 A
5627873 Hanover et al. May 1997 A
5628315 Vilsmeier et al. May 1997 A
5630431 Taylor May 1997 A
5636644 Hart et al. Jun 1997 A
5638819 Manwaring et al. Jun 1997 A
5640170 Anderson Jun 1997 A
5642395 Anderton et al. Jun 1997 A
5643268 Vilsmeier et al. Jul 1997 A
5645065 Shapiro et al. Jul 1997 A
5646524 Gilboa Jul 1997 A
5647361 Damadian Jul 1997 A
5662111 Cosman Sep 1997 A
5664001 Tachibana et al. Sep 1997 A
5674296 Bryan et al. Oct 1997 A
5676673 Ferre et al. Oct 1997 A
5681260 Ueda et al. Oct 1997 A
5682886 Delp et al. Nov 1997 A
5682890 Kormos et al. Nov 1997 A
5690108 Chakeres Nov 1997 A
5690117 Gilbert Nov 1997 A
5694945 Ben-Haim Dec 1997 A
5695500 Taylor et al. Dec 1997 A
5695501 Carol et al. Dec 1997 A
5696500 Diem Dec 1997 A
5697377 Wittkampf Dec 1997 A
5702406 Vilsmeier et al. Dec 1997 A
5711299 Manwaring et al. Jan 1998 A
5713946 Ben-Haim Feb 1998 A
5715822 Watkins et al. Feb 1998 A
5715836 Kliegis et al. Feb 1998 A
5718241 Ben-Haim et al. Feb 1998 A
5727552 Ryan Mar 1998 A
5727553 Saad Mar 1998 A
5729129 Acker Mar 1998 A
5730129 Darrow et al. Mar 1998 A
5730130 Fitzpatrick et al. Mar 1998 A
5732703 Kalfas et al. Mar 1998 A
5733259 Valcke et al. Mar 1998 A
5735278 Hoult et al. Apr 1998 A
5735814 Elsberry et al. Apr 1998 A
5738096 Ben-Haim Apr 1998 A
5740802 Nafis et al. Apr 1998 A
5740808 Panescu et al. Apr 1998 A
5741214 Ouchi et al. Apr 1998 A
5742394 Hansen Apr 1998 A
5744953 Hansen Apr 1998 A
5748767 Raab May 1998 A
5749362 Funda et al. May 1998 A
5749835 Glantz May 1998 A
5752513 Acker et al. May 1998 A
5755725 Druais May 1998 A
RE35816 Schulz Jun 1998 E
5758667 Slettenmark Jun 1998 A
5762064 Polvani Jun 1998 A
5767669 Hansen et al. Jun 1998 A
5767699 Bosnyak et al. Jun 1998 A
5767960 Orman Jun 1998 A
5769789 Wang et al. Jun 1998 A
5769843 Abela et al. Jun 1998 A
5769861 Vilsmeier Jun 1998 A
5772594 Barrick Jun 1998 A
5772661 Michelson Jun 1998 A
5775322 Silverstein et al. Jul 1998 A
5776064 Kalfas et al. Jul 1998 A
5782765 Jonkman Jul 1998 A
5787886 Kelly et al. Aug 1998 A
5792055 McKinnon Aug 1998 A
5795294 Luber et al. Aug 1998 A
5797849 Vesely et al. Aug 1998 A
5799055 Peshkin et al. Aug 1998 A
5799099 Wang et al. Aug 1998 A
5800352 Ferre et al. Sep 1998 A
5800535 Howard, III Sep 1998 A
5802719 O'Farrell, Jr. et al. Sep 1998 A
5803089 Ferre et al. Sep 1998 A
5807252 Hassfeld et al. Sep 1998 A
5810008 Dekel et al. Sep 1998 A
5810728 Kuhn Sep 1998 A
5810735 Halperin et al. Sep 1998 A
5820553 Hughes Oct 1998 A
5823192 Kalend et al. Oct 1998 A
5823958 Truppe Oct 1998 A
5824048 Tuch Oct 1998 A
5828725 Levinson Oct 1998 A
5828770 Leis et al. Oct 1998 A
5829444 Ferre et al. Nov 1998 A
5831260 Hansen Nov 1998 A
5833608 Acker Nov 1998 A
5834759 Glossop Nov 1998 A
5836954 Heilbrun et al. Nov 1998 A
5840024 Taniguchi et al. Nov 1998 A
5840025 Ben-Haim Nov 1998 A
5843076 Webster, Jr. et al. Dec 1998 A
5848967 Cosman Dec 1998 A
5851183 Bucholz Dec 1998 A
5865846 Bryan et al. Feb 1999 A
5868674 Glowinski et al. Feb 1999 A
5868675 Henrion et al. Feb 1999 A
5871445 Bucholz Feb 1999 A
5871455 Ueno Feb 1999 A
5871487 Warner et al. Feb 1999 A
5873822 Ferre et al. Feb 1999 A
5882304 Ehnholm et al. Mar 1999 A
5884410 Prinz Mar 1999 A
5889834 Vilsmeier et al. Mar 1999 A
5891034 Bucholz Apr 1999 A
5891157 Day et al. Apr 1999 A
5904691 Barnett et al. May 1999 A
5907395 Schulz et al. May 1999 A
5913820 Bladen et al. Jun 1999 A
5920395 Schulz Jul 1999 A
5921992 Costales et al. Jul 1999 A
5923727 Navab Jul 1999 A
5928248 Acker Jul 1999 A
5938603 Ponzi Aug 1999 A
5938694 Jaraczewski et al. Aug 1999 A
5947980 Jensen et al. Sep 1999 A
5947981 Cosman Sep 1999 A
5950629 Taylor et al. Sep 1999 A
5951475 Gueziec et al. Sep 1999 A
5951571 Audette Sep 1999 A
5954647 Bova et al. Sep 1999 A
5954796 McCarty et al. Sep 1999 A
5957844 Dekel et al. Sep 1999 A
5967980 Ferre et al. Oct 1999 A
5967982 Barnett Oct 1999 A
5968047 Reed Oct 1999 A
5970499 Smith et al. Oct 1999 A
5971997 Guthrie et al. Oct 1999 A
5976156 Taylor et al. Nov 1999 A
5980535 Barnett et al. Nov 1999 A
5983126 Wittkampf et al. Nov 1999 A
5987349 Schulz Nov 1999 A
5987960 Messner et al. Nov 1999 A
5999837 Messner et al. Dec 1999 A
5999840 Grimson et al. Dec 1999 A
6001130 Bryan et al. Dec 1999 A
6006126 Cosman Dec 1999 A
6006127 Van Der Brug et al. Dec 1999 A
6013087 Adams et al. Jan 2000 A
6014580 Blume et al. Jan 2000 A
6016439 Acker Jan 2000 A
6019725 Vesely et al. Feb 2000 A
6024695 Taylor et al. Feb 2000 A
6026316 Kucharczyk et al. Feb 2000 A
6050724 Schmitz et al. Apr 2000 A
6059718 Taniguchi et al. May 2000 A
6063022 Ben-Haim May 2000 A
6064904 Yanof et al. May 2000 A
6071288 Carol et al. Jun 2000 A
6073043 Schneider Jun 2000 A
6076008 Bucholz Jun 2000 A
6096050 Audette Aug 2000 A
6104944 Martinelli Aug 2000 A
6118845 Simon et al. Sep 2000 A
6122538 Sliwa, Jr. et al. Sep 2000 A
6122541 Cosman et al. Sep 2000 A
6131396 Duerr et al. Oct 2000 A
6139183 Graumann Oct 2000 A
6147480 Osadchy et al. Nov 2000 A
6149592 Yanof et al. Nov 2000 A
6156067 Bryan et al. Dec 2000 A
6161032 Acker Dec 2000 A
6165181 Heilbrun et al. Dec 2000 A
6167296 Shahidi Dec 2000 A
6172499 Ashe Jan 2001 B1
6175756 Ferre et al. Jan 2001 B1
6178345 Vilsmeier et al. Jan 2001 B1
6194639 Botella et al. Feb 2001 B1
6198794 Peshkin et al. Mar 2001 B1
6201387 Govari Mar 2001 B1
6203497 Dekel et al. Mar 2001 B1
6206885 Ghahremani et al. Mar 2001 B1
6211666 Acker Apr 2001 B1
6223067 Vilsmeier et al. Apr 2001 B1
6233476 Strommer et al. May 2001 B1
6246231 Ashe Jun 2001 B1
6259942 Westermann et al. Jul 2001 B1
6273896 Franck et al. Aug 2001 B1
6283951 Flaherty et al. Sep 2001 B1
6285902 Kienzle, III et al. Sep 2001 B1
6298262 Franck et al. Oct 2001 B1
6314310 Ben-Haim et al. Nov 2001 B1
6332089 Acker et al. Dec 2001 B1
6341231 Ferre et al. Jan 2002 B1
6348058 Melkent et al. Feb 2002 B1
6351659 Vilsmeier Feb 2002 B1
6381485 Hunter et al. Apr 2002 B1
6385483 Uber, III et al. May 2002 B1
6390097 Chandra May 2002 B1
6402689 Scarantino et al. Jun 2002 B1
6406426 Reuss et al. Jun 2002 B1
6424856 Vilsmeier et al. Jul 2002 B1
6427314 Acker Aug 2002 B1
6428547 Vilsmeier et al. Aug 2002 B1
6434415 Foley et al. Aug 2002 B1
6437567 Schenck et al. Aug 2002 B1
6445943 Ferre et al. Sep 2002 B1
6464662 Raghavan et al. Oct 2002 B1
6470207 Simon et al. Oct 2002 B1
6474341 Hunter et al. Nov 2002 B1
6478802 Kienzle, III et al. Nov 2002 B2
6482182 Carroll et al. Nov 2002 B1
6484049 Seeley et al. Nov 2002 B1
6490475 Seeley et al. Dec 2002 B1
6491699 Henderson et al. Dec 2002 B1
6493573 Martinelli et al. Dec 2002 B1
6498944 Ben-Haim et al. Dec 2002 B1
6499488 Hunter et al. Dec 2002 B1
6516046 Frohlich et al. Feb 2003 B1
6516212 Bladen et al. Feb 2003 B1
6526415 Smith et al. Feb 2003 B2
6527443 Vilsmeier et al. Mar 2003 B1
6531152 Lerner et al. Mar 2003 B1
6549803 Raghavan et al. Apr 2003 B1
6551325 Neubauer et al. Apr 2003 B2
6567690 Giller et al. May 2003 B2
6584174 Schubert et al. Jun 2003 B2
6609022 Vilsmeier et al. Aug 2003 B2
6611700 Vilsmeier et al. Aug 2003 B1
6640128 Vilsmeier et al. Oct 2003 B2
6658396 Tang et al. Dec 2003 B1
6671538 Ehnholm et al. Dec 2003 B1
6694162 Hartlep et al. Feb 2004 B2
6701179 Martinelli et al. Mar 2004 B1
6740883 Stodilka et al. May 2004 B1
6755789 Stringer et al. Jun 2004 B2
6828966 Gavriliu et al. Dec 2004 B1
6901287 Davis et al. May 2005 B2
6973718 Sheppard, Jr. et al. Dec 2005 B2
6979348 Sundar Dec 2005 B2
6982282 Lambert et al. Jan 2006 B2
7011814 Suddarth et al. Mar 2006 B2
7047235 Yang et al. May 2006 B2
7072705 Miga et al. Jul 2006 B2
7092748 Valdes Sosa et al. Aug 2006 B2
7103399 Miga et al. Sep 2006 B2
7167180 Shibolet Jan 2007 B1
7194295 Vilsmeier Mar 2007 B2
7313430 Urquhart et al. Dec 2007 B2
7599730 Hunter et al. Oct 2009 B2
20010007918 Vilsmeier et al. Jul 2001 A1
20020095081 Vilsmeier et al. Jul 2002 A1
20030078485 Hartlep Apr 2003 A1
20030101081 Putnam et al. May 2003 A1
20030114752 Henderson et al. Jun 2003 A1
20030191408 Montgomery Oct 2003 A1
20040024309 Ferre et al. Feb 2004 A1
20040039259 Krause et al. Feb 2004 A1
20040068172 Nowinski et al. Apr 2004 A1
20040092809 DeCharms May 2004 A1
20040097806 Hunter et al. May 2004 A1
20040107210 Yang et al. Jun 2004 A1
20040138551 Hartlep et al. Jul 2004 A1
20040158313 Altman Aug 2004 A1
20040210124 Nowinski et al. Oct 2004 A1
20040215071 Frank et al. Oct 2004 A1
20040215162 Putz Oct 2004 A1
20040236554 Raghavan et al. Nov 2004 A1
20040240753 Hu et al. Dec 2004 A1
20050002918 Strauss et al. Jan 2005 A1
20050004617 Dawant et al. Jan 2005 A1
20050018885 Chen et al. Jan 2005 A1
20050031210 Shen et al. Feb 2005 A1
20050049486 Urquhart et al. Mar 2005 A1
20050070781 Dawant et al. Mar 2005 A1
20050084146 Watson et al. Apr 2005 A1
20050085714 Foley et al. Apr 2005 A1
20050085720 Jascob et al. Apr 2005 A1
20050101855 Miga et al. May 2005 A1
20050111621 Riker et al. May 2005 A1
20050148859 Miga et al. Jul 2005 A1
20050171558 Abovitz et al. Aug 2005 A1
20050245814 Anderson et al. Nov 2005 A1
20060017749 McIntyre et al. Jan 2006 A1
20060084867 Tremblay et al. Apr 2006 A1
20060182321 Hu et al. Aug 2006 A1
20060217733 Plassky et al. Sep 2006 A1
20070021668 Boese et al. Jan 2007 A1
20080081982 Simon et al. Apr 2008 A1
Foreign Referenced Citations (107)
Number Date Country
964149 Mar 1975 CA
3042343 Jun 1982 DE
3508730 Sep 1986 DE
3717871 Dec 1988 DE
3831278 Mar 1989 DE
3838011 Jul 1989 DE
4213426 Oct 1992 DE
4225112 Dec 1993 DE
4233978 Apr 1994 DE
19715202 Oct 1998 DE
19751761 Oct 1998 DE
19832296 Feb 1999 DE
19747427 May 1999 DE
10085137 Nov 2002 DE
0062941 Oct 1982 EP
0119660 Sep 1984 EP
0155857 Sep 1985 EP
0319844 Jun 1989 EP
0326768 Aug 1989 EP
0350996 Jan 1990 EP
0419729 Apr 1991 EP
0427358 May 1991 EP
0456103 Nov 1991 EP
0469966 Feb 1992 EP
0581704 Feb 1994 EP
0651968 May 1995 EP
0655138 May 1995 EP
0894473 Feb 1999 EP
0908146 Apr 1999 EP
0930046 Jul 1999 EP
1306050 May 2003 EP
1344187 Sep 2003 EP
1396233 Mar 2004 EP
1406203 Apr 2004 EP
1442715 Aug 2004 EP
1474782 Nov 2004 EP
1597701 Nov 2005 EP
1603076 Dec 2005 EP
1691687 Aug 2006 EP
1692633 Aug 2006 EP
1692657 Aug 2006 EP
1713015 Oct 2006 EP
2417970 Sep 1979 FR
2618211 Jan 1989 FR
2094590 Sep 1982 GB
2164856 Apr 1986 GB
62327 Jun 1983 JP
61-94639 Oct 1984 JP
63240851 Oct 1988 JP
2765738 Apr 1991 JP
3267054 Nov 1991 JP
WO-8809151 Dec 1988 WO
WO-8905123 Jun 1989 WO
WO-9005494 May 1990 WO
WO-9103982 Apr 1991 WO
WO-9104711 Apr 1991 WO
WO-9107726 May 1991 WO
WO-9203090 Mar 1992 WO
WO-9206645 Apr 1992 WO
WO-9404938 Mar 1994 WO
WO-9423647 Oct 1994 WO
WO-9424933 Nov 1994 WO
WO-9507055 Mar 1995 WO
WO-9611624 Apr 1996 WO
WO-9632059 Oct 1996 WO
WO-9736192 Oct 1997 WO
WO-9749453 Dec 1997 WO
WO-9808554 Mar 1998 WO
WO-9838908 Sep 1998 WO
WO-9915097 Apr 1999 WO
WO-9921498 May 1999 WO
WO-9923956 May 1999 WO
WO-9926549 Jun 1999 WO
WO-9927839 Jun 1999 WO
WO-9929253 Jun 1999 WO
WO-9933406 Jul 1999 WO
WO-9937208 Jul 1999 WO
WO-9938449 Aug 1999 WO
WO-9952094 Oct 1999 WO
WO-9960939 Dec 1999 WO
WO-0007652 Feb 2000 WO
WO-0010034 Feb 2000 WO
WO-0130437 May 2001 WO
WO-0243003 May 2002 WO
WO-02093292 Nov 2002 WO
WO-02097735 Dec 2002 WO
WO-02098292 Dec 2002 WO
WO-03039600 May 2003 WO
WO-03060827 Jul 2003 WO
WO-2004077359 Sep 2004 WO
WO-2004096018 Nov 2004 WO
WO-2005002444 Jan 2005 WO
WO-2005048844 Jun 2005 WO
WO-2005052838 Jun 2005 WO
WO-2005057493 Jun 2005 WO
WO-2005057498 Jun 2005 WO
WO-2005084542 Sep 2005 WO
WO-2005096227 Oct 2005 WO
WO-2005111931 Nov 2005 WO
WO-2006011850 Feb 2006 WO
WO-2006017053 Feb 2006 WO
WO-2006017392 Feb 2006 WO
WO-2006028416 Mar 2006 WO
WO-2006028474 Mar 2006 WO
WO-2006069250 Jun 2006 WO
WO-2006083236 Aug 2006 WO
WO-2006088429 Aug 2006 WO
Related Publications (1)
Number Date Country
20100081914 A1 Apr 2010 US