Computer assisted surgical navigation system for spine procedures

Information

  • Patent Grant
  • 12178523
  • Patent Number
    12,178,523
  • Date Filed
    Friday, March 25, 2022
    3 years ago
  • Date Issued
    Tuesday, December 31, 2024
    4 months ago
Abstract
A surgical system for computer assisted navigation during surgery, includes at least one processor that obtains a 3D radiological representation of a targeted anatomical structure of a patient and a set of fiducials of a registration fixture. The operations attempt to register locations of the set of fiducials in the 3D radiological representation to a 3D imaging space tracked by a camera tracking system. Based on determining one of the fiducials of the set has a location that was not successfully registered to the 3D imaging space, the operations display at least one view of the 3D radiological representation with a graphical overlay indicating the fiducial has not been successfully registered to the 3D imaging space, receive user-supplied location information identifying where the fiducial is located in the 3D radiological representation, and register the location of the fiducial to the 3D imaging space based on the user-supplied location information.
Description
TECHNICAL FIELD

The present disclosure relates to computer assisted navigation of equipment and operations during surgery.


BACKGROUND

Surgical operating rooms can contain a diverse range of medical equipment, which can include computer assisted surgical navigation systems, medical imaging devices (e.g., computerized tomography (CT) scanners, fluoroscopy imaging, etc.), surgical robots, etc.


A computer assisted surgical navigation system can provide a surgeon with computerized visualization of the present pose of a surgical tool relative to medical images of a patient's anatomy. Camera tracking systems for computer assisted surgical navigation typically use a set of cameras to track pose of a reference array on a surgical tool, which is being positioned by a surgeon during surgery, relative to a patient reference array (also “dynamic reference base” (DRB)) attached to a patient. The reference arrays allow the camera tracking system to determine a pose of the surgical tool relative to anatomical structure imaged by a medical image of the patient and relative to the patient. The surgeon can thereby use real-time visual feedback of the pose to navigate the surgical tool during a surgical procedure on the patient.


Many surgical workflows using computer assisted surgical navigation systems require image scans, such as CT scans or magnetic resonance imaging scans, during the surgical procedure. Perpendicular scan slices (axial, sagittal, and coronal) may be used to enable operators to visualize the patient's anatomy alongside the relative poses of surgical instruments. The surgical workflows may be challenging for surgeons and other surgical team members to recall, interpret, and follow under the time constraints and other pressures of a surgery environment. Improved surgical workflows and computer implemented operations to reduce the workload on the surgery team and to ensure processes and best practices are followed.


SUMMARY

Some embodiments of the present disclosure are directed to a surgical system for computer assisted navigation during surgery. The surgical system includes at least one processor that is operative to obtain a three-dimensional (3D) radiological representation of a targeted anatomical structure of a patient and a set of fiducials of a registration fixture. The operations attempt to register locations of the set of fiducials in the 3D radiological representation to a 3D imaging space tracked by a camera tracking system. Based on determining one of the fiducials of the set has a location that was not successfully registered to the 3D imaging space, the operations display at least one view of the 3D radiological representation with a graphical overlay indicating the fiducial has not been successfully registered to the 3D imaging space, receive user-supplied location information identifying where the fiducial is located in the 3D radiological representation, and register the location of the fiducial to the 3D imaging space based on the user-supplied location information.


In some further embodiments, the operation to attempt to register locations of the set of fiducials in the 3D radiological representation to the 3D imaging space tracked by the camera tracking system, includes to obtain, from at least one camera of the camera tracking system, an optical image of a reference array fixated to the patient. The reference array includes a set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space. The operations attempt to register locations of a pattern of the set of optical markers to locations of a pattern of the set of fiducials in the 3D radiological representation, and identify any of the optical markers of the reference array that are not successfully registered to any of the fiducials of the registration fixture.


In some further embodiments, the operations include to display a virtual implant device as an overlay on a view of the 3D radiological representation of the targeted anatomical structure. The operations display a graphical indication of a trajectory of the virtual implant device representing an implantation trajectory of the virtual implant device into the targeted anatomical structure. The operations update pose of the graphical indication of the trajectory of the virtual implant device displayed in the view of the 3D radiological representation, to track steering inputs received through a user interface of the surgical system. The operations store as a planned trajectory of the virtual implant device, a user-designated one of the poses of the graphical indication of the trajectory.


In some further embodiments, prior to obtaining the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture, the operations include to obtain, from at least one camera of the camera tracking system, optical images of a reference array fixated to the patient and of a registration fixture attached to a radiological imaging device. The reference array includes a first set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space, and the registration fixture includes a second set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space. The operations obtain a fluoroscopic image of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture. The operations determine whether a first condition is satisfied based on a defined number of the optical markers in the first set being detected by the at least one camera of the camera tracking system in the 3D imaging space. The operations determine whether a second condition is satisfied based on a defined number of the optical markers in the second set being detected by the at least one camera of the camera tracking system in the 3D imaging space. The operations determine whether a third condition is satisfied based on a defined number of the set of fiducials of the registration fixture being visible in the fluoroscopic image. When one of the first, second, and third conditions is not satisfied, the operations display an indication of the not satisfied one of the first, second, and third conditions, and inhibit capture by a radiological imaging process of the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture. In contrast, when each of the first, second, and third conditions are satisfied, the operations enable capture by the radiological imaging process of the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture.


Still other surgical systems, methods, and computer program products according to embodiments of the inventive subject matter will be or become apparent to one with skill in the art upon review of the following drawings and detailed description. It is intended that all such surgical systems, methods, and computer program products be included within this description, be within the scope of the present inventive subject matter, and be protected by the accompanying claims. Moreover, it is intended that all embodiments disclosed herein can be implemented separately or combined in any way and/or combination.





DESCRIPTION OF THE DRAWINGS

The accompanying drawings, which are included to provide a further understanding of the disclosure and are incorporated in a constitute a part of this application, illustrate certain non-limiting embodiments of inventive concepts. In the drawings:



FIG. 1 illustrates the ExcelsiusHub components according to some embodiments;



FIG. 2 illustrates an example user interface of the CONFIGURE tab during Procedure Setup according to some embodiments;



FIG. 3 illustrates an example user interface of the WORKFLOW tab during Procedure Setup according to some embodiments;



FIG. 4 illustrates an example user interface of the VERIFY tab during Procedure Setup according to some embodiments;



FIG. 5 illustrates an instrument verification using a verification divot according to some embodiments;



FIG. 6 illustrates an example user interface during instrument verification according to some embodiments;



FIG. 7 illustrates a process for securing a Dynamic Reference Base (DRB) to a patient attachment instrument according to some embodiments;



FIG. 8 illustrates a process for tightening a DRB knob using a clamp driver according to some embodiments;



FIG. 9 illustrates a placement of a DRB and surveillance marker according to some embodiments;



FIG. 10 illustrates a process for removing a quattro spike with removal tool according to some embodiments;



FIG. 11 illustrates a process for securing a pivoting arm starburst on a registration fixture according to some embodiments;



FIG. 12 illustrates a process for securing a registration fixture to a patient attachment instrument post according to some embodiments;



FIG. 13 illustrates a complete intra-op CT registration assembly and release button according to some embodiments;



FIG. 14 illustrates an example user interface of the IMAGE tab during an intra-operative CT imaging workflow according to some embodiments;



FIG. 15 illustrates an example user interface of the manual registration screen during an intra-operative CT imaging workflow according to some embodiments;



FIG. 16 illustrates an example user interface of the PLAN tab during an intra-operative CT imaging workflow according to some embodiments;



FIG. 17 illustrates an example user interface of the NAVIGATE tab during an intra-operative CT imaging workflow according to some embodiments;



FIG. 18 illustrates an example user interface of the IMAGE tab during a pre-operative CT imaging workflow according to some embodiments;



FIG. 19 illustrates an example user interface of the PLAN tab during a pre-operative CT imaging workflow according to some embodiments;



FIG. 20 illustrates an example user interface of the NAVIGATE tab during a pre-operative CT imaging workflow according to some embodiments;



FIG. 21 illustrates a fluoroscopy registration fixture attached to an image intensifier according to some embodiments;



FIG. 22 illustrates an example user interface of image acquisition during a pre-operative CT imaging workflow according to some embodiments;



FIG. 23 illustrates an example user interface after level selection during a pre-operative CT imaging workflow according to some embodiments;



FIG. 24 illustrates an example user interface after successful registration during a pre-operative CT imaging workflow according to some embodiments;



FIG. 25 illustrates an example user interface of a real-time instrument/implant trajectory and navigated instruments during a pre-operative CT imaging workflow according to some embodiments;



FIG. 26 illustrates an example user interface of the IMAGE tab during a fluoroscopic imaging workflow according to some embodiments;



FIG. 27 illustrates an example user interface of image acquisition during a fluoroscopic imaging workflow according to some embodiments;



FIG. 28 illustrates an example user interface of the PLAN tab during a fluoroscopic imaging workflow according to some embodiments;



FIG. 29 illustrates an example user interface of the NAVIGATE tab during a fluoroscopic imaging workflow according to some embodiments;



FIG. 30 is an overhead view of personnel optionally wearing extended reality (XR) headsets during a surgical procedure in a surgical room that includes a camera tracking system for navigated surgery and a surgical robot system for robotic assistance and configured in accordance with some embodiments;



FIG. 31 illustrates a block diagram of surgical system which includes a camera tracking system and navigation system, and further optionally includes a surgical robot and XR headset, which are each operative in accordance with some embodiments; and



FIGS. 32-34 illustrate operations that may be performed by a surgical system in accordance with some embodiments.





DETAILED DESCRIPTION

The following discussion is presented to enable a person skilled in the art to make and use embodiments of the present disclosure. Various modifications to the illustrated embodiments will be readily apparent to those skilled in the art, and the principles herein can be applied to other embodiments and applications without departing from embodiments of the present disclosure. Thus, the embodiments are not intended to be limited to embodiments shown, but are to be accorded the widest scope consistent with the principles and features disclosed herein. The following detailed description is to be read with reference to the figures, in which like elements in different figures have like reference numerals. The figures, which are not necessarily to scale, depict selected embodiments and are not intended to limit the scope of the embodiments. Skilled artisans will recognize the examples provided herein have many useful alternatives and fall within the scope of the embodiments.


System Overview


The ExcelsiusHub, by Globus Medical, Inc. (hereinafter “Globus Medical” or “Globus”), enables real-time surgical navigation and visualization using radiological patient images, and guides compatible surgical instruments to a precise location and trajectory based on implant planning or provides visualization for assisting with free-hand navigation. The software reformats patient-specific CT images acquired before or during surgery, or fluoroscopic images acquired during surgery, and displays them on-screen based on the preferred perspective. Prior to operating, the surgeon may create, store, access, and simulate instrument and other trajectories relative to patient anatomy captured in the CT images. During surgery, the system recognizes the instrument in use and aids the user by way of free-hand navigation to place implants with consistent accuracy. ExcelsiusHub tracks the position of surgical instruments in or on the patient anatomy, and continuously updates the instrument position on these images. The surgery is performed by the surgeon, using various Globus Medical specialized surgical instruments.


Although various embodiments are described in the context of operational extensions to Excelsius system products and other Globus Medical products, these and other embodiments are not limited thereto and can be used with any surgical procedure navigation system.


Device Description


The ExcelsiusHub is a surgical procedure navigation system that enables real-time surgical visualization using radiological patient images (e.g., preoperative CT, intraoperative CT and/or fluoroscopy), a patient dynamic reference base, and an advanced camera tracking system. The system is mapped based on the registration between the virtual patient (points on the patient images) and the physical patient (corresponding points on the patient's anatomy). Once this registration is created, the software displays the relative position of a tracked instrument on the patient images. This visualization can help guide the surgeon's planning and approach for implant placement and other surgical procedures. The patient's scan coupled with the registration provides guidance assistance to the surgeon when using the system independently for free hand navigation or can provide robotic guidance and align the end effector when used with the ExcelsiusGPS Robotic System. During surgery, the system uses the camera tracking system to track the position of compatible instruments, including the end effector on the robotic arm, in or on the patient anatomy and continuously updates the instrument position on patient images utilizing optical tracking. System software may be responsible for navigation functions, data storage, network connectivity, user management, case management, and safety functions. The ExcelsiusHub surgical instruments are typically non-sterile, re-usable instruments that can be operated manually.


The ExcelsiusHub freehand instrumentation includes registration instruments, patient reference instruments, and implant-specific surgical instruments. The system can also be used with the ExcelsiusGPS robotic system, including actively tracked end-effectors. Registration instruments have incorporated arrays of reflective markers, which are used to track patient anatomy and surgical instruments and implants. Components include the verification probe, surveillance marker, surgical instrument arrays, intra-op CT registration fixture, fluoroscopy registration fixture, and dynamic reference base (DRB). Patient reference instruments are either clamped or driven into any appropriate rigid anatomy that is considered safe and provides a point of rigid fixation for the DRB. Surgical instruments are used to prepare the implant site or implant the device, and include awls, drills, drivers, taps, and probes.


Indications for Use


The ExcelsiusHub can be used as an aid for precisely locating anatomical structures in open or percutaneous procedures and for precisely positioning compatible surgical instruments or implants during surgery. The ExcelsiusHub is indicated for any medical condition in which the use of stereotactic surgery may be appropriate, and where reference to a rigid anatomical structure, such as the spine or pelvis, can be identified relative to a CT, X-ray, and/or MRI based model of the anatomy. The ExcelsiusHub supports pre-operative CT, intra-operative CT, and/or intra-operative Fluoroscopic procedures.


Further System Overview


The ExcelsiusHub Visualization System provides standalone navigation and guidance for previously cleared posterior fixation and interbody implant placement. The ExcelsiusHub can operate with the ExcelsiusGPS Robotic System to provide camera and tracking system functionality. The ExcelsiusHub can also provide a universal viewing station and registration operations for intraoperative mobile CT systems. Excelsius enabling technologies can be supported and communicate through the ExcelsiusHub, which can provide desired integral components for operating rooms.


Hardware


A camera tracking system 200 The ExcelsiusHub components can be divided into three subassemblies as shown in FIG. 1: a camera arm assembly 100, a display assembly 120, and a housing base assembly 130, according to some embodiments. The camera arm assembly 100 includes spaced apart stereo cameras 102 and an articulating arm 104. The housing base assembly 130 includes a processing platform with at least one processor and memory, an input and/or output user interface, and communication circuitry which is configured to communicate through wired (e.g., connector panel 107) and/or wireless connections with other system components. The cameras 102 operating with the processing platform are configured to detect pose of markers, e.g., reflective markers, on instruments, the patient (e.g., Dynamic Reference Base (DRB)), a surgical robot, etc.


Instruments


Navigated Instruments


The ExcelsiusHub works with all preexisting Globus navigated, arrayed instrumentation. This includes the drills, awls, probes, taps, and drivers for placement of Globus Screws and the dilators, disc preparation instruments (curettes, Cobb elevators, rasps, scrapers, etc.), trials, and inserters for navigated placement of Globus Interbodies. Each instrument is identified by a unique array pattern that is recognized by the camera.


Patient Attachment and Registration Instruments


The ExcelsiusHub may be used with existing patient fixation instruments and the current DRB.


Registration Fixtures


The ExcelsiusHub may be used with existing intra-op registration fixtures and the fluoroscopy registration.


System Software


The software operating with the ExcelsiusHub may include existing Spine software available on the ExcelsiusGPS Robotic System. The system software may be responsible for all navigation functions, data storage, network connectivity, user management, case management, and safety functions.


Applications


Spine surgical procedures are supported by the ExcelsiusHub System. The CONFIGURE tab displays procedure types. The spine procedural steps are the same as seen on ExcelsiusGPS Robotic System as the software is the same on both pieces of hardware.


Spine Procedures


Various Spinal surgical procedures Spinal surgical procedures supported by the ExcelsiusHub System are listed in the table below.












Supported Spine Procedures










Procedures
Patient Position







Posterior Cervical
Prone



Posterior Thoracic
Prone



Anterolateral Thoracic
Lateral



Posterior Lumbar
Prone



Lateral Lumbar
Lateral



Lateral Lumbar Interbody Fusion
Lateral



Transforaminal Lumbar Interbody Fusion
Prone



Posterior Lumbar Interbody Fusion
Prone










Globus spinal implant systems that are compatible with the ExcelsiusHub System include those listed in the table below.












Compatible Spinal Implant Systems

















CREO Stabilization System



REVERE Stabilization System



REVOLVE Stabilization System



ELLIPSE Occipito-Cervico-Throacic Spinal System



QUARTEX Occipito-Cervico-Throacic Spinal System



SUSTAIN Spacers (Oblique, Small) (Posterior Interbody)



ALTERA Expandable Spacers (Posterior Interbody)



RISE Expandable Spacers (Posterior Interbody)



CALIBER Expandable Spacers (Posterior Interbody)



RISE-L Spacers (Lateral Interbody)



CALIBER-L Expandable Spacers (Lateral Interbody)



ELSA Integrated Expandable Spacers (Lateral Interbody)











Procedure Setup


Configure Tab


After selecting a case, the CONFIGURE tab is displayed on the monitor. An example user interface is shown in FIG. 2 according to some embodiments. Using the CONFIGURE tab select the surgeon from among a list of registered surgeons, the imaging modality and the procedure type. The imaging modality can include intra-operative CT, fluoroscopy, and pre-operative CT. Click the right arrows to advance to the next tab.


Workflow Tab


Using the WORKFLOW tab, select the desired stage of the procedure (e.g., interbody, or screws) in the desired order of operation (e.g., interbody first). For each stage, select the imaging modality, interbody implant system, and desired interbody level on the anatomical model, which may include cervical, thoracic, lumbar, and sacroiliac. Add stages to the workflow by clicking the “Add Stage” button. Click “Verify Instruments” to proceed to advance to the next tab. An example user interface is shown in FIG. 3 according to some embodiments.


Verify Tab


The VERIFY tab displays navigation details including visibility, location and verification status of the instruments selected on the WORKFLOW tab. Verification is used to indicate whether one or more instruments may be damaged, e.g., during handling. All instruments with arrays should be verified prior to use, either with a verification adapter, instrument, implant, or dilator, as appropriate.


An example user interface is shown in FIG. 4 according to some embodiments. The VERIFY tab shows CAMERA VIEW and INSTRUMENT STATUS listing a set of identified instruments.


CAMERA VIEW can be operationally updated in real-time from the perspective of the camera with, e.g., color circles being displayed indicating instrument location. A solid colored circle can displayed to indicate that the corresponding instrument is visible to the cameras 102, while a hollow circle indicates that it is not visible to the cameras 102. Size of the colored circle can be dynamically updated to change size to indicate distance from the physical cameras 102. In one embodiment, the circle size is adapted to grow larger as the instrument is moved closer to the cameras 102 and smaller as the instrument is moved away from the cameras 102. An ideal distance from the cameras 102 may be approximately 2 meters or 6 feet, in accordance with some embodiments.


INSTRUMENT STATUS lists each instrument and its verification status, with corresponding color circles to identify each instrument. Verification status symbols are shown in FIG. 4, in accordance with some embodiments.


Instrument Verification


With reference to FIG. 5, each instrument can be verified by placing the tip of the instrument into a verification dot provided at a known location on another piece of equipment, e.g., on a registered instruments array or on a registered robot art of a surgical robot system, or by placing a verification adapter to be verified into a verification divot provided at a known location on the instrument, according to some embodiments. Divots can be formed at defined locations on navigated instrument arrays for use in verifying the instruments. If using a ExcelsiusGPS Robotic System, a divot can also be located on the top surface of the End Effector or other known location which can be tracked by the cameras 102. Next, ensure both instruments are visible to the cameras 102 and held steady. A pop-up screen then is displayed on the VERIFY tab to indicate operations of the verification progress.


As shown in FIG. 6, once verification is complete, verification status is indicated on the screen with the tip error displayed in mm, according to some embodiments. If verification has failed a corresponding indication 600 is displayed, e.g., a red crossed circle is displayed and verification operations can be repeated until verification is completed successfully. In contrast, successful verification is indicated by another displayed indication 601, e.g., a green circle. When all instruments are successfully verified, the user can click the right arrow to advance to the next tab.


Patient Attachment Instruments


Patient attachment instruments are secured to rigid bony anatomy neighboring the surgical site. A user selects the desired instrument. Patient attachment instruments should be placed no more than 185 mm from the center of the surgical site to maintain accuracy, in accordance with some embodiments. Bone clamps are clamped onto anatomical structures such as the spinous process, iliac crest, long bone, or any rigid bony structure that can be safely clamped. Quattro spikes are inserted into the iliac crest or a long bone. Rod attachments are secured to an existing spinal rod, e.g., 4.5 mm to 6.35 mm in diameter. Example recommended anatomic locations for the various patient attachment instruments are described in the table below according to some embodiments.












Patient Attachment Instruments - Recommended Anatomic Locations













Recommended Patient



Patient
Patient Attachment
Attachment Instrument


Spine Procedures
Position
Instrument
Location





Posterior Cervical
Prone
Bone Clamp
Spinous Process C2-T3




Rod Attachment
Existing Rod


Posterior Thoracic
Prone
Bone Clamp
Spinous Process T1-L1




Rod Attachment
Existing Rod


Anterolateral Thoracic
Lateral
Bone Clamp
Spinous Process T1-L1


Posterior Lumbar
Prone
Quattro Spike
Iliac Crest




Low Profile Quattro Spike
Iliac Crest




Bone Clamp
Spinous Process T12-L5




Rod Attachment
Existing Rod


Lateral Lumbar
Lateral
Quattro Spike
Iliac Crest




Low Profile Quattro Spike
Iliac Crest




Bone Clamp
Spinous Process T12-L5




Rod Attachment
Existing Rod










Dynamic Reference Base Insertion


A user positions a compression clamp on the Dynamic Reference Base (DRB) over the patient attachment instrument and tightens the knob, as shown in FIG. 7 according to some embodiments. If needed, a clamp driver can be used to further tighten the DRB knob, as shown in FIG. 8 according to some embodiments. The user positions the reflective markers on the DRB in the direction of the cameras 102. Care should be taken with initial placement of the patient reference instrument as to not interfere with the surgical procedure. Following navigation, the patient attachment instrument is removed.


Surveillance Marker


A surveillance marker is inserted into rigid bony anatomy to enable the camera tracking system to use the cameras 102 to track the relative distance to the DRB, e.g., to identify unwanted shifts in the DRB during the procedure. FIG. 9 illustrates a placement of a DRB 900 spaced apart from a surveillance marker 902 according to some embodiments.


Surveillance markers may be inserted into the iliac crest or long bone, or may be attached to the spinous process using a bone clamp. The user verifies that the clamp is rigidly secured. The surveillance marker should be placed no more than 185 mm from the Dynamic Reference Base in some embodiments. Example recommended anatomic locations for the various surveillance markers are described in the table below, according to some embodiments.












Surveillance Marker - Recommended Anatomic Locations













Recommended



Patient

Surveillance Marker


Procedures
Position
Marker
Location





Posterior Cervical
Prone
On Bone Clamp
Spinous Process C2-T3


Posterior Thoracic
Prone
Single
Iliac Crest




On Bone Clamp
Spinous Process T1-L1


Anterolateral Thoracic
Lateral
On Bone Clamp
Spinous Process T1-L1


Posterior Lumbar
Prone
Single
Iliac Crest




On Bone Clamp
Spinous Process T12-L5


Lateral Lumbar
Lateral
Single
Iliac Crest




On Bone Clamp
Spinous Process T12-L5









The user attaches a disposable reflective marker to the marker post of the surveillance marker 902. [0082] The user attaches the impaction cap, designed to fit over the reflective marker sphere 904, onto the surveillance marker 902. The user inserts the surveillance marker 902 into rigid bony anatomy near the surgical site, and may gently impact with a mallet. The user removes the impaction cap. The user removes the reflective marker 904 prior to using the removal tool. To use a bone clamp with the marker, the user attaches a disposable marker onto the tip of the bone clamp. The user may use the clamp driver to secure the bone clamp. The user then verifies that the clamp is rigidly secured.


Removal


The quattro spikes and surveillance marker are removed from bony anatomy manually or using the removal tool. The bone clamp is removed by loosening the clamp with the clamp driver, attaching the removal tool and lifting up the bone clamp, as shown in FIG. 10 according to some embodiments.


Intra-Operative Ct Imaging Workflow


Intra-Op Ct Registration Fixture Setup


A pivoting arm starburst 1100 is placed over a starburst post of the registration fixture post 1102 of a registration fixture 1110 and rotated 90° to secure, as shown in FIG. 11 according to some embodiments. The registration fixture 1110 is positioned on a patient attachment instrument post 1202 and a compression clamp 1200 knob is tightened, as shown in FIG. 12 according to some embodiments. If needed, a clamp driver can be used to further tighten the knob on the compression clamp 1200. The completed assembly is shown in FIG. 13 according to some embodiments. To release the pivoting arm, a release button 1300 on the fixture is pushed, the pivoting arm 90° is rotated and pull up. The intra-op CT registration fixture has six degrees of freedom and can be moved by adjusting one of the three joints so that it is stable and hovering over the surgical site. Only the metal fiducials embedded in the fixture need to be in the 3D scan (not the reflective markers). It may be operationally helpful or necessary for the intra-op CT registration fixture to not move between image acquisition and performing an anatomical landmark check.


Loading the Image


Selection of the IMAGE tab shown in FIG. 14 displays the steps needed to load a CT scan image according to some embodiments. The image can be loaded from, e.g., a USB drive, hard drive, and/or a networked device. If the image is transferred via the Ethernet, it may automatically appear on the hard drive when the transfer is complete.


To view images on a USB drive, the USB drive is inserted into a USB port on the connector panel 107. To trigger loading of an image, the hard drive or USB drive icon can be selected followed by selection of the desired patient image. The right arrows can be selected to load the patient images and advance to the next tab.


Manual Registration


Automatic registration can be performed when loading images. However, if automatic registrations operations fail or otherwise not utilized, then a manual registration screen can be displayed to allow manual registration as shown in FIG. 15 according to some embodiments. The image on the left panel of the registration screen is a full scan with a depiction of the intra-operative CT.


A registration fixture and seven fiducials, in some embodiments of the fixture, should be visible below the image. Fiducials that are not registered need to be adjusted by the operator. On the screen, a fiducial is selected which is not yet registered; causing that image to then appear on the right. A colored circle 1500 is moved on the screen by a user until the user determines that it surrounds a displayed fiducial marker. The three small boxes 1510 shown in FIG. 15 at the bottom of the right panel show the x, y and z direction of the fiducial and all should be adjusted until the blue circle is centered. In the example embodiment of FIG. 15, a registered fiducial 1500 is displayed with a solid color filled-in center, another fiducial 1504 which is not visible to the system is displayed with a non-filled-in center, and yet another fiducial 1502 which is partially visible to the system is displayed with a partially filled-in center.


Ensure that all seven fiducials are properly identified by viewing the 3D model of the intra-op registration fixture. A fiducial may be deleted by selecting the delete icon on the right panel. Click the right arrows to confirm that the fiducials have been properly identified before proceeding to the next step.


Corresponding operations that may be performed are now described in the context of FIG. 32. FIG. 32 illustrates operations that may be performed by a surgical system in accordance with some embodiments. The surgical system includes at least one processor that may reside in a component of the computer platform 3600, such as the camera tracking system 10 and/or the navigation system 3604. Referring to FIG. 32, the operations include to obtain 3200 a three-dimensional (3D) radiological representation, e.g., CT scan(s) and/or fluoroscopy scan(s), of a targeted anatomical structure of a patient and a set of fiducials of a registration fixture, e.g., 900 in FIGS. 9 and/or 2100 in FIG. 21. The operations attempt 3202 to register locations of the set of fiducials in the 3D radiological representation to a 3D imaging space tracked by a camera tracking system, e.g., system 10 in FIG. 31. The registration may include attempting to correlate the pose (e.g., location and orientation) set of fiducials in the 3D radiological representation to the pose of optical markers detected in the 3D imaging space. Based on determining 3204 one of the fiducials of the set has a location that was not successfully registered to the 3D imaging space, the operations display 3206 at least one view of the 3D radiological representation with a graphical overlay indicating the fiducial has not been successfully registered to the 3D imaging space, receive 3208 user-supplied location information identifying where the fiducial is located in the 3D radiological representation, and register 3210 the location of the fiducial to the 3D imaging space based on the user-supplied location information.


The operation to receive 3208 user-supplied location information identifying where the fiducial is located in the 3D radiological representation, may include to display three orthogonal views of the fiducial in the 3D radiological representation and/or the 3D imaging space, and display a graphical object overlaid on an initial location in the three orthogonal views. The operation moves location of where the graphical object is displayed in the three orthogonal views responsive to input from the user through a user interface, and determine location of the fiducial in the 3D radiological representation and/or the 3D imaging space based on the location of where the graphical object is displayed in the three orthogonal views. The operation registers the location of the fiducial to the 3D imaging space is based on the determined location of the fiducial in the 3D radiological representation.


The operation to receive 3208 may display the graphical object overlaid on the initial location in the three orthogonal views, may include to determine the initial location to correspond to a predicted location of the fiducial based on relative locations of fiducials defined by a registration fixture template.


The operation to receive 3208 may further include to move the location where the graphical object is displayed in the three orthogonal views to track directional inputs received through the user interface of the surgical system.


The operation to attempt 3202 to register locations of the set of fiducials in the 3D radiological representation to the 3D imaging space tracked by the camera tracking system, may include to obtain, from at least one camera of the camera tracking system, an optical image of a reference array fixated to the patient. The reference array including a set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space. The reference array may also be connected to the registration fixture. The operation may then attempt to register locations of a pattern of the set of optical markers to locations of a pattern of the set of fiducials in the 3D radiological representation, and identify any of the optical markers of the reference array that are not successfully registered to any of the fiducials of the registration fixture.


Landmark Check


After registration has been completed, a landmark check can be performed to ensure that the registration was calculated successfully. Using the verification probe, an anatomical landmark or a fiducial is touched on the registration fixture to trigger verification that the corresponding location is shown on the system monitor. This process can be repeated using other, e.g., 2 to 3 other landmarks.


Corresponding operations that may be performed by the surgical system may include to track locations of a tool captured in video from a camera of the camera tracking system while the tool is being moved by a user toward one of the fiducials that was successfully registered to the 3D image space, and display updated representations of the tool according to the tracked locations in the 3D imaging space. The operations confirm registration accuracy of the one of the fiducials that was successfully registered to the 3D image space based on comparison of a designated one of the tracked locations of the tool to the location of the one of the fiducials registered to the 3D image space.


Removing Registration Fixture


The Intra-op CT Registration Fixture can then be removed while ensuring the patient attachment instrument does not move.


PLAN Tab


The PLAN tab allows the user to plan all screw insertion trajectories (e.g., 1600) and interbody placement (e.g., 1610) on the patient image, such as shown in the example user interface of FIG. 16 according to some embodiments. Implants are preloaded (e.g., with defined characteristics) in the system and displayed on the right-hand side of the screen, based on selections made in the PREPLAN tab.


To add an implant onto the planning page, a user can drag and drop the appropriate implant label on the image at the desired slice. The active plan is shown in a defined color. Details of the active screw plan are shown on the lower right of the screen, including screw family, diameter, and length. The right arrow can be selected to advance to the next tab once plans are complete for all screws.


Once the implant is dropped on the image, the implant planning features are used to adjust implant location by, e.g., dragging the implant image on the touch screen. The user selects or otherwise defines the specific implant size (width, length, height, lordosis) on the right panel of the screen.


Corresponding operations that may be performed are now described in the context of FIG. 33. FIG. 33 illustrates operations that may be performed by the surgical system in accordance with some embodiments. Referring to FIG. 33, the operations include to display 3300 a virtual implant device as an overlay on a view of the 3D radiological representation of the targeted anatomical structure, and display 3302 a graphical indication of a trajectory of the virtual implant device representing an implantation trajectory of the virtual implant device into the targeted anatomical structure. The operations update 3304 pose of the graphical indication of the trajectory of the virtual implant device displayed in the view of the 3D radiological representation, to track steering inputs received through a user interface of the surgical system. The operations store 3306 as a planned trajectory of the virtual implant device, a user-designated one of the poses of the graphical indication of the trajectory.


The operations may further include to display a set of implant devices which are selectable by a user for implant planning, and generate a graphical representation of the virtual implant device based on a template of one of the set of user-selectable implant devices which is selected by a user through the user interface.


NAVIGATE Tab


The NAVIGATE tab allows the user to visualize the navigated instrument trajectory and the planned trajectory with respect to patient anatomy. An example user interface is shown in FIG. 17 according to some embodiments. The desired implant label can be selected on the right of the screen.


The real-time instrument/implant trajectory 1700a, 1700b (actual plan) is displayed on the patient images, e.g., in orthogonal image slice views, along with the planned screw, allowing the user to confirm the desired trajectory. If the real-time trajectory is not acceptable, the user can return to the PLAN tab to select another trajectory. If the real-time trajectory is acceptable, the user inserts the screw according to the instrument's current trajectory to the desired depth.


Navigated instruments are displayed as they are advanced to the planned position. While navigating the instruments, the user repetitively observes the monitor and surgical site to ensure consistency between tactile and navigation feedback.


Corresponding operations by the surgical system may include to display the planned trajectory of virtual implant device as an overlay on the view of the 3D radiological representation of the targeted anatomical structure. The operations obtain, from at least one camera of the camera tracking system, optical images of a reference array fixated to a real-implant device corresponding to the virtual implant device, the reference array including a set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space. The operations track pose of the real-implant device in the 3D imaging space based on pose of the reference array in the optical images while the real implant device is being positioned by a user relative to the targeted anatomical structure of the patient. The operations display updated graphical representations of the real-implant device relative to the planned trajectory of the virtual implant device according to the tracked pose in the 3D imaging space.


Pre-Operative Ct Imaging Workflow


IMAGE Tab


Loading the Image


The IMAGE tab shows the steps needed to load a CT scan image. An example user interface is shown in FIG. 18 according to some embodiments. The image can be loaded from, e.g., a USB drive, hard drive, or networked device. If the image is transferred through the Ethernet, it may automatically appear on the hard drive when the transfer is complete.


To view images on a USB drive, the USB drive is inserted into the USB port on the connector panel. To load an image, the user selects the hard drive or USB drive icon and selects the desired patient image. The right arrows can be selected to load the patient images and advance to the next tab.


PLAN Tab


The PLAN tab allows the user to plan all screw trajectories and interbody placement on the patient image. An example user interface is shown in FIG. 19 according to some embodiments. Implants are preloaded (e.g., characteristics predefined in the system) on the right side of the screen, based on selections made in the PREPLAN tab.


To add an implant onto the planning page, a user may drag and drop the appropriate implant label on the image at the desired slice. The active plan is shown in a defined color. Details of the active screw plan are shown on the lower right of the screen, including screw family, diameter, and length. The right arrows can be selected to advance to the next tab once plans are complete for all screws.


Once the implant is dropped on the image, the implant planning features a performed to adjust implant location by, e.g., dragging the implant image on the touch screen. The specific implant size (width, length, height, lordosis) is selected or defined on the right panel of the screen.


NAVIGATE Tab


The NAVIGATE tab allows the user to visualize the navigated instruments and trajectory alignment with respect to patient anatomy, according to the implant plan.


Registration Setup


Another display screen, e.g., as shown in FIG. 20, highlights the three steps to complete before the fluoroscopy images can be taken to register the pre-operative CT image, according to some embodiments. The steps may be to insert the DRB, position the C-ARM of an C-arm imaging device, and register a surveillance marker with the camera navigation system. Animation may be used to visually depict the steps.


A Fluoroscopy Registration Fixture 2100 is attached to an image intensifier 2110 on the C-arm, as shown in FIG. 21 according to some embodiments, by turning the clamp clockwise until tight. New optical markers are installed on the fixture 2100 prior to orienting the fixture 2100 such that the optical markers are facing the cameras 102. A video capture cable is connected to the C-arm viewing station. A video capture USB cable is inserted into one of the USB ports on ExcelsiusHub connector panel 107.


The user may ensure that the Dynamic Reference Base is visible to the cameras 102 after the C-Arm is in place.


The surveillance marker is registered with the camera tracking system by, e.g., placing an instrument close to the reflective sphere 904 on the surveillance marker 902 but not touching. The box is then displayed in a defined color when it is activated.


The right arrows can be selected to advance to the next tab.


Registration


Operations acquire the intra-operative fluoroscopic images, one anteroposterior (AP) and one lateral for each level planned. The same image may be used for multiple levels.


In some embodiments, the operations verify that the following three conditions are met prior to enable acquisition of the images: 1) the DRB is visible by the cameras 102; 2) the Fluoroscopy Registration Fixture is visible by the cameras 102; and 3) a valid fluoroscopic image was taken.


An example user interface is shown in FIG. 22 according to some embodiments. Each of the three images on the left of the screen turn to a defined color when ready for image capture. When all three conditions are met, the intra-operative fluoroscopic image is acquired and then the CAPTURE button is selected to transfer the image to the system. Once both images are successfully captured, the spinal level on the right side of the screen displays a check mark. The right arrows can be selected to advance to the next tab.


An example user interface for selecting a desired level is shown in FIG. 23 according to some embodiments. The planned screw may be dragged-and-dropped onto the fluoroscopic images. A displayed graphical object, e.g., circle 2300a and/or 2300b, can be controlled, via user input through a user interface, to roughly position the screw within the vertebral body. The user input may correspond to a user touch-selecting or clicking on a desired location on the display and/or providing steering commands through a keyboard. The screw shank is confirmed to be positioned correctly, with the head and tail of the screws in the desired direction, and with the left/right correctly oriented. The register button can be selected when the confirmation is complete to allow registration.


A check mark is shown next to the active level when registration is successful. An example user interface is shown in FIG. 24 according to some embodiments. The right arrows are selected when registration is completed.


Corresponding operations that may be performed are now described in the context of FIG. 34. FIG. 34 illustrates operations that may be performed by the surgical system in accordance with some embodiments. Referring to FIG. 34, prior to obtaining the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture, the operations include to obtain 3400, from at least one camera of the camera tracking system, optical images of a reference array (e.g., DRB 900 in FIG. 9) fixated to the patient and of a registration fixture (e.g., fixture 2100 in FIG. 21) attached to a radiological imaging device, the reference array including a first set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space, and the registration fixture including a second set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space. The operations obtain 3402 a fluoroscopic image of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture. The operations determine 3404 whether a first condition is satisfied based on a defined number of the optical markers in the first set being detected by the at least one camera of the camera tracking system in the 3D imaging space. The operations determine 3406 whether a second condition is satisfied based on a defined number of the optical markers in the second set being detected by the at least one camera of the camera tracking system in the 3D imaging space. The operations determine 3408 whether a third condition is satisfied based on a defined number of the set of fiducials of the registration fixture being visible in the fluoroscopic image. A determination 3410 is made whether any of the three conditions is not satisfied. When one of the first, second, and third conditions is not satisfied, the operations display 3412 an indication of the not satisfied one of the first, second, and third conditions, and inhibit capture by a radiological imaging process of the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture. In contrast, when each of the first, second, and third conditions are satisfied, the operations enable 3414 capture by the radiological imaging process of the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture.


The operations may trigger capture of anteroposterior and lateral fluoroscopic images at a plurality of defined locations of the targeted anatomical structure of the patient based on determining each of the first, second, and third conditions are satisfied. The operations may then compute the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture based on the captured anteroposterior and lateral fluoroscopic images at the plurality of defined locations of the targeted anatomical structure of the patient.


Landmark Check


After registration has been completed, a landmark check, or verification, can be performed to operationally ensure that the registration was calculated successfully. Using the verification probe, touch an anatomical landmark and verify that the corresponding location is shown on the system monitor. This process may be repeated using, e.g., 2 to 3 landmarks.


Navigation


With reference to FIG. 25, the user selects the desired implant label on the right of the screen. The real-time instrument/implant trajectory (actual plan) is updated to be displayed on the patient images along with the planned screw, allowing the user to confirm the desired trajectory. If the real-time trajectory is not acceptable, the user can return to the PLAN tab to select another trajectory. If the real-time trajectory is acceptable, the user inserts the screw according to the instrument's current trajectory to the desired depth.


Navigated instruments are displayed as they are advanced to the planned position. While navigating the instruments, the user can repetitively observe the monitor and surgical site to ensure consistency between tactile and navigation feedback.


Corresponding operations by the surgical system may include to display the planned trajectory of virtual implant device as an overlay on the view of the 3D radiological representation of the targeted anatomical structure. The operations obtain, from at least one camera of the camera tracking system, optical images of a reference array fixated to a real-implant device corresponding to the virtual implant device, the reference array including a set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space. The operations track pose of the real-implant device in the 3D imaging space based on pose of the reference array in the optical images while the real implant device is being positioned by a user relative to the targeted anatomical structure of the patient. The operations display updated graphical representations of the real-implant device relative to the planned trajectory of the virtual implant device according to the tracked pose in the 3D imaging space.


Removing Registration Fixture


Carefully remove the Fluoroscopic Registration Fixture. Ensure the patient attachment instrument does not move.


Fluoroscopic Imaging Workflow


IMAGE Tab


Registration Setup


One screen shown in FIG. 26, which may substantially correspond to the screen of FIG. 20, highlights the three steps to complete before fluoroscopic images can be taken to register the patient, such as described above for FIG. 20.


The Fluoroscopy Registration Fixture can be attached to the image intensifier on the C-arm, as shown in FIG. 21 according to some embodiments, by turning the clamp clockwise until tight. Install new optical markers on the fixture prior to orienting the fixture such that the optical markers are facing the camera.


The user ensures that the Dynamic Reference Base is still visible to the cameras 102 after the C-Arm is in place. The surveillance marker may be registered by placing an instrument close to the reflective sphere on the surveillance marker but not touching. The box turns to a defined color when it is activated. The right arrows can be selected to advance to the next tab.


Image Acquisition


Intra-operative fluoroscopic images are acquired, such as one AP and one lateral.


In some embodiments, the operations verify that the following three conditions are met prior to enable acquisition of the images: 1) the DRB is visible by the cameras 102; 2) the Fluoroscopy Registration Fixture is visible by the cameras 102; and 3) a valid fluoroscopic image was taken.


An example user interface is shown in FIG. 27 according to some embodiments. Each of the three images on the left of the screen turn to a defined color when ready for image capture. When all three conditions are met, the intra-operative fluoroscopic image is acquired and then the CAPTURE button is selected to transfer the image to the system. Once both images are successfully captured, the level on the right side of the screen displays a check mark. Once the appropriate images have been loaded and selected, the right arrows can be selected to proceed.


Landmark Check


After registration has been completed, a landmark check, or verification, can be performed to operationally ensure that the registration was calculated successfully. Using the navigated verification probe, touch an anatomical landmark and verify that the corresponding location is shown on the system monitor. This process may be repeated using, e.g., 2 to 3 landmarks.


Removing Registration Fixture


The Fluoroscopic Registration Fixture is removed while ensuring the patient attachment instrument does not move.


PLAN Tab


The PLAN tab allows the user to plan all screw trajectories and interbody placement on the patient image. An example user interface is shown in FIG. 28 according to some embodiments. Implants are preloaded on the right-hand side of the screen, based on selections made in the PREPLAN tab.


To add an implant onto the planning page, the user may drag and drop the appropriate implant label on the image at the desired slice. The active plan is shown in a defined color. Details of the active screw plan are shown on the lower right of the screen, including screw family, diameter, and length. A user may select (e.g., click) on the right arrows to advance to the next tab once plans are complete for all screws.


Once the implant is dropped on the image, the implant planning features are used to adjust implant location by, e.g., dragging the implant image on the touch screen. The selects the specific implant size (width, length, height, lordosis) on the right panel of the screen. Alternatively or additionally, software of the planning system may perform automatic adjustment of the implant location so that the dropped implant satisfies one or more defined rules with respect to anatomy in the image.


NAVIGATE Tab


The NAVIGATE tab allows the user to visualize the navigated instrument trajectory and the planned trajectory with respect to patient anatomy. An example user interface is shown in FIG. 29 according to some embodiments.


The user selects the desired implant label on the right of the screen. The real-time instrument/implant trajectory (actual plan) is updated to be displayed on the patient images along with the planned screw, e.g., as graphical objects 2900a and 2900b, allowing the user to confirm the desired trajectory. If the real-time trajectory is not acceptable, the user can return to the PLAN tab to select another trajectory. If the real-time trajectory is acceptable, the user inserts the screw according to the instrument's current trajectory to the desired depth.


Navigated instruments are displayed as they are advanced to the planned position. While navigating the instruments, the repetitively observes the monitor and surgical site to ensure consistency between tactile and navigation feedback.


Example Surgical Room Equipment Layout and Operation



FIG. 30 is an overhead view of personnel optionally wearing extended reality (XR) headsets 3150a-3150b during a surgical procedure in a surgical room that includes a camera tracking system 10 for navigated surgery and optionally includes a surgical robot system for robotic assistance, and each of which is configured in accordance with some embodiments.


Referring to FIG. 30, the robot system 3000 may include, for example, a surgical robot 3102, one or more robotic arms 104, an end-effector 3112, for example, configured to attach to a joint manipulation arm, and an end-effector reference array which can include one or more tracking markers. The robot system 3000 may further include one or more displays. The DRB 900 includes a plurality of tracking markers and is adapted to be secured directly to a patient 3130 (e.g., to a bone of the patient 3130). Another reference array 3170 is attached or formed on an instrument, etc. The camera tracking system 10 can have any suitable configuration to move, orient, and support the tracking cameras 102 in a desired position, and may contain a computer operable to track pose of reference arrays.


The tracking cameras 102 may include any suitable camera or cameras, such as one or more infrared cameras (e.g., bifocal or stereophotogrammetric cameras), able to identify, for example, active and passive tracking markers for various reference arrays attached as the patient 3130 (e.g., DRB 900), end-effector 3112 (end-effector reference array), instrument(s) (e.g., instrument array 3170), extended reality (XR) headset(s) 3150a-3150b worn by a surgeon 3120 and/or a surgical assistant 3126, etc. in a given measurement volume viewable from the perspective of the tracking cameras 102. The tracking cameras 102 may track markers attached or formed on the robot arm 140 manipulated by a user (surgeon) and/or the robot system 3000. The tracking cameras 102 may scan the given measurement volume and detect light that is emitted or reflected from the reference arrays in order to identify and determine poses of the reference arrays in three-dimensions. For example, active reference arrays may include infrared-emitting markers that are activated by an electrical signal (e.g., infrared light emitting diodes (LEDs)), and passive reference arrays may include retro-reflective markers that reflect infrared light (e.g., they reflect incoming IR radiation into the direction of the incoming light), for example, emitted by illuminators on the tracking cameras 102 or other suitable device.


The XR headsets 3150a and 3150b (also referred to as an XR headset 150) may each include tracking cameras that can track poses of reference arrays within their camera field-of-views (FOVs) 3152 and 3154, respectively. Accordingly, as illustrated in FIG. 30, the poses of reference arrays attached to various objects can be tracked while in the FOVs 3152 and 3154 of the XR headsets 3150a and 3150b and/or a FOV 3600 of the tracking cameras 102.


An XR headset may be configured to augment a real-world scene with computer generated XR images. The XR headset may be configured to provide an augmented reality (AR) viewing environment by displaying the computer generated XR images on a see-through display screen that allows light from the real-world scene to pass therethrough for combined viewing by the user. Alternatively, the XR headset may be configured to provide a virtual reality (VR) viewing environment by preventing or substantially preventing light from the real-world scene from being directly viewed by the user while the user is viewing the computer generated AR images on a display screen. An XR headset can be configured to provide both AR and VR viewing environments. Thus, the term XR headset can referred to as an AR headset or a VR headset.



FIG. 30 illustrates a potential configuration for the placement of the camera tracking system 10 and the surgical robot system 3000 (when present) in an operating room environment. Computer-aided navigated surgery can be provided by the camera tracking system 10 controlling the XR headsets 3150a and 3150b, the display 110, and/or the display 3034 to display surgical procedure navigation information. Including the surgical robot system 3000 in an operating room is optional for computer-aided navigated surgery.


The camera tracking system 10 may use tracking information and other information from the camera tracking system 10 along with other tracking information and information from one or more XR headsets 3150a and 3150b, e.g., inertial tracking information and optical tracking information as well as (optional) microphone information. The XR headsets 3150a and 3150b operate to display visual information and may play-out audio information to the wearer. This information can be from local sources (e.g., the surgical robot 3102 and/or other operating room equipment), remote sources (e.g., patient medical image server), and/or other electronic equipment. The XR headsets 3150a and 3150b may be used to track poses of instruments, patient references, and/or a robot end-effector in 6 degrees-of-freedom (6DOF), and may track the hands of the wearer. The XR headsets 3150a and 3150b may also operate to track hand poses and gestures to enable gesture-based interactions with “virtual” buttons and interfaces displayed through the XR headsets 3150a and 3150b and may interpret hand or finger pointing or gesturing as triggering operation of various defined commands. Additionally, the XR headsets 3150a and 3150b may have a 1-10× magnification digital color camera sensor called a digital loupe.


An “outside-in” machine vision navigation bar (tracking cameras 102) may track pose of the joint manipulation arm using monochrome and/or color camera(s). The machine vision navigation bar generally has a more stable view of the environment because it does not move as often or as quickly as the XR headsets 3150a and 3150b tend to move while positioned on wearers' heads. The patient reference array 900 is generally rigidly attached to the patient with stable pitch and roll relative to gravity. This local rigid patient reference 900 can serve as a common reference for reference frames relative to other tracked arrays, such as a reference array on the end-effector 3112, instrument reference array 3170, and reference arrays on the XR headsets 3150a and 3150b.


In some embodiments, one or more of the XR headsets 3150a and 3150b are minimalistic XR headsets that display local or remote information but include fewer sensors and are therefore more lightweight.


The robot system 3000 may be positioned near or next to patient 3130. The tracking camera 102 may be separated from the robot system 3000 and positioned at the foot of patient 3130. This location allows the tracking camera 102 to have a direct visual line of sight to the surgical field 208. It is contemplated that the robot system 3000 and the tracking camera 102 will be located at any suitable position. In the configuration shown, the surgeon 3120 may be positioned across from the robot 3102, but is still able to manipulate the end-effector 3112 (and joint manipulation arm) and the display 110. A surgical assistant 3126 may be positioned across from the surgeon 3120 again with access to both the end-effector 3112 and the display 110. If desired, the locations of the surgeon 3120 and the assistant 3126 may be reversed. The traditional areas for the anesthesiologist 3122 and the nurse or scrub tech 124 remain unimpeded by the locations of the robot 3102 and camera 102. The anesthesiologist 3122 can operate anesthesia equipment which can include a display 3034.


The end-effector 3112 may be releasably coupled to the robotic arm 104 and movement of the end-effector 3112 can be controlled by at least one motor based on input from the camera tracking system 10. In some embodiments, the end-effector 3112 can be connectable to a joint manipulation arm 104 and may include a guide tube 114 configured to receive and orient a surgical instrument, tool, or implant used to perform a surgical procedure on the patient 3130.


As used herein, the term “end-effector” is used interchangeably with the terms “end-effectuator” and “effectuator element.” The term “instrument” is used in a non-limiting manner and can be used interchangeably with “tool” and “implant” to generally refer to any type of device that can be used during a surgical procedure in accordance with embodiments disclosed herein. Example instruments, tools, and implants include, without limitation, joint manipulation arms, drills, screwdrivers, saws, dilators, retractors, probes, implant inserters, and implant devices such as screws, spacers, interbody fusion devices, plates, rods, etc. In some embodiments, the end-effector 3112 can comprise any structure for effecting the movement of a surgical instrument in a desired manner.


The surgical robot 3102 is operable to control the translation and orientation of the end-effector 3112. The robot 3102 may be operable to move end-effector 3112 under computer control along x-, y-, and z-axes, for example. The end-effector 3112 can be configured for selective rotation under computer control about one or more of the x-, y-, and z-axis, and a Z Frame axis (such that one or more of the Euler Angles (e.g., roll, pitch, and/or yaw) associated with end-effector 3112 can be selectively computer controlled). In some embodiments, selective control of the translation and orientation of end-effector 3112 can permit performance of medical procedures with significantly improved accuracy compared to conventional robots that utilize, for example, a six degree of freedom robotic arm comprising only rotational axes. For example, the surgical robot system 3000 may be used to operate on patient 3130, and robotic arm 104 can be positioned above the body of patient 3130, with end-effector 3112 selectively angled relative to the z-axis toward the body of patient 3130.


In some example embodiments, the XR headsets 3150a and 3150b can be controlled to dynamically display an updated graphical indication of the pose of the surgical instrument so that the user can be aware of the pose of the surgical instrument at all times during the procedure.


As used herein, the term “pose” refers to the position and/or the rotational angle of one object (e.g., dynamic reference array, end-effector, surgical instrument, anatomical structure, etc.) relative to another object and/or to a defined coordinate system. A pose may therefore be defined based on only the multidimensional position of one object relative to another object and/or relative to a defined coordinate system, based on only the multidimensional rotational angles of the object relative to another object and/or to a defined coordinate system, or based on a combination of the multidimensional position and the multidimensional rotational angles. The term “pose” therefore is used to refer to position, rotational angle, or combination thereof.


In some further embodiments, the surgical robot 3102 can be configured to correct the path of the joint manipulation arm being moved by the surgeon with guidance by the robotic arm 104. In some example embodiments, surgical robot 3102 can be configured to perform stoppage, modification, and/or manual control of the movement of end-effector 3112. Thus, in use, in example embodiments, a surgeon or other user can operate the system 3000, and has the option to stop, modify, or manually control the autonomous movement of end-effector 3112.


Reference arrays can be formed on or connected to the robotic arm 104, the end-effector 3112, patient 3130, and/or the surgical instrument. The camera tracking system 10 can track poses of the reference arrays in, e.g., 6 degree-of-freedom (e.g., position along 3 orthogonal axes and rotation about the axes). In some embodiments, a reference array including a plurality of tracking markers can be provided thereon (e.g., formed-on or connected-to) to an outer surface of the robot 3102, such as on the robot arm 104 and/or on the end-effector 3112. A patient reference array 900 including one or more tracking markers can further be provided on the patient 3130 (e.g., formed-on or connected-to). An instrument reference array 3170 including one or more tracking markers can be provided on surgical instruments (e.g., a screwdriver, dilator, implant inserter, or the like). The reference arrays enable each of the marked objects (e.g., the end-effector 3112, the patient 3130, and the surgical instruments) to be tracked by the camera tracking system 10, and the tracked poses can be used to provide navigation guidance to a user for performance of a surgical procedure and/or can be used to control movement of the surgical robot 3102 for guiding the end-effector 3112.


Example Surgical System:



FIG. 31 illustrates a block diagram of surgical system which includes a camera tracking system 10 and navigation system 3604, and further optionally includes a surgical robot 3102, imaging device(s) 3620, and an XR headset 3150, which are each operative in accordance with some embodiments.


The imaging devices 3620 may include a C-arm imaging device, an O-arm imaging device, and/or a patient image database. A computer platform 3600 includes at least one processor, at least one memory storing program instructions executable by the at least one processor to perform operations. The computer platform 3600 may perform operations of the camera tracking system 10 and/or the navigation system 3604. The XR headset 3150 provides an improved human interface for performing navigated surgical procedures. The XR headset 3150 can be configured to provide functionalities, e.g., via the computer platform 3600, that include without limitation any one or more of: display camera tracking information and surgical procedure navigation information, identify hand gesture-based commands, etc. A display device 3112 may include a video projector, flat panel display, etc. The user can view XR graphical objects as an overlay anchored to particular real-world objects viewed through a see-through display screen. The XR headset 3150 may additionally or alternatively be configured to display on the display device 3112 video streams from cameras mounted to one or more XR headsets 3150 and other cameras, and/or medical images obtained from the imaging device(s) 3620.


Electrical components of the XR headset 3150 can include a plurality of cameras 3122, a microphone 3121, a gesture sensor 3108, a pose sensor (e.g., inertial measurement unit (IMU)) 3116, the display device 3112, and a wireless/wired communication interface 3124. The cameras 3122 of the XR headset 3150 may be visible light capturing cameras, near infrared capturing cameras, or a combination of both.


The cameras 3122 may be configured to operate as the gesture sensor 3108 by tracking user hand gestures performed within the field of view of the camera(s) 3122. Alternatively, the gesture sensor 3108 may be a proximity sensor and/or a touch sensor that senses hand gestures performed proximately to the gesture sensor 3108 and/or senses physical contact, e.g. tapping on the sensor 3108 or its enclosure. The pose sensor 3116, e.g., IMU, may include a multi-axis accelerometer, a tilt sensor, and/or another sensor that can sense rotation and/or acceleration of the XR headset 3150 along one or more defined coordinate axes. Some or all of these electrical components may be contained in a head-worn component enclosure or may be contained in another enclosure configured to be worn elsewhere, such as on the hip or shoulder.


As explained above, a surgical system includes a camera tracking system 10 which may be part of the computer platform 3600 which may also provide functionality of the navigation system 3604 and/or of the XR headset controller 3110. The surgical system may include the imaging devices and/or the surgical robot 3102. The navigation system 3604 can be configured to provide visual navigation guidance to an operator for moving and positioning an instrument relative and/or an end effector relative to patient anatomy (e.g., relative to the DRB 900) based on a surgical plan, e.g., from a surgical planning function, defining where a surgical procedure is to be performed using the instrument on the anatomy and based on a pose of the anatomy determined by the camera tracking system 10. The navigation system 3604 may be further configured to generate navigation information based on a target pose for the instrument, a present pose of the patient anatomy, and a present pose of the instrument and/or an end-effector of the surgical robot 3102, where the steering information is used to display information through the XR headset 3150 and/or another display to indicate where the instrument and/or the end-effector of the surgical robot 3102 should be moved to perform the surgical plan.


The electrical components of the XR headset 3150 can be operatively connected to the electrical components of the computer platform 3600 through a wired/wireless interface 3124. The electrical components of the XR headset 3150 may be operatively connected, e.g., through the computer platform 3600 or directly connected, to various imaging devices 3620, e.g., the C-arm imaging device, the I/O-arm imaging device, the patient image database, and/or to other medical equipment through the wired/wireless interface 3124.


The surgical system further includes at least one XR headset controller 3110 that may reside in the XR headset 3150, the computer platform 3600, and/or in another system component connected via wired cables and/or wireless communication links. Various functionality is provided by software executed by the XR headset controller 3110. The XR headset controller 3110 is configured to receive information from the camera tracking system 10 and the navigation controller 3604, and to generate an XR objects based on the information for display on the display device 3112.


The XR headset controller 3110 can be configured to operationally process signaling from the cameras 3122, the microphone 3121, and/or the pose sensor 3116, and be connected to display XR images on the display device 3112 for user viewing. Thus, the XR headset controller 3110 illustrated as a circuit block within the XR headset 3150 is to be understood as being operationally connected to other illustrated components of the XR headset 3150 but not necessarily residing within a common housing or being otherwise transportable by the user. For example, the XR headset controller 3110 may reside within the computer platform 3600 which, in turn, may reside within a housing of the surgical robot 3102, the camera tracking system 10, etc.


Further Definitions and Embodiments:


In the above-description of various embodiments of the present disclosure, aspects of the present disclosure may be illustrated and described herein in any of a number of patentable classes or contexts including any new and useful process, machine, manufacture, or composition of matter, or any new and useful improvement thereof. Accordingly, aspects of the present disclosure may be implemented in entirely hardware, entirely software (including firmware, resident software, micro-code, etc.) or combining software and hardware implementation that may all generally be referred to herein as a “circuit,” “module,” “component,” or “system.” Furthermore, aspects of the present disclosure may take the form of a computer program product comprising one or more computer readable media having computer readable program code embodied thereon.


Any combination of one or more computer readable media may be used. The computer readable media may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an appropriate optical fiber with a repeater, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.


A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable signal medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.


Computer program code for carrying out operations for aspects of the present disclosure may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C++, C#, VB.NET, Python or the like, conventional procedural programming languages, such as the “C” programming language, Visual Basic, Fortran 2003, Perl, COBOL 2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider) or in a cloud computing environment or offered as a service such as a Software as a Service (SaaS).


Aspects of the present disclosure are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the disclosure. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable instruction execution apparatus, create a mechanism for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.


These computer program instructions may also be stored in a computer readable medium that when executed can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions when stored in the computer readable medium produce an article of manufacture including instructions which when executed, cause a computer to implement the function/act specified in the flowchart and/or block diagram block or blocks. The computer program instructions may also be loaded onto a computer, other programmable instruction execution apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatuses or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.


It is to be understood that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of this specification and the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.


The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various aspects of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.


The terminology used herein is for the purpose of describing particular aspects only and is not intended to be limiting of the disclosure. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. Like reference numbers signify like elements throughout the description of the figures.


The corresponding structures, materials, acts, and equivalents of any means or step plus function elements in the claims below are intended to include any disclosed structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present disclosure has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the disclosure in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the disclosure. The aspects of the disclosure herein were chosen and described in order to best explain the principles of the disclosure and the practical application, and to enable others of ordinary skill in the art to understand the disclosure with various modifications as are suited to the particular use contemplated.

Claims
  • 1. A surgical system for computer assisted navigation during surgery, the surgical system comprising: a camera tracking system including one or more cameras;a display assembly; anda housing base assembly including a processing platform having a memory, a user interface, communication circuitry configured to communicate with the camera tracking system and the display assembly, and at least one processor configured to: obtain a three-dimensional (3D) radiological representation of a targeted anatomical structure of a patient containing a set of radiopaque fiducials of a registration fixture, the 3D radiological representation being derived from a 3D scan of the patient with a radiologic imaging device;attempt to register locations of the set of fiducials contained in the 3D radiological representation to a 3D imaging space tracked by the camera tracking system;based on determining that one of the fiducials of the set has a location that was not successfully registered to the 3D imaging space,display, via the display assembly, at least one two-dimensional (2D) view of the 3D radiological representation with a graphical overlay indicating that the fiducial has not been successfully registered to the 3D imaging space,receive user-supplied location information, via the user interface, identifying where the fiducial is located in the displayed 2D view of the 3D radiological representation, andregister the location of the fiducial to the 3D imaging space based on the user-supplied location information; andobtain, from at least one camera of the camera tracking system, optical images of a reference array fixated to the patient and of the registration fixture, the reference array including a first set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space, and the registration fixture including a second set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space;wherein the operation to receive user-supplied location information identifying where the fiducial is located in the 3D radiological representation, comprises to:display three orthogonal views of the fiducial in the 3D radiological representation and/or the 3D imaging space;display a graphical object overlaid on an initial location in the three orthogonal views;move location of where the graphical object is displayed in the three orthogonal views responsive to input from the user through a user interface;determine location of the fiducial in the 3D radiological representation and/or the 3D imaging space based on the location of where the graphical object is displayed in the three orthogonal views; andthe operation to register the location of the fiducial to the 3D imaging space is based on the determined location of the fiducial in the 3D radiological representation;wherein the operation to display the graphical object overlaid on the initial location in the three orthogonal views, comprises:determining the initial location to correspond to a predicted location of the fiducial based on relative locations of fiducials defined by a registration fixture template.
  • 2. The surgical system of claim 1, wherein the at least one processor is further operative to move the location where the graphical object is displayed in the three orthogonal views to track directional inputs received through the user interface of the surgical system.
  • 3. The surgical system of claim 1, wherein the at least one processor is further operative to: track locations of a tool captured in video from a camera of the camera tracking system while the tool is being moved by a user toward one of the fiducials that was successfully registered to the 3D image space;display updated representations of the tool according to the tracked locations in the 3D imaging space; andconfirm registration accuracy of the one of the fiducials that was successfully registered to the 3D image space based on comparison of a designated one of the tracked locations of the tool to the location of the one of the fiducials registered to the 3D image space.
  • 4. The surgical system of claim 1, wherein the operation to attempt to register locations of the set of fiducials in the 3D radiological representation to the 3D imaging space tracked by the camera tracking system, comprises: obtain, from at least one camera of the camera tracking system, an optical image of a reference array fixated to the patient, the reference array including a set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space;attempt to register locations of a pattern of the set of optical markers to locations of a pattern of the set of fiducials in the 3D radiological representation; andidentify any of the optical markers of the reference array that are not successfully registered to any of the fiducials of the registration fixture.
  • 5. The surgical system of claim 1, wherein the operations further comprise to: display a virtual implant device as an overlay on a view of the 3D radiological representation of the targeted anatomical structure;display a graphical indication of a trajectory of the virtual implant device representing an implantation trajectory of the virtual implant device into the targeted anatomical structure;update pose of the graphical indication of the trajectory of the virtual implant device displayed in the view of the 3D radiological representation, to track steering inputs received through a user interface of the surgical system; andstore as a planned trajectory of the virtual implant device, a user-designated one of the poses of the graphical indication of the trajectory.
  • 6. The surgical system of claim 5, wherein the operations further comprise to: display a set of implant devices which are selectable by a user for implant planning; andgenerate a graphical representation of the virtual implant device based on a template of one of the set of user-selectable implant devices which is selected by a user through the user interface.
  • 7. The surgical system of claim 5, wherein the operations further comprise to: display the planned trajectory of virtual implant device as an overlay on the view of the 3D radiological representation of the targeted anatomical structure;obtain, from at least one camera of the camera tracking system, optical images of a reference array fixated to a real-implant device corresponding to the virtual implant device, the reference array including a set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space;track pose of the real-implant device in the 3D imaging space based on pose of the reference array in the optical images while the real implant device is being positioned by a user relative to the targeted anatomical structure of the patient; anddisplay updated graphical representations of the real-implant device relative to the planned trajectory of the virtual implant device according to the tracked pose in the 3D imaging space.
  • 8. The surgical system of claim 1, wherein prior to obtaining the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture, the at least one processor is further operative to: obtain a fluoroscopic image of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture;determine whether a first condition is satisfied based on a defined number of the optical markers in the first set being detected by the at least one camera of the camera tracking system in the 3D imaging space;determine whether a second condition is satisfied based on a defined number of the optical markers in the second set being detected by the at least one camera of the camera tracking system in the 3D imaging space;determine whether a third condition is satisfied based on a defined number of the set of fiducials of the registration fixture being visible in the fluoroscopic image;when one of the first, second, and third conditions is not satisfied, display an indication of the not satisfied one of the first, second, and third conditions, and inhibit capture by a radiological imaging process of the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture; andwhen each of the first, second, and third conditions are satisfied, enable capture by the radiological imaging process of the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture.
  • 9. The surgical system of claim 8, wherein the at least one processor is further operative to: trigger capture of anteroposterior and lateral fluoroscopic images at a plurality of defined locations of the targeted anatomical structure of the patient based on determining each of the first, second, and third conditions are satisfied; andcompute the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture based on the captured anteroposterior and lateral fluoroscopic images at the plurality of defined locations of the targeted anatomical structure of the patient.
  • 10. A method by a surgical system for computer assisted navigation during surgery, the method comprising: providing the surgical system for computer assisted navigation during surgery, the system including:a camera tracking system including one or more cameras;a display assembly; anda housing base assembly including a processing platform having a memory, a user interface, communication circuitry configured to communicate with the camera tracking system and the display assembly, and at least one processor;obtaining a three-dimensional (3D) radiological representation of a targeted anatomical structure of a patient containing a set of fiducials of a registration fixture;attempting to register locations of the set of fiducials contained in the 3D radiological representation to a 3D imaging space tracked by the camera tracking system;based on determining that one of the fiducials of the set has a location that was not successfully registered to the 3D imaging space,displaying, via the display device, at least one two-dimensional (2D) view of the 3D radiological representation with a graphical overlay indicating that the fiducial has not been successfully registered to the 3D imaging space,receiving user-supplied location information, via the user interface, identifying where the fiducial is located in the displayed 2D view of the 3D radiological representation, andregistering the location of the fiducial to the 3D imaging space based on the user-supplied location information; andobtaining, from at least one camera of the camera tracking system, optical images of a reference array fixated to the patient and of the registration fixture, the reference array including a first set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space, and the registration fixture including a second set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space;wherein: the receiving user-supplied location information identifying where the fiducial is located in the 3D radiological representation, comprises:displaying three orthogonal views of the fiducial in the 3D radiological representation and/or the 3D imaging space;displaying a graphical object overlaid on an initial location in the three orthogonal views;moving location of where the graphical object is displayed in the three orthogonal views responsive to input from the user through a user interface;determining location of the fiducial in the 3D radiological representation and/or the 3D imaging space based on the location of where the graphical object is displayed in the three orthogonal views; andthe registering the location of the fiducial to the 3D imaging space is performed based on the determined location of the fiducial in the 3D radiological representation;wherein the operation to display the graphical object overlaid on the initial location in the three orthogonal views, comprises:determining the initial location to correspond to a predicted location of the fiducial based on relative locations of fiducials defined by a registration fixture template.
  • 11. The method of claim 10, further comprising: tracking locations of a tool captured in video from a camera of the camera tracking system while the tool is being moved by a user toward one of the fiducials that was successfully registered to the 3D image space;displaying updated representations of the tool according to the tracked locations in the 3D imaging space; andconfirming registration accuracy of the one of the fiducials that was successfully registered to the 3D image space based on comparison of a designated one of the tracked locations of the tool to the location of the one of the fiducials registered to the 3D image space.
  • 12. The method of claim 10, wherein the attempting to register locations of the set of fiducials in the 3D radiological representation to the 3D imaging space tracked by the camera tracking system, comprises: obtaining, from at least one camera of the camera tracking system, an optical image of a reference array fixated to the patient, the reference array including a set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space;attempting to register locations of a pattern of the set of optical markers to locations of a pattern of the set of fiducials in the 3D radiological representation; andidentifying any of the optical markers of the reference array that are not successfully registered to any of the fiducials of the registration fixture.
  • 13. The method of claim 10, further comprising: displaying a virtual implant device as an overlay on a view of the 3D radiological representation of the targeted anatomical structure;displaying a graphical indication of a trajectory of the virtual implant device representing an implantation trajectory of the virtual implant device into the targeted anatomical structure;updating pose of the graphical indication of the trajectory of the virtual implant device displayed in the view of the 3D radiological representation, to track steering inputs received through a user interface of the surgical system; andstoring as a planned trajectory of the virtual implant device, a user-designated one of the poses of the graphical indication of the trajectory.
  • 14. The method of claim 10, further comprising: displaying a set of implant devices which are selectable by a user for implant planning; andgenerating a graphical representation of the virtual implant device based on a template of one of the set of user-selectable implant devices which is selected by a user through the user interface.
  • 15. The method of claim 10, further comprising: displaying the planned trajectory of virtual implant device as an overlay on the view of the 3D radiological representation of the targeted anatomical structure;obtaining, from at least one camera of the camera tracking system, optical images of a reference array fixated to a real-implant device corresponding to the virtual implant device, the reference array including a set of optical markers detectable by the at least one camera of the camera tracking system in the 3D imaging space;tracking pose of the real-implant device in the 3D imaging space based on pose of the reference array in the optical images while the real implant device is being positioned by a user relative to the targeted anatomical structure of the patient; anddisplaying updated graphical representations of the real-implant device relative to the planned trajectory of the virtual implant device according to the tracked pose in the 3D imaging space.
  • 16. The method of claim 10, further comprising, prior to obtaining the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture: obtaining a fluoroscopic image of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture;determining whether a first condition is satisfied based on a defined number of the optical markers in the first set being detected by the at least one camera of the camera tracking system in the 3D imaging space;determining whether a second condition is satisfied based on a defined number of the optical markers in the second set being detected by the at least one camera of the camera tracking system in the 3D imaging space;determining whether a third condition is satisfied based on a defined number of the set of fiducials of the registration fixture being visible in the fluoroscopic image;when one of the first, second, and third conditions is not satisfied, displaying an indication of the not satisfied one of the first, second, and third conditions, and inhibiting capture by a radiological imaging process of the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture; andwhen each of the first, second, and third conditions are satisfied, enabling capture by the radiological imaging process of the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture.
  • 17. The method of claim 16, further comprising: triggering capture of anteroposterior and lateral fluoroscopic images at a plurality of defined locations of the targeted anatomical structure of the patient based on determining each of the first, second, and third conditions are satisfied; andcomputing the 3D radiological representation of the targeted anatomical structure of the patient and the set of fiducials of the registration fixture based on the captured anteroposterior and lateral fluoroscopic images at the plurality of defined locations of the targeted anatomical structure of the patient.
RELATED APPLICATIONS

This application claims the benefit of U.S. provisional patent application Ser. No. 63/176,424, filed Apr. 19, 2021, the disclosure of which is hereby incorporated herein by reference in its entirety.

US Referenced Citations (686)
Number Name Date Kind
4150293 Franke Apr 1979 A
5246010 Gazzara et al. Sep 1993 A
5354314 Hardy et al. Oct 1994 A
5397323 Taylor et al. Mar 1995 A
5598453 Baba et al. Jan 1997 A
5772594 Barrick Jun 1998 A
5791908 Gillio Aug 1998 A
5820559 Ng et al. Oct 1998 A
5825982 Wright et al. Oct 1998 A
5887121 Funda et al. Mar 1999 A
5911449 Daniele et al. Jun 1999 A
5951475 Gueziec et al. Sep 1999 A
5987960 Messner et al. Nov 1999 A
6012216 Esteves et al. Jan 2000 A
6031888 Ivan et al. Feb 2000 A
6033415 Mittelstadt et al. Mar 2000 A
6080181 Jensen et al. Jun 2000 A
6106511 Jensen Aug 2000 A
6122541 Cosman et al. Sep 2000 A
6144875 Schweikard et al. Nov 2000 A
6157853 Blume et al. Dec 2000 A
6167145 Foley et al. Dec 2000 A
6167292 Badano et al. Dec 2000 A
6201984 Funda et al. Mar 2001 B1
6203196 Meyer et al. Mar 2001 B1
6205411 DiGioia, III et al. Mar 2001 B1
6212419 Blume et al. Apr 2001 B1
6231565 Tovey et al. May 2001 B1
6236875 Bucholz et al. May 2001 B1
6246900 Cosman et al. Jun 2001 B1
6301495 Gueziec et al. Oct 2001 B1
6306126 Montezuma Oct 2001 B1
6312435 Wallace et al. Nov 2001 B1
6314311 Williams et al. Nov 2001 B1
6320929 Von Der Haar Nov 2001 B1
6322567 Mittelstadt et al. Nov 2001 B1
6325808 Bernard et al. Dec 2001 B1
6340363 Bolger et al. Jan 2002 B1
6377011 Ben-Ur Apr 2002 B1
6379302 Kessman et al. Apr 2002 B1
6402762 Hunter et al. Jun 2002 B2
6424885 Niemeyer et al. Jul 2002 B1
6447503 Wynne et al. Sep 2002 B1
6451027 Cooper et al. Sep 2002 B1
6477400 Barrick Nov 2002 B1
6484049 Seeley et al. Nov 2002 B1
6487267 Wolter Nov 2002 B1
6490467 Bucholz et al. Dec 2002 B1
6490475 Seeley et al. Dec 2002 B1
6499488 Hunter et al. Dec 2002 B1
6501981 Schweikard et al. Dec 2002 B1
6507751 Blume et al. Jan 2003 B2
6535756 Simon et al. Mar 2003 B1
6560354 Maurer, Jr. et al. May 2003 B1
6565554 Niemeyer May 2003 B1
6587750 Gerbi et al. Jul 2003 B2
6614453 Suri et al. Sep 2003 B1
6614871 Kobiki et al. Sep 2003 B1
6619840 Rasche et al. Sep 2003 B2
6636757 Jascob et al. Oct 2003 B1
6645196 Nixon et al. Nov 2003 B1
6666579 Jensen Dec 2003 B2
6669635 Kessman et al. Dec 2003 B2
6701173 Nowinski et al. Mar 2004 B2
6757068 Foxlin Jun 2004 B2
6782287 Grzeszczuk et al. Aug 2004 B2
6783524 Anderson et al. Aug 2004 B2
6786896 Madhani et al. Sep 2004 B1
6788018 Blumenkranz Sep 2004 B1
6804581 Wang et al. Oct 2004 B2
6823207 Jensen et al. Nov 2004 B1
6827351 Graziani et al. Dec 2004 B2
6837892 Shoham Jan 2005 B2
6839612 Sanchez et al. Jan 2005 B2
6856826 Seeley et al. Feb 2005 B2
6856827 Seeley et al. Feb 2005 B2
6879880 Nowlin et al. Apr 2005 B2
6892090 Verard et al. May 2005 B2
6920347 Simon et al. Jul 2005 B2
6922632 Foxlin Jul 2005 B2
6968224 Kessman et al. Nov 2005 B2
6978166 Foley et al. Dec 2005 B2
6988009 Grimm et al. Jan 2006 B2
6991627 Madhani et al. Jan 2006 B2
6996487 Jutras et al. Feb 2006 B2
6999852 Green Feb 2006 B2
7007699 Martinelli et al. Mar 2006 B2
7016457 Senzig et al. Mar 2006 B1
7043961 Pandey et al. May 2006 B2
7062006 Pelc et al. Jun 2006 B1
7063705 Young et al. Jun 2006 B2
7072707 Galloway, Jr. et al. Jul 2006 B2
7083615 Peterson et al. Aug 2006 B2
7097640 Wang et al. Aug 2006 B2
7099428 Clinthorne et al. Aug 2006 B2
7108421 Gregerson et al. Sep 2006 B2
7130676 Barrick Oct 2006 B2
7139418 Abovitz et al. Nov 2006 B2
7139601 Bucholz et al. Nov 2006 B2
7155316 Sutherland et al. Dec 2006 B2
7164968 Treat et al. Jan 2007 B2
7167738 Schweikard et al. Jan 2007 B2
7169141 Brock et al. Jan 2007 B2
7172627 Fiere et al. Feb 2007 B2
7194120 Wicker et al. Mar 2007 B2
7197107 Arai et al. Mar 2007 B2
7231014 Levy Jun 2007 B2
7231063 Naimark et al. Jun 2007 B2
7239940 Wang et al. Jul 2007 B2
7248914 Hastings et al. Jul 2007 B2
7301648 Foxlin Nov 2007 B2
7302288 Schellenberg Nov 2007 B1
7313430 Urquhart et al. Dec 2007 B2
7318805 Schweikard et al. Jan 2008 B2
7318827 Leitner et al. Jan 2008 B2
7319897 Leitner et al. Jan 2008 B2
7324623 Heuscher et al. Jan 2008 B2
7327865 Fu et al. Feb 2008 B2
7331967 Lee et al. Feb 2008 B2
7333642 Green Feb 2008 B2
7339341 Oleynikov et al. Mar 2008 B2
7366562 Dukesherer et al. Apr 2008 B2
7379790 Toth et al. May 2008 B2
7386365 Nixon Jun 2008 B2
7422592 Morley et al. Sep 2008 B2
7435216 Kwon et al. Oct 2008 B2
7440793 Chauhan et al. Oct 2008 B2
7460637 Clinthorne et al. Dec 2008 B2
7466303 Yi et al. Dec 2008 B2
7493153 Ahmed et al. Feb 2009 B2
7505617 Fu et al. Mar 2009 B2
7533892 Schena et al. May 2009 B2
7542791 Mire et al. Jun 2009 B2
7555331 Viswanathan Jun 2009 B2
7567834 Clayton et al. Jul 2009 B2
7594912 Cooper et al. Sep 2009 B2
7606613 Simon et al. Oct 2009 B2
7607440 Coste-Maniere et al. Oct 2009 B2
7623902 Pacheco Nov 2009 B2
7630752 Viswanathan Dec 2009 B2
7630753 Simon et al. Dec 2009 B2
7643862 Schoenefeld Jan 2010 B2
7660623 Hunter et al. Feb 2010 B2
7661881 Gregerson et al. Feb 2010 B2
7683331 Chang Mar 2010 B2
7683332 Chang Mar 2010 B2
7689320 Prisco et al. Mar 2010 B2
7691098 Wallace et al. Apr 2010 B2
7702379 Avinash et al. Apr 2010 B2
7702477 Tuemmler et al. Apr 2010 B2
7711083 Heigl et al. May 2010 B2
7711406 Kuhn et al. May 2010 B2
7720523 Omernick et al. May 2010 B2
7725253 Foxlin May 2010 B2
7726171 Langlotz et al. Jun 2010 B2
7742801 Neubauer et al. Jun 2010 B2
7751865 Jascob et al. Jul 2010 B2
7760849 Zhang Jul 2010 B2
7762825 Burbank et al. Jul 2010 B2
7763015 Cooper et al. Jul 2010 B2
7787699 Mahesh et al. Aug 2010 B2
7796728 Bergfjord Sep 2010 B2
7813838 Sommer Oct 2010 B2
7818044 Dukesherer et al. Oct 2010 B2
7819859 Prisco et al. Oct 2010 B2
7824401 Manzo et al. Nov 2010 B2
7831294 Viswanathan Nov 2010 B2
7834484 Sartor Nov 2010 B2
7835557 Kendrick et al. Nov 2010 B2
7835778 Foley et al. Nov 2010 B2
7835784 Mire et al. Nov 2010 B2
7840253 Tremblay et al. Nov 2010 B2
7840256 Lakin et al. Nov 2010 B2
7843158 Prisco Nov 2010 B2
7844320 Shahidi Nov 2010 B2
7853305 Simon et al. Dec 2010 B2
7853313 Thompson Dec 2010 B2
7865269 Prisco et al. Jan 2011 B2
D631966 Perloff et al. Feb 2011 S
7879045 Gielen et al. Feb 2011 B2
7881767 Strommer et al. Feb 2011 B2
7881770 Melkent et al. Feb 2011 B2
7886743 Cooper et al. Feb 2011 B2
RE42194 Foley et al. Mar 2011 E
RE42226 Foley et al. Mar 2011 E
7900524 Calloway et al. Mar 2011 B2
7907166 Lamprecht et al. Mar 2011 B2
7909122 Schena et al. Mar 2011 B2
7925653 Saptharishi Apr 2011 B2
7930065 Larkin et al. Apr 2011 B2
7935130 Williams May 2011 B2
7940999 Liao et al. May 2011 B2
7945012 Ye et al. May 2011 B2
7945021 Shapiro et al. May 2011 B2
7953470 Vetter et al. May 2011 B2
7954397 Choi et al. Jun 2011 B2
7971341 Dukesherer et al. Jul 2011 B2
7974674 Hauck et al. Jul 2011 B2
7974677 Mire et al. Jul 2011 B2
7974681 Wallace et al. Jul 2011 B2
7979157 Anvari Jul 2011 B2
7983733 Viswanathan Jul 2011 B2
7988215 Seibold Aug 2011 B2
7996110 Lipow et al. Aug 2011 B2
8004121 Sartor Aug 2011 B2
8004229 Nowlin et al. Aug 2011 B2
8010177 Csavoy et al. Aug 2011 B2
8019045 Kato Sep 2011 B2
8021310 Sanborn et al. Sep 2011 B2
8035685 Jensen Oct 2011 B2
8046054 Kim et al. Oct 2011 B2
8046057 Clarke Oct 2011 B2
8052688 Wolf, II Nov 2011 B2
8054184 Cline et al. Nov 2011 B2
8054752 Druke et al. Nov 2011 B2
8057397 Li et al. Nov 2011 B2
8057407 Martinelli et al. Nov 2011 B2
8062288 Cooper et al. Nov 2011 B2
8062375 Glerum et al. Nov 2011 B2
8066524 Burbank et al. Nov 2011 B2
8073335 Labonville et al. Dec 2011 B2
8079950 Stern et al. Dec 2011 B2
8086299 Adler et al. Dec 2011 B2
8092370 Roberts et al. Jan 2012 B2
8098914 Liao et al. Jan 2012 B2
8100950 St. Clair et al. Jan 2012 B2
8105320 Manzo Jan 2012 B2
8108025 Csavoy et al. Jan 2012 B2
8109877 Moctezuma de la Barrera et al. Feb 2012 B2
8112292 Simon Feb 2012 B2
8116430 Shapiro et al. Feb 2012 B1
8120301 Goldberg et al. Feb 2012 B2
8121249 Wang et al. Feb 2012 B2
8123675 Funda et al. Feb 2012 B2
8133229 Bonutti Mar 2012 B1
8142420 Schena Mar 2012 B2
8147494 Leitner et al. Apr 2012 B2
8150494 Simon et al. Apr 2012 B2
8150497 Gielen et al. Apr 2012 B2
8150498 Gielen et al. Apr 2012 B2
8165658 Waynik et al. Apr 2012 B2
8170313 Kendrick et al. May 2012 B2
8179073 Farritor et al. May 2012 B2
8182476 Julian et al. May 2012 B2
8184880 Zhao et al. May 2012 B2
8202278 Orban, III et al. Jun 2012 B2
8208708 Homan et al. Jun 2012 B2
8208988 Jensen Jun 2012 B2
8219177 Smith et al. Jul 2012 B2
8219178 Smith et al. Jul 2012 B2
8220468 Cooper et al. Jul 2012 B2
8224024 Foxlin et al. Jul 2012 B2
8224484 Swarup et al. Jul 2012 B2
8225798 Baldwin et al. Jul 2012 B2
8228368 Zhao et al. Jul 2012 B2
8231610 Jo et al. Jul 2012 B2
8263933 Hartmann et al. Jul 2012 B2
8239001 Verard et al. Aug 2012 B2
8241271 Millman et al. Aug 2012 B2
8248413 Gattani et al. Aug 2012 B2
8256319 Cooper et al. Sep 2012 B2
8271069 Jascob et al. Sep 2012 B2
8271130 Hourtash Sep 2012 B2
8281670 Larkin et al. Oct 2012 B2
8282653 Nelson et al. Oct 2012 B2
8301226 Csavoy et al. Oct 2012 B2
8311611 Csavoy et al. Nov 2012 B2
8320991 Jascob et al. Nov 2012 B2
8332012 Kienzle, III Dec 2012 B2
8333755 Cooper et al. Dec 2012 B2
8335552 Stiles Dec 2012 B2
8335557 Maschke Dec 2012 B2
8348931 Cooper et al. Jan 2013 B2
8353963 Glerum Jan 2013 B2
8358818 Miga et al. Jan 2013 B2
8359730 Burg et al. Jan 2013 B2
8374673 Adcox et al. Feb 2013 B2
8374723 Zhao et al. Feb 2013 B2
8379791 Forthmann et al. Feb 2013 B2
8386019 Camus et al. Feb 2013 B2
8392022 Ortmaier et al. Mar 2013 B2
8394099 Patwardhan Mar 2013 B2
8395342 Prisco Mar 2013 B2
8398634 Manzo et al. Mar 2013 B2
8400094 Schena Mar 2013 B2
8414957 Enzerink et al. Apr 2013 B2
8418073 Mohr et al. Apr 2013 B2
8450694 Baviera et al. May 2013 B2
8452447 Nixon May 2013 B2
RE44305 Foley et al. Jun 2013 E
8462911 Vesel et al. Jun 2013 B2
8465476 Rogers et al. Jun 2013 B2
8465771 Wan et al. Jun 2013 B2
8467851 Mire et al. Jun 2013 B2
8467852 Csavoy et al. Jun 2013 B2
8469947 Devengenzo et al. Jun 2013 B2
RE44392 Hynes Jul 2013 E
8483434 Buehner et al. Jul 2013 B2
8483800 Jensen et al. Jul 2013 B2
8486532 Enzerink et al. Jul 2013 B2
8489235 Moll et al. Jul 2013 B2
8500722 Cooper Aug 2013 B2
8500728 Newton et al. Aug 2013 B2
8504201 Moll et al. Aug 2013 B2
8506555 Ruiz Morales Aug 2013 B2
8506556 Schena Aug 2013 B2
8508173 Goldberg et al. Aug 2013 B2
8512318 Tovey et al. Aug 2013 B2
8515576 Lipow et al. Aug 2013 B2
8518120 Glerum et al. Aug 2013 B2
8521331 Itkowitz Aug 2013 B2
8526688 Groszmann et al. Sep 2013 B2
8526700 Isaacs Sep 2013 B2
8527094 Kumar et al. Sep 2013 B2
8528440 Morley et al. Sep 2013 B2
8532741 Heruth et al. Sep 2013 B2
8541970 Nowlin et al. Sep 2013 B2
8548563 Simon et al. Oct 2013 B2
8549732 Burg et al. Oct 2013 B2
8551114 Ramos de la Pena Oct 2013 B2
8551116 Julian et al. Oct 2013 B2
8556807 Scott et al. Oct 2013 B2
8556979 Glerum et al. Oct 2013 B2
8560118 Green et al. Oct 2013 B2
8561473 Blumenkranz Oct 2013 B2
8562594 Cooper et al. Oct 2013 B2
8571638 Shoham Oct 2013 B2
8571710 Coste-Maniere et al. Oct 2013 B2
8573465 Shelton, IV Nov 2013 B2
8574303 Sharkey et al. Nov 2013 B2
8585420 Burbank et al. Nov 2013 B2
8594841 Zhao et al. Nov 2013 B2
8597198 Sanborn et al. Dec 2013 B2
8600478 Verard et al. Dec 2013 B2
8603077 Cooper et al. Dec 2013 B2
8611985 Lavallee et al. Dec 2013 B2
8613230 Blumenkranz et al. Dec 2013 B2
8621939 Blumenkranz et al. Jan 2014 B2
8624537 Nowlin et al. Jan 2014 B2
8630389 Kato Jan 2014 B2
8634897 Simon et al. Jan 2014 B2
8634957 Toth et al. Jan 2014 B2
8638056 Goldberg et al. Jan 2014 B2
8638057 Goldberg et al. Jan 2014 B2
8639000 Zhao et al. Jan 2014 B2
8641726 Bonutti Feb 2014 B2
8644907 Hartmann et al. Feb 2014 B2
8657809 Schoepp Feb 2014 B2
8660635 Simon et al. Feb 2014 B2
8666544 Moll et al. Mar 2014 B2
8675939 Moctezuma de la Barrera Mar 2014 B2
8678647 Gregerson et al. Mar 2014 B2
8679125 Smith et al. Mar 2014 B2
8679183 Glerum et al. Mar 2014 B2
8682413 Lloyd Mar 2014 B2
8684253 Giordano et al. Apr 2014 B2
8685098 Glerum et al. Apr 2014 B2
8693730 Umasuthan et al. Apr 2014 B2
8694075 Groszmann et al. Apr 2014 B2
8696458 Foxlin et al. Apr 2014 B2
8700123 Okamura et al. Apr 2014 B2
8706086 Glerum Apr 2014 B2
8706185 Foley et al. Apr 2014 B2
8706301 Zhao et al. Apr 2014 B2
8717430 Simon et al. May 2014 B2
8727618 Maschke et al. May 2014 B2
8734432 Tuma et al. May 2014 B2
8738115 Amberg et al. May 2014 B2
8738181 Greer et al. May 2014 B2
8740882 Jun et al. Jun 2014 B2
8746252 McGrogan et al. Jun 2014 B2
8749189 Nowlin et al. Jun 2014 B2
8749190 Nowlin et al. Jun 2014 B2
8761930 Nixon Jun 2014 B2
8764448 Yang et al. Jul 2014 B2
8771170 Mesallum et al. Jul 2014 B2
8781186 Clements et al. Jul 2014 B2
8781630 Banks et al. Jul 2014 B2
8784385 Boyden et al. Jul 2014 B2
8786241 Nowlin et al. Jul 2014 B2
8787520 Baba Jul 2014 B2
8792704 Isaacs Jul 2014 B2
8798231 Notohara et al. Aug 2014 B2
8800838 Shelton, IV Aug 2014 B2
8808164 Hoffman et al. Aug 2014 B2
8812077 Dempsey Aug 2014 B2
8814793 Brabrand Aug 2014 B2
8816628 Nowlin et al. Aug 2014 B2
8818105 Myronenko et al. Aug 2014 B2
8820605 Shelton, IV Sep 2014 B2
8821511 Von Jako et al. Sep 2014 B2
8823308 Nowlin et al. Sep 2014 B2
8827996 Scott et al. Sep 2014 B2
8828024 Farritor et al. Sep 2014 B2
8830224 Zhao et al. Sep 2014 B2
8834489 Cooper et al. Sep 2014 B2
8834490 Bonutti Sep 2014 B2
8838270 Druke et al. Sep 2014 B2
8844789 Shelton, IV et al. Sep 2014 B2
8855822 Bartol et al. Oct 2014 B2
8858598 Seifert et al. Oct 2014 B2
8860753 Bhandarkar et al. Oct 2014 B2
8864751 Prisco et al. Oct 2014 B2
8864798 Weiman et al. Oct 2014 B2
8864833 Glerum et al. Oct 2014 B2
8867703 Shapiro et al. Oct 2014 B2
8870880 Himmelberger et al. Oct 2014 B2
8876866 Zappacosta et al. Nov 2014 B2
8880223 Raj et al. Nov 2014 B2
8882803 Iott et al. Nov 2014 B2
8883210 Truncale et al. Nov 2014 B1
8888821 Rezach et al. Nov 2014 B2
8888853 Glerum et al. Nov 2014 B2
8888854 Glerum et al. Nov 2014 B2
8894652 Seifert et al. Nov 2014 B2
8894688 Suh Nov 2014 B2
8894691 Iott et al. Nov 2014 B2
8906069 Hansell et al. Dec 2014 B2
8964934 Ein-Gal Feb 2015 B2
8992580 Bar et al. Mar 2015 B2
8996169 Lightcap et al. Mar 2015 B2
9001963 Sowards-Emmerd et al. Apr 2015 B2
9002076 Khadem et al. Apr 2015 B2
9044190 Rubner et al. Jun 2015 B2
9107683 Hourtash et al. Aug 2015 B2
9125556 Zehavi et al. Sep 2015 B2
9131986 Greer et al. Sep 2015 B2
9215968 Schostek et al. Dec 2015 B2
9308050 Kostrzewski et al. Apr 2016 B2
9380984 Li et al. Jul 2016 B2
9393039 Lechner et al. Jul 2016 B2
9398886 Gregerson et al. Jul 2016 B2
9398890 Dong et al. Jul 2016 B2
9414859 Ballard et al. Aug 2016 B2
9420975 Gutfleisch et al. Aug 2016 B2
9492235 Hourtash et al. Nov 2016 B2
9592096 Maillet et al. Mar 2017 B2
9750465 Engel et al. Sep 2017 B2
9757203 Hourtash et al. Sep 2017 B2
9795354 Menegaz et al. Oct 2017 B2
9814535 Bar et al. Nov 2017 B2
9820783 Donner et al. Nov 2017 B2
9833265 Donner et al. Nov 2017 B2
9848922 Tohmeh et al. Dec 2017 B2
9925011 Gombert et al. Mar 2018 B2
9931025 Graetzel et al. Apr 2018 B1
10034717 Miller et al. Jul 2018 B2
11553969 Lang Jan 2023 B1
20010036302 Miller Nov 2001 A1
20020035321 Bucholz et al. Mar 2002 A1
20040068172 Nowinski et al. Apr 2004 A1
20040076259 Jensen et al. Apr 2004 A1
20050096502 Khalili May 2005 A1
20050143651 Verard et al. Jun 2005 A1
20050171558 Abovitz et al. Aug 2005 A1
20060100610 Wallace et al. May 2006 A1
20060173329 Marquart et al. Aug 2006 A1
20060184396 Dennis et al. Aug 2006 A1
20060241416 Marquart et al. Oct 2006 A1
20060291612 Nishide et al. Dec 2006 A1
20070015987 Benlloch Baviera et al. Jan 2007 A1
20070021738 Hasser et al. Jan 2007 A1
20070038059 Sheffer et al. Feb 2007 A1
20070073133 Schoenefeld Mar 2007 A1
20070156121 Millman et al. Jul 2007 A1
20070156157 Nahum et al. Jul 2007 A1
20070167712 Keglovich et al. Jul 2007 A1
20070233238 Huynh et al. Oct 2007 A1
20080004523 Jensen Jan 2008 A1
20080013809 Zhu et al. Jan 2008 A1
20080033283 Dellaca et al. Feb 2008 A1
20080046122 Manzo et al. Feb 2008 A1
20080082109 Moll et al. Apr 2008 A1
20080108912 Node-Langlois May 2008 A1
20080108991 Von Jako May 2008 A1
20080109012 Falco et al. May 2008 A1
20080144906 Allred et al. Jun 2008 A1
20080161680 Von Jako et al. Jul 2008 A1
20080161682 Kendrick et al. Jul 2008 A1
20080177203 von Jako Jul 2008 A1
20080214922 Hartmann et al. Sep 2008 A1
20080228068 Viswanathan et al. Sep 2008 A1
20080228196 Wang et al. Sep 2008 A1
20080235052 Node-Langlois et al. Sep 2008 A1
20080269596 Revie et al. Oct 2008 A1
20080287771 Anderson Nov 2008 A1
20080287781 Revie et al. Nov 2008 A1
20080300477 Lloyd et al. Dec 2008 A1
20080300478 Zuhars et al. Dec 2008 A1
20080302950 Park et al. Dec 2008 A1
20080306490 Lakin et al. Dec 2008 A1
20080319311 Hamadeh Dec 2008 A1
20090012509 Csavoy et al. Jan 2009 A1
20090030428 Omori et al. Jan 2009 A1
20090080737 Battle et al. Mar 2009 A1
20090185655 Koken et al. Jul 2009 A1
20090198121 Hoheisel Aug 2009 A1
20090216113 Meier et al. Aug 2009 A1
20090228019 Gross et al. Sep 2009 A1
20090259123 Navab et al. Oct 2009 A1
20090259230 Khadem et al. Oct 2009 A1
20090264899 Appenrodt et al. Oct 2009 A1
20090281417 Hartmann et al. Nov 2009 A1
20100022874 Wang et al. Jan 2010 A1
20100039506 Sarvestani et al. Feb 2010 A1
20100125286 Wang et al. May 2010 A1
20100130986 Mailloux et al. May 2010 A1
20100228117 Hartmann Sep 2010 A1
20100228265 Prisco Sep 2010 A1
20100249571 Jensen et al. Sep 2010 A1
20100274120 Heuscher Oct 2010 A1
20100280363 Skarda et al. Nov 2010 A1
20100331858 Simaan et al. Dec 2010 A1
20110022229 Jang et al. Jan 2011 A1
20110077504 Fischer et al. Mar 2011 A1
20110098553 Robbins et al. Apr 2011 A1
20110137152 Li Jun 2011 A1
20110213384 Jeong Sep 2011 A1
20110224684 Larkin et al. Sep 2011 A1
20110224685 Larkin et al. Sep 2011 A1
20110224686 Larkin et al. Sep 2011 A1
20110224687 Larkin et al. Sep 2011 A1
20110224688 Larkin et al. Sep 2011 A1
20110224689 Larkin et al. Sep 2011 A1
20110224825 Larkin et al. Sep 2011 A1
20110230967 O'Halloran et al. Sep 2011 A1
20110238080 Ranjit et al. Sep 2011 A1
20110276058 Choi et al. Nov 2011 A1
20110282189 Graumann Nov 2011 A1
20110286573 Schretter et al. Nov 2011 A1
20110295062 Solsona et al. Dec 2011 A1
20110295370 Suh et al. Dec 2011 A1
20110306986 Lee et al. Dec 2011 A1
20120035507 George et al. Feb 2012 A1
20120046668 Gantes Feb 2012 A1
20120051498 Koishi Mar 2012 A1
20120053597 Anvari et al. Mar 2012 A1
20120059248 Holsing et al. Mar 2012 A1
20120071753 Hunter et al. Mar 2012 A1
20120108954 Schulhauser et al. May 2012 A1
20120136372 Amat Girbau et al. May 2012 A1
20120143084 Shoham Jun 2012 A1
20120184839 Woerlein Jul 2012 A1
20120197182 Millman et al. Aug 2012 A1
20120226145 Chang et al. Sep 2012 A1
20120235909 Birkenbach et al. Sep 2012 A1
20120245596 Meenink Sep 2012 A1
20120253332 Moll Oct 2012 A1
20120253360 White et al. Oct 2012 A1
20120256092 Zingerman Oct 2012 A1
20120294498 Popovic Nov 2012 A1
20120296203 Hartmann et al. Nov 2012 A1
20130006267 Odermatt et al. Jan 2013 A1
20130016889 Myronenko et al. Jan 2013 A1
20130030571 Ruiz Morales et al. Jan 2013 A1
20130035583 Park et al. Feb 2013 A1
20130060146 Yang et al. Mar 2013 A1
20130060337 Petersheim et al. Mar 2013 A1
20130094742 Feilkas Apr 2013 A1
20130096574 Kang et al. Apr 2013 A1
20130113791 Isaacs et al. May 2013 A1
20130116706 Lee et al. May 2013 A1
20130131695 Scarfogliero et al. May 2013 A1
20130144307 Jeong et al. Jun 2013 A1
20130158542 Manzo et al. Jun 2013 A1
20130165937 Patwardhan Jun 2013 A1
20130178867 Farritor et al. Jul 2013 A1
20130178868 Roh Jul 2013 A1
20130178870 Schena Jul 2013 A1
20130204271 Brisson et al. Aug 2013 A1
20130211419 Jensen Aug 2013 A1
20130211420 Jensen Aug 2013 A1
20130218142 Tuma et al. Aug 2013 A1
20130223702 Holsing et al. Aug 2013 A1
20130225942 Holsing et al. Aug 2013 A1
20130225943 Holsing et al. Aug 2013 A1
20130231556 Holsing et al. Sep 2013 A1
20130237995 Lee et al. Sep 2013 A1
20130245375 DiMaio et al. Sep 2013 A1
20130261640 Kim et al. Oct 2013 A1
20130272488 Bailey et al. Oct 2013 A1
20130272489 Dickman et al. Oct 2013 A1
20130274761 Devengenzo et al. Oct 2013 A1
20130281821 Liu et al. Oct 2013 A1
20130296884 Taylor et al. Nov 2013 A1
20130303887 Holsing et al. Nov 2013 A1
20130307955 Deitz et al. Nov 2013 A1
20130317521 Choi et al. Nov 2013 A1
20130325033 Schena et al. Dec 2013 A1
20130325035 Hauck et al. Dec 2013 A1
20130331686 Freysinger et al. Dec 2013 A1
20130331858 Devengenzo et al. Dec 2013 A1
20130331861 Yoon Dec 2013 A1
20130342578 Isaacs Dec 2013 A1
20130345717 Markvicka et al. Dec 2013 A1
20130345757 Stad Dec 2013 A1
20140001235 Shelton, IV Jan 2014 A1
20140012131 Heruth et al. Jan 2014 A1
20140031664 Kang et al. Jan 2014 A1
20140046128 Lee et al. Feb 2014 A1
20140046132 Hoeg et al. Feb 2014 A1
20140046340 Wilson et al. Feb 2014 A1
20140049629 Siewerdsen et al. Feb 2014 A1
20140058406 Tsekos Feb 2014 A1
20140073914 Lavallee et al. Mar 2014 A1
20140080086 Chen Mar 2014 A1
20140081128 Verard et al. Mar 2014 A1
20140088612 Bartol et al. Mar 2014 A1
20140094694 Moctezuma de la Barrera Apr 2014 A1
20140094851 Gordon Apr 2014 A1
20140096369 Matsumoto et al. Apr 2014 A1
20140100587 Farritor et al. Apr 2014 A1
20140121676 Kostrzewski et al. May 2014 A1
20140128882 Kwak et al. May 2014 A1
20140135796 Simon et al. May 2014 A1
20140142591 Alvarez et al. May 2014 A1
20140142592 Moon et al. May 2014 A1
20140148692 Hartmann et al. May 2014 A1
20140163581 Devengenzo et al. Jun 2014 A1
20140171781 Stiles Jun 2014 A1
20140171900 Stiles Jun 2014 A1
20140171965 Loh et al. Jun 2014 A1
20140180308 von Grunberg Jun 2014 A1
20140180309 Seeber et al. Jun 2014 A1
20140187915 Yaroshenko et al. Jul 2014 A1
20140188132 Kang Jul 2014 A1
20140194699 Roh et al. Jul 2014 A1
20140130810 Azizian et al. Aug 2014 A1
20140221819 Sarment Aug 2014 A1
20140222023 Kim et al. Aug 2014 A1
20140228631 Kwak et al. Aug 2014 A1
20140234804 Huang et al. Aug 2014 A1
20140257328 Kim et al. Sep 2014 A1
20140257329 Jang et al. Sep 2014 A1
20140257330 Choi et al. Sep 2014 A1
20140275760 Lee et al. Sep 2014 A1
20140275985 Walker et al. Sep 2014 A1
20140276931 Parihar et al. Sep 2014 A1
20140276940 Seo Sep 2014 A1
20140276944 Farritor et al. Sep 2014 A1
20140288413 Hwang et al. Sep 2014 A1
20140299648 Shelton, IV et al. Oct 2014 A1
20140303434 Farritor et al. Oct 2014 A1
20140303643 Ha et al. Oct 2014 A1
20140305995 Shelton, IV et al. Oct 2014 A1
20140309659 Roh et al. Oct 2014 A1
20140316436 Bar et al. Oct 2014 A1
20140323803 Hoffman et al. Oct 2014 A1
20140324070 Min et al. Oct 2014 A1
20140330288 Date et al. Nov 2014 A1
20140364720 Darrow et al. Dec 2014 A1
20140371577 Maillet et al. Dec 2014 A1
20150039034 Frankel et al. Feb 2015 A1
20150085970 Bouhnik et al. Mar 2015 A1
20150146847 Liu May 2015 A1
20150150524 Yorkston et al. Jun 2015 A1
20150196261 Funk Jul 2015 A1
20150213633 Chang et al. Jul 2015 A1
20150335480 Alvarez et al. Nov 2015 A1
20150342647 Frankel et al. Dec 2015 A1
20160005194 Schretter et al. Jan 2016 A1
20160166329 Langan et al. Jun 2016 A1
20160235480 Scholl et al. Aug 2016 A1
20160249990 Glozman et al. Sep 2016 A1
20160302871 Gregerson et al. Oct 2016 A1
20160320322 Suzuki Nov 2016 A1
20160331335 Gregerson et al. Nov 2016 A1
20170135770 Scholl et al. May 2017 A1
20170143284 Sehnert et al. May 2017 A1
20170143426 Isaacs et al. May 2017 A1
20170156816 Ibrahim Jun 2017 A1
20170202629 Maillet et al. Jul 2017 A1
20170209225 Wu Jul 2017 A1
20170212723 Atarot et al. Jul 2017 A1
20170215825 Johnson et al. Aug 2017 A1
20170215826 Johnson et al. Aug 2017 A1
20170215827 Johnson et al. Aug 2017 A1
20170231710 Scholl et al. Aug 2017 A1
20170258426 Risher-Kelly et al. Sep 2017 A1
20170273748 Hourtash et al. Sep 2017 A1
20170296277 Hourtash et al. Oct 2017 A1
20170360493 Zucher et al. Dec 2017 A1
20190209080 Gullotti Jul 2019 A1
20200197102 Shekhar Jun 2020 A1
20200253640 Mullaney Aug 2020 A1
20210045813 Wickham et al. Feb 2021 A1
Foreign Referenced Citations (8)
Number Date Country
2676627 Dec 2013 EP
3431032 Jan 2019 EP
2019022658 Feb 2019 JP
2020096795 Jun 2020 JP
20200109292 Sep 2020 KR
WO-2020163316 Aug 2020 WO
WO-2021062001 Apr 2021 WO
WO-2021062373 Apr 2021 WO
Non-Patent Literature Citations (1)
Entry
US 8,231,638 B2, 07/2012, Swarup et al. (withdrawn)
Related Publications (1)
Number Date Country
20230081244 A1 Mar 2023 US
Provisional Applications (1)
Number Date Country
63176424 Apr 2021 US