The present disclosure relates generally to surgical methods, and more particularly to registration and/or tracking of a patient's bone employing a patient specific bone jig along with imaging and/or point sampling.
Typically, computer assisted navigation procedures, pre-operative surgical imaging is frequently used to digitally capture an anatomical feature of interest such as a patient's bone. The process of registration and tracking is used, firstly, to provide the navigation system information about the actual starting position of the object of interest and secondly, to update the navigation system with movements of the tracked object over time. The registration process commonly involves a skilled user sampling clinically known anatomical landmarks with a tracked stylus. For example, with computer assisted navigation procedures for a hip, these points may be the apex of the greater or lesser trochanter. In other conventional registration processes, a tracked probe with fiducial markers is manually placed on the object of interest and sampled points are integrated into algorithms that infer the position of the object. By way of nonlimiting example, the fiducial markers may be visible to IR cameras, optical cameras or radar. The registration algorithms associate sets of data into a common coordinate system by minimizing the alignment error. The registration parameters define the transformation between the coordinate frame of the surface model and the coordinate frame of the robot.
Shortcomings of the prior art are overcome and additional advantages are provided through the provision, in one embodiment, of a computer implemented method which includes for example, obtaining, via one or more processors, three-dimensional data representing a patient's bone, obtaining, via the one or more processors, three-dimensional data representing at least portions of a patient specific bone jig, the patient specific bone jig having an inner surface portion matched to an outer surface portion of the patient's bone, obtaining, via the one or more processors, image data image data representing the at least portions of the patient specific bone jig registered to the patient's bone, and generating, via the one or more processors, data representing a location and an orientation of the patient's bone based on the obtained image data, the obtained three-dimensional data representing the patient's bone, and the obtained three-dimensional data representing the at least portions of the patient specific bone jig.
In another embodiment, a computer implemented method includes for example, obtaining, via one or more processors, three-dimensional data representing a first object, obtaining, via the one or more processors, three-dimensional data representing at least portions of a second object, the second object having an inner surface portion matched to an outer surface portion of the first object, obtaining, via the one or more processors, image data representing the at least portions of the second object registered to the first object, and generating, via the one or more processors, data representing a location and/or an orientation of the second object based on the obtained image data, the obtained three-dimensional data representing the first object, and the obtained three-dimensional data representing at least portions of the second object.
In another embodiment, a computer implemented method includes for example, obtaining, via one or more processors, three-dimensional data representing the patient's bone, obtaining, via the one or more processors, three-dimensional data representing at least portions of a patient specific bone jig having predetermined spatial indicia, the patient specific bone jig having an inner surface portion matched to an outer surface portion of the patient's bone, obtaining, via the one or more processors, point sampling data representing the predetermined spatial indicia of the patient specific bone jig registered to the patient's bone, and generating data representing a location and an orientation of the patient's bone based on the obtained point sampling data, the obtained three-dimensional data representing the patient's bone, and the obtained three-dimensional data representing the at least portions of the patient specific bone jig having predetermined spatial indicia.
In another embodiment, a computer implemented method includes for example, obtaining, via one or more processors, three-dimensional data representing a first object, obtaining, via the one or more processors, three-dimensional data representing at least portions of a second object having predetermined spatial indicia, the second object having an inner surface portion matched to an outer surface portion of the first object, obtaining, via the one or more processors, point sampling data representing the predetermined spatial indicia of the second object registered to the first object, and generating, via the one or more processors, data representing a location and an orientation of the first object based on the obtained point sampling data, the obtained three-dimensional data representing the first object, and the obtained data three-dimensional data representing at least portions of the second object having the predetermined spatial indicia.
The subject matter which is regarded as the disclosure is particularly pointed out and distinctly claimed in the concluding portion of the specification. The disclosure, however, may best be understood by reference to the following detailed description of various embodiments and the accompanying drawings in which:
Generally stated, the present disclosure is directed to surgical methods, and more particularly to registration and tracking of a patient's bone employing a patient specific bone jig along with imaging and/or minimal point sampling of the patient's bone.
In this detailed description and the following claims, the words proximal, distal, anterior, posterior, medial, lateral, superior, and inferior are defined by their standard usage for indicating a particular part of a bone or implant according to the relative disposition of the natural bone or directional terms of reference. Positions or directions may be used herein with reference to anatomical structures or surfaces. For example, as the current devices and methods are described herein with reference to use with the bones of the hip, the bones of the hip may be used to describe the surfaces, positions, directions or orientations of the implant apparatus, implant installation apparatus, and surgical methods. Further, the devices and surgical methods, and the aspects, components, features and the like thereof, disclosed herein are described with respect to one side of the body for brevity purposes. However, as the human body is relatively symmetrical or mirrored about a line of symmetry (midline), it is hereby expressly contemplated that the device and surgical methods, and the aspects, components, features and the like thereof, described and/or illustrated herein may be changed, varied, modified, reconfigured or otherwise altered for use or association with another side of the body for a same or similar purpose without departing from the spirit and scope of the disclosure. For example, the tools and methods, and the aspects, components, features and the like thereof, described herein with respect to a right femur may be mirrored so that they likewise function with a left femur and vice versa.
Registration is the process of matching actual anatomical structures such as an actual patient's bone with radiograph, fluoroscopy, and/or computed tomography (CT) scan data that has been entered into a computer. The present disclosure, in some embodiments, is directed generally to processes for reducing the time and increasing the accuracy of a registration process of aligning a pre-operative computer model to its real-time position or “pose” in the surgical field, and/or generally to processes for navigated tracking or updating the registered object's real-time position (generally once a known starting pose has been determined, although the pose can be captured at every sampled position). Mathematically, a transform may be determined to find the starting pose and this transform may then be applied to all of the real-time tracked positions. A benefit of the technique of the present disclosure may include elimination of the need for a manual point sampling registration method of the patient's bone itself while increasing the registering accuracy by sampling more points than may be possible in a comparable time. After registering the actual patient's bone with the computer model or data representing the patient's bone, further automated or robotic procedures may be performed including resection of the patient's bone.
In this illustrated embodiment, the method 10 may be operable to determine a location and an orientation of a patient's bone, and may include, for example, at 20 obtaining, via one or more processors, three-dimensional data representing a patient's bone, at 30 obtaining, via the one or more processors, three-dimensional data representing at least portions of a patient specific bone jig, the patient specific bone jig matched to a portion of the patient's bone, at 40 obtaining, via the one or more processors, image data representing the at least portions of the patient specific bone jig secured to the patient's bone, and at 50 generating, via the one or more processors, data representing the location and the orientation of the patient's bone based on the obtained image data, the obtained three-dimensional data representing the at least portions of the patient specific bone jig, and the obtained three-dimensional data representing the patient's bone. The method may further include at 60 controlling a robotic system having a tool to resect at least a portion of the patient's bone based on the generated data that represents the location and the orientation of the patient's bone.
In some embodiments, one or more features or indicia of the patient specific bone jig 100 may be designed to optimize or aid the performance of a system that utilizes one or more depth cameras for registration and tracking of the patient's bone 12 in a clinical setting. For example, in some embodiments, the patient specific bone jig 100 may be based on features or indicia that include a color of the patient specific bone jig 100. In a surgical incision, there may be numerous colors present, for example skin tissue, fat tissue, muscle tissue, blood, cartilage, surgical retractors, sterilization drapes, etc., that have various colors. Furthermore, the lighting conditions present may be highly variable. The patient specific bone jig 100 may be designed with a non-organic, highly differentiated color such that it is likely to contrast with the surrounding scene. For example, the patient specific bone jig 100 may have a bright orange color or bright green color. In some embodiments, the patient specific bone jig may extend over an outer surface portion of the patient's bone and have a generally constant cross-sectional thickness T.
With reference to
As will be appreciated, using a color filter, a “noisy” scene may be simplified for further processing and computations. Accordingly, a colored patient specific bone jig 100 aids with object identification. The present technique may simplify and reduce the processing time necessary compared to the image 400 (
In some embodiments, a patient specific bone jig may utilize features or indicia such as a surface that is highly irregular with multiple sharp corners 120, such as a polygonal surface coating such as shown in
The patient specific bone jig may be used for registration, or for registration and tracking. For example, training an algorithm to identify the patient specific bone jig having known features or indicia such as a color and/or surface feature may simplify the challenge of utilizing a depth camera for registration and/or tracking. The depth camera may sample thousands of points at a time, which may increase the potential accuracy of the registration.
The position accuracy of robotic systems can be leveraged to calibrate the tool center point (TCP) to a known point on the robot, for example the flange. The end effector can be anodized or otherwise colored to allow for filtering with the RGBD camera. Known visible landmarks can be embedded at known positions on the robot and to which the TCP can be registered.
The RGBD camera can be used to identify the tool type that is engaged with the robot end-effector. For example, if the system calls for a 6 mm burr but a 9 mm burr is engaged, the RGBD camera can use fast feature detection algorithms to identify the object and return this information to the user. By way of nonlimiting example, if a spherical burr is engaged but a non-spherical burr is required, likewise the software algorithms can return this information to the user.
In another embodiment, the RGBD camera may be tracked such that the position of the camera at every sampled point is known. The TCP position can be inferred from multiple readings at multiple known positions (localization).
The handheld depth camera 250 may be used with a fast feature detection algorithm to accurately and quickly find the end-effector and tool tip pose and position. With the ability to accurately measure the special position of the tool tip, the robot can be moved to an assortment of positions (for example XYZ-4 point calibration) and quickly and accurately infer the tool offsets for the control algorithm. For example, the theoretical position may be compared to the actual position and infer a transform. Confirmation may be made of the proper tool (including diameter and tool type) and that the tool is fully engaged and properly seated within the end-effector of the robot for proper application.
Conventional methods for calibrating a robot may include touching reference parts, using distance sensors, and employing laser interferometry. A robot may also be calibrated using external sensors, such as camera systems, that can be attached to various robot locations to acquire the precise position of a reference object. The present disclosure using a camera capable of fast feature identification localizing position with localization algorithms may overcome problems associated with these conventional methods, which conventional methods are time-consuming and complicated.
Notably, the same color RGB filter, as previously described, can be used here. The end-effector, burr guide and tool for example may be colored or anodized to simplify the computational problem of identifying objects in the field. Notably, an advantage of this system is the ability to move the camera to more accurately localize object positions. In other setups, the sensors are often fixed and the robot is moved.
In some embodiments of the present disclosure, the depth camera registers the tool center point (TCP) to the robot flange or some known point on the robot. For example, industrial automation robots may include joint sensing (for example encoders) that return the exact position of each joint. The position of the flange in space can be known with high accuracy. The present disclosure proposes registering the TCP to the position of the flange in space. The depth camera would not need to be tracked by a navigation system to return to its position in space because the registration is to objects within its field of view (the tool tip to the flange). As a redundancy, the RGBD camera may then be tracked to confirm the TCP position is within an acceptable range of the expected position.
The depth camera may not only be used for pose estimation of tracked objects in the surgical field, but it can be used for calibration of the robot tool itself as well as confirmation that the proper cutting tool type is engaged (for example correct diameter and geometry burr) and properly seated in the cutting instrument. The robot desirably knows the exact location of the tool while it works. Each time a different tool is attached, the robot may need to be precisely re-calibrated. The TCP determines the exact working point of the tool in robot tool measurement.
Another technique of the present disclosure is directed to methods of utilizing a patient specific bone jig to improve the performance of conventional point sampling-based registration methods. Conventional point sampling methods are generally a two-step process including: 1) an approximation of the object's pose is determined by sampling points of anatomical importance; and 2) a more accurate position is determined by sampling more points and running an iterative closest point algorithm. The technique of the present disclosure may include patient specific bone jigs having indicia or features that allow readily coordinating the patient specific bone jig and the patient's bone with reduced point sampling.
In some embodiments, the patient specific bone jig 900 may be employed in a registration process that does not rely on a tracked stylus to sample points as described below. For example, the data representing the patient specific bone jig with the fiducials is uploaded into a database that can be referenced by the navigation system. If the navigation system uses IR cameras, as an example, the array that matches the bone jig can be uploaded such that once identified by the navigation software, the patient's bone may be tracked. Similarly, a beacon may be attached to the jig such that the object may be tracked with radar. Once a user places the patient specific bone jig on the proper bone surface location, and the patent specific bone jig is properly identified and tracked by the navigation cameras, the user can sample the position of the patient specific bone jig and an initial pose estimate can be obtained.
A mathematical concept that is important with registration is the concept of the initial pose transform. The position of the tracked object (the patient's bone) will be inferred from the position of the patient specific bone jig. The navigation cameras may track the position of the patient specific bone jig, and not track the position of the patient's bone. The position of the patient's bone may be inferred, determined, and/or generated from the position of the patent specific bone jig, which is inferred, determined, and/or generated from the position of the tracked markers, for example the fiducial array.
To simplify these processes, attention may be made to the origin location for the patient's bone and the patient specific bone jig. When the pre-operative image data of the bone (for example the Dicom file from the CT scan) is generated, the data may be converted into a digital form that can be manipulated with computer aided design software. The digital files may have an origin location. The patient specific bone jig may be designed with the same origin frame as was used as the origin frame for the digital file of the patient's bone. When the navigation cameras track the patient specific bone jig, the origin point of the patient specific bone jig will correspond to the origin point of the patient's bone when the patient specific bone jig is properly placed on the exterior of the patient's bone.
For example, patient bone data 1202 such as three-dimensional data representing at least a portion of a patient's bone such as a proximal portion of a femur of the patient (block 20 in
Patient specific bone jig data 1204 such as three-dimensional data or model representing at least a portion of a patient's specific bone jig (block 30 in
Tool data 1206 such as three-dimensional data or a model representing at least a portion of a robotic tool, e.g., a working end such as a cutting end, may be obtained or inputted to the system 1200. The tool data 1202 may include three-dimensional data previously generated, for example, based on the scanning of the tool or data for fabricating the tool. For example, the tool data 1206 may be limited to sufficient data representing portions of the patient specific bone jig allowing for use in the robotic cut plan.
Image input 1207 such as image data may be obtained from an imager, camera, depth camera, RGB camera, IR sensor, or other suitable imager or scanning device representing at least a portion of a patient's specific bone jig (block 40 in
Pointer input 1208 such as pointer data can be obtained from a pointer or other suitable device for use in locating and orientating at least a portion of a patient's specific bone jig, while the patient's specific bone jig is secured to the patient's bone. The pointer data 1208 may be inputted to the system 1200.
Further input data may include surgeon input 1209 such as desired general data regarding location, orientation, tool, patient specific bone jig, or other data.
The processor 1210 may be a computer operating, for example, WINDOWS, OSX, UNIX or Linux operating system. In some embodiments, the processor 1210 may be a portable or handheld computing device. In other embodiments, the processing unit 1210 may be one or more operably connected processing units, computing devices, servers, linked or operating over one or more networks such as a global communications network, e.g., the Internet.
The memory 1230 may include various modules for processing the input data. For example, the memory 1230 may include the patient specific bone jig generator 1240, a bone registration generator 1250, a tool registration generator 1260, and a robotic bone cut plan generator 1270.
The patient specific bone jig generator 1240 may be operable to determine a configuration of a patient specific bone jig having, for example, an interior surface based on the patient specific bone data 1202, and coordinated or associated outer spatial indicia or fiducials. For example, the patient specific bone jig 1240 may determine the body of the patient specific bone jig having an interior surface that matches or corresponds to a portion of the patient's bone. In some embodiments, a 3D model of the patient specific bone jig may be used by the 3D printer 1295 or other manufacturing device known in the art for generating a patient specific bone jig. The 3D printer or manufacturing device may be operable to form patient specific bone jigs, as described above, as a one-piece, monolithic, or integral body, or formed from one or more components, and formed from a standard metallic material, polymeric material, or other suitable materials.
The bone registration generator 1250 may be operable to receive the patient bone data 1202, the patient specific bone jig data, and the imager input 1207 and/or the pointer input 1208 to determine the actual location and orientation of the patient's bone (block 50,
The tool registration generator 1260 may be operable to receive the tool data 1206, and the imager input 1207, and/or the pointer input 1208 to determine the actual location and orientation of the tool such as when attached to the end of the robot with the patient disposed on an operating table and the patient specific bone jig secured to the patient's bone. For example, using the tool data or tool model, suitable programing may be provided for locating and orienting the actual tool based on the tool data and the observed (using imager or pointer) actual tool attached to the robot. The tool registration generator 1260 may be operable to also confirm or check that the proper tool is engaged for a specific procedure.
The robotic bone cut plan generator 1270 may be operable to determine data or instructions for operating a surgical robot 1290 or other automated devices to, for example, resect a patient's bone. In some embodiments, a 3D model of the resected bone, such as a resected proximal portion of the patient's femur or tibia may be uploaded to the surgical robot 1290 to allow the surgical robot to be operable to effect a bone cut plan to resize the proximal portion of the femur or tibia autonomously, or semi-autonomously to form, for example, a resection and/or form one or more cavities in the patient's bone. The data or instructions may be combined with data received by the surgical robot 1290 such as data from local cameras imagers or sensors. A suitable surgical robot may be an LBR iiwa Kuka robot manufactured by KUKA ROBOTICS Corporation of Shelby Township, Mich., and may be operable with one or more bone saws, rasps, saws, drills, and/or other devices. The robotic bone cut plan generator 1270 may include various modules such as a resection surface or surface generator, a cavity or cavities generator, and an optimizing generator. The robotic bone cut plan generator 1270 may allow for a surgeon to indicate, for example, a resection plane or such plane may be automatically generated provided, e.g., by input from a surgeon, or based on or utilizing predetermined data. The robotic bone cut plan generator 1270 may include receiving initial inputs from a surgeon such as locations, widths, lengths, depths, or may be based on or utilizing predetermined data.
As will be appreciated by one skilled in the art, aspects of the technique may be embodied as a system, method, or computer program product. Accordingly, aspects of the technique may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.), or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module,” or “system.”
It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. Each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which includes one or more executable instructions for implementing the specified logical function(s).
These computer program instructions, also referred to as software and/or program code, may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks. For example, in a particular arrangement, a desktop or workstation computer may be employed using a commercially available operating system, e.g., Windows®, OSX®, UNIX or Linux based implementation.
The computer readable storage medium 1340 may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. The storage 1340 may include an internal storage device, an attached storage device and/or a network accessible storage device. More specific examples (a non-exhaustive list) of the computer readable storage medium include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain or store a program for use by or in connection with an instruction execution system, apparatus, or device.
Computer program code for carrying out operations for aspects of the present technique may be written in any combination of one or more programming languages, including an object oriented programming language, such as Java, Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language, PHP, ASP, assembler or similar programming languages, as well as functional programming languages and languages for technical computing. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). Furthermore, more than one computer can be used for implementing the program code, including, but not limited to, one or more resources in a cloud computing environment.
The Input/Output or I/O devices 1360 (including, but not limited to, keyboards, displays, pointing devices, DASD, tape, CDs, DVDs, thumb drives and other memory media, etc.) can be coupled to the system either directly or through intervening I/O controllers. Network adapters may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modems, and Ethernet cards are just a few of the available types of network adapters.
Data relating to a patient, e.g., the patient's pelvis and hip, may be created by, or accessed from, a medical sensor/diagnostic device. For example, previous medical scans of an extremity, such as those obtained from a computerized axial tomography (CAT or CT) or magnetic resonance imaging (MRI) scan may be stored in a medical record storage apparatus, in storage 1340, or accessed by system 1300. Such patient data may include other data for a given patient (e.g., bone density, type, length, medical conditions etc.). By way of a non-limiting example, the patient data may include a scan data set containing a series of two-dimensional images obtained from the scanning device (e.g., CT scan slices). As such, the scan data set is a 3D dimensional representation of the scan data.
It will be appreciated that the technique of the present disclosure may overcome the time and processing constraints of current state-of-the-art registration and tracking which employs manually sampling a large number of points of interest. In particular, the technique of the present disclosure may overcome the disadvantage of current registration methods that include the needed time required to manually sample a large number of points to achieve a high accuracy registration and the need for the registration to be performed by a skilled user.
For example, with the use of a camera, the present technique may allow for thousands of points to be “sampled” in a very short time, increasing registration accuracy and reducing time. The registration can also be performed by a less skilled user.
From the present description, it will be appreciated that the technique of the present disclosure may provide methods for optimizing the scene and a target to enhance the utility of depth cameras in the surgical field. To utilize a depth camera for registration and tracking in the surgical field, the camera may be operable to 1) identify the clinically relevant object to track, and 2) determine the pose or spatial position and orientation of the object.
The present technique employing patient specific bone jigs with imaging and/or minimal point sampling overcomes problems associated with current registration and tracking. To algorithmically identify anatomical structures is a non-trivial computational problem that may involve machine learning or complex geometric modelling techniques. Determining the pose or spatial position and orientation of an object is also challenging. In clinical practice exposures are often small and the features of a surgical site can be highly variable. Algorithmic feature detection with which to determine spatial orientation can prove a challenge. The availability of data on which to “train” or validate these algorithms is another major impediment to development. Conventional feature detection algorithms on which depth cameras rely struggle correlating to models with small exposures with limited surface anomalies (smooth) in “noisy” environments (cartilage, blood, surgical tools and other soft tissues). Data sets of real-world procedures to train the algorithms are difficult to obtain. Furthermore, unlike spinal anatomy, identifying key anatomical landmarks with knees and hips in limited exposure can be quite challenging.
For example, the technique of the present disclosure may overcome the problems of conventional registration where a user sampling clinically known anatomical landmarks is subjective and prone to error. The present technique reduces the need in conventional registrations of having to sample a high number of points to increase accuracy, which increased sampling increases surgery time. Additionally, because the sampling of points requires a highly skilled user, such as a surgeon, it limits the ability of lower skilled users to support the task. Generating high accuracy registrations in a timely manner continues to be a challenge in the industry.
As may be recognized by those skilled in the art based on the teachings herein, numerous changes and modifications may be made to the above-described and other embodiments of the present invention without departing from the scope of the invention. The jigs, programming, other components of the systems, devices and/or apparatus as disclosed in the specification, including the accompanying abstract and drawings, may be replaced by alternative component(s) or feature(s), such as those disclosed in another embodiment, which serve the same, equivalent or similar purpose as known by those skilled in the art to achieve the same, equivalent or similar results by such alternative component(s) or feature(s) to provide a similar function for the intended purpose. In addition, the devices and apparatus may include more or fewer components or features than the embodiments as described and illustrated herein. Accordingly, this detailed description of the currently-preferred embodiments is to be taken as illustrative, as opposed to limiting the disclosure.
The technique of the present disclosure may be employed with the one or more components described in U.S. provisional patent application no. 62/854,648, filed May 30, 2019, by Sexson et al., entitled “Robot Mounted Camera Registration and Tracking System for Orthopedic and Neurological Surgery” (atty. dock. no. 5247.005P), the entire contents of which are incorporated herein by reference. For example, the patient specific bone jig may be employed as the marker.
The technique of the present disclosure may be employed with the one or more components described in U.S. provisional patent application no. 62/850,050, filed May 20, 2019, by Sexson et al., entitled “A System And Method For Interaction And Definition Of Tool Pathways For A Robotic Cutting Tool” (atty. dock. no. 5247.003P), the entire contents of which are incorporated herein by reference.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprise” (and any form of comprise, such as “comprises” and “comprising”), “have” (and any form of have, such as “has”, and “having”), “include” (and any form of include, such as “includes” and “including”), and “contain” (and any form of contain, such as “contains” and “containing”) are open-ended linking verbs. As a result, a method or device that “comprises,” “has,” “includes,” or “contains” one or more steps or elements possesses those one or more steps or elements, but is not limited to possessing only those one or more steps or elements. Likewise, a step of a method or an element of a device that “comprises,” “has,” “includes,” or “contains” one or more features possesses those one or more features, but is not limited to possessing only those one or more features. Furthermore, a device or structure that is configured in a certain way is configured in at least that way, but may also be configured in ways that are not listed.
The invention has been described with reference to the preferred embodiments. It will be understood that the architectural and operational embodiments described herein are exemplary of a plurality of possible arrangements to provide the same general features, characteristics, and general apparatus operation. Modifications and alterations will occur to others upon a reading and understanding of the preceding detailed description. It is intended that the invention be construed as including all such modifications and alterations.
This application is a Bypass Continuation Application to PCT International Application No. PCT/US2021/022524, filed Mar. 16, 2021, and entitled “Registration and/or Tracking of a Patient's Bone Employing a Patient Specific Bone Jig,” which claims priority benefit of U.S. Provisional Patent Application No. 62/990,827, filed Mar. 17, 2020, entitled “Registration and/or Tracking of a Patient's Bone Employing a Patient Specific Bone Jig,” which both applications are hereby incorporated herein by reference in their entirety.
Number | Date | Country | |
---|---|---|---|
62990627 | Mar 2020 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/US2021/022524 | Mar 2021 | US |
Child | 17932839 | US |