The field of the invention is the use of augmented reality in surgery. More particularly, the present invention can be applied to navigation systems using augmented reality with a servo-controlled robotic arm in order to improve the precision of surgical gestures.
Robotic-based navigation systems for orthopedics currently available on the market are bulky setups that require a huge robotic arm unit, a cumbersome station composed of one or more viewing screens and a perception system generally of the infrared stereoscopic camera type.
The visualization of information useful to the surgeon, such as position and orientation measurements, operative data, is performed on a remote screen. The surgeon's attention is therefore focused on this screen rather than on the real scene, which can disturb the surgeon, complicate his gestures and require an adaptation of his practice compared to more conventional techniques. Moreover, the perception of the three dimensions of a scene can be difficult on a 2D screen, whereas it is essential for the precision of surgical gestures, for example the good orientation of prosthetic components, on which the success of an operation depends.
In one aspect of the present invention, an augmented reality system includes the use of a wearable mobile computing device. Perception sensors can be used to detect anatomical and/or instrument data in the surgical scene. A display component can overlay useful information in the surgeon's field of vision.
In another aspect, a wearable computer is connected to a robotic effector, such as a robotic arm which may carry an instrument, to assist the surgeon in performing some of the gestures of the surgical technique.
In another aspect, a wearable computer is equipped with a perception device, advantageously adapted for visual perception, typically one or a plurality of optical sensors, such as one or more cameras, which may each be secured to the wearable computer or to a different support. Sensors of various technologies can be used, such as RGB camera, infrared camera, monochrome camera, stereovision system by passive or active camera, LiDAR, RADAR, ultrasound devices, etc. A single sensor can be used, or a plurality of sensors using the same or different technologies.
In another aspect, the wearable computer is mobile and adapted to be worn or carried on the user's head, for example, in the form of or including smart glasses. 3D models of anatomical and/or instrumental components may be shown by the display component, alone or in superposition with a view of at least a portion of the surgical scene, which may be an image or a 3D model of at least a portion of the surgical scene.
In another aspect, the robotic effector is controlled to be constrained in displacement in the operating scene accordance with rules or restrictions based on the anatomical and/or instrument data acquired by the perception sensors. Movements of the robotic effector can be slaved to the anatomical and/or instrumental data, so as to delimit an area of action within which the movements of the robotic arm can be restricted.
In another aspect, a program for controlling the system can be stored on a non-transitory computer-readable data storage medium, such as a memory of the system.
A system according to the invention can be used, for example, in the context of total or partial knee replacement surgery, where certain practical imprecisions may cause the femoral and tibial components orientations to differ from the values initially planned by the surgeon, such as incidence angles, hold of the tool, fixation of the instruments, etc. A robotic arm can assist in achieving optimal control of such orientations, which depend on the precision of the cuts performed on the patient's bones. An augmented reality system can assist in determining or approximating the values planned by the surgeon pre-operatively. Sensor(s) can complement measurement made by a perception device of an augmented reality system.
In at least some embodiments, a system combines the qualities of the robotic arm with the small size of a portable navigation system that relies on augmented reality, offering the surgeon an improved visualization of the information and assistance in controlling the robotic arm. At least one processor, at least one memory, or both are included in the wearable computer or hosted in a separate component of the system, such as a station carrying the robotic effector, or an additional device, such as a controller of the system.
In at least some embodiments, some measurements, or even all measurements, are carried out by the perception system of the wearable computer using markers present on the instrumentation and the patient's bones.
Further aspects of the invention will be described in reference to the appended drawings which illustrate non-limiting exemplary embodiments, and among which:
In a surgical operation using a robot, in order to be able to control the robot effectively, it is advantageous to provide constraints to guide the gesture of the surgeon. To this effect, it is necessary to locate the robot and in particular its effector, e.g., a robotic arm, such as an articulated robotic arm, possibly including an instrument carried by the robotic arm, in the reference frame of the perception system of the wearable computer.
For this purpose, at least one marker detectable by the perception system can be positioned on the robot. For example, a marker can be positioned on the effector, such as at the tip of a robotic arm.
As shown in
Alternatively, or additionally, as shown in
In still another embodiment, the effector is associated with at least two cameras synchronized with the perception system. In this configuration, the localization of the robot's effector in the 3D reference frame of the cameras can be obtained by matching visual primitives in the images coming from the at least two cameras, in an approach similar to stereoscopy. In this embodiment, the cameras may include a first camera 1a mounted on the effector, as shown in
The measurements thus taken make it possible to define anatomical landmarks on the patient's anatomy in the 3D surgical scene. The robotic arm can then navigate through the surgical scene and models created by these measurements.
The robot's effector can accommodate different instruments depending on the type of acquisition and the stage of surgery. The tools can be provided on the effector successively or together, for successive or simultaneous use. For example, for the modeling of a bone surface, such as the acquisition of a points cloud and registration with the 3D model of the patient, or for the acquisition of anatomical landmarks, a pointer-type instrument can be used.
Alternatively, additionally, or complementarily, the acquisition of the points cloud can be based on the data generated by a depth perception device, such as a LIDAR or a RGBD camera, advantageously installed on the robot's effector, for example, like other surgical tools.
In the context of surgery, particularly knee surgery, a robotic saw can be attached to the robot's effector and navigated by the wearable computer to assist the surgeon's gesture in performing cuts, such as a cut of the femur and/or of the tibia in knee surgery. For shoulder surgery, a pin or a drill can be attached to the robot's effector and navigated via the wearable computer to assist the surgeon's action in drilling the glenoid, for example.
The robotic arm is manipulated by the surgeon to perform the desired operation(s), such as cutting or drilling. The data acquired via the perception system and the planning defined by the surgeon make it possible to delimit one or more cutting or drilling area, if desired. The coordinates obtained constrain the movement of the robotic arm in space for the surgical gesture, setting a framework outside of which robot interventions are not possible.
Movements of the robot can be fully controlled by the system in accordance with predetermined parameters, or they can be co-manipulated by the surgeon. For example, the movements of the robot can accompany movements of the surgeon when the latter is in the cutting zone, and constrain him if he moves away from it.
A control of the robotic arm using the position of the robot and the patient's bones is calculated thanks to the measurements of the perception system, making it possible to determine the action to be carried out. In a particular embodiment, the system control of the movements can be deactivated temporarily or permanently, during which time the effector can be freely controlled and manipulated only by the surgeon.
In a first stage 11, visual sensors, such as cameras 1a, 1b, etc., are used to capture images i) of the robot or portions of the robot, such as all or part of the robotic arm, and ii) of the patient or portions of the patient, such as all or part of the operation zone of the patient. Reference markers can be used or not. The reference markers, if used, can be placed on the robot and/or on the patient. Thus, the images may include visual reference markers or no visual reference markers.
In a second stage 12, the images are processed to obtain 1) a localization of the robot, and advantageously, more precisely a location of the robotic arm and/or a localization of a surgical tool carried by the robotic arm, and ii) a localization of the patient's bones, and advantageously, more precisely of an operation area of the patient.
As indicated above, the localization of the robotic effector can be obtained on the basis of a geometric and/or cinematic model of the robot and/or of the robotic arm and/or of the surgical tool, preferably via 3D model(s).
Thus, in a further stage 13, based on the localization of the robotic effector, a localization of the surgical tool can be obtained, optionally using a geometric and/or cinematic model of the robot and/or of the robotic arm and/or of the surgical tool, preferably via 3D model(s).
In a next stage 14, a set of anatomical values are obtained, based on the localization of the patient's bones and optionally of the cutting instrument.
Next, in a stage 15, the anatomical values can be validated by the user, typically the surgeon. For example, the values are shown to the surgeon via the augmented reality display of the wearable computer worn by the surgeon. The surgeon then validates the values, for example, by pressing a key or button, or by performing a predetermined movement or providing another visual or vocal instruction detectable by the system.
Optionally, the system then is adapted to provide or to store in one or more memories anatomical values useful to the surgery, such as varus and/or valgus, resection data, etc., as shown by stage 16 on
Depending on the type of robot used, several surgical assistance options can be offered to the surgeon. For example, in the context of surgery including cutting bone, such as partial or total knee surgery:
The robot can be equipped with a cutting guide (a metal block with a thin slot), in which case the robot is automatically positioned in front of the patient's bones according to the steps of the operating protocol. The surgeon then performs the bone cut by manipulating an oscillating saw whose blade is guided by the position of the robot cutting guide.
The robot is directly equipped with a motor with an oscillating saw blade, in which case the robot is co-manipulated by the surgeon who is then free to move the instrument at the end of the robotic arm, but constrained in the cutting plane defined during the operating protocol.
The surgical gesture can be carried out with precision on the patient's bone, even if it is in motion because the perception system permanently locates the instruments and the bone structures, which allows real-time servoing of the robot with respect to defined cutting planes.
In this example, the oscillating saw shown on
The combination of augmented reality visualization and robotic assistance provides additional safety to the surgeon as the system can display critical areas related to the robot's movements directly in the surgical scene, present the surgeon with virtual control elements displayed directly on the robot, and display “transparently” through the patient's soft tissues instruments that would normally be hidden to prevent the risk of injury (rupture of a ligament, nerve or artery, for example).
In a variant embodiment, the use of markers is avoided, in that spatial localization is obtained based on 3D models of anatomical features and/or measuring and/or surgical instruments. This approach makes it possible to locate specific objects in the space of the operating scene, for example, anatomical surfaces such as articular surfaces, instruments such as those carried by the robotic arm, etc., as well as the position of the camera itself in the space of the operating scene based solely on their geometric and/or cinematic characteristics.
In another variant embodiment, the robotic arm is also used to carry out acquisitions, making it possible to digitize reference points and articular surfaces on the patient's bones. This is made possible by the installation of appropriate instruments at the end of the robotic arm, such as a probe or an active sensor of the 3D scanner type. The robot's internal localization system can be used. A potential advantage of this embodiment is the possibility of bringing the instrument very close to the measured zone. Another potential advantage is the possibility of visualizing directly on the surface of the patient's bone the joint areas already digitized and those still to be acquired, or of indicating by visual indicators the level of detail or precision of the acquired surfaces. Another advantage is to avoid the occlusion of the markers installed on the patient's bones and surgical instruments from the perception system, which can occur often during the surgical protocol.
Of course, the principle of a navigation system coupling augmented reality and a robotic arm can be applied to other operating contexts, for example to other joints: shoulder, hip, etc., according to the same principles as those previously defined for surgery of the knee. For example, in various applications, the system can be used to assist a surgeon in drilling and/or cutting bones for which the orientation of prosthetic components represents a challenge, for example, narrow approach, bulky soft tissues, hidden bone structure, etc.
Augmented reality makes it possible to display virtual information in the operating scene directly in the surgeon's field of vision, either superimposed on the reality perceived by the surgeon (“mixed reality” vision) or superimposed on a video stream displayed in his field of vision (“augmented reality” vision). The robotic arm makes it possible to reduce inaccuracies induced during the manipulation of the instruments by the surgeon, and/or to guide his gesture through co-manipulation constraints. The combination of augmented reality visualization and robotic guidance can provide additional safety to the surgeon because the system can display critical areas related to robot movements directly in the operating scene, and/or present the surgeon with virtual control elements displayed directly on the robot, and/or display in “transparency” through the patient's soft tissues instruments that would normally be hidden to prevent the risk of injuring the patient, such as through rupture of a ligament, nerve or artery, for example.
Thus, a major potential advantage of this navigation solution compared to existing systems lies in the combination of advantages of augmented reality and of robotics, as described above. For example, the system can integrate augmented reality glasses by Vuzix, Rochester, NY, for example, a model M400, which will cooperate with robot effectors from various manufacturers, such as Stäubli, Pf{dot over (a)}ffikon (Switzerland).
The augmented reality display can be a single 2D image view with superposed information, or an actual 3D view by the eye of the user, with superposed stereoscopic information provided by the device, or even a virtual simulation of an actual view with superposed information, or any combination of these augmented reality displays. The virtual simulation of an actual view can be created based on an actual view, on a model, such as a 2D or 3D model, of the actual view, from a predefined model, such as a 2D or 3D model, or any combination of these simulations. A portion of an actual view or of an image of an actual view may be excluded or occluded in favor of the augmented reality information, such as by framing and/or covering.
The above disclosures, embodiments and examples are nonlimitative and for illustrative purposes only.
This application claims priority of provisional application No. 63/229,312 filed Aug. 4, 2021, whose content is hereby incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5721566 | Rosenberg | Feb 1998 | A |
5959613 | Rosenberg | Sep 1999 | A |
6020876 | Rosenberg | Feb 2000 | A |
7056123 | Gregorio | Jun 2006 | B2 |
7155316 | Sutherland | Dec 2006 | B2 |
7404716 | Gregorio | Jul 2008 | B2 |
7660623 | Hunter | Feb 2010 | B2 |
8831779 | Ortmaier | Sep 2014 | B2 |
9060796 | Seo | Jun 2015 | B2 |
9220570 | Kim | Dec 2015 | B2 |
9480534 | Bowling | Nov 2016 | B2 |
9539726 | Simaan | Jan 2017 | B2 |
9566121 | Staunton | Feb 2017 | B2 |
9630323 | Kishi | Apr 2017 | B2 |
9937014 | Bowling | Apr 2018 | B2 |
9956042 | Simaan | May 2018 | B2 |
10034675 | Pack | Jul 2018 | B2 |
10251709 | Cho | Apr 2019 | B2 |
10350017 | Bowling | Jul 2019 | B2 |
10390890 | Jagga | Aug 2019 | B2 |
10420619 | Bowling | Sep 2019 | B2 |
10463440 | Bowling | Nov 2019 | B2 |
10695134 | Barral | Jun 2020 | B2 |
10806527 | Bar | Oct 2020 | B2 |
11033296 | Carlson | Jun 2021 | B2 |
11045958 | Bowling | Jun 2021 | B2 |
11471232 | Bowling | Oct 2022 | B2 |
11750794 | Benishti | Sep 2023 | B2 |
11766296 | Wolf | Sep 2023 | B2 |
11801115 | Elimelech | Oct 2023 | B2 |
11896445 | Gera | Feb 2024 | B2 |
11911277 | Ratron | Feb 2024 | B2 |
11974887 | Elimelech | May 2024 | B2 |
11980429 | Wolf | May 2024 | B2 |
11980506 | Wolf | May 2024 | B2 |
20050162383 | Rosenberg | Jul 2005 | A1 |
20060261770 | Kishi | Nov 2006 | A1 |
20070171200 | Rosenberg | Jul 2007 | A1 |
20090037033 | Phillips | Feb 2009 | A1 |
20090088774 | Swarup | Apr 2009 | A1 |
20100331855 | Zhao | Dec 2010 | A1 |
20100331859 | Omori | Dec 2010 | A1 |
20110015649 | Anvari | Jan 2011 | A1 |
20110130761 | Plaskos | Jun 2011 | A1 |
20110152676 | Groszmann | Jun 2011 | A1 |
20110208256 | Zuhars | Aug 2011 | A1 |
20110263971 | Nikou | Oct 2011 | A1 |
20110264107 | Nikou | Oct 2011 | A1 |
20110301500 | Maguire | Dec 2011 | A1 |
20120053597 | Anvari | Mar 2012 | A1 |
20120059378 | Farrell | Mar 2012 | A1 |
20120071752 | Sewell | Mar 2012 | A1 |
20120071893 | Smith | Mar 2012 | A1 |
20120083801 | Nixon | Apr 2012 | A1 |
20130325029 | Hourtash | Dec 2013 | A1 |
20130345718 | Crawford | Dec 2013 | A1 |
20140039517 | Bowling | Feb 2014 | A1 |
20140039681 | Bowling | Feb 2014 | A1 |
20140052153 | Griffiths | Feb 2014 | A1 |
20140135795 | Yanagihara | May 2014 | A1 |
20140148818 | Komuro | May 2014 | A1 |
20140222207 | Bowling | Aug 2014 | A1 |
20140276952 | Hourtash | Sep 2014 | A1 |
20140277742 | Wells | Sep 2014 | A1 |
20140350571 | Maillet et al. | Nov 2014 | A1 |
20140378999 | Crawford | Dec 2014 | A1 |
20150032164 | Crawford | Jan 2015 | A1 |
20150081098 | Kogan | Mar 2015 | A1 |
20180071026 | Malackowski | Mar 2018 | A1 |
20210169605 | Calloway | Jun 2021 | A1 |
20230076894 | Agustinos | Mar 2023 | A1 |
Entry |
---|
Zimmer Biomet, “Rosa(R) Knee System”, https://www.zimmerbiomet.com/en/products-and-solutions/specialties/knee/rosa--knee-system.html, visited Aug. 3, 2022 (total 13 pages). |
Stryker, “Mako SmartRobotics Overview”, https://www.stryker.com/us/en/joint-replacement/systems/Mako_SmartRobotics_Overview.html#know-more, visited Aug. 3, 2022 (total 6 pages). |
Number | Date | Country | |
---|---|---|---|
20230076894 A1 | Mar 2023 | US |
Number | Date | Country | |
---|---|---|---|
63229312 | Aug 2021 | US |