The present invention relates to the field of robotic surgery, especially systems involving the coordination of imaging of the patient with robotic surgical procedures.
There exist a number of surgical robotic systems, which enable accurate positioning or operation of surgical tools in the performance of surgical procedures. One of the primary problems in such systems is the ability to relate the position and orientation, i.e. the pose, of the surgical tools, with the preoperative surgical plan, or with intraoperative images taken during the surgery. Relating the pose of surgical tools with a preoperative surgical plan is a problem which is solved by using a registration procedure, usually between the features on which it is desired to operate according to the surgical plan, as shown in the preoperative images, and the real-life intraoperative position of the robotically guided tools, and the patient on the operating bed. The preoperative images are usually three dimensional images, such as CT or MRI image sets, while the intraoperative surgical situation is generally obtained by use of fluoroscopic images obtained using a C-arm imaging system.
There further exist a number of systems in which the historically conventional C-arm system, having a rigid support member circumscribing the patient, with an X-ray source on one end of the member and an X-ray collector or detector plate on the other end of the member, has been superseded by “virtual C-arms” having robotically aligned elements. In such systems, an X-ray source and a detector element are carried on the ends of separate robotically activated arms, such that they can be positioned at opposite sides of the patient. The exact pose of the X-ray fluoroscopic image can be selected by means of robotic control of the arms respectively carrying the source and the imaging element. Such systems have been described in U.S. Pat. No. 6,582,121 for “X-Ray Positioner with Side-Mounted Independently Articulated Arms” to M. Crain et al, assigned to GE Medical Systems Global Technology, and in U.S. Pat. No. 6,644,852 for “Automatically Reconfigurable X-Ray Positioner” also to M. Crain and also assigned to GE Medical Systems. In U.S. Pat. No. 8,611,495 for “Biplane X-Ray Imaging System” to M. Maschke, assigned to Siemens AG, there is described an imaging system having two recording units disposed on C-arms in different planes. This patent also mentions the possibility of having each X-ray tube assembly and X-ray detector disposed individually on retaining elements such as robotic arms. Similar systems using X-ray source and X-ray detector panel mounted on separate robotic arms are also shown in U.S. Pat. No. 6,200,024 for “Virtual C-Arm Robotic Positioning System for Use in Radiographic Imaging Equipment” to D. Negrelli, assigned to Picker International. Patent application No. DE 10 2015 217059 A1 by M. Hörning, filed by Siemens Healthcare GmbH, also describes an X-ray system having the X-ray emitter and X-ray receiver on separately controlled support arms. US published patent application No. 2011/0069818 to M. Muller, assigned to Kuka Roboter GmbH, also describes an X-ray system with the source and receiver on separate robotic arms. U.S. Pat. No. 6,435,715 to R. Betz et al, for “Radiography Device” assigned to Siemens AG, also describes a system having an X-ray source and an X-ray receiver mounted on different robotic arms, to generate a virtual C-arm arrangement. The robotic control ensures that the X-ray source is always oriented opposite to the X-ray receiver for the purpose of acquiring images. The patient position may be related to the robotic co-ordinate system by means of image analysis of markers using an external camera system connected to the robotic control system.
U.S. Pat. No. 8,781,630 for “Imaging Platform to Provide Integrated Navigation Capabilities for Surgical Guidance” to S. A. Banks et al, assigned to the University of Florida Research Foundation Inc., describes a system which includes two robotic arm systems, one holding an imaging source, and the other holding an imaging sensor, like the systems previously mentioned above. A third robotic arm system can be included as a surgeon guided tool-holder. The tool is manually guided by the surgeon using haptic control and optional force feedback. The spatial poses of the imaging source robotic arm, and the tool-holder robotic arm can be related to each other by using X-ray calibration targets on the imaging source arm and on the tool-holding arm, or a tracking system, using an electromagnetic or optical tracker which can locate a localizing marker on the imaging source arm and on the tool-holding arm. The tool tip can be registered with the navigation reference frames using a touch device.
However, the above described systems are limited mostly to imaging functionality, or to haptic systems, in which the systems mimic robotically controlled conventional C-arm imaging functionality
The disclosures of each of the publications mentioned in this section and in other sections of the specification, are hereby incorporated by reference, each in its entirety.
There exists a need for a more comprehensive robotically controlled virtual C-arm system that overcomes at least some of the disadvantages of prior art systems and methods. Modern robotic-guided surgery requires the ability to automatically relate any imaging function, to the current execution of the surgical plan. The present disclosure describes new exemplary systems for executing robotically guided surgery, in combination with an imaging system that enables a robotically guided tool to be related to the image or images generated with the imaging system. The system is based on a robotically controlled virtual C-arm, having a frame of reference that is relatable to the frame of reference of a surgical tool, such that it can autonomously guide the surgical tool using intraoperative images generated using the virtual C-arm. Such a system, has more versatile functionality and higher accuracy than robotic systems which use registration or navigation or tracking to define the surgical tool for performing the surgical plan.
The systems are based on a twofold combination of at least two robotic arms to enable these novel features to be achieved during surgery. The robotic arms are mounted on a common base, this term being understood to include both having all of the arms mounted on a single base or alternatively having at least some of the arms mounted on separate bases, with the mutual spatial positions of those bases being known to each other. The term “common base” is also thuswise claimed. Because of this feature of a common base, the coordinate system in which each robotic arm operates, can be uniquely related to the coordinate systems of all of the other robotic arms in the system. This feature is important because it enables all of the arms to function co-operatively.
A first pair of arms is provided, in which one arm holds the imaging system source, and the other arm holds the collector or detector element for generating the image itself of the surgical region. The arms are robotically guided, such that their absolute position and orientation in space, and relative to each other, is known at all times. Consequently, they can adopt any source-sensor position desired, according to the robotic control commands issued to them. The reference frame of such an imaging system is known relative to the common base of all of the robotic arms, and hence also to the reference frames of any other robotic arms mounted on the common base. Since the robotic arms are an integral part of the robotic surgical suite, there is no need for a separate C-arm imager, which would take up valuable space in the operating room, and requires a medical technician and alignment time to operate it. The imaging system can be of any suitable type, including an X-ray imager, an ultrasound imager, or any other type of imaging system having separate source and detection elements for mounting on separate robotic arms of the system. However, since the most commonly used such imaging system is the X-ray system, as used on the usual C-arm systems, the disclosure herewith uses the X-ray example for describing the imaging system and its operation, though it must be emphasized that the system and methods are not intended to be limited to X-ray imaging.
The current system differs, however, from previously proposed virtual C-arm systems, this term being used to describe imaging systems in which the X-ray source and the X-ray detection device are mounted on co-related controlled robotic arms, in that a surgical tool is positioned, and its operation optionally also controlled, by a robotic arm whose coordinate system is known to the coordinate system of the X-ray imaging system. One method of achieving this is by using a third robotic arm to carry and position or activate the surgical tool. The third arm may even be an active robotic arm, which can be equipped with a surgical scalpel, a surgical drill, a guide tube, a retractor, or some other surgical tool, and which can then perform a robotically controlled surgical action on the patient. The use of an active robotic arm in this system is advantageous, since the pose of the surgical tool is then known with high accuracy relative to the fluoroscopic imaged data, since the coordinate systems of the third robotic arm carrying the tool is known to the control system of the co-ordinate system of the imaging stage, the co-ordinate systems of all of the arms being co-related to each other. Thus the position of the surgical tool is achieved with the same robotic control as was used in aligning the imaging stage, and can be guided to a position which is known on the fluoroscopic images obtained by the system.
According to a second and generally simpler implementation of such systems, the surgical tool holder is mounted on the same robotic arm as one of the imaging components, preferably the detection plate since it is substantially lighter than the X-ray source, such that the pose of the surgical tool is known relative to the X-ray detection plate with high accuracy. The tool can most advantageously be mounted on a holder which is attached to the same robotic arm as the detection plate, such that the comparative pose of the tool and the detection plate are known and fixed. This provides the optimum accuracy of position between the tool and the other robotically moved imaging elements. Alternatively, the detection plate can be provided with an accurate demountable attachment fixture to its robotic arm, and the tool-holder equipped with a similar demountable attachment fixture, such that the tool can be mounted on the arm instead of the detection plate in much the same way as the tool exchanger in a machine tool operates. In either of these two implementations the surgical tool performs the surgical operation in the same reference frame as that in which the images were acquired. Once the required imaging has been performed using the imaging pair of robotic arms, the detection plate can be demounted and the tool assembly mounted in its place, such that the surgical tool can then perform a surgical action on the patient, with the same advantages as those mentioned in the previous paragraph. Both of these implementations provide improved tool position accuracy than types of registration procedure using preoperative images to define the tool insertion pose, in which the tool position is not directly related to the co-ordinates of the imaging system.
In such an arrangement of using only two robotically controlled arms, a third robotic arm can then also be used to perform another associated surgical action, such as a retraction or holding action in conjunction with the surgical operation procedure being performed by the now “vacated” arm of the imaging pair of arms.
One exemplary procedure by which the system is able to align the tool using only fluoroscopic images obtained intraoperatively using the robotically controlled arms, could be as follows:
In the above described procedure, use is made only of the two dimensional fluoroscope images to align the tool or its guide along the correct trajectory, and there is no need for registration at all since there is no preoperative planning—the “planning” is based on direct vision of the subject's anatomy in the fluoroscope images. Although present day registration techniques have good accuracy, the elimination of the need for this step eliminates any potential inaccuracy which may arise with such a registration procedure.
Another advantages of using the imaging robot of the present disclosure is the ability to automatically detect the correct trajectory according to an algorithm that calculates the correct trajectory from images in a machine vision and machine learning mode. In such procedures, the robot takes one or more x-ray images, and, based on this algorithm moves the detector to the expected straight-on view of the region of interest, and takes successive images until it converges to an image approved by the surgeon to apply the surgical act.
According to further implementations of the above described systems, input from preoperative three-dimensional image data, showing either fiducial markers or anatomical details, can also be used in order to register a preoperative surgical plan produced on the preoperative three-dimensional image data, with real time fluoroscopic images obtained from the X-ray fluoroscopic imager mounted on the imaging pair of robotic arms. Such a registration will then enable monitoring the action of the surgical tool and of its accuracy relative to the preoperative plan. This procedure thus enables high resolution preoperative three-dimensional images, such as CT or MRI image sets, to be used as the basis for defining the pose of the tool held in one of the robotic arms. This procedure enables the system to perform a surgical procedure with the advantages both of real-time intraoperative imaging and of high resolution preoperative three-dimensional images. Though less convenient and potentially having somewhat reduced accuracy than the use of the intraoperative surgical planning methods described previously in this disclosure, in which intraoperative imaging is used using combinations of fluoroscope images, this use of planning on the basis of preoperative high resolution images may be advantageous when the region of interest contains features which do not enable the provision of high quality fluoroscope images, such as when there are only soft tissues, or when there is a preponderance of opaque bone structures in the region of interest.
There is thus provided in accordance with an exemplary implementation of the devices described in this disclosure, a system for a robotic surgical system comprising:
In such a system, the surgical tool or tool holder may be carried on one of the first or second robotic arms in addition to the source or detection element, or it may be carried on one of the first or second robotic arms in place of the source or detection element.
Furthermore, in such a system, the robotic arm configured to carry a surgical tool or a surgical tool holder may be the second robotic arm. This second robotic arm may be adapted to carry simultaneously, both the surgical tool or tool holder, and the detection element in a known spatial relationship. Additionally, the second robotic arm may comprise an attachment element adapted to be attached either to the surgical tool or tool holder, or to the detection element, the attachment element being such that the spatial relationship between the surgical tool or tool holder, and the detection element is accurately known.
In any of the above described systems, the known relationship between the tool or tool holder, and at least one of the robotic arms may be configured to enable autonomous guidance of the surgical tool to a position on the images obtained by use of the first and second robotic arms, determined by a user.
Furthermore, any of the above described systems may further comprise a third robotic arm whose co-ordinate system is co-related to those of the first and second robotic arms, and which is adapted to hold additional surgical elements.
In all of the above described systems, the imager may be either an X-ray imager or an ultrasound imager.
There is further provided according to other implementations of the robotic surgical systems of the present disclosure, a system comprising at least first, second and third robotic arms, the robotic arms being mutually mounted such that their co-ordinate systems are known relative to each other, at least the first and the second robotic arms being disposed on opposite sides of a support element on which a subject may be positioned, and which are configured to carry respectively an imager source and an imager detection element such that images of a portion of the subject's anatomy can be generated,
wherein the third robotic arm may be configured to carry a surgical tool holder or tool, such that the pose of the surgical tool or tool holder is known relative to images generated by the first and second robotic arms. In such a system, the imager may be either an X-ray imager or an ultrasound imager.
According to yet further implementations, there is provided a method of performing a surgical procedure on a region of a subject, comprising:
In such a method, if the surgical tool is carried on one of the first or second robotic arms, it may be carried either in addition to the imager source or detector element, or may be carried in place of the imager source or detector element. Additionally, if it is carried on the third robotic arm, the trajectory necessary for performing the procedure is assured by virtue of the commonly related coordinate systems of the third robotic arm to those of the first and second robotic arms. In any of these methods, the procedure may be performed using intraoperative alignment of the tool trajectory in at least one image generated using the imaging system having a co-ordinate system common to that of the tool.
There is further provided a method of performing a surgical procedure on a region of a subject, comprising:
According to yet another exemplary implementation of the present systems, there is provided a robotic surgical system comprising:
Finally, according to yet another implementation of the systems of the present disclosure, there is proposed a robotic surgical system comprising:
The present invention will be understood and appreciated more fully from the following detailed description, taken in conjunction with the drawings in which:
Reference is now made to
A third robotic arm 14, can carry a surgical tool 17, or a surgical tool holder, and since the frame of reference of this third robotic arm 14 is known relative to those of the first and second arms 12,13, the position and orientation of the surgical tool is known relative to the co-ordinate system of the fluoroscopic images generated on the imager arranged on the first 12 and second 13 arms. Consequently, the relative position and progress of the surgical tool 17 during the operation can be directly related to the fluoroscopic images of the surgical site of the patient, without the need for further external registration.
According to an alternative mode of operation, once the fluoroscopic image or images have been generated to define the features of the surgical site of the patient, at least one element of the X-ray imaging equipment may be removed from its supporting robotic arm—preferably the detector plate 16 from the robotic arm 13, since that is the lighter element—and that robotic arm 13, is then free to be equipped with a surgical tool or tool holder, whose pose is known relative to the previously generated X-ray images, by virtue of the tool being attached at a known position and angle to the robotic arm 13, whose position is known relative to the X-ray images axis. Alignment and progress of the tool using the robotic control system of the robotic arm 13, can therefore be directly related to the fluoroscopic images previously obtained, without the need for any registration transformation. The third robotic arm 14, since it is no longer required to hold the tool holder or tool, can then be used to perform an additional surgical task, such as retraction or holding of the patient's tissue.
Reference is now made to
Reference is now made to
In any of the above described implementations, image processing of the X-ray images can be used to define the position of the patient, or anatomical features of the patient to be operated on, and the known relationship between the imaging frame of reference and the tool frame of reference, enables accurate positioning of the tool relative to the patient or the anatomical feature of the patient to be operated on. Such a system can be configured to autonomously guide a surgical tool to a position which is known on the X-ray images obtained by the system itself, without the need for any other alignment, since the frame of reference used for generating the images can be spatially and angularly related to the frame of reference in which the tool is mounted.
Reference is now made to
The processor 46 controls the entire controller operation, including input-output and calculations. The input output units include a user interface 43, a robotic arm drive interface 47, a robotic arm position sensors interface 48 and a network interface 49.
It is appreciated by persons skilled in the art that the present invention is not limited by what has been particularly shown and described hereinabove. Rather the scope of the present invention includes both combinations and subcombinations of various features described hereinabove as well as variations and modifications thereto which would occur to a person of skill in the art upon reading the above description and which are not in the prior art.
This application is a national stage application under 35 U.S.C. 371 and claims the benefit of PCT Application No. PCT/IB2019/058795 having an international filing date of 15 Oct. 2019, which designated the United States, which PCT application claimed the benefit of U.S. Provisional Patent Application No. 62/745,497 filed 15 Oct. 2018, the entire disclosures of each of which are incorporated herein by reference.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/IB2019/058795 | 10/15/2019 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2020/079596 | 4/23/2020 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5329933 | Graf | Jul 1994 | A |
6152890 | Kupfer et al. | Nov 2000 | A |
6200024 | Negrelli | Mar 2001 | B1 |
6435715 | Betz et al. | Aug 2002 | B1 |
6582121 | Crain et al. | Jun 2003 | B2 |
6644852 | Crain et al. | Nov 2003 | B2 |
7198630 | Lipow | Apr 2007 | B2 |
7567834 | Clayton et al. | Jul 2009 | B2 |
7708741 | Bonutti | May 2010 | B1 |
7899226 | Pescatore et al. | Mar 2011 | B2 |
7907166 | Lamprecht et al. | Mar 2011 | B2 |
8073528 | Zhao et al. | Dec 2011 | B2 |
8108072 | Zhao et al. | Jan 2012 | B2 |
8147503 | Zhao et al. | Apr 2012 | B2 |
8392022 | Ortmaier et al. | Mar 2013 | B2 |
8568317 | Gharib et al. | Oct 2013 | B1 |
8611495 | Maschke | Dec 2013 | B2 |
8721566 | Connor et al. | May 2014 | B2 |
8781630 | Banks et al. | Jul 2014 | B2 |
8870889 | Frey | Oct 2014 | B2 |
9050728 | Ban et al. | Jun 2015 | B2 |
9259276 | Mintz et al. | Feb 2016 | B2 |
9259278 | Jensen | Feb 2016 | B2 |
9259282 | Azizian et al. | Feb 2016 | B2 |
9307894 | von Grunberg et al. | Apr 2016 | B2 |
9402689 | Prisco et al. | Aug 2016 | B2 |
9592096 | Maillet et al. | Mar 2017 | B2 |
9649072 | Ragnarsdottir et al. | May 2017 | B2 |
9662174 | Taylor et al. | May 2017 | B2 |
9687301 | Lee et al. | Jun 2017 | B2 |
9767608 | Lee et al. | Sep 2017 | B2 |
9789313 | Lipani | Oct 2017 | B2 |
9888966 | Farritor et al. | Feb 2018 | B2 |
9895813 | Blumenkranz et al. | Feb 2018 | B2 |
9969090 | Warashina et al. | May 2018 | B2 |
10076385 | Shoham et al. | Sep 2018 | B2 |
10149729 | Smaby et al. | Dec 2018 | B2 |
10271832 | O'Neil et al. | Apr 2019 | B2 |
10299880 | Ramirez Luna et al. | May 2019 | B2 |
10368054 | Panescu et al. | Jul 2019 | B2 |
10383765 | Alvarez et al. | Aug 2019 | B2 |
10390895 | Henderson et al. | Aug 2019 | B2 |
10456076 | Liu | Oct 2019 | B2 |
10456211 | McAfee | Oct 2019 | B2 |
10512413 | Schepis et al. | Dec 2019 | B2 |
10561384 | Tanaka et al. | Feb 2020 | B2 |
10668625 | Kuroda et al. | Jun 2020 | B2 |
10675098 | Zhao et al. | Jun 2020 | B2 |
10675107 | Steger et al. | Jun 2020 | B2 |
10812778 | Wang et al. | Oct 2020 | B1 |
11298195 | Ye et al. | Apr 2022 | B2 |
20040122427 | Holmes | Jun 2004 | A1 |
20050096502 | Khalili | May 2005 | A1 |
20060241414 | Nowlin et al. | Oct 2006 | A1 |
20060258938 | Hoffman et al. | Nov 2006 | A1 |
20070106307 | Bodduluri et al. | May 2007 | A1 |
20070232958 | Donofrio et al. | Oct 2007 | A1 |
20090088634 | Zhao | Apr 2009 | A1 |
20110069818 | Muller | Mar 2011 | A1 |
20110087238 | Wang et al. | Apr 2011 | A1 |
20110106102 | Balicki | May 2011 | A1 |
20110172566 | Kawchuk | Jul 2011 | A1 |
20110172720 | Metcalf, Jr. et al. | Jul 2011 | A1 |
20110276179 | Banks | Nov 2011 | A1 |
20120029694 | Mueller | Feb 2012 | A1 |
20120041562 | Shachar et al. | Feb 2012 | A1 |
20130012831 | Schmitz et al. | Jan 2013 | A1 |
20130096574 | Kang et al. | Apr 2013 | A1 |
20130211419 | Jensen | Aug 2013 | A1 |
20130268007 | Rezach et al. | Oct 2013 | A1 |
20140163736 | Azizian et al. | Jun 2014 | A1 |
20140179997 | von Grunberg et al. | Jun 2014 | A1 |
20140188132 | Kang | Jul 2014 | A1 |
20140253684 | Kumar et al. | Sep 2014 | A1 |
20140309659 | Roh et al. | Oct 2014 | A1 |
20140350571 | Maillet | Nov 2014 | A1 |
20150297177 | Boctor et al. | Oct 2015 | A1 |
20160081754 | Kostrzewski et al. | Mar 2016 | A1 |
20160184032 | Romo et al. | Jun 2016 | A1 |
20170273614 | Giphart et al. | Sep 2017 | A1 |
20180064497 | Hussain et al. | Mar 2018 | A1 |
20180193101 | Hashimoto | Jul 2018 | A1 |
20180263530 | Jung | Sep 2018 | A1 |
20180307314 | Connor | Oct 2018 | A1 |
20180333061 | Pracyk et al. | Nov 2018 | A1 |
20180368931 | Hongo et al. | Dec 2018 | A1 |
20190069957 | Barral et al. | Mar 2019 | A1 |
20190125460 | Maillet et al. | May 2019 | A1 |
20190206565 | Shelton, IV | Jul 2019 | A1 |
20190307519 | Popovic | Oct 2019 | A1 |
20190357884 | Williams et al. | Nov 2019 | A1 |
20190374299 | Peine | Dec 2019 | A1 |
20200015806 | Scheib et al. | Jan 2020 | A1 |
20200015923 | Scheib et al. | Jan 2020 | A1 |
20200054215 | Roche | Feb 2020 | A1 |
20200060775 | Bonutti | Feb 2020 | A1 |
20200069377 | Finley et al. | Mar 2020 | A1 |
20200113637 | Ida et al. | Apr 2020 | A1 |
20200169673 | King et al. | May 2020 | A1 |
20200179065 | Crawford et al. | Jun 2020 | A1 |
20200205900 | Buckland et al. | Jul 2020 | A1 |
20200261160 | Peine et al. | Aug 2020 | A1 |
20200261297 | Strydom et al. | Aug 2020 | A1 |
20200360099 | Smaby et al. | Nov 2020 | A1 |
20200405403 | Shelton, IV et al. | Dec 2020 | A1 |
20210029307 | King et al. | Jan 2021 | A1 |
20210298590 | Ayvali et al. | Sep 2021 | A1 |
20210402603 | Murphy et al. | Dec 2021 | A1 |
20220096188 | Ellman et al. | Mar 2022 | A1 |
20220160445 | Meglan et al. | May 2022 | A1 |
20220241026 | Shoham et al. | Aug 2022 | A1 |
20220241032 | Zucker et al. | Aug 2022 | A1 |
20220395342 | Weiss et al. | Dec 2022 | A1 |
Number | Date | Country |
---|---|---|
2958013 | Apr 2017 | CA |
108186120 | Jun 2018 | CN |
109171745 | Jan 2019 | CN |
10042599 | Mar 2002 | DE |
102012004506 | Sep 2013 | DE |
102015217059 | Mar 2017 | DE |
2289452 | Mar 2011 | EP |
2757948 | Jul 2014 | EP |
2467074 | Jan 2019 | EP |
3492046 | Jun 2019 | EP |
3492047 | Jun 2019 | EP |
2739231 | Sep 2019 | EP |
3551983 | Oct 2019 | EP |
3060157 | Dec 2019 | EP |
3609422 | Feb 2020 | EP |
2879608 | Mar 2020 | EP |
2156983 | Oct 1985 | GB |
5207490 | Jun 2013 | JP |
2017-512299 | May 2017 | JP |
2019-524230 | Sep 2019 | JP |
10-2019-114052 | Oct 2019 | KR |
132416 | Mar 2018 | RO |
WO 9501757 | Jan 1995 | WO |
WO 9629930 | Oct 1996 | WO |
WO 0035366 | Jun 2000 | WO |
WO 2007025218 | Mar 2007 | WO |
WO 2007038510 | Apr 2007 | WO |
WO 2010068005 | Jun 2010 | WO |
WO 2011149845 | Dec 2011 | WO |
WO 2012168534 | Dec 2012 | WO |
WO 2014106262 | Jul 2014 | WO |
WO 2017002143 | Jan 2017 | WO |
WO 2018203903 | Nov 2018 | WO |
WO 2019036006 | Feb 2019 | WO |
WO 2019206340 | Oct 2019 | WO |
WO 2020072255 | Apr 2020 | WO |
WO 2020079596 | Apr 2020 | WO |
WO 2020118244 | Jun 2020 | WO |
WO 2020201353 | Oct 2020 | WO |
Entry |
---|
Invitation to Pay Additional Fees for International (PCT) Patent Application No. PCT/IL2022/050603, dated Oct. 6, 2022, 13 pages. |
International Search Report and Written Opinion for International (PCT) Patent Application No. PCT/IL2022/050603, dated Nov. 28, 2022, 21 pages. |
International Preliminary Report on Patentability for International (PCT) Patent Application No. PCT/IB2021/058420, dated Apr. 13, 2023 10 pages. |
Official Action with English Translation for Japan Patent Application No. 2021-514567, dated May 11, 2023, 12 pages. |
“Robotic Assisted Systems,” Intuitive Surgical, Jul. 2019, 5 pages. |
Abdelaal et al. “A multi-camera, multi-view system for training and skill assessment for robot-assisted surgery,” International Journal of Computer Assisted Radiology and Surgery, May 2020, vol. 15, pp. 1369-1377. |
Boctor et al. “A Dual-Armed Robotic System for Intraoperative Ultrasound Guided Hepatic Ablative Therapy: A Prospective Study,” IEEE, Proceedings of the 2004 IEEE International Conference on Robotics & Automation, New Orleans, LA, Apr. 2004, pp. 2517-2522. |
Joskowicz “Computer-aided surgery meets predictive, preventive, and personalized medicine,” EPMA Journal, 2017, vol. 8, 4 pages. |
Khandalavala “Emerging surgical robotic technology: a progression toward microbots,” Annals of Laparoscopic and Endoscopic Surgery, Jan. 2020, vol. 5, Article 3, 18 pages. |
Kim et al. “Robot-Assisted Cardiac Surgery Using the Da Vinci Surgical System: A Single Center Experience,” Korean Journal of Thoracic and Cardiovascular Surgery, 2015, vol. 48, pp. 99-104. |
Kumar et al. “Emerging role of robotics in urology, Journal of Minimal Access Surgery,” Oct. 2005, vol. 1, No. 4, pp. 202-210. |
Li et al. “Design of a Multi-Arm Surgical Robotic System for Dexterous Manipulation,” Journal of Mechanisms and Robotics, Dec. 2016, vol. 8, article 061017, 10 pages. |
Staub “Micro Endoscope based Fine Manipulation in Robotic Surgery,” Technische Universitat Munchen Lehrstuhl Robotic und Echtzeitsysteme, Dissertation, Apr. 2013, 146 pages. |
Invitation to Pay Additional Fees for International (PCT) Patent Application No. PCT/IL2022/050126, dated May 27, 2022, 11 pages. |
International Search Report and Written Opinion for International (PCT) Patent Application No. PCT/IL2022/050126, dated Jul. 18, 2022, 18 pages. |
Extended Search Report for European Patent Application No. 19872498.1, dated Jun. 9, 2022, 9 pages. |
International Search Report and Written Opinion prepared by the Israel Patent Office on Dec. 31, 2019, for International Application No. PCT/IB2019/058795. |
International Search Report and Written Opinion for International (PCT) Patent Application No. PCT/IB2021/058420, dated Nov. 29, 2021 16 pages. |
“Robot-Assisted Laparoscopic Procedures,” EBESCO Information Services, Updated May 2014, 6 pages [retrieved online from: www.wkhs.com/cancer/cancer-treatment-services/surgery/robotic-surgery/robot-assisted-laparoscopic-procedures]. |
Cao et al. “A Novel Robotic Suturing System for Flexible Endoscopic Surgery,” IEEE, 2019 International Conference on Robotics and Automation (ICRA), May 20-24, 2019, Montreal, Canada, 7 pages. |
He et al. “A Multi-Function Force Sensing Instrument for Variable Admittance Robot Control in Retinal Microsurgery,” 2014 IEEE International Conference on Robotics and Automation (ICRA), 2014, pp. 1411-1418. |
Kong et al. “Da Vinci Tool Torque Mapping over 50,000 Grasps and its Implications on Grip Force Estimation Accuracy,” 2018 International Symposium on Medical Robotics (ISMR), 2018, 6 pages. |
Matthews “How better data, 5G, and surgical robots will improve healthcare,” The Robot Report, Sep. 29, 2019, 12 pages [retrieved online from: www.therobotreport.com/better-data-advancing-healthcare-robotics]. |
Seibold “An Advanced Force Feedback Tool Design for Minimally Invasive Robotic Surgery,” Technische Universität München, May 15, 2012, Doctoral Engineers dissertation, 218 pages. |
Tian et al. “A Robot-Assisted Surgical System Using a Force-Image Control Method for Pedicle Screw Insertion,” PLOS One, Jan. 2014, vol. 9, No. 1, article e86346, 9 pages. |
Valeo “Scarless Surgery: The benefits and drawbacks of robotic thryroidectomy,” ENTtoday, Apr. 1, 2010, 4 pages [retrieved online from: www.enttoday.org/article/scarless-surgery-the-benefits-and-drawbacks-of-robotic-thryroidectomy/]. |
Vivek et al. “Study of Neuroarm and Force Sensing Grippers in Robo-Assisted Neurosurgery,” International Journal of Current Engineering and Technology, Mar. 2016, Special Issue 4, pp. 444-447. |
Official Action with English Translation for Japan Patent Application No. 2021-514567, dated Sep. 26, 2023, 12 pages. |
Official Action for U.S. Appl. No. 17/344,658, dated Aug. 4, 2023 6 pages Restriction Requirement. |
Official Action for U.S. Appl. No. 17/344,658, dated Oct. 25, 2023 9 pages. |
Official Action for U.S. Appl. No. 17/575,404, dated Jun. 21, 2024 13 pages. |
Official Action for U.S. Appl. No. 17/344,658, dated Jul. 15, 2024 10 pages. |
Aranda-Valera et al. “Measuring Spinal Mobility Using an Inertial Measurement Unit System: A Validation Study in Axial Spondyloarthritis,” Diagnostics, 2020, vol. 10, No. 6, Article 426, 13 pages. |
Dai et al. “Vibration-Based Milling Condition Monitoring in Robot-Assisted Spine Surgery,” IEEE/AMSE Transactions on Mechatronics, Dec. 2015, vol. 20, No. 6, pp. 3028-3039. |
Haratian et al. “Toward Flexibility in Sensor Placement for Motion Capture Systems: A Signal Processing Approach,” IEEE Sensors Journal, Mar. 2014, vol. 14, No. 3, pp. 701-709. |
Hennersperger et al. “Towards MRI-Based Autonomous Robotic US Acquisitions: A First Feasibility Study,” IEEE Transactions on Medical Imaging, Feb. 2017, vol. 36, No. 2, pp. 538-548. |
Leroy “Analysis of the spine through a multibody model and IMU technology,” Ecole polytechnique de Louvain, Université catholique de Louvain, 2019, 90 pages [retrieved online Oct. 29, 2020 from: hdl.handle.net/2078.1/thesis:19491]. |
Stolka et al. “Improving Navication Precision of Milling Operations in Surgical Robotics,” Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robotics and Systems, Oct. 9-15, 2008, Beijing, China, pp. 2351-2357. |
Vandini et al. “Unified Tracking and Shape Estimation for Concentric Tube Robots,” IEEE Transactions on Robotics, Aug. 2017, vol. 33, No. 4, pp. 901-915. |
Voinea et al. “Measurement and Geometric Modelling of Human Spine Posture for Medical Rehabilitation Purposes Using a Wearable Monitoring System Based on Inertial Sensors,” Sensors, 2017, vol. 17, No. 1, Article 3, 19 pages. |
International Search Report and Written Opinion for International (PCT) Patent Application No. PCT/IL2022/050141, dated Jun. 8, 2022, 16 pages. |
Official Action for U.S. Appl. No. 17/464,300, dated Aug. 20, 2024 31 pages. |
Official Action for U.S. Appl. No. 17/583,788, dated Aug. 19, 2024 6 pages Restriction Requirement. |
Official Action with Machine Translation for China Patent Application No. 201980067844.9, dated Mar. 1, 2024, 27 pages. |
Official Action for U.S. Appl. No. 17/344,658, dated Mar. 21, 2024 12 pages. |
Li et al. “Robotic Systems for MIR-Guided Stereotactic Neurosurgery,” IEEE Transactions on Biomedical Engineering, Apr. 2015, vol. 62, No. 4, pp. 1077-1088. |
Notice of Allowance for U.S. Appl. No. 17/464,300, dated Nov. 20, 2024 10 pages. |
Official Action for U.S. Appl. No. 17/575,404, dated Oct. 25, 2024 20 pages. |
Official Action for U.S. Appl. No. 17/344,658, dated Dec. 16, 2024 15 pages. |
Official Action with English Translation for Japan Patent Application No. 2024-009920, dated Jan. 15, 2025, 6 pages. |
Number | Date | Country | |
---|---|---|---|
20210338348 A1 | Nov 2021 | US |
Number | Date | Country | |
---|---|---|---|
62745497 | Oct 2018 | US |