The technology disclosed in the present description (hereinafter, “the present disclosure”) relates to a surgical system and a surgical support method for supporting a surgical operation by applying a robotics technology.
In general, surgical operation is a difficult task performed with sensory movement of an operator. In particular, in the case of surgery using a fine surgical tool in a small and fragile environment, such as ophthalmic surgery, it is necessary for the operator to perform a micron-order operation while suppressing tremor of the hands. Therefore, a surgical system that achieves restriction of tremor of the hands of an operator, absorption of the difference in skill between operators by operation support, and the like using a robotics technology is becoming widespread.
For example, a master slave system has been developed in which an operator operates a surgical tool while observing an operative field of a fundus portion through a microscope (see Non-Patent Document 1). According to this master slave system, the operator can perform the precise operation in the ophthalmic surgery by operating the slave robot supporting the surgical tool depending on an operation amount of the master robot held with the right hand (or the dominant hand).
Furthermore, there has been proposed a surgical system that controls scanning of a surgical laser beam on the basis of information of an optical coherence tomography (OCT) image to perform corneal incision, anterior capsule incision, and disruption of a crystalline lens in cataract surgery (see Patent Document 1).
In the surgical system as described above, an operator performs surgery while viewing an image of a target tissue captured by an observation device such as a microscope or an OCT. At that time, the operator operates the master robot (or the surgical tool) while imagining, in the brain, the spatial positional relationship between the captured image and the surgical tool. The operator needs to perform sufficient training to be proficient in hand-eye coordination between the surgical tool and the captured image.
An object of the present disclosure is to provide a surgical system and a surgical support method for supporting a surgical operation using an observation device such as a microscope or an OCT, and a surgical robot.
The present disclosure has been made in view of the above problems, and a first aspect thereof is a surgical system including:
However, a “system” described here refers to a logical assembly of a plurality of apparatuses (or functional modules that implement specific functions), and it does not matter whether or not each of the apparatuses or functional modules is in a single housing.
For example, the observation device includes a microscope and a pre-lens, and the fixing portion is configured to fix a relationship of relative position and posture of the surgical robot with respect to the pre-lens.
Furthermore, for example, the fixing portion is configured to fix a relationship of relative position and posture of the surgical robot with respect to a retainer that has a marker and retains a state of the surgical site. The retainer is, for example, an eyelid speculum.
Furthermore, in a case where the surgical system includes a first surgical robot and a second surgical robot that support surgical tools, respectively, and the observation device includes a microscope and a pre-lens, the fixing portion may be configured to fix a relationship of relative position and posture of each of the first surgical robot and the second surgical robot with respect to the pre-lens.
Furthermore, a second aspect of the present disclosure is a surgical support method using a surgical system including an observation device that observes an operative field and a surgical robot that supports a surgical tool, a relationship of relative position and posture between the observation device and the surgical robot being fixed, the surgical support method including:
According to the present disclosure, it is possible to provide a surgical system and a surgical support method for supporting execution of an accurate surgery by making an observation device and a surgical robot cooperate with each other.
Note that the effects described in the present description are merely examples, and the effects brought by the present disclosure are not limited thereto. Furthermore, the present disclosure may further provide additional effects in addition to the effects described above.
Still another object, feature, and advantage of the present disclosure will become clear by further detailed description with reference to an embodiment as described later and the attached drawings.
Hereinafter, the present disclosure will be described in the following order with reference to the drawings.
In the present description, an embodiment in which the present disclosure is mainly applied to a surgical system of a master-slave system will be primarily described. In such a surgical system, a user such as an operator performs an operation on the master side, and performs surgery on the slave side by controlling driving of a robot according to the user's operation. Examples of the purpose of incorporating the robotics technology into the surgical system include restriction of tremor of the hands of an operator, operation support, absorption of a difference in skill between operators, and execution of surgery from a remote site.
The master device 110 includes a master-side control unit 111, an operation user interface (UI) unit 112, a presentation unit 113, and a master-side communication unit 114. The master device 110 operates under the overall control of the master-side control unit 111.
The operation UI unit 112 includes a device to which a user (operator or the like) inputs an instruction for a slave robot 112 (described later) that operates a surgical tool such as forceps in the slave device 120. The operation UI unit 112 includes, for example, a dedicated input device such as a controller or a joystick, and a general-purpose input device such as a GUI screen to which a mouse operation or a touch operation with a fingertip is input. Furthermore, a “medical device” configured by supporting a gripping interface by a parallel link as disclosed in Patent Document 2 can be used as the operation UI unit 112.
The presentation unit 113 presents information regarding the surgery performed on the slave device 120, to a user (operator) operating the operation UI unit 112, on the basis of sensor information mainly acquired by a sensor unit 123 (described later) on the slave device 120 side.
For example, in a case where the sensor unit 123 is equipped with an observation device such as an RGB camera or an OCT for capturing a microscopic image for observing the surface of an affected site, or is equipped with an interface for capturing an image captured by these observation devices, and the image data is transferred to the master device 110 with low latency via the transmission path 130, the presentation unit 113 displays on the screen a real-time microscopic image or OCT image of the affected site using a monitor display or the like.
Furthermore, in a case where the sensor unit 123 is equipped with a function of measuring an external force and a moment acting on the surgical tool operated by the slave robot 112, and such force sense information is transferred to the master device 110 with low latency via the transmission path 130, the presentation unit 113 presents the force sense to the user (operator). For example, the presentation unit 113 may present the force sense to the user (operator) using the operation UI unit 112.
The master-side communication unit 114 performs a signal transmission/reception process with the slave device 120 via the transmission path 130 under the control of the master-side control unit 111. For example, in a case where the transmission path 130 includes an optical fiber, the master-side communication unit 114 includes an electro-optical conversion unit that converts an electric signal transmitted from the master device 110 into an optical signal, and a photoelectric conversion unit that converts an optical signal received from the transmission path 130 into an electric signal.
The master-side communication unit 114 transfers an operation command for the slave robot 122 input by the user (operator) via the operation UI unit 112 to the slave device 120 via the transmission path 130. Furthermore, the master-side communication unit 114 receives the sensor information transmitted from the slave device 120 via the transmission path 130.
On the other hand, the slave device 120 includes a slave-side control unit 121, a slave robot 122, a sensor unit 123, and a slave-side communication unit 124. The slave device 120 performs an operation depending on an instruction from the master device 110 under the overall control of the slave-side control unit 121.
The slave robot 122 is, for example, an arm type robot having a multi-link structure, and a surgical tool such as forceps is mounted as an end effector on a tip end (or the distal end). The slave-side control unit 121 interprets an operation command transmitted from the master device 110 via the transmission path 130, converts the operation command into a drive signal of an actuator that drives the slave robot 122, and outputs the drive signal. Then, the slave robot 122 operates on the basis of the drive signal from the slave-side control unit 121.
The sensor unit 123 includes a plurality of sensors for detecting a situation in an affected site of the surgery performed by the slave robot 122 or the slave robot 122, and further includes an interface for acquiring sensor information from various sensor devices installed in an operating room.
For example, the sensor unit 123 includes a force torque sensor (FTS) for measuring an external force and a moment applied during surgery on a surgical tool mounted on the tip end (distal end) of the slave robot 122.
Furthermore, the sensor unit 123 is provided with an interface through which, during the surgery, the slave robot 122 captures a microscopic image of the surface of an affected site or an OCT image obtained by scanning a cross section of the affected site (eyeball).
The slave-side communication unit 124 performs a signal transmission/reception process with the master device 110 via the transmission path 130 under the control of the slave-side control unit 121. For example, in a case where the transmission path 130 includes an optical fiber, the slave-side communication unit 124 includes an electro-optical conversion unit that converts an electric signal transmitted from the slave device 120 into an optical signal, and a photoelectric conversion unit that converts an optical signal received from the transmission path 130 into an electric signal.
The slave-side communication unit 124 transfers the force sense data of a surgical tool acquired by the sensor unit 123, a microscopic image of the affected site, the OCT image obtained by scanning the cross section of an affected site, and the like to the master device 110 via the transmission path 130. Furthermore, the slave-side communication unit 124 receives an operation command for the slave robot 122 transmitted from the master device 110 via the transmission path 130.
As illustrated in
As illustrated in
In the example illustrated in
An eyelid speculum (not illustrated in
Note that, it is sufficient that a movable range of the surgical tool 411 required for the fundus operation be small, so that it is assumed that the robot arm is a microrobot having a total length or a total height of about several centimeters and a mass of about several grams to several tens grams.
Furthermore, for the information of the eyeball 420, an observation device (a stereo video microscope in the example illustrated in
The operator operates the surgical tool 411 while observing an operative field such as the surface and the fundus of the eyeball via the captured image of the observation device 430. In a case where the surgical system 100 is used, the slave robot 122 supporting the surgical tool 411 operates depending on an operation amount of the operation UI unit 112 operated by the operator with the right hand (or the dominant hand), to perform the fundus surgery.
At that time, the operator operates the operation UI unit 112 while imagining, in the brain, the spatial positional relationship between the captured image of the observation device 430 and the surgical tool 411. The operator needs to perform sufficient training to be proficient in hand-eye coordination between the surgical tool 411 and the captured image. In general, the relationship of relative position and posture between the captured image of the observation device 430 and the robot arm is unknown. In such a case, even if the operator tries to instruct the slave robot 122 to operate the surgical tool 411 through an operation of the operation UI unit 112 on the basis of the captured image of the observation device 430, an operation amount of the operation UI unit 112 based on the captured image and the motion of the surgical tool 411 intended by the operator deviate from each other, making it difficult to perform a precise surgery. Furthermore, in a case where many components are mechanically connected between the slave robot 122 and the observation device 430, the accuracy of the position and posture of the surgical tool 411 is also affected by the deflection of the robot arm and the processing accuracy of the components.
Therefore, in the present disclosure, the relationship of relative position and posture between the observation device that observes an operative field and the robot arm that supports a surgical tool is fixed. As a result, for example, an operation amount of the operation UI unit 112 on a captured image of the observation device on the master side can be coordinate-transformed into a movement of the distal end (alternatively, the surgical tool 411 mounted on the distal end) of the robot arm. Therefore, according to the present disclosure, even if the operator is not proficient in hand-eye coordination, the operator can perform accurate surgery by making the observation device and the surgical system of the master-slave system cooperate with each other.
The observation device 500 is, for example, a stereo video microscope equipped with an OCT, and corresponds to the sensor unit 123 in the surgical system 100 illustrated in
Furthermore, the surgical robot 510 corresponds to the slave robot 122 in the surgical system 100 illustrated in
Then, in the example illustrated in
In the first place, the positional relationship between the observation device 500 and the pre-lens 501 is known. Then, since the surgical robot 510 is attached onto the pre-lens 501, the positional relationship between the captured image of the observation device 500 and the surgical robot 510 is known. Assuming that the coordinate system of the captured image of the observation device 500 is (xv, yv, zv) and the coordinate system of the surgical robot 510 is (xr, yr, zr), the coordinate system (xv, yv, zv) of the captured image of the observation device 500 can be converted into the coordinate system (xr, yr, zr) of the surgical robot 510 using a conversion matrix A1 as illustrated in the following equation (1). Since the positional relationship between the observation device 500 and the surgical robot 510 is known, the conversion matrix A1 can be obtained.
[Equation 1]
(xr,yr,zr,1)=(xv,yv,zv,1)A1 (1)
On the other hand, since the configuration information of the surgical robot 510 (the configuration information of each link and joint of the robot arm) and the configuration information of the surgical tool 511 attached to the distal end of the surgical robot 510 are known, the positional relationship between the surgical robot 510 and the surgical tool 511 is known. Assuming that the coordinate system of the tip end of the surgical tool 511 is (xe, ye, ze), the coordinate system (xr, yr, zr) of the surgical robot 510 can be converted into the coordinate system (xe, ye, ze) of the tip end of the surgical tool 511 using a conversion matrix A2 as illustrated in the following equation (2). The conversion matrix A2 can be obtained on the basis of the configuration information of the surgical robot 510 and the configuration information of the surgical tool 511.
[Equation 2]
(xe,ye,ze,1)=(xr,yr,zr,1)A2 (2)
Therefore, the coordinate relationship between the captured image of the observation device 500 and the tip end of the surgical tool 511 is determined as illustrated in the following equation (3).
[Equation 3]
(xe,ye,ze,1)=(xv,yv,zv,1)A1A2 (3)
A case where the arrangement of the observation device 500 and the surgical robot 510 as illustrated in
The hand-eye coordination in the surgical system 100 in this case is that the operator visually views the captured image of the observation device 500, accurately grasps the position information of the tip end of the surgical tool 511 with respect to the captured image, predicts a trajectory of the tip end of the surgical tool 511, and performs operation using the operation UI unit 112. As described above, according to the first example of the present disclosure, the operator views the captured image of the observation device 500 and performs the input to the operation UI unit 112, and the operation of the surgical tool 511 by the surgical robot 510 can be performed smoothly. That is, according to the first example of the present disclosure, even if the operator is not fully trained and proficient in hand-eye coordination, the operator can perform accurate surgery by making the observation device and the surgical system of the master-slave system cooperate with each other.
A condition under which the optimum hand-eye coordination is established can be defined as in the following equation (4) (see Non-Patent Document 2). From the above equation (3), it can be seen that, according to the first example of the present disclosure, the conditional equation (4) is satisfied.
[Equation 4]
C
S
R=
S
M
R (4)
The observation device 600 is, for example, a stereo video microscope equipped with an OCT, and corresponds to the sensor unit 123 in the surgical system 100 illustrated in
Furthermore, the surgical robot 610 corresponds to the slave robot 122 in the surgical system 100 illustrated in
Then, in the example illustrated in
Assuming that the coordinate system of the captured image of the observation device 600 is (xv, yv, zv) and the coordinate system of the eyelid speculum 620 is (xs, ys, zs), the coordinate system (xv, yv, zv) of the captured image of the observation device 500 can be converted into the coordinate system (xs, ys, zs) of the eyelid speculum 620 using a conversion matrix B1 as illustrated in the following equation (5). As described above, the observation device 600 simultaneously images the operative field and the markers 621, 622, and 623 attached to the eyelid speculum 620, and thus the conversion matrix B1 can be obtained on the basis of the relationship of relative position and posture between the observation device 600 and the eyelid speculum 620 calculated on the basis of the positional relationship among the markers 621, 622, and 623 on the captured image.
[Equation 5]
(xs,ys,zs,1)=(xv,yv,zv,1)B1 (5)
Then, since the surgical robot 610 is attached onto the eyelid speculum 620, the positional relationship between the eyelid speculum 620 and the surgical robot 610 is known. Assuming that the coordinate system of the surgical robot 610 is (xr, yr, zr), the coordinate system (xv, yv, zv) of the captured image of the eyelid speculum 620 can be converted into the coordinate system (xr, yr, zr) of the surgical robot 610 using a conversion matrix B2 as illustrated in the following equation (6). Since the positional relationship between the eyelid speculum 620 and the surgical robot 610 is known, the conversion matrix B2 can be obtained.
[Equation 6]
(xr,yr,zr,1)=(xs,ys,zs,1)B2 (6)
Furthermore, since the configuration information of the surgical robot 610 (the configuration information of each link and joint of the robot arm) and the configuration information of the surgical tool 611 attached to the distal end of the surgical robot 610 are known, the positional relationship between the surgical robot 610 and the surgical tool 611 is known. Assuming that the coordinate system of the tip end of the surgical tool 611 is (xe, ye, ze), the coordinate system (xr, yr, zr) of the surgical robot 610 can be converted into the coordinate system (xe, ye, ze) of the tip end of the surgical tool 611 using a conversion matrix B3 as illustrated in the following equation (7). The conversion matrix B3 can be obtained on the basis of the configuration information of the surgical robot 510 and the configuration information of the surgical tool 511.
[Equation 7]
(xe,ye,ze,1)=(xr,yr,zr,1)B3 (7)
Therefore, the coordinate relationship between the captured image of the observation device 600 and the tip end of the surgical tool 611 is determined as illustrated in the following equation (8).
[Equation 8]
(xe,ye,ze,1)=(xv,yv,zv,1)B1B2B3 (8)
A case where the arrangement of the observation device 600 and the surgical robot 610 as illustrated in
As described above, also in the second example of the present disclosure, the operator views the captured image of the observation device 600 and performs the input to the operation UI unit 112, and the operation of the surgical tool 611 by the surgical robot 610 can be performed smoothly.
From the above equation (8), it can be seen that, according to the second example of the present disclosure, the conditional equation (4) described above in which the optimum hand-eye coordination is established is satisfied. Therefore, also in the second example of the present disclosure, even if the operator is not fully trained and proficient in hand-eye coordination, the operator can perform accurate surgery by making the observation device and the surgical system of the master-slave system cooperate with each other.
The observation device 800 is, for example, a stereo video microscope equipped with an OCT, and corresponds to the sensor unit 123 in the surgical system 100 illustrated in
The first surgical robot 810 and the second surgical robot 820 correspond to the slave robot 122 in the surgical system 100 illustrated in
Then, in the example illustrated in
In the first place, the positional relationship between the observation device 800 and the pre-lens 801 is known. Then, since the first surgical robot 810 and the second surgical robot 820 are attached onto the pre-lens 801, the positional relationship among the captured image of the observation device 800, the first surgical robot 810, and the second surgical robot 820 is known. Assuming that the coordinate system of the captured image of the observation device 800 is (xv, yv, zv) and the coordinate system of the first surgical robot 80 is (xr1, yr1, zr1), the coordinate system (xv, yv, zv) of the captured image of the observation device 800 can be converted into the coordinate system (xr, yr, zr) of the first surgical robot 810 using a conversion matrix A11 as illustrated in the following equation (9). Similarly, as illustrated in the following equation (10), the coordinate system (xv, yv, zv) of the captured image of the observation device 800 can be converted into the coordinate system (xr, yr, zr) of the second surgical robot 820 using a conversion matrix A21. Since the positional relationship among the observation device 800, the first surgical robot 810, and the second surgical robot 820 is known, the conversion matrices A11 and A21 can be obtained.
[Equation 9]
(xr1,yr1,zr1,1)=(xv,yv,zv,1)A11 (9)
[Equation 10]
(xr2,yr2,zr2,1)=(xv,yv,zv,1)A21 (10)
On the other hand, since the configuration information of the first surgical robot 810 (the configuration information of each link and joint of the robot arm) and the configuration information of the surgical tool 811 attached to the distal end of the first surgical robot 810 are known, the positional relationship between the first surgical robot 810 and the surgical tool 811 is known. Similarly, the positional relationship between the second surgical robot 820 and the surgical tool 821 is known. Assuming that the coordinate system of the tip end of the surgical tool 811 is (xe1, ye1, ze1) and the coordinate system of the tip end of the surgical tool 821 is (xe2, ye2, ze2), as illustrated in the following equations (11) and (12), the coordinate system (xr1, yr1, zr1) of the first surgical robot 810 can be converted into the coordinate system (xe1, ye1, ze1) of the tip end of the surgical tool 811 using a conversion matrix A12, and the coordinate system (xr2, yr2, zr2) of the second surgical robot 820 can be converted into the coordinate system (xe2, ye2, ze2) of the tip end of the surgical tool 811 using a conversion matrix A22. The conversion matrix A12 can be obtained on the basis of the configuration information of the first surgical robot 810 and the configuration information of the surgical tool 811, and the conversion matrix A22 can also be obtained on the basis of the configuration information of the second surgical robot 820 and the configuration information of the surgical tool 821.
[Equation 11]
(xe1,ye1,ze1,1)=(xr1,yr1,zr1,1)A12 (11)
[Equation 12]
(xe2,ye2,ze2,1)=(xr2,yr2,zr2,1)A22 (12)
Therefore, the coordinate relationship between the captured image of the observation device 800 and the tip end of each of the surgical tools 811 and 812 is determined as illustrated in the following equations (13) and (14).
[Equation 13]
(xe1,ye1,ze1,1)=(xv,yv,zv,1)A11A12 (13)
[Equation 14]
(xe2,ye2,ze2,1)=(xv,yv,zv,1)A21A22 (14)
A case where the arrangement of the observation device 800, the first surgical robot 810, and the second surgical robot 820 as illustrated in
As described above, also in the third example of the present disclosure, the operator views the captured image of the observation device 800 and performs the input to the operation UI unit 112, and the operation of the surgical tool 811 by the first surgical robot 810 and the operation of the surgical tool 821 by the second surgical robot 820 can be performed smoothly.
From the above equations (13) and (14), it can be seen that, according to the third example of the present disclosure, the conditional equation (4) described above in which the optimum hand-eye coordination is established is satisfied. Therefore, also in the third example of the present disclosure, even if the operator is not fully trained and proficient in hand-eye coordination, the operator can perform accurate surgery by making the observation device and the surgical system of the master-slave system cooperate with each other.
In this item F, effects brought about by applying the present disclosure to the surgical system 100 will be described.
According to the present disclosure, it is possible to provide a structure that highly accurately obtains a relative positional relationship between an observation device such as a microscope that observes an operative field and a surgical tool supported by a surgical robot. Therefore, it is possible to achieve precise manipulation, hand-eye coordination, and surgery support by cooperative operations of the observation device, the surgical robot, and further a plurality of surgical robots.
The present disclosure is heretofore described in detail with reference to the specific embodiment. However, it is obvious that those skilled in the art can make modifications and substitutions of the embodiment without departing from the gist of the present disclosure.
In this description, the embodiment in which the surgical system according to the present disclosure is applied to ophthalmic surgery has been mainly described, but the gist of the present disclosure is not limited thereto. The present disclosure can be similarly applied to various types of surgical systems that support surgery using observation devices and surgical robots.
In short, the present disclosure is heretofore described in a form of an example and the content described in this specification should not be interpreted in a limited manner. In order to determine the gist of the present disclosure, the claims should be taken into consideration.
Note that the present disclosure can have the following configurations.
(2) The surgical system according to (1) described above, in which
Number | Date | Country | Kind |
---|---|---|---|
2021-060413 | Mar 2021 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2021/048962 | 12/28/2021 | WO |