The disclosure relates to surgical systems, and more particularly, to systems and methods of performing endoscopic thoracic imaging and treatment.
Current monocular optical devices (e.g. endoscope, bronchoscope, colonoscope) used for viewing surgical fields during minimally invasive surgery (e.g, laparoscopy) and visual diagnostic procedures (e.g, colonoscopy, bronchoscopy) provide limited reference information on the absolute position of surgical tools and anatomical features because the image has no depth of field. To compensate, surgeons may advance the surgical tool until it comes in contact with a feature or another tool. This leads to inefficient motion and the potential for perforation of critical anatomical structures. Binocular (also known as stereoscopic) optical devices provide limited depth of field affording the surgeon visual information on the distance between items within the optical device's field of view. The accuracy of distance information is limited based on the amount of parallax provided by the optical paths, determined by the distance between the optical paths, and the amount of overlap between the two optical paths.
The disclosure is directed to derivation and display of distance references between objects both within the current optical view and previously viewed objects that may no longer be in view. This information may be displayed in multiple modalities: point to point distance from a tool to a specific anatomical reference or other tool, point to point distance from a tool to the closest anatomical feature, guard band fencing keep-out regions around critical anatomy, distance to all surfaces in the optical view through usage of a gradient display via color, numerical scale or audio feedback, and user placed scales used started and oriented from a fixed anatomical location (e.g. a ruler starting at the splenic flexure running to the hepatic flexure, a hatch marked arc of defined radius from a fixed point denoting range of lymph node harvesting). The combined distance information and optical endoscope view may be projected on two-dimensional monitors, through augmented reality (AR) displays, or mixed into virtual reality (VR) displays.
In an aspect of the disclosure, a method for enhanced surgical navigation is provided. The method includes generating a 3D spatial map of a surgical site using a scanning 3D endoscope including a camera source and a scan source, receiving a selection of an anatomy in the surgical site, receiving a selection of a threshold minimum distance from the received selection of the anatomy, detecting a position of a surgical tool in the generated 3D spatial map, and measuring a distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy. Additionally, the method includes determining whether the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy is greater than the threshold minimum distance from the received selection of the anatomy, and generating a warning when it is determined that the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy is equal to or not greater than the threshold minimum distance from the received selection of the anatomy. In an aspect, the scan source is an IR scan source.
In an aspect, the method further includes displaying a value representing the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy in a first form when it is determined that the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy is greater than the threshold minimum distance from the received selection of the anatomy. Generating the warning may include displaying the value in a second form different from the first form. Generating the warning may include generating at least one of an audible notification, a tactile notification (e.g., a shake, vibration, or other imposition on the user's hand), or a visual notification.
In an aspect, receiving the selection of the anatomy in the surgical site includes receiving the selection from pre-surgical imagery during a planning phase, the pre-surgical imagery including at least one of CT, MRI, fluoroscopic, ultrasound, or any combinations thereof, Additionally, or alternatively, receiving the selection of the anatomy in the surgical site may include receiving the selection from surgical imagery during a surgical phase from images generated from the 3D endoscope.
Receiving the selection of the threshold minimum distance from the received selection of the anatomy may include receiving the selection from pre-surgical imagery during a planning phase, the pre-surgical imagery including at least one of CT, MRI, fluoroscopic, ultrasound, or any combinations thereof. Additionally, or alternatively, receiving the selection of the threshold minimum distance from the received selection of the anatomy includes receiving the selection from surgical imagery during a surgical phase from images generated from the 3D endoscope. Additionally, or alternatively, the selection is a preset value based on user (e.g., surgeon) preference or defined by procedure type.
In an aspect, the 3D spatial map includes a matrix of equidistant data points representing fixed points in a current view of the surgical site and a value of a data point represents an existence of an object at the data point in space. Additionally, or alternatively, measuring the distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy may include at least one of calculating a difference between coordinates of two data points in the matrix or following a contour of a surface between two data points in the matrix and calculating a distance along the contour.
In another aspect of the disclosure, a system for enhanced surgical navigation is provided including a scanning 3D endoscope including a camera source and a scan source and a computing device operably coupled to the 3D endoscope. The 3D endoscope is configured to be used to generate a 3D spatial map of a surgical site. The computing device is configured to display the 3D spatial map of the surgical site on a graphical user interface, receive a selection of an anatomy in the surgical site, receive a selection of a threshold minimum distance from the received selection of the anatomy, detect a position of a surgical tool in the generated 3D spatial map, and measure a distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy. The computing device further determines whether the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy is greater than the threshold minimum distance from the received selection of the anatomy. Additionally, the computing device generates a warning when it is determined that the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy is equal to or not greater than the threshold minimum distance from the received selection of the anatomy. In an aspect, the scan source is an IR scan source.
In an aspect, the computing device displays, on a graphic user interface, a value representing the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy in a first form when it is determined that the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy is greater than the threshold minimum distance from the received selection of the anatomy, and displays, on a graphic user interface, the value in a second form different from the first form when it is determined that the measured distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy is equal to or not greater than the threshold minimum distance from the received selection of the anatomy.
The generated warning may include at least one of an audible notification, a tactile notification (e.g., a shake, vibration, or other imposition on the user's hand), or a visual notification.
In an aspect, the computing device receives the selection of the anatomy in the surgical site by receiving the selection from pre-surgical imagery during a planning phase, the pre-surgical imagery including at least one of CT, MRI, fluoroscopic, ultrasound, or any combinations thereof. Additionally, or alternatively, the computing device may receive the selection of the anatomy in the surgical site by receiving the selection from surgical imagery during a surgical phase from images generated from the 3D endoscope.
In an aspect, the computing device may receive a selection of the threshold minimum distance from the received selection of the anatomy by receiving the selection from pre-surgical imagery during a planning phase, the pre-surgical imagery including at least one of CT, MRI, fluoroscopic, ultrasound, or any combinations thereof. Additionally, or alternatively, the computing device may receive a selection of the threshold minimum distance from the received selection of the anatomy by receiving the selection from surgical imagery during a surgical phase from images generated from the 3D endoscope. Additionally, or alternatively, the selection is a pre-set value based on user (e.g., surgeon) preference or defined by procedure type.
In an aspect, the 3D spatial map includes a matrix of equidistant data points representing fixed points in a current view of the surgical site and a value of a data point represents an existence of an object at the data point in space. The computing device may measure the distance between the detected position of the surgical tool in the generated 3D spatial map and the received selection of the anatomy by calculating a difference between coordinates of two data points in the matrix or following a contour of a surface between two data points in the matrix and calculating a distance along the contour.
In yet another aspect of the disclosure a non-transitory computer-readable storage medium encoded with a program, that when executed by a processor, causes the processor to generate a 3D spatial map of a surgical site, detect a position of a surgical tool in the generated 3D spatial map, determine whether a distance between the detected position of the surgical tool in the generated 3D spatial map and a location of an anatomy in the 3D spatial map is greater than a threshold minimum distance. A warning is generated when it is determined that the distance between the detected position of the surgical tool in the generated 3D spatial map and the location of the anatomy is equal to or not greater than the threshold minimum distance.
The 3D spatial map may include a matrix of equidistant data points representing fixed points in a current view of the surgical site. Additionally, a value of a data point may represent an existence of an object at the data point in space. The distance between the detected position of the surgical tool in the generated 3D spatial map and the location of the anatomy may be measured by at least one of calculating a difference between coordinates of two data points in the matrix or following a contour of a surface between two data points in the matrix and calculating a distance along the contour.
Various aspects and features of the disclosure are described hereinbelow with references to the drawings, wherein:
One aspect of the disclosure is directed to a 3D endoscope and systems that support organ matching to preoperative images, for example images of a lung, other anatomy or anatomical features within a surgical site. The 3D endoscope can provide both visual imaging and also surface mapping and is used to generate a 3D spatial map, either by the 3D endoscope or by another component of the system such as a computing device. The computing device utilizes the 3D spatial map to provide enhanced navigational guidance including distance measurements and warnings or other notifications when a surgical device is placed near or is approaching a portion of the anatomy within the surgical site.
In accordance with the disclosure, as will be described in greater detail below, the 3D endoscope (also referred to herein as the “endoscope”) includes a structured light (or laser) scanner. As can be appreciated, the structured light scanner may employ infrared light so as to avoid interference from visible light sources, although it is contemplated that the structured light scanner may emit light in the visible spectrum, or any other wavelength, depending upon the tissue being scanned during the procedure. The structured light source includes a known position relative to a camera and permits the calculation of the exact location of the intersection between the light ray from the structured light source and the camera. This information can be scanned as single points, lines, or arrays to create topologic maps of surfaces. In embodiments, the structured light source is that of an LED or LED infrared laser that is dispersed into a scan pattern (line, mesh, or dots), buy rotating mirror, beam splitter, or diffraction grating. In one non-limiting embodiment, the structured light source may be a LED laser having collimated light. The laser scanner will enable visualization systems to achieve accurate surface maps of the lung needed in order to match preoperative computed images to the operative image delivered to the endoscopic camera. Having both in one endoscope offers additional advantage of matching the preoperative computed image to the current camera view as the camera offset is known relative to the surface scan.
In particular applications, the endoscope position will also be tracked by intraoperative instrument tracking systems for example electromagnetic navigation systems. The locational information obtained by the intraoperative instrument tracking system aids in simplifying the algorithms needed to produce large-scale spatial surface maps from segmental sized scans taken from an endoscope. Further, this immediate intraoperative guidance of the optical image location to the surface map and preoperative computed images provides even greater clarity of location and orientation of the endoscope.
In certain embodiments, the 3D endoscope is positionable by a robotic system. The robotic system provides precise six-axis orientation of the endoscope in a similar manner to the navigation systems but benefited by active positioning as well as locational knowledge of the endoscope within the patient. As can be appreciated, the robot may be utilized to autonomously move the endoscope to complete scans of larger areas or whole organs.
In one embodiment, the endoscope includes a visual-light optical camera, a light source of preferably at least one light-emitting diode (LED), a scanning laser, and a second camera used to map the laser. In some embodiments, the scanning laser (and/or visual light) may be detected by the same optical camera in near to mid infrared imaging as optical sensor technology continues to advance. In its simplest form, as detailed below, the endoscope uses a typical arrangement of these components on the distal end of the endoscope. In order to reduce the required distal end diameter of the instrument and to improve triangulation between the laser and the second camera, these four components may have a location on at least one extensible surface. This enables the four components to be arranged along the side of the extensible surface such that the needed space for the individual components is provided by having a cross section equal or slightly larger than any single component and sufficient length to align the components side by side.
The computation of the topology viewed by the endoscope may require a calibration source to detail the alignment of the laser with the second camera. Anticipated is that the calibration may be conducted at the time of manufacture and stored within a memory coupled to a suitable computer, as will be described in detail hereinbelow, or by targeting a calibration surface at the time of use. The calibration will be used internally with the device anticipating the computational topology that may be created with the endoscope and transmitted for clinician via common video transmission means or the raw camera data along with the calibration may be transmitted to an external graphics processor creating the computational topology.
In some embodiments, at least the laser and the second camera may be spaced along the length of the instrument shaft to enable triangulation where the laser and second camera are directed at an angle from the centerline of the instrument.
One advantage of the disclosure is to enable 3D surfacing of organs and other anatomical features and objects in a surgical site, which can be matched to preoperative computational imaging needed for operative guidance to target lesions with particular special knowledge of adjacent structures and anatomic boundaries such as in sublobar resection or lung cancer. Primary use for this system is thoracic but one can vision equal value in deep pelvic, rectal surgery, or other surgical applications. These and further aspects of the disclosure are detailed herein below.
The 3D endoscope 200 includes an elongate body 202 configured to be advanced within a suitable thoracic trocar (not shown) or other device capable of penetrating the chest cavity and receiving an endoscope therein or a thoracic catheter or the like. In embodiments, the elongate body 202 may include segments capable of being manipulated relative to one another. In this manner, the 3D endoscope 200 may be positioned in close proximity to the chest wall to navigate the shallow portions of the surgical site “S” (e.g., the thoracic cavity) between the lungs or other anatomy “A” (
It is contemplated that the second camera 212 may be any thermographic camera known in the art, such as such as ferroelectric, silicon microbolometer, or uncooled focal plane array or may be any other suitable visible light camera such as a charge-coupled device (CCD), complementary metal-oxide-semiconductor (CMOS), N-type metal-oxide-semiconductor (NMOS), or other suitable camera known in the art where the light emitted from the laser 210 is in the visible or detectable spectrum. In embodiments, the distal surface 204 may include a suitable transparent protective cover (not shown) capable of inhibiting fluids or other contaminants from coming into contact with each of the optical camera 206, light source 208, laser 210, and second camera 212. Since the distance between the laser 210 and second camera 212 relative to the optical camera 206 is fixed (i.e., the offset of the optical camera 206 relative to the laser 210 and second camera 212), the images obtained by the optical camera 206 can more accurately be matched with a pre-operative image, as will be described in further detail hereinbelow.
In operation, initially, the patient “P” (
After the patient “P” is imaged, the clinician penetrates the chest of a patient “P” using a trocar (not shown) or other suitable device. The distal portion of the 3D endoscope 200 is advanced within the trocar, and thereafter, within the surgical site “S” (e.g., the thoracic cavity) of the patient “P” (
Once facing the surface of the anatomy “A”, for example the lung “L” (e.g., incident the lung surface), the laser 210 emits ER light, which is reflected off the surface of the anatomy “A” and detected by the second camera 212. The 3D endoscope 200 is advanced over the surface of the anatomy “A” in a caudal, cephalad, or lateral direction, or combinations thereof. The data obtained by the second camera 212 is processed by the computing device 150 to generate a 3D spatial map of the surface of the surgical site “S” including the anatomy “A” and any objects present therein, such as surgical tools, using any suitable means, such as stitching or the like. In an aspect, the clinician advances the 3D endoscope 200 over the entire surface of the anatomy “A” in order to obtain as complete a map as possible.
The light source 208 and the optical camera 206 are simultaneously operated with the laser 210 and second camera 212 to permit correlation of the images received from the optical camera 206 with the previously acquired MRI (or other modality identified above) images. The correlation between the images obtained by the optical camera 206 and the previously acquired MRI images permits the clinician, and the computing device 150, to more accurately map the surface of the anatomy “A” and the surgical site “S”, As can be appreciated, the accuracy of the correlation may be further improved using tracking software to track the distal tip of the 3D endoscope 200.
Referring now to
In step 403, a selection of an anatomy “A” is made. This selection in step 403 may be carried out in real time using the images or data captured by the 3D endoscope 200 or during a planning phase using images or data captured by some other imaging device such as a fluoroscope, CT, MRI, or PET scanner. For example, referring briefly to
In step 405, a selection of a threshold minimum distance from the received selection of the anatomy “A” (selected in step 403) is made. The selection in step 405 may be carried out in real time using the images or data captured by the 3D endoscope 200 or during a planning phase using images, data captured by some other imaging device such as a fluoroscope, CT, MRI, or PET scanner, or preset preferences. For example, referring briefly to
Once the anatomy “A” is selected (in step 403) and the threshold minimum distance is selected (step 405), method 400 proceeds to step 407 where a real time position of a surgical tool “ST” within the surgical site “S” in the 3D spatial map is determined.
In step 409, a distance between the detected position of the surgical tool “ST” in the generated 3D spatial map and the received selection of the anatomy “A” is measured by computing device 150. In an aspect, the 3D spatial map includes a matrix of equidistant data points representing fixed points in a current view of the surgical site “S” and a value of a data point represents an existence of an object at the data point in space. For example, one object can be any portion of the surgical tool “ST” and another object can be the anatomy “A” selected. The distance between the detected position of the surgical tool “ST” in the 3D spatial map and the received selection of the anatomy “A” in the 3D spatial map may be achieved by calculating a difference between coordinates of two data points (e.g., the point representing the surgical tool “ST” and the point representing the anatomy “A”) in the matrix. In an aspect, the surgical tool “ST” and the anatomy “A” can be represented by a line of points where the distance can be the smallest distance between any point of the surgical tool “ST” and any point of the anatomy “A.”
Additionally, or alternatively, the distance between the detected position of the surgical tool “ST” in the 3D spatial map and the received selection of the anatomy “A” in the 3D spatial map may be achieved by following a contour of a surface between two data points the point representing the surgical tool “ST” and the point representing the anatomy “A”) in the matrix and calculating a distance along the contour. In an aspect, two contours are considered where two points on the anatomy “A” and two points on the surgical tool “ST” are factors, and where the method 400 further includes determining the distance between the two contours and/or determining the distance between each contour and the surgical tool “ST.”
Following the measurement in step 409, in step 411, a determination is made as to whether or not the measured distance (from step 409) is greater than the threshold selected in step 405. If the measured distance is greater than the selected threshold (yes in step 411), then method 400 proceeds to step 413, where the value of the measured distance is displayed in a first (normal) form 605a (
On the other hand, if the measured distance is equal to or not greater than the selected threshold (no in step 411), then method 400 proceeds to step 412 where the user is warned that the surgical tool “ST” is currently too close to the anatomy “A”. This warning may be in the form of displaying the value of the measured distance in a second form 605b (
In step 711, if the measured distance is greater than the selected threshold (yes in step 711), then method 700 proceeds to step 713, where at least one guard band 805a (
On the other hand, if the measured distance is equal to or not greater than the selected threshold (no in step 711), then method 700 proceeds to step 712 where the user is warned that the surgical tool “ST” is currently too close to the anatomy “A”. This warning may be in the form of displaying at least one of the guard bands 805a (
In step 911, if the measured distance is greater than the selected threshold (yes in step 911), then method 900 proceeds to step 913, where at least one guard band 1005a (
On the other hand, if the measured distance is equal to or not greater than the selected threshold (no in step 911), then method 900 proceeds to step 912 where the user is warned that the surgical tool “ST” is currently too close to the anatomy “A”. This warning may be in the form of displaying at least one of the guard bands 1005a (
In addition to the above-described methods performable by system 100, system 100 may detect coordinate mismatches and notify the user or modify the display of the graphical user interface based on the detected mismatch. The anatomy being scanned is not static and will change over time due to elements such as manipulation by the surgeon, natural biological rhythms (e.g. cardiac, pulmonary), and gravity. Detection of such changes by system 100 can include 3D coordinate mismatches between current scanned locations of objects in the current field of view and those from a previous view (e.g., a surface that-extends out of the current view where the Y coordinates of the surface in view differs from that outside). Previously scanned structures completely outside of the current view may change as well. In an aspect, system 100 indicates to the user that all items outside the current field of view may have changed. To this end, system 100 may modify the displayed image of all elements outside the current field of view via blurring, removing 3D effects (e.g., flattening the image), and removal of color or fading of color. Additionally, or alternatively, the items within the field of view may continue to be updated in real-time by system 100.
Surgical instruments such as the endoscopes, computing devices, and other components of system 100 described herein may also be configured to work with robotic surgical systems and what is commonly referred to as “Telesurgery.” Such systems employ various robotic elements to assist the surgeon and allow remote operation (or partial remote operation) of surgical instrumentation. Various robotic arms, gears, cams, pulleys, electric and mechanical motors, etc. may be employed for this purpose and may be designed with a robotic surgical system to assist the surgeon during the course of an operation or treatment. Such robotic systems may include remotely steerable systems, automatically flexible surgical systems, remotely flexible surgical systems, remotely articulating surgical systems, wireless surgical systems, modular or selectively configurable remotely operated surgical systems, etc.
The robotic surgical systems may be employed with one or more consoles that are next to the operating theater or located in a remote location. In this instance, one team of surgeons or nurses may prep the patient for surgery and configure the robotic surgical system with one or more of the instruments disclosed herein while another surgeon (or group of surgeons) remotely control the instruments via the robotic surgical system. As can be appreciated, a highly skilled surgeon may perform multiple operations in multiple locations without leaving his/her remote console which can be both economically advantageous and a benefit to the patient or a series of patients.
The robotic arms of the surgical system are typically coupled to a pair of master handles by a controller. The handles can be moved by the surgeon to produce a corresponding movement of the working ends of any type of surgical instrument (e.g., end effectors, graspers, knifes, scissors, endoscopes, etc.) which may complement the use of one or more of the embodiments described herein. The movement of the master handles may be scaled so that the working ends have a corresponding movement that is different, smaller or larger, than the movement performed by the operating hands of the surgeon. The scale factor or gearing ratio may be adjustable so that the operator can control the resolution of the working ends of the surgical instrument(s).
It is contemplated that the endoscopes described herein may be positioned by the robotic system and the precise position of the endoscope transmitted to the computer to construct the 3D image of the scanned organ or operative field. The robotic system has the ability to autonomously scan the surgical field and construct a complete 3D model of the field to aid the surgeon in directing the robotic arms or to provide necessary 3D information for the robotic system to further conduct surgical steps autonomously. In embodiments, where the endoscope includes a camera and a structured light source that are independent of one another, the robotic system may direct the camera and a structured light source separately. The robotic system provides the relative coordinates between respective endoscopes needed to triangulate the points in the structured light and camera views to construct a 3D surface of the operative field. In this manner, the robotic system has a specific advantage of being able to autonomously position the structure light source onto the field of view of the camera or camera endoscope, Additionally, or alternatively, with the robot controlling the camera location (or other component location), the robot may move the camera (or other component) to expand the size of the scanned anatomy (e.g, the amount scanned) to create a larger view for the user (e.g., surgeon) without input or knowledge by the user.
The master handles may include various sensors to provide feedback to the surgeon relating to various tissue parameters or conditions, e.g., tissue resistance due to manipulation, cutting or otherwise treating, pressure by the instrument onto the tissue, tissue temperature, tissue impedance, etc. As can be appreciated, such sensors provide the surgeon with enhanced tactile feedback simulating actual operating conditions. The master handles may also include a variety of different actuators for delicate tissue manipulation or treatment further enhancing the surgeon's ability to mimic actual operating conditions.
Referring to
Each of the robot arms 1102, 1103 may include a plurality of members, which are connected through joints, and an attaching device 1109, 1111, to which may be attached, for example, a surgical tool “ST” supporting an end effector 1120, in accordance with any one of several embodiments disclosed herein, as will be described in greater detail below.
Robot arms 1102, 1103 may be driven by electric drives (not shown) that are connected to control device 1104. Control device 1104 (e.g., a computer) may be set up to activate the drives, in particular by means of a computer program, in such a way that robot arms 1102, 1103, their attaching devices 1109, 1111 and thus the surgical tool (including end effector 1120) execute a desired movement according to a movement defined by means of manual input devices 1107, 1108. Control device 1104 may also be set up in such a way that it regulates the movement of robot arms 1102, 1103 and/or of the drives.
Medical workstation 1100 may be configured for use on a patient “F” lying on a patient table 1112 to be treated in a minimally invasive manner by means of end effector 1120. Medical workstation 1100 may also include more than two robot arms 1102, 1103, the additional robot arms likewise being connected to control device 1104 and being telemanipulatable by means of operating console 1105. A medical instrument or surgical tool (including an end effector 1120) may also be attached to the additional robot arm. Medical workstation 1100 may include a database 1114, in particular coupled to with control device 1104, in which are stored, for example, pre-operative data from patient/living being “P” and/or anatomical atlases.
While several embodiments of the disclosure have been shown in the drawings, it is not intended that the disclosure be limited thereto, as it is intended that the disclosure be as broad in scope as the art will allow and that the specification be read likewise. Therefore, the above description should not be construed as limiting, but merely as exemplifications of particular embodiments.
As used hereinabove, the term “clinician” refers to a doctor, a nurse, or any other care provider and may include support personnel. Throughout this description, the term “proximal” refers to the portion of the device or component thereof that is closer to the clinician and the term “distal” refers to the portion of the device or component thereof that is farther from the clinician. Additionally, in the drawings and in the description above, terms such as front, rear, upper, lower, top, bottom, and similar directional terms are used simply for convenience of description and are not intended to limit the disclosure. In the description hereinabove, well-known functions or constructions are not described in detail to avoid obscuring the disclosure in unnecessary detail.
This application is a continuation of U.S. patent application Ser. No. 16/682,351, filed on Nov. 13, 2019, which claims the benefit of U.S. Provisional Application Ser. No. 62/779,229, filed on Dec. 13, 2018, and is related to, and claims the benefit of, U.S. Provisional Application Ser. No. 62/779,242, filed on Dec. 13, 2018 and U.S. Provisional Application No. 62/782,683, filed on Dec. 20, 2018, the entire contents of each of which being incorporated by reference herein.
| Number | Name | Date | Kind |
|---|---|---|---|
| 3928925 | Gay | Dec 1975 | A |
| 5057494 | Sheffield | Oct 1991 | A |
| 5321113 | Cooper et al. | Jun 1994 | A |
| 5687737 | Branham et al. | Nov 1997 | A |
| 6003517 | Sheffield et al. | Dec 1999 | A |
| 7474407 | Gutin | Jan 2009 | B2 |
| 7538859 | Tearney et al. | May 2009 | B2 |
| 7559895 | Stetten et al. | Jul 2009 | B2 |
| 7756305 | Price | Jul 2010 | B2 |
| 7857756 | Warren et al. | Dec 2010 | B2 |
| 7949385 | Khamene et al. | May 2011 | B2 |
| 7952718 | Li et al. | May 2011 | B2 |
| 8335359 | Fidrich et al. | Dec 2012 | B2 |
| 8335557 | Maschke | Dec 2012 | B2 |
| 8382662 | Soper et al. | Feb 2013 | B2 |
| 8384909 | Yun et al. | Feb 2013 | B2 |
| 8460195 | Courtney et al. | Jun 2013 | B2 |
| 8494794 | Dutta et al. | Jul 2013 | B2 |
| 8706184 | Mohr et al. | Apr 2014 | B2 |
| 8784321 | Courtney et al. | Jul 2014 | B2 |
| 8827934 | Chopra et al. | Sep 2014 | B2 |
| 8983580 | Boppart et al. | Mar 2015 | B2 |
| 9375268 | Long | Jun 2016 | B2 |
| 9398936 | Razzaque et al. | Jul 2016 | B2 |
| 9554774 | Moore et al. | Jan 2017 | B2 |
| 9861338 | Kanade et al. | Jan 2018 | B2 |
| 9918659 | Chopra et al. | Mar 2018 | B2 |
| 10004558 | Long et al. | Jun 2018 | B2 |
| 10194897 | Cedro et al. | Feb 2019 | B2 |
| 10350009 | Panescu et al. | Jul 2019 | B2 |
| 10368054 | Panescu et al. | Jul 2019 | B2 |
| 10373719 | Soper et al. | Aug 2019 | B2 |
| 10376178 | Chopra | Aug 2019 | B2 |
| 10391277 | Rahimian et al. | Aug 2019 | B2 |
| 10405753 | Sorger | Sep 2019 | B2 |
| 10478162 | Barbagli et al. | Nov 2019 | B2 |
| 10480926 | Froggatt et al. | Nov 2019 | B2 |
| 10524866 | Srinivasan et al. | Jan 2020 | B2 |
| 10555788 | Panescu et al. | Feb 2020 | B2 |
| 10569071 | Harris et al. | Feb 2020 | B2 |
| 10603106 | Weide et al. | Mar 2020 | B2 |
| 10610306 | Chopra | Apr 2020 | B2 |
| 10638953 | Duindam et al. | May 2020 | B2 |
| 10639114 | Schuh et al. | May 2020 | B2 |
| 10674970 | Averbuch et al. | Jun 2020 | B2 |
| 10682070 | Duindam | Jun 2020 | B2 |
| 10702137 | Deyanov | Jul 2020 | B2 |
| 10706543 | Donhowe et al. | Jul 2020 | B2 |
| 10709506 | Coste-Maniere et al. | Jul 2020 | B2 |
| 10772485 | Schlesinger et al. | Sep 2020 | B2 |
| 10796432 | Mintz et al. | Oct 2020 | B2 |
| 10823627 | Sanborn et al. | Nov 2020 | B2 |
| 10827913 | Ummalaneni et al. | Nov 2020 | B2 |
| 10835153 | Rafii-Tari et al. | Nov 2020 | B2 |
| 10885630 | Li et al. | Jan 2021 | B2 |
| 11172184 | Komp et al. | Nov 2021 | B2 |
| 20020147462 | Mair et al. | Oct 2002 | A1 |
| 20030013972 | Makin | Jan 2003 | A1 |
| 20030164952 | Deichmann et al. | Sep 2003 | A1 |
| 20040120981 | Nathan | Jun 2004 | A1 |
| 20040210105 | Hale et al. | Oct 2004 | A1 |
| 20060084860 | Geiger et al. | Apr 2006 | A1 |
| 20080045938 | Weide et al. | Feb 2008 | A1 |
| 20100217117 | Glossop et al. | Aug 2010 | A1 |
| 20110282188 | Burnside et al. | Nov 2011 | A1 |
| 20120209069 | Popovic et al. | Aug 2012 | A1 |
| 20130018255 | Kitamura et al. | Jan 2013 | A1 |
| 20130303945 | Blumenkranz et al. | Nov 2013 | A1 |
| 20140035798 | Kawada et al. | Feb 2014 | A1 |
| 20140163359 | Sholev et al. | Jun 2014 | A1 |
| 20140243596 | Yoon et al. | Aug 2014 | A1 |
| 20140336461 | Reiter et al. | Nov 2014 | A1 |
| 20150148690 | Chopra et al. | May 2015 | A1 |
| 20150235373 | Kato et al. | Aug 2015 | A1 |
| 20150238276 | Atarot et al. | Aug 2015 | A1 |
| 20150265368 | Chopra et al. | Sep 2015 | A1 |
| 20160081653 | Masuda et al. | Mar 2016 | A1 |
| 20160151119 | Michel et al. | Jun 2016 | A1 |
| 20160157939 | Larkin et al. | Jun 2016 | A1 |
| 20160183841 | Duindam et al. | Jun 2016 | A1 |
| 20160192860 | Allenby et al. | Jul 2016 | A1 |
| 20160287344 | Donhowe et al. | Oct 2016 | A1 |
| 20170035380 | Barak et al. | Feb 2017 | A1 |
| 20170112571 | Thiel et al. | Apr 2017 | A1 |
| 20170112576 | Coste-Maniere et al. | Apr 2017 | A1 |
| 20170172662 | Panescu et al. | Jun 2017 | A1 |
| 20170209071 | Zhao et al. | Jul 2017 | A1 |
| 20170265952 | Donhowe et al. | Sep 2017 | A1 |
| 20170280970 | Sartor et al. | Oct 2017 | A1 |
| 20170311844 | Zhao et al. | Nov 2017 | A1 |
| 20170319165 | Averbuch | Nov 2017 | A1 |
| 20180078318 | Barbagli et al. | Mar 2018 | A1 |
| 20180144092 | Flitsch et al. | May 2018 | A1 |
| 20180153621 | Duindam et al. | Jun 2018 | A1 |
| 20180200001 | Erbe | Jul 2018 | A1 |
| 20180235709 | Donhowe et al. | Aug 2018 | A1 |
| 20180240237 | Donhowe et al. | Aug 2018 | A1 |
| 20180256262 | Duindam et al. | Sep 2018 | A1 |
| 20180263527 | Kitamura | Sep 2018 | A1 |
| 20180263706 | Averbuch | Sep 2018 | A1 |
| 20180279852 | Rafii-Tari et al. | Oct 2018 | A1 |
| 20180310831 | Cheng et al. | Nov 2018 | A1 |
| 20180325419 | Zhao et al. | Nov 2018 | A1 |
| 20180356493 | Stapert et al. | Dec 2018 | A1 |
| 20190000559 | Berman et al. | Jan 2019 | A1 |
| 20190000560 | Berman et al. | Jan 2019 | A1 |
| 20190008413 | Duindam et al. | Jan 2019 | A1 |
| 20190038365 | Soper et al. | Feb 2019 | A1 |
| 20190065209 | Mishra et al. | Feb 2019 | A1 |
| 20190110839 | Rafii-Tari et al. | Apr 2019 | A1 |
| 20190175062 | Rafii-Tari et al. | Jun 2019 | A1 |
| 20190175799 | Hsu et al. | Jun 2019 | A1 |
| 20190183318 | Froggatt et al. | Jun 2019 | A1 |
| 20190183585 | Rafii-Tari et al. | Jun 2019 | A1 |
| 20190183587 | Rafii-Tari et al. | Jun 2019 | A1 |
| 20190192234 | Gadda et al. | Jun 2019 | A1 |
| 20190209016 | Herzlinger et al. | Jul 2019 | A1 |
| 20190209043 | Zhao et al. | Jul 2019 | A1 |
| 20190216548 | Ummalaneni | Jul 2019 | A1 |
| 20190239723 | Duindam et al. | Aug 2019 | A1 |
| 20190239831 | Chopra | Aug 2019 | A1 |
| 20190250050 | Sanborn et al. | Aug 2019 | A1 |
| 20190254649 | Walters et al. | Aug 2019 | A1 |
| 20190269470 | Barbagli et al. | Sep 2019 | A1 |
| 20190269818 | Dhanaraj et al. | Sep 2019 | A1 |
| 20190269819 | Dhanaraj et al. | Sep 2019 | A1 |
| 20190272634 | Li et al. | Sep 2019 | A1 |
| 20190298160 | Ummalaneni et al. | Oct 2019 | A1 |
| 20190298451 | Wong et al. | Oct 2019 | A1 |
| 20190320878 | Duindam et al. | Oct 2019 | A1 |
| 20190320937 | Duindam et al. | Oct 2019 | A1 |
| 20190336238 | Yu et al. | Nov 2019 | A1 |
| 20190343424 | Blumenkranz et al. | Nov 2019 | A1 |
| 20190350659 | Wang et al. | Nov 2019 | A1 |
| 20190365199 | Zhao et al. | Dec 2019 | A1 |
| 20190365479 | Rafii-Tari | Dec 2019 | A1 |
| 20190365486 | Srinivasan et al. | Dec 2019 | A1 |
| 20190380787 | Ye et al. | Dec 2019 | A1 |
| 20200000319 | Saadat et al. | Jan 2020 | A1 |
| 20200000526 | Zhao | Jan 2020 | A1 |
| 20200008655 | Schlesinger et al. | Jan 2020 | A1 |
| 20200020127 | Hirakawa | Jan 2020 | A1 |
| 20200030044 | Wang et al. | Jan 2020 | A1 |
| 20200030461 | Sorger | Jan 2020 | A1 |
| 20200038750 | Kojima | Feb 2020 | A1 |
| 20200043207 | Lo et al. | Feb 2020 | A1 |
| 20200046431 | Soper et al. | Feb 2020 | A1 |
| 20200046436 | Tzeisler et al. | Feb 2020 | A1 |
| 20200054398 | Kovtun et al. | Feb 2020 | A1 |
| 20200054399 | Duindam et al. | Feb 2020 | A1 |
| 20200054408 | Schuh et al. | Feb 2020 | A1 |
| 20200060771 | Lo et al. | Feb 2020 | A1 |
| 20200069192 | Sanborn et al. | Mar 2020 | A1 |
| 20200077870 | Dicarlo et al. | Mar 2020 | A1 |
| 20200078023 | Cedro et al. | Mar 2020 | A1 |
| 20200078095 | Chopra et al. | Mar 2020 | A1 |
| 20200078103 | Duindam et al. | Mar 2020 | A1 |
| 20200085514 | Blumenkranz | Mar 2020 | A1 |
| 20200107886 | Govari et al. | Apr 2020 | A1 |
| 20200109124 | Pomper et al. | Apr 2020 | A1 |
| 20200129045 | Prisco | Apr 2020 | A1 |
| 20200129239 | Bianchi et al. | Apr 2020 | A1 |
| 20200138514 | Blumenkranz et al. | May 2020 | A1 |
| 20200138515 | Wong | May 2020 | A1 |
| 20200142013 | Wong | May 2020 | A1 |
| 20200155116 | Donhowe et al. | May 2020 | A1 |
| 20200155232 | Wong | May 2020 | A1 |
| 20200170623 | Averbuch | Jun 2020 | A1 |
| 20200170720 | Ummalaneni | Jun 2020 | A1 |
| 20200179058 | Barbagli et al. | Jun 2020 | A1 |
| 20200188021 | Wong et al. | Jun 2020 | A1 |
| 20200188038 | Donhowe et al. | Jun 2020 | A1 |
| 20200205903 | Srinivasan et al. | Jul 2020 | A1 |
| 20200205904 | Chopra | Jul 2020 | A1 |
| 20200214664 | Zhao et al. | Jul 2020 | A1 |
| 20200229679 | Zhao et al. | Jul 2020 | A1 |
| 20200242767 | Zhao et al. | Jul 2020 | A1 |
| 20200275860 | Duindam | Sep 2020 | A1 |
| 20200297442 | Adebar et al. | Sep 2020 | A1 |
| 20200315554 | Averbuch et al. | Oct 2020 | A1 |
| 20200330795 | Sawant et al. | Oct 2020 | A1 |
| 20200352427 | Deyanov | Nov 2020 | A1 |
| 20200364865 | Donhowe et al. | Nov 2020 | A1 |
| 20200383750 | Kemp et al. | Dec 2020 | A1 |
| 20210000524 | Barry et al. | Jan 2021 | A1 |
| 20210220078 | Godhani et al. | Jul 2021 | A1 |
| Number | Date | Country |
|---|---|---|
| 0013237 | Jul 2003 | BR |
| 0116004 | Jun 2004 | BR |
| 0307259 | Dec 2004 | BR |
| 0412298 | Sep 2006 | BR |
| 112018003862 | Oct 2018 | BR |
| 103565484 | Feb 2014 | CN |
| 103796597 | May 2014 | CN |
| 107198567 | Sep 2017 | CN |
| 108289598 | Jul 2018 | CN |
| 1644519 | Dec 2008 | CZ |
| 486540 | Sep 2016 | CZ |
| 2709512 | Aug 2017 | CZ |
| 2884879 | Jan 2020 | CZ |
| 1644519 | Dec 2008 | EP |
| 2141497 | Jan 2010 | EP |
| 3225151 | Oct 2017 | EP |
| 3413830 | Sep 2019 | EP |
| 3478161 | Feb 2020 | EP |
| 3641686 | Apr 2020 | EP |
| 3644885 | May 2020 | EP |
| 3644886 | May 2020 | EP |
| 3749239 | Dec 2020 | EP |
| 2008018172 | Jan 2008 | JP |
| PA03005028 | Jan 2004 | MX |
| PA03000137 | Sep 2004 | MX |
| PA03006874 | Sep 2004 | MX |
| 225663 | Jan 2005 | MX |
| 226292 | Feb 2005 | MX |
| PA03010507 | Jul 2005 | MX |
| PA05011725 | May 2006 | MX |
| 06011286 | Mar 2007 | MX |
| 246862 | Jun 2007 | MX |
| 2007006441 | Aug 2007 | MX |
| 265247 | Mar 2009 | MX |
| 284569 | Mar 2011 | MX |
| Entry |
|---|
| Notification to Grant Patent Right for Invention issued in Chinese Patent Application No. 201911278581.4 dated Jan. 26, 2024 with English translation. |
| Communication pursuant to Article 94(3) EPC issued in European Patent Application No. 19215514.1 dated Dec. 4, 2023. |
| European Search Report issued in European Patent Application No. EP 19215517 dated Apr. 14, 2020, 9 pages. |
| Notice of Allowance issued in U.S. Appl. No. 17/521,714 dated Apr. 7, 2023. |
| First Office Action issued in Chinese Patent Application No. 201911278464.8 dated Jan. 26, 2024 with English translation. |
| Number | Date | Country | |
|---|---|---|---|
| 20240033033 A1 | Feb 2024 | US |
| Number | Date | Country | |
|---|---|---|---|
| 62782683 | Dec 2018 | US | |
| 62779242 | Dec 2018 | US | |
| 62779229 | Dec 2018 | US |
| Number | Date | Country | |
|---|---|---|---|
| Parent | 16682351 | Nov 2019 | US |
| Child | 18379443 | US |