Robotic surgical systems may be used in minimally invasive surgical procedures. During a robotic surgical procedure, a surgeon controls a robotic surgical arm with a user interface at a remote surgeon console. The user interface allows the surgeon to manipulate a surgical instrument coupled to the robotic arm and to control a camera to receive images of a surgical site within a patient.
The surgeon console may include a stereoscopic display, sometimes referred to as a three-dimensional (3D) display. In this regard, sometimes in conjunction with a corresponding pair of stereoscopic eyeglasses worn by the surgeon, such displays facilitate depth perception from an image by presenting the image to the surgeon as a pair of distinct images separately provided to the left and right eyes, respectively. The stereoscopic display may display images provided by a stereoscopic endoscope. Stereoscopic endoscopes employ two signal paths, usually a left-eye view and a right-eye view, which are matched and interdigitated to generate a stereoscopic image. As does typically occur during surgical procedures, biological material or other procedure-related material may occlude one of the lenses of the stereoscopic endoscope, thereby degrading the images provided to the display. In the case of a stereoscopic endoscope, this degradation has potential side effects upon the surgeon through the now mismatched stereoscopic image pairs which can cause perception issues that tax the surgeon's visual and cognitive pathways without the surgeon's awareness. This can result in degraded performance of the surgeon in perceiving and responding to the observed stereoscopic information. Thus, it is useful to be able to detect these mismatch situations and inform the surgeon of the need to correct the situation.
Disclosed according to embodiments of the present disclosure are methods for detecting image degradation during a surgical procedure. In an aspect of the present disclosure, an illustrative method includes receiving images of a surgical instrument, obtaining baseline images of an edge of the surgical instrument, comparing a characteristic of the images of the surgical instrument to a characteristic of the baseline images of the edge of the surgical instrument, the images of the surgical instrument being received subsequent to the obtaining of the baseline images of the edge of the surgical instrument and being received while the surgical instrument is disposed at a surgical site in a patient, determining whether the images of the surgical instrument are degraded, based on the comparing of the characteristic of the images of the surgical instrument and the characteristic of the baseline images of the edge of the surgical instrument, and generating an image degradation notification, in response to a determination that the images of the surgical instrument are degraded.
In a further aspect of the present disclosure, the images of the surgical instrument are received by an image capture device.
In another aspect of the present disclosure, the image capture device is a stereoscopic endoscope including a left-eye lens and a right-eye lens.
In a further aspect of the present disclosure, the characteristic of the baseline images of the edge of the surgical instrument is obtained during an initial image capture device calibration.
In a further aspect of the present disclosure, the method further includes periodically receiving images of the edge of the surgical instrument at a predetermined interval.
In another aspect of the present disclosure, the determination that the images of the surgical instrument are degraded is based at least in part on a difference between the characteristic of the received images of the surgical instrument and the characteristic of the baseline images of the edge of the surgical instrument being greater than a threshold value.
In yet another aspect, the method further includes determining the characteristic of the images of the surgical instrument by computing a modulation transfer function derived from the received images of the surgical instrument.
Disclosed according to embodiments of the present disclosure are systems for detecting image degradation during a surgical procedure. In an aspect of the present disclosure, an illustrative system includes a surgical instrument including at least one edge, an image capture device configured to capture images of the surgical instrument, the images of the surgical instrument including a characteristic, a display device, at least one processor coupled to the image capture device and the display device, and a memory coupled to the at least one processor and having stored thereon a characteristic of baseline images of the edge of the surgical instrument, and instructions which, when executed by the at least one processor, cause the at least one processor to obtain the characteristic of the baseline images of the edge of the surgical instrument, receive the images of the surgical instrument, compare a characteristic of the images of the surgical instrument to the characteristic of the baseline images of the edge of the surgical instrument, the images of the surgical instrument being received subsequent to the obtaining of the characteristic of the baseline images of the edge of the surgical instrument and being received while the surgical instrument is disposed at a surgical site in a patient, determine whether the images of the surgical instrument are degraded, based on the comparing of the characteristic of the images of the surgical instrument and the characteristic of the baseline images of the edge of the surgical instrument, and generate an image degradation notification, in response to a determination that the images of the surgical instrument are degraded.
Disclosed according to embodiments of the present disclosure are non-transitory computer-readable media storing instructions for detecting image degradation during a surgical procedure. In an aspect of the present disclosure, an illustrative non-transitory computer-readable medium stores instructions which, when executed by a processor, cause the processor to receive images of a surgical instrument, obtain baseline images of an edge of the surgical instrument, compare a characteristic of the images of the surgical instrument to a characteristic of the baseline images of the edge of the surgical instrument, the images of the surgical instrument being received subsequent to obtaining the baseline images of the edge of the surgical instrument and being received while the surgical instrument is disposed at a surgical site in a patient, determine whether the images of the surgical instrument are degraded, based on the comparison of the characteristic of the images of the surgical instrument and the characteristic of the baseline images of the edge of the surgical instrument, and generate an image degradation notification, in response to a determination that the images of the surgical instrument are degraded.
Any of the above aspects and embodiments of the present disclosure may be combined without departing from the scope of the present disclosure.
Objects and features of the presently disclosed systems, methods, and computer-readable media will become apparent to those of ordinary skill in the art when descriptions of various embodiments thereof are read with reference to the accompanying drawings, wherein:
The present disclosure generally relates to dynamic detection of image degradation, and providing associated notifications, during a surgical procedure. In order to determine an amount of degradation occurring during a surgical procedure, endoscopic calibration techniques may be used. Prior to its use, and/or at the start of a surgical procedure, an endoscopic imaging system may be calibrated. Calibration, prior to use, includes the process of determining and recording base parameters, at peak or near-peak operating conditions for the imaging system by using a calibration target. Calibration prior to the use of the endoscopic system thus provides a baseline metric of the endoscopic system before the occurrence of degradation. During a surgical procedure, a similar technique as that employed during calibration may be used to determine current parameters of the endoscopic imaging system. By automatically and dynamically comparing the current parameters with those of the base parameters, endoscopic image degradation can be determined.
To that end, the present disclosure relates to systems, methods, and computer-readable media for enabling dynamic detection of image degradation of images of a surgical site during a surgical procedure, and for generating and displaying image degradation notifications during the surgical procedure. In this regard, during calibration of the image capture device, one or more baseline parameters of the image capture device, based on calibration targets such as test patterns and/or edges of tools within the image capture device's field of view, are determined and recorded. During the surgical procedure, images of the surgical site are captured by the image capture device and provided to a computing device, such as a control console, for processing, using a similar technique as that employed during calibration, to determine one or more current parameters of the image capture device. By dynamically comparing the current parameter(s) with the baseline parameter(s) of the image capture device, a determination regarding image degradation can be made.
As used herein, the terms “clinician,” “surgeon,” “observer,” and/or “viewer” generally refer to a user of a stereoscopic display device described herein. Additionally, although the terms “first eye” and “second eye” are used herein to refer to a left eye and a right eye, respectively, of a user, this use is provided by way of example and should not be construed as limiting. Throughout this description, the term “proximal” refers to the portion of the device or component thereof that is farthest away from the patient (and thus closest to the clinician and/or surgical robot) and the term “distal” refers to the portion of the device or component thereof that is closest to the patient (and thus furthest away from the clinician and/or surgical robot). Further, as referred herein, the term “signal path” (whether right-eye or left-eye) refers to an optical-electrical-optical signal path whereby images are captured optically, converted to an electrical/digital signal to be transmitted, and again converted back to an optical image when received by a computing or display device. While the illustrative embodiments below describe a robotic surgical system, those skilled in the art will recognize that the systems, methods, and computer-readable media described herein may also be used in other surgical procedures, for example minimally-invasive surgical procedures, where a patient image capture device is used to capture images of a surgical site. Thus, the present disclosure is not intended to be limited to the exemplary embodiments using a robotic surgical system, as described hereinbelow.
With reference to
Controller 30 includes, and/or is communicatively coupled to, the at least one processor 32 and memory 35, and may be integrated with user interface 40 or provided as a standalone device within the operating theater. As described in further detail below, processor 32 executes instructions (not shown) stored in memory 35 to perform steps and/or procedures of the various embodiments described herein. As will be appreciated, the implementation of processor 32 and memory 35 is provided by way of example only and should not be construed as limiting. For instance, steps and/or procedures of any of the embodiments of the present disclosure may be implemented by hardware components, firmware components, software components, and/or any combination thereof.
User interface 40 communicates with base 18 through controller 30 and includes a display device 44 which is configured to display stereoscopic images of the surgical site “S.” The images are captured by an imaging device (also referred to as “image capture device”) 16 and/or captured by imaging devices that are positioned about the surgical theater (e.g., an imaging device positioned adjacent patient “P,” and/or an imaging device 56 positioned at a distal end of an imaging arm 52). Imaging devices (e.g., imaging devices 16, 56) may capture optical images, infra-red images, ultrasound images, X-ray images, thermal images, and/or any other known real-time images of surgical site “S.” Imaging devices 16, 56 transmit captured images to controller 30 for processing, such as by processor 32, and transmits the captured and/or processed images to display device 44 for display. In one embodiment, one or both of imaging devices 16, 56 are stereoscopic endoscopes capable of capturing images of surgical site “S” via a right-eye lens 210 and a left-eye lens 220, as further described in the description of
In further embodiments, user interface 40 may include or be associated with a portable display device 45, which, similar to display device 44, is configured to permit the user to view the stereoscopic images in a manner that the user perceives a three-dimensional and/or depth effect from the stereoscopic images. Portable display device 45 may be goggles, glasses, or any other portable or semi-portable display device, which may be used to allow the user to view stereoscopic images.
User interface 40 further includes input handles attached to gimbals 70 which allow a clinician to manipulate surgical robot 10 (e.g., move robotic arms 12, ends 14 of robotic arms 12, and/or surgical instrument 250). Each of gimbals 70 is in communication with controller 30 and processor 32 to transmit control signals thereto and to receive feedback signals therefrom. Additionally or alternatively, each of gimbals 70 may include control interfaces or input devices (not shown) which allow the surgeon to manipulate (e.g., clamp, grasp, fire, open, close, rotate, thrust, slice, etc.) surgical instrument 250 supported at ends 14 of robotic arms 12.
Each of gimbals 70 is moveable to move ends 14 of robotic arms 12 within surgical site “S.” The stereoscopic images displayed on display device 44 are oriented such that movement of gimbals 70 moves ends 14 of robotic arms 12 as viewed on display device 44. It will be appreciated that the orientation of the stereoscopic images on display device 44 may be mirrored or rotated relative to a view from above patient “P.” In addition, it will be appreciated that the size of the stereoscopic images displayed on display device 44 may be scaled to be larger or smaller than the actual structures of surgical site “S” permitting the surgeon to have a better view of structures within surgical site “S.” As gimbal 70 is moved, surgical instrument 250 are moved within surgical site “S.” Movement of surgical instrument 250 may also include movement of ends 14 of robotic arms 12 which support surgical instrument 250. In addition to gimbals 70, one or more additional input devices may be included as part of user interface 40, such as a handle including a clutch switch, a touchpad, joystick, keyboard, mouse, or other computer accessory, and/or a foot switch, pedal, trackball, or other actuatable device configured to translate physical movement from the clinician into signals sent to processor 32.
As noted briefly above, to provide the user with a view of surgical site “S” during a surgical procedure, one or more of imaging devices 16, 56 may be a stereoscopic endoscope disposed about surgical site “S,” such as adjacent to surgical instrument 250, and configured to capture images of surgical site “S” to be displayed as stereoscopic images on display device 44.
Turning now to
Surgical instrument 250 is illustrated as a vessel sealing device, which includes a body 242 having a surface 252. Those skilled in the art will recognize that this illustrative surgical instrument 250 is provided merely as an example, and that any other surgical tool or device may be substituted for the illustrated vessel sealing device without departing from the scope of the present disclosure. In some embodiments, one or more test patterns (for example, test pattern 255a) are included on surgical instrument 250. The test pattern is an identifier, for example a unique identifier, that can be used to distinguish surgical instrument 250 from a background during image processing. For example, as depicted in
In another embodiment, as shown in
In still another embodiment, test patterns 255a, 255b extend along an entire outer area of surface 252 of surgical instrument 250. Alternatively, it is contemplated that test patterns 255a, 255b may be located at discrete locations on surface 252 of surgical instrument 250. In a further embodiment, it is contemplated that any of test patterns 255a, 255b may include specific markers, such as specialized shapes, which enhance the ability of image capture device 200 to determine that test patterns 255a, 255b are present within the received image of surgical site “S.” In embodiments, it is contemplated that test patterns 255a, 255b correspond to a type of surgical instrument so that each different surgical instrument 250 or type of surgical instrument 250 (for example, ablation device, dissection device, stapler, vessel sealing device, etc.) has a unique test pattern, which can be used to, where necessary, identify surgical instrument 250.
In still another embodiment, a pseudo test pattern is generated using a contrast between surgical instrument 250 and the background of an image of surgical instrument 250. The pseudo test pattern may be a proxy or substitute for an actual test pattern, such as test pattern 255a, 255b. In some embodiments, one or more geometric characteristics of surgical instrument 250 may be used as a pseudo test pattern and/or may be used to define a region of an image that acts as a pseudo test pattern. In one illustrative embodiment, one or more edges of surgical tool 250 are used to define a pseudo test pattern, e.g., the regions of the image on both sides of an edge of surgical tool 250 define a pseudo test pattern. For example, referring now to
Referring now to
In an embodiment, the image of test pattern 255b′ is used to determine a value of a modulation transfer function (“MTF”). For example, the MTF is used to provide a measure of the transfer of contrast of test pattern 255b′ and how well lenses 210, 220 of image capture device 200 reproduce (or transfer) the detail of test pattern 255b in a captured image. By obtaining an image of test pattern 255b or other test patterns (for example, test pattern 255a, pseudo test pattern 255c, and/or the like) during calibration of image capture device 200, baseline characteristics of the image of test pattern 255b or other test patterns can be determined. During use of image capture device 200, the ability of lenses 210, 220 of image capture device 200 to continue to reproduce (or transfer) the detail of the test pattern is determined by applying the MTF to the received image of the test pattern to yield one or more determined characteristics of the image of the test pattern. Over time, the determined characteristic(s) of the image of the test pattern may change due to degradation and/or occlusion of lenses 210, 220 of image capture device 200, for example, as shown in the image of test pattern 255b′. As described below with reference to
In order to calculate the MTF for both the baseline characteristics of the test pattern and determined characteristic(s) of the test pattern, processor 32 is used to differentiate the black bars and white bars of test pattern 255b, or in the case of pseudo test pattern 255c, the darker sections and the brighter sections of the captured image proximate edges 251 of surgical instrument 250. Based on the differences between the black bars and white bars, or darker sections and brighter sections, a sinusoidal graph can be plotted. Turning now to
Additionally, as the widths of the black bars and white bars of test pattern 255b′ decrease, the ability of processor 32 to continue to differentiate between black bars and white bars may likewise decrease. As such, the peak amplitudes of the graph may no longer approach peak amplitudes of −100 and 100 such that, eventually, the width of the black bars and white bars becomes so thin that processor 32 can no longer distinguish the black bars from white bars of test pattern 225b′ and the peak amplitude settles at 0. In some embodiments, a value of 100% may be assigned where the peak amplitude is between −100 and 100 and a value of 0% may be assigned where the peak amplitudes settles at 0. For peak amplitudes between −100 and 100, a corresponding percentage between 100% and 0% can be assigned. The MTF is typically expressed as a percentage of the distinguishable contrast between black bars and white bars based on the line widths per picture height (LW/PH). Thus, as the line widths (width of black bars and white bars) of test pattern 255b′ become increasingly thinner, the percentage representing the contrast between the black and white bars decreases.
By assigning a value to the groups of increasingly thinner widths of black bars and white bars in the form of the LW/PH, the MTF percentage may correspond to the LW/PH. For example, if a portion of the graph of
In further embodiments, the MTF percentages may be converted from percentages ranging between 0% to 100% to corresponding values ranging between 0 and 1, wherein 0 corresponds to processor 32 being incapable of distinguishing black bars from white bars, and 1 corresponds to processor 32 being able to completely distinguish the black bars from white bars. It is further contemplated that processor 32 is capable of determining the MTF of test pattern 255b′ as it is received by way of each of right-eye lens 210 and left-eye lens 220, independently. Thus, image degradation for images captured by way of each of right-eye lens 210 and left-eye lens 220 may be detected.
While the above description of
Generally, prior to the execution of method 700, calibration of image capture device 200 will have been performed. For example, during a factory calibration process, image capture device 200 may receive, or the memory 32 may have stored therein, left-eye and right-eye images of one or more test patterns. The test patterns may be similar to test patterns 255a, 255b disposed on surface 252 of surgical device 250, or may be data related to the contrast between the edges of surgical device 250 and a surrounding environment, thereby generating a pseudo test pattern 255c. In another example, the calibration process may be performed at the start of a surgical procedure. In an embodiment in which images of the test patterns (“baseline images”) are received by image capture device 200 during calibration, a pattern analysis function, such as the modulation transfer function (MTF), is applied to the test pattern to calculate output values. The output values may represent the sharpness or clarity of a transition across features making up the test pattern and may be expressed as line widths per picture height (LW/PH). The calculated output values may be included as one or more characteristics of the baseline images captured by the right-eye lens 210 and the left-eye lens 220 of the image capture device 200 (“baseline characteristics”), and can be stored in the memory 32 for later use, as will be described in detail below. Alternatively, these baseline characteristics may be known values that are stored in the memory 32. Additionally, during a calibration process, image capture device 200 may store the baseline characteristics of the test pattern in memory.
In any case, surgical system 10 is configured to permit the user to begin the surgical procedure within surgical site “S,” at step 705. For example, in the case of a robotic surgical procedure, the user moves the gimbals 70 to thereby position image capture device 200 and surgical instrument 250 about surgical site “S”. In some embodiments, the field of view of image capture device 200 may initially be aligned with surgical instrument 250 to enable image capture device 200 to capture images of surgical instrument 250, and particularly test patterns 255a, 255b, via right-eye lens 210 and left-eye lens 220, respectively. Alternatively, as will be appreciated by those skilled in the art, in non-robotic minimally-invasive surgical procedures, image capture device 200 and surgical instrument 250 may be positioned manually about surgical site “S.”
Once suitably positioned, image capture device 200 captures images of surgical site “S,” and transmits the captured images to controller 30 at step 710. In addition to tissue and surrounding anatomical material 230 on which the surgical procedure is being performed, the captured images show surgical instrument 250 as it is being manipulated by the user. The captured images may be stereoscopic images, that is, left-eye images and right-eye images.
After receiving the captured images from image capture device 200, processor 32 processes the captured images to identify a test pattern, at step 715. As described herein, it is contemplated that the test pattern may be a pattern disposed on surface 252 of surgical instrument 250 or a test pattern formed by the contrast been surgical instrument 250 and surrounding anatomical material 230 about a slanted edge 251 of surgical instrument 250. As noted above, examples of test patterns include but are not limited to test patterns 255a, 255b and pseudo test pattern 255c, and/or the like. In an embodiment, a suitable algorithm is applied to the left-eye images and the right-eye images of surgical site “S” to output a result, which is analyzed by processor 32 to detect a presence of the test pattern.
Optionally, in an embodiment where test pattern 255b is disposed on surgical instrument 250, at step 720, after identifying test pattern 255b′, a determination is made as to whether test pattern 255b′ matches a known test pattern. For example, a database of known test patterns may be stored in memory 35, for example, in a look-up table. The images of test pattern 255b′ captured by image capture device 200 is compared with the known test pattern images stored in memory 35. In an embodiment, each of the known test patterns is associated with a different surgical instrument. As such, matching test pattern 255b′ with the known test pattern further includes identifying the surgical instrument corresponding to the known test pattern. In an embodiment, the identification of the surgical instrument is provided to the user via display device 44 or through an audio device. For illustrative purposes and provided by way of example, pseudo test pattern 255c, that is, the slanted edges 251 of surgical instrument 250, is used as the exemplary test pattern for the remaining description of
At step 725, the one or more baseline characteristics of the test pattern (“characteristic(s) of the baseline image of the test pattern”), generated and/or calculated during calibration, are obtained. In embodiments where the calibration process is not performed prior to the start of the surgical procedure, the calibration process may be performed at step 725. In other embodiments, as noted above, the characteristic(s) of the baseline images of the test pattern may be stored in memory 35. In such embodiments, the corresponding characteristic(s) of the baseline image of the test pattern are retrieved from a lookup table and/or database stored in memory 35.
At step 730, the image of pseudo test pattern 255c received at step 715 is analyzed and one or more characteristics are determined from the images of pseudo test pattern 255c. In an embodiment, a MTF is calculated for the images of pseudo test pattern 255c in order to determine the characteristics of the images of pseudo test pattern 255c. The determined characteristic(s) of the images of pseudo test pattern 255c may be in the form of a percentage at LW/PH. For example, at step 730, the MTF may yield values of 100% at 105 mm and 45% at 50 mm. In other embodiments, various other types of analysis functions may be applied to the images of pseudo test pattern 255c. In any case, step 730 is continuously reiterated, so that changes in the determined characteristic(s) of the images of pseudo test pattern 255c may be detected from the images of pseudo test pattern 255c over time.
In step 735, the determined characteristic(s) of the images of pseudo test pattern 255c are compared to the one or more characteristics of the baseline images of pseudo test pattern 255c. In one embodiment, it is contemplated that either the percentage or the LW/PH are selected at a specific value for the characteristic(s) of the baseline images of pseudo test pattern 255c to be compared with percentages or LW/PH of the determined characteristic(s) of images of pseudo test pattern 255c. For example, if the LW/PH is selected for the characteristic(s) of the baseline images of pseudo test pattern 255c at a specific value of 50 mm and the corresponding percentage is 50% (thus yielding a MTF graph amplitude ranging from 50 to −50 at LW/PH of 50 mm), as described above in the description of
In an embodiment, at step 740, a determination is made as to whether the difference between the characteristic(s) of the baseline images of pseudo test pattern 255c and the determined characteristic(s) of the images of pseudo test pattern 255c is greater than a predetermined threshold. The predetermined threshold is a percentage decrease between the characteristic(s) of the baseline images of pseudo test pattern 255c and the determined characteristic(s) of pseudo test pattern 255c. For example, a predetermined threshold of the modulation function may translate to a decrease of a 15% of line widths per picture height (LW/PH), which may indicate image marrying from material on the lens of one or both images. If the difference is not greater than a predetermined threshold (“NO” at step 740), method 700 proceeds to step 745, where new images of surgical site “S” are received. Next, at step 747, similar to step 715, the new images of surgical site “S” are processed to identify pseudo test pattern 255c. Following step 747, method 700 returns to step 730 where the received images of pseudo test pattern 255c are analyzed by processor 32 and one or more characteristics of the images of pseudo test pattern 255c are determined.
It is contemplated that steps 730 through step 740 and returning to step 730 may be performed iteratively and repeated at regular intervals. In one embodiment, it is contemplated that processor 32 will proceed from steps 730 through step 740 returning to step 730 processing newly received images of surgical site “S” at 60 Hz, possibly 30 Hz, and even as low as 10 Hz. Thus, relatively frequesly, new images of surgical site “S” is received, processed, and a determination made as to whether the difference between the characteristic(s) of the baseline image of pseudo test pattern 255c and the determined characteristic(s) of the images of pseudo test pattern 255c, is greater than a predetermined threshold. In other embodiments, based on the surgical procedure, the intervals from steps 730 through step 740 returning to step 730 may be shortened in order to increase the frequency of the determination of image degradation, for example, every half, quarter, or one-tenth of a second.
In a further embodiment, a determination is made as to whether a trend is detected from the differences determined at step 735. It is contemplated that the differences from step 735 are stored in memory 35, and the data is monitored to determine the trend. For example, processor 32 may determine that image degradation has likely occurred due to tissue occluding one or both of lenses 210, 220, where image degradation occurs rapidly following a small number of passes from steps 730 through 740 and returning to step 730. Alternatively, processor 32 may determine that image degradation has occurred due to a gradual build-up of fluid or other anatomical material where image degradation occurs more slowly.
If, at step 740, it is determined that the difference between the characteristic(s) of the baseline images of pseudo test pattern 255c and the determined characteristic(s) of the images of pseudo test pattern 255c is greater than a predetermined threshold (“YES” at step 740), method 700 proceeds to step 742. At step 742, a determination is made as to whether image degradation has occurred, based on the result of the determination at step 740. For example, image degradation may include the images being distorted, out of focus, partially or wholly occluded, and/or a mismatch between the images captured by left-eye lens 210 and right-eye lens 220, thereby causing a stereoscopic visual distortion even if the images, when viewed separately, do not appear distorted. Thereafter, method 700 proceeds to step 750, where a notification is generated and provided to the user indicating that image degradation may have occurred. The notification may be displayed, for example, via display device 44 or portable display device 45 of user interface 40, and/or be provided audibly or tactilely via gimbals 70.
Following step 750, method 700 proceeds to step 755 where feedback is provided to the user indicating how image quality may be improved. For example, the feedback provided may be in the form of a notification via display device 44 or portable display device 45, that the user should remove and clean one or both of lenses 210, 220 in order to improve image quality. After feedback is provided indicating how to improve image quality at step 755, the method 700 proceeds to step 760 where it is determined whether the image quality can be improved during the surgical procedure. For example, it may be determined whether image capture device 200 needs to be removed from surgical site “S” and cleaned, or be replaced. Those skilled in the art will envision various other actions that may be taken to improve the image quality of images captured by image capture device 200, and thus, for purpose of brevity, all such alternative actions will not be described here. If it is determined at step 760 that the image quality cannot be improved during the surgical procedure (“NO” at step 760), the method 700 proceeds to step 775 where the surgical procedure ends. Alternatively, if it is determined at step 760 that the image quality can be improved during the surgical procedure (“YES” at step 760), the method 700 proceeds to step 765.
At step 765 it is determined whether the image quality has been improved. For example, the processes described above with reference to steps 745, 747, 730, 735, and 740 may be repeated to determine whether the image quality of images captured by image capture device 200 has been improved. If it is determined at step 765 that the image quality has not been improved (“NO” at step 765), processing returns to step 755. Alternatively, if it is determined at step 765 that the image quality has been improved (“YES” at step 765), the method 700 proceeds to step 770.
At step 770 it is determined whether the surgical procedure has been completed. For example, it may be determined whether the user has provided an instruction and/or indication that the surgical procedure has been completed. If it is determined at step 770 that the surgical procedure has not been completed (“NO” at step 770), processing returns to step 745. Alternatively, if it is determined at step 770 that the surgical procedure has been completed (“YES” at step 770), processing ends.
Referring back to the computer-readable media of
Detailed embodiments of devices, systems incorporating such devices, and methods using the same have been described herein. However, these detailed embodiments are merely examples of the disclosure, which may be embodied in various forms. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a basis for the claims and as a representative basis for allowing one skilled in the art to employ the present disclosure in virtually any appropriately detailed structure.
The present application is a Continuation of U.S. patent application Ser. No. 17/256,422, filed Dec. 28, 2020, which is a U.S. National Stage Application filed under 35 U.S.C. § 371(a) of International Patent Application Serial No. PCT/US2019/038869, filed Jun. 25, 2019, which claims the benefit of and priority to U.S. Provisional Patent Application Ser. No. 62/693,530, filed Jul. 3, 2018, the entire disclosure of each of which is incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
6069691 | Rosow et al. | May 2000 | A |
6132368 | Cooper | Oct 2000 | A |
6206903 | Ramans | Mar 2001 | B1 |
6219182 | McKinley | Apr 2001 | B1 |
6246200 | Blumenkranz et al. | Jun 2001 | B1 |
6312435 | Wallace et al. | Nov 2001 | B1 |
6331181 | Tierney et al. | Dec 2001 | B1 |
6388742 | Duckett | May 2002 | B1 |
6394998 | Wallace et al. | May 2002 | B1 |
6424885 | Niemeyer et al. | Jul 2002 | B1 |
6441577 | Blumenkranz et al. | Aug 2002 | B2 |
6459926 | Nowlin et al. | Oct 2002 | B1 |
6491691 | Morley et al. | Dec 2002 | B1 |
6491701 | Tierney et al. | Dec 2002 | B2 |
6493608 | Niemeyer | Dec 2002 | B1 |
6565554 | Niemeyer | May 2003 | B1 |
6645196 | Nixon et al. | Nov 2003 | B1 |
6659939 | Moll | Dec 2003 | B2 |
6671581 | Niemeyer et al. | Dec 2003 | B2 |
6676684 | Morley et al. | Jan 2004 | B1 |
6685698 | Morley et al. | Feb 2004 | B2 |
6699235 | Wallace et al. | Mar 2004 | B2 |
6714839 | Salisbury, Jr. et al. | Mar 2004 | B2 |
6716233 | Whitman | Apr 2004 | B1 |
6728599 | Wang et al. | Apr 2004 | B2 |
6746443 | Morley et al. | Jun 2004 | B1 |
6766204 | Niemeyer et al. | Jul 2004 | B2 |
6770081 | Cooper et al. | Aug 2004 | B1 |
6772053 | Niemeyer | Aug 2004 | B2 |
6783524 | Anderson et al. | Aug 2004 | B2 |
6793652 | Whitman et al. | Sep 2004 | B1 |
6793653 | Sanchez et al. | Sep 2004 | B2 |
6799065 | Niemeyer | Sep 2004 | B1 |
6837883 | Moll et al. | Jan 2005 | B2 |
6839612 | Sanchez et al. | Jan 2005 | B2 |
6840938 | Morley et al. | Jan 2005 | B1 |
6843403 | Whitman | Jan 2005 | B2 |
6846309 | Whitman et al. | Jan 2005 | B2 |
6866671 | Tierney et al. | Mar 2005 | B2 |
6871117 | Wang et al. | Mar 2005 | B2 |
6879880 | Nowlin et al. | Apr 2005 | B2 |
6899705 | Niemeyer | May 2005 | B2 |
6902560 | Morley et al. | Jun 2005 | B1 |
6936042 | Wallace et al. | Aug 2005 | B2 |
6951535 | Ghodoussi et al. | Oct 2005 | B2 |
6974449 | Niemeyer | Dec 2005 | B2 |
6991627 | Madhani et al. | Jan 2006 | B2 |
6994708 | Manzo | Feb 2006 | B2 |
7048745 | Tierney et al. | May 2006 | B2 |
7066926 | Wallace et al. | Jun 2006 | B2 |
7118582 | Wang et al. | Oct 2006 | B1 |
7125403 | Julian et al. | Oct 2006 | B2 |
7155315 | Niemeyer et al. | Dec 2006 | B2 |
7189000 | Miyauchi et al. | Mar 2007 | B2 |
7239940 | Wang et al. | Jul 2007 | B2 |
7277120 | Gere et al. | Oct 2007 | B2 |
7306597 | Manzo | Dec 2007 | B2 |
7357774 | Cooper | Apr 2008 | B2 |
7373219 | Nowlin et al. | May 2008 | B2 |
7379790 | Toth et al. | May 2008 | B2 |
7386365 | Nixon | Jun 2008 | B2 |
7391173 | Schena | Jun 2008 | B2 |
7398707 | Morley et al. | Jul 2008 | B2 |
7413565 | Wang et al. | Aug 2008 | B2 |
7453227 | Prisco et al. | Nov 2008 | B2 |
7524320 | Tierney et al. | Apr 2009 | B2 |
7574250 | Niemeyer | Aug 2009 | B2 |
7594912 | Cooper et al. | Sep 2009 | B2 |
7607440 | Coste-Maniere et al. | Oct 2009 | B2 |
7666191 | Orban, II et al. | Feb 2010 | B2 |
7682357 | Ghodoussi et al. | Mar 2010 | B2 |
7689320 | Prisco et al. | Mar 2010 | B2 |
7695481 | Wang et al. | Apr 2010 | B2 |
7695485 | Whitman et al. | Apr 2010 | B2 |
7699855 | Anderson et al. | Apr 2010 | B2 |
7713263 | Niemeyer | May 2010 | B2 |
7725214 | Diolaiti | May 2010 | B2 |
7727244 | Orban, III et al. | Jun 2010 | B2 |
7741802 | Prisco | Jun 2010 | B2 |
7756036 | Druke et al. | Jul 2010 | B2 |
7757028 | Druke et al. | Jul 2010 | B2 |
7762825 | Burbank et al. | Jul 2010 | B2 |
7778733 | Nowlin et al. | Aug 2010 | B2 |
7803151 | Whitman | Sep 2010 | B2 |
7806891 | Nowlin et al. | Oct 2010 | B2 |
7819859 | Prisco et al. | Oct 2010 | B2 |
7819885 | Cooper | Oct 2010 | B2 |
7824401 | Manzo et al. | Nov 2010 | B2 |
7835823 | Sillman et al. | Nov 2010 | B2 |
7843158 | Prisco | Nov 2010 | B2 |
7865266 | Moll et al. | Jan 2011 | B2 |
7865269 | Prisco et al. | Jan 2011 | B2 |
7886743 | Cooper et al. | Feb 2011 | B2 |
7899578 | Prisco et al. | Mar 2011 | B2 |
7907166 | Lamprecht et al. | Mar 2011 | B2 |
7935130 | Williams | May 2011 | B2 |
7963913 | Devengenzo et al. | Jun 2011 | B2 |
7983793 | Toth et al. | Jul 2011 | B2 |
8002767 | Sanchez | Aug 2011 | B2 |
8004229 | Nowlin et al. | Aug 2011 | B2 |
8012170 | Whitman et al. | Sep 2011 | B2 |
8054752 | Druke et al. | Nov 2011 | B2 |
8062288 | Cooper et al. | Nov 2011 | B2 |
8079950 | Stern et al. | Dec 2011 | B2 |
8100133 | Mintz et al. | Jan 2012 | B2 |
8108072 | Zhao et al. | Jan 2012 | B2 |
8120301 | Goldberg et al. | Feb 2012 | B2 |
8142447 | Cooper et al. | Mar 2012 | B2 |
8147503 | Zhao et al. | Apr 2012 | B2 |
8151661 | Schena et al. | Apr 2012 | B2 |
8155479 | Hoffman et al. | Apr 2012 | B2 |
8182469 | Anderson et al. | May 2012 | B2 |
8202278 | Orban, III et al. | Jun 2012 | B2 |
8206406 | Orban, III | Jun 2012 | B2 |
8210413 | Whitman et al. | Jul 2012 | B2 |
8216250 | Orban, III et al. | Jul 2012 | B2 |
8220468 | Cooper et al. | Jul 2012 | B2 |
8223193 | Zhao et al. | Jul 2012 | B2 |
8256319 | Cooper et al. | Sep 2012 | B2 |
8285517 | Sillman et al. | Oct 2012 | B2 |
8315720 | Mohr et al. | Nov 2012 | B2 |
8335590 | Costa et al. | Dec 2012 | B2 |
8347757 | Duval | Jan 2013 | B2 |
8374723 | Zhao et al. | Feb 2013 | B2 |
8418073 | Mohr et al. | Apr 2013 | B2 |
8419717 | Diolaiti et al. | Apr 2013 | B2 |
8423182 | Robinson et al. | Apr 2013 | B2 |
8452447 | Nixon | May 2013 | B2 |
8454585 | Whitman | Jun 2013 | B2 |
8499992 | Whitman et al. | Aug 2013 | B2 |
8508173 | Goldberg et al. | Aug 2013 | B2 |
8528440 | Morley et al. | Sep 2013 | B2 |
8529582 | Devengenzo et al. | Sep 2013 | B2 |
8540748 | Murphy et al. | Sep 2013 | B2 |
8551116 | Julian et al. | Oct 2013 | B2 |
8562594 | Cooper et al. | Oct 2013 | B2 |
8594841 | Zhao et al. | Nov 2013 | B2 |
8597182 | Stein et al. | Dec 2013 | B2 |
8597280 | Cooper et al. | Dec 2013 | B2 |
8600551 | Itkowitz et al. | Dec 2013 | B2 |
8608773 | Tierney et al. | Dec 2013 | B2 |
8620473 | Diolaiti et al. | Dec 2013 | B2 |
8624537 | Nowlin et al. | Jan 2014 | B2 |
8634957 | Toth et al. | Jan 2014 | B2 |
8638056 | Goldberg et al. | Jan 2014 | B2 |
8638057 | Goldberg et al. | Jan 2014 | B2 |
8644988 | Prisco et al. | Feb 2014 | B2 |
8666544 | Moll et al. | Mar 2014 | B2 |
8668638 | Donhowe et al. | Mar 2014 | B2 |
8746252 | McGrogan et al. | Jun 2014 | B2 |
8749189 | Nowlin et al. | Jun 2014 | B2 |
8749190 | Nowlin et al. | Jun 2014 | B2 |
8758352 | Cooper et al. | Jun 2014 | B2 |
8761930 | Nixon | Jun 2014 | B2 |
8768516 | Diolaiti et al. | Jul 2014 | B2 |
8786241 | Nowlin et al. | Jul 2014 | B2 |
8790243 | Cooper et al. | Jul 2014 | B2 |
8808164 | Hoffman et al. | Aug 2014 | B2 |
8816628 | Nowlin et al. | Aug 2014 | B2 |
8821480 | Burbank | Sep 2014 | B2 |
8823308 | Nowlin et al. | Sep 2014 | B2 |
8827989 | Niemeyer | Sep 2014 | B2 |
8828023 | Neff et al. | Sep 2014 | B2 |
8838270 | Druke et al. | Sep 2014 | B2 |
8852174 | Burbank | Oct 2014 | B2 |
8858547 | Brogna | Oct 2014 | B2 |
8862268 | Robinson et al. | Oct 2014 | B2 |
8864751 | Prisco et al. | Oct 2014 | B2 |
8864752 | Diolaiti et al. | Oct 2014 | B2 |
8903546 | Diolaiti et al. | Dec 2014 | B2 |
8903549 | Itkowitz et al. | Dec 2014 | B2 |
8911428 | Cooper et al. | Dec 2014 | B2 |
8912746 | Reid et al. | Dec 2014 | B2 |
8939894 | Morrissette et al. | Jan 2015 | B2 |
8944070 | Guthart | Feb 2015 | B2 |
8989903 | Weir et al. | Mar 2015 | B2 |
9002518 | Manzo | Apr 2015 | B2 |
9014856 | Manzo et al. | Apr 2015 | B2 |
9016540 | Whitman et al. | Apr 2015 | B2 |
9019345 | O'Grady et al. | Apr 2015 | B2 |
9043027 | Durant et al. | May 2015 | B2 |
9050120 | Swarup et al. | Jun 2015 | B2 |
9055961 | Manzo et al. | Jun 2015 | B2 |
9068628 | Solomon et al. | Jun 2015 | B2 |
9078684 | Williams | Jul 2015 | B2 |
9084623 | Gomez et al. | Jul 2015 | B2 |
9095362 | Dachs, II et al. | Aug 2015 | B2 |
9096033 | Holop et al. | Aug 2015 | B2 |
9101381 | Burbank et al. | Aug 2015 | B2 |
9113877 | Whitman et al. | Aug 2015 | B1 |
9134150 | Zhao | Sep 2015 | B2 |
9138284 | Krom et al. | Sep 2015 | B2 |
9144456 | Rosa et al. | Sep 2015 | B2 |
9198730 | Prisco et al. | Dec 2015 | B2 |
9204923 | Manzo et al. | Dec 2015 | B2 |
9226648 | Saadat et al. | Jan 2016 | B2 |
9226750 | Weir et al. | Jan 2016 | B2 |
9226761 | Burbank | Jan 2016 | B2 |
9232984 | Guthart et al. | Jan 2016 | B2 |
9241766 | Duque et al. | Jan 2016 | B2 |
9241767 | Prisco et al. | Jan 2016 | B2 |
9241769 | Larkin et al. | Jan 2016 | B2 |
9259275 | Burbank | Feb 2016 | B2 |
9259277 | Rogers et al. | Feb 2016 | B2 |
9259281 | Griffiths et al. | Feb 2016 | B2 |
9259282 | Azizian et al. | Feb 2016 | B2 |
9261172 | Solomon et al. | Feb 2016 | B2 |
9265567 | Orban, III et al. | Feb 2016 | B2 |
9265584 | Itkowitz et al. | Feb 2016 | B2 |
9283049 | Diolaiti et al. | Mar 2016 | B2 |
9301811 | Goldberg et al. | Apr 2016 | B2 |
9314307 | Richmond et al. | Apr 2016 | B2 |
9317651 | Nixon | Apr 2016 | B2 |
9345546 | Toth et al. | May 2016 | B2 |
9393017 | Flanagan et al. | Jul 2016 | B2 |
9402689 | Prisco et al. | Aug 2016 | B2 |
9417621 | Diolaiti | Aug 2016 | B2 |
9424303 | Hoffman et al. | Aug 2016 | B2 |
9433418 | Whitman et al. | Sep 2016 | B2 |
9446517 | Burns et al. | Sep 2016 | B2 |
9452020 | Griffiths et al. | Sep 2016 | B2 |
9474569 | Manzo et al. | Oct 2016 | B2 |
9480533 | Devengenzo et al. | Nov 2016 | B2 |
9503713 | Zhao et al. | Nov 2016 | B2 |
9526587 | Zhao | Dec 2016 | B2 |
9550300 | Danitz et al. | Jan 2017 | B2 |
9554859 | Nowlin et al. | Jan 2017 | B2 |
9566124 | Prisco et al. | Feb 2017 | B2 |
9579164 | Itkowitz et al. | Feb 2017 | B2 |
9585641 | Cooper et al. | Mar 2017 | B2 |
9615883 | Schena et al. | Apr 2017 | B2 |
9623563 | Nixon | Apr 2017 | B2 |
9623902 | Griffiths et al. | Apr 2017 | B2 |
9629520 | Diolaiti | Apr 2017 | B2 |
9662177 | Weir et al. | May 2017 | B2 |
9664262 | Donlon et al. | May 2017 | B2 |
9687312 | Dachs, II et al. | Jun 2017 | B2 |
9700334 | Hinman et al. | Jul 2017 | B2 |
9718190 | Larkin et al. | Aug 2017 | B2 |
9730719 | Brisson et al. | Aug 2017 | B2 |
9737199 | Pistor et al. | Aug 2017 | B2 |
9795446 | DiMaio et al. | Oct 2017 | B2 |
9797484 | Solomon et al. | Oct 2017 | B2 |
9801690 | Larkin et al. | Oct 2017 | B2 |
9814530 | Weir et al. | Nov 2017 | B2 |
9814536 | Goldberg et al. | Nov 2017 | B2 |
9814537 | Itkowitz et al. | Nov 2017 | B2 |
9820823 | Richmond et al. | Nov 2017 | B2 |
9827059 | Robinson et al. | Nov 2017 | B2 |
9830371 | Hoffman et al. | Nov 2017 | B2 |
9839481 | Blumenkranz et al. | Dec 2017 | B2 |
9839487 | Dachs, II | Dec 2017 | B2 |
9850994 | Schena | Dec 2017 | B2 |
9855102 | Blumenkranz | Jan 2018 | B2 |
9855107 | Labonville et al. | Jan 2018 | B2 |
9872737 | Nixon | Jan 2018 | B2 |
9877718 | Weir et al. | Jan 2018 | B2 |
9883920 | Blumenkranz | Feb 2018 | B2 |
9888974 | Niemeyer | Feb 2018 | B2 |
9895813 | Blumenkranz et al. | Feb 2018 | B2 |
9901408 | Larkin | Feb 2018 | B2 |
9918800 | Itkowitz et al. | Mar 2018 | B2 |
9943375 | Blumenkranz et al. | Apr 2018 | B2 |
9948852 | Lilagan et al. | Apr 2018 | B2 |
9949798 | Weir | Apr 2018 | B2 |
9949802 | Cooper | Apr 2018 | B2 |
9952107 | Blumenkranz et al. | Apr 2018 | B2 |
9956044 | Gomez et al. | May 2018 | B2 |
9980778 | Ohline et al. | May 2018 | B2 |
10008017 | Itkowitz et al. | Jun 2018 | B2 |
10028793 | Griffiths et al. | Jul 2018 | B2 |
10033308 | Chaghajerdi et al. | Jul 2018 | B2 |
10034719 | Richmond et al. | Jul 2018 | B2 |
10052167 | Au et al. | Aug 2018 | B2 |
10085811 | Weir et al. | Oct 2018 | B2 |
10092344 | Mohr et al. | Oct 2018 | B2 |
10123844 | Nowlin | Nov 2018 | B2 |
10188471 | Brisson | Jan 2019 | B2 |
10201390 | Swarup et al. | Feb 2019 | B2 |
10213202 | Flanagan et al. | Feb 2019 | B2 |
10258416 | Mintz et al. | Apr 2019 | B2 |
10278782 | Jarc et al. | May 2019 | B2 |
10278783 | Itkowitz et al. | May 2019 | B2 |
10282881 | Itkowitz et al. | May 2019 | B2 |
10335242 | Devengenzo et al. | Jul 2019 | B2 |
10405934 | Prisco et al. | Sep 2019 | B2 |
10433922 | Itkowitz et al. | Oct 2019 | B2 |
10464219 | Robinson et al. | Nov 2019 | B2 |
10485621 | Morrissette et al. | Nov 2019 | B2 |
10500004 | Hanuschik et al. | Dec 2019 | B2 |
10500005 | Weir et al. | Dec 2019 | B2 |
10500007 | Richmond et al. | Dec 2019 | B2 |
10507066 | DiMaio et al. | Dec 2019 | B2 |
10510267 | Jarc et al. | Dec 2019 | B2 |
10524871 | Liao | Jan 2020 | B2 |
10548459 | Itkowitz et al. | Feb 2020 | B2 |
10575909 | Robinson et al. | Mar 2020 | B2 |
10592529 | Hoffman et al. | Mar 2020 | B2 |
10595946 | Nixon | Mar 2020 | B2 |
10881469 | Robinson | Jan 2021 | B2 |
10881473 | Itkowitz et al. | Jan 2021 | B2 |
10898188 | Burbank | Jan 2021 | B2 |
10898189 | McDonald, II | Jan 2021 | B2 |
10905506 | Itkowitz et al. | Feb 2021 | B2 |
10912544 | Brisson et al. | Feb 2021 | B2 |
10912619 | Jarc et al. | Feb 2021 | B2 |
10918387 | Duque et al. | Feb 2021 | B2 |
10918449 | Solomon et al. | Feb 2021 | B2 |
10932873 | Griffiths et al. | Mar 2021 | B2 |
10932877 | Devengenzo et al. | Mar 2021 | B2 |
10939969 | Swarup et al. | Mar 2021 | B2 |
10939973 | DiMaio et al. | Mar 2021 | B2 |
10952801 | Miller et al. | Mar 2021 | B2 |
10965933 | Jarc | Mar 2021 | B2 |
10966742 | Rosa et al. | Apr 2021 | B2 |
10973517 | Wixey | Apr 2021 | B2 |
10973519 | Weir et al. | Apr 2021 | B2 |
10984567 | Itkowitz et al. | Apr 2021 | B2 |
10993773 | Cooper et al. | May 2021 | B2 |
10993775 | Cooper et al. | May 2021 | B2 |
11000331 | Krom et al. | May 2021 | B2 |
11013567 | Wu et al. | May 2021 | B2 |
11020138 | Ragosta | Jun 2021 | B2 |
11020191 | Diolaiti et al. | Jun 2021 | B2 |
11020193 | Wixey et al. | Jun 2021 | B2 |
11026755 | Weir et al. | Jun 2021 | B2 |
11026759 | Donlon et al. | Jun 2021 | B2 |
11040189 | Vaders et al. | Jun 2021 | B2 |
11045077 | Stern et al. | Jun 2021 | B2 |
11045274 | Dachs, II et al. | Jun 2021 | B2 |
11058501 | Tokarchuk et al. | Jul 2021 | B2 |
11076925 | DiMaio et al. | Aug 2021 | B2 |
11090119 | Burbank | Aug 2021 | B2 |
11096687 | Flanagan et al. | Aug 2021 | B2 |
11098803 | Duque et al. | Aug 2021 | B2 |
11109925 | Cooper et al. | Sep 2021 | B2 |
11116578 | Hoffman et al. | Sep 2021 | B2 |
11129683 | Steger et al. | Sep 2021 | B2 |
11135029 | Suresh et al. | Oct 2021 | B2 |
11147552 | Burbank et al. | Oct 2021 | B2 |
11147640 | Jarc et al. | Oct 2021 | B2 |
11154373 | Abbott et al. | Oct 2021 | B2 |
11154374 | Hanuschik et al. | Oct 2021 | B2 |
11160622 | Goldberg et al. | Nov 2021 | B2 |
11160625 | Wixey et al. | Nov 2021 | B2 |
11161243 | Rabindran et al. | Nov 2021 | B2 |
11166758 | Mohr et al. | Nov 2021 | B2 |
11166770 | DiMaio et al. | Nov 2021 | B2 |
11166773 | Ragosta et al. | Nov 2021 | B2 |
11173597 | Rabindran et al. | Nov 2021 | B2 |
11185378 | Weir et al. | Nov 2021 | B2 |
11191596 | Thompson et al. | Dec 2021 | B2 |
11197729 | Thompson et al. | Dec 2021 | B2 |
11213360 | Hourtash et al. | Jan 2022 | B2 |
11221863 | Azizian et al. | Jan 2022 | B2 |
11234700 | Ragosta et al. | Feb 2022 | B2 |
11241274 | Vaders et al. | Feb 2022 | B2 |
11241290 | Waterbury et al. | Feb 2022 | B2 |
11259870 | DiMaio et al. | Mar 2022 | B2 |
11259884 | Burbank | Mar 2022 | B2 |
11272993 | Gomez et al. | Mar 2022 | B2 |
11272994 | Saraliev et al. | Mar 2022 | B2 |
11291442 | Wixey et al. | Apr 2022 | B2 |
11291513 | Manzo et al. | Apr 2022 | B2 |
11576739 | Meglan | Feb 2023 | B2 |
20030007672 | Harman et al. | Jan 2003 | A1 |
20040159773 | Fein | Aug 2004 | A1 |
20070156021 | Morse et al. | Jul 2007 | A1 |
20110301447 | Park et al. | Dec 2011 | A1 |
20160117823 | Isaacs et al. | Apr 2016 | A1 |
20160210518 | Script | Jul 2016 | A1 |
20180168737 | Ren | Jun 2018 | A1 |
20190378301 | Lee | Dec 2019 | A1 |
Number | Date | Country |
---|---|---|
106204523 | Dec 2016 | CN |
2020009830 | Jan 2020 | WO |
Entry |
---|
International Search Report mailed Oct. 15, 2019 and Written Opinion completed Oct. 15, 2019 corresponding to counterpart Int'l Patent Application PCT/US2019/038869. |
International Search Report dated Oct. 15, 2019 and Written Opinion completed Oct. 15, 2019 corresponding to counterpart Int'l Patent Application PCT/US2019/038869. |
Extended European Search Report dated Mar. 22, 2022 corresponding to counterpart Patent Application EP 19829924.0. |
Number | Date | Country | |
---|---|---|---|
20230005150 A1 | Jan 2023 | US |
Number | Date | Country | |
---|---|---|---|
62693530 | Jul 2018 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17256422 | US | |
Child | 17941053 | US |