The present invention relates to devices used in endoscopic surgery and, more particularly, to endoscopic devices with field stop encoding allowing automatic identification of an endoscope that is coupled to an endoscopic video camera.
Endoscopic camera systems provide a surgeon with a view of an area of the body being worked on without the necessity of fully opening up the area, thereby allowing for less invasive surgical procedures. Generally, these devices consist of a camera head coupled to an endoscope and a camera control unit. Different endoscopes are used depending on the area of the body being worked on and the procedures being performed. Likewise, optimal camera settings are different depending on the endoscope being used, the area of the body being worked on and the procedures being performed. It would be desirable to have a system for automatically determining the endoscope being used for optimizing one or more settings for the display or processing of images that improves upon the prior art.
The present disclosure relates to an endoscopic camera system, which is usable with multiple different endoscopes. A field stop of the endoscope is encoded with detectable endoscope information for customization of the camera system based on the endoscope being used.
In an exemplary embodiment, the endoscopic camera system has an endoscope having a field stop encoded with endoscope information; a camera head coupleable to the endoscope, the camera head generating image data including the endoscope information; and a camera control unit coupleable to the camera head to receive image data from the camera head; and wherein the camera control unit is configured to analyze the endoscope information to identify the endoscope coupled to the camera head. The endoscope information may include an endoscope type. Additionally, the endoscope information may include a unique endoscope identification number.
The system may be usable with more than one endoscope and the endoscope information in each of the plurality of endoscopes may have a same start feature and a same end feature. The field stop may be encoded with endoscope information using notches or slots having varying widths. The field stop may also be encoded with endoscope information using notches or slots having varying lengths. The field stop may be encoded with endoscope information at more than one location. Optionally, the field stop has a direction of view indicator and the camera control unit infers location of the direction of view indicator from the locations of the endoscope information.
The camera control unit may automatically adjust operational settings of the camera head in accordance with the endoscope coupled to the camera head. Optionally, the camera control unit automatically adjusts the operational settings of the camera control unit in accordance with the endoscope coupled to the camera head.
In a further exemplary embodiment of the present invention, the system has a memory coupled to the camera control unit and the camera control unit obtains endoscope usage information from the memory. The camera control unit may also track endoscope usage and send endoscope usage information to the memory.
In a further exemplary embodiment, the present invention is directed to a method of communicating endoscope information from an endoscope having a field stop to a camera head, and communicating endoscope information from the camera head to a camera control unit comprising the steps of: coupling the camera head to the endoscope; generating image data with the camera head including the endoscope information encoded in the field stop; transmitting the image data from the camera head to the camera control unit; and analyzing the image data from the camera head to identify the endoscope coupled to the camera head.
The method may also have the step of validating whether the endoscope is appropriate for use with the camera head. Optionally, the method has the step of storing endoscope usage information in a storage device. Optionally, the method has the step of modifying image data processing based upon the identified endoscope. Optionally, the method has the step of modifying image data display based upon the identified endoscope. Optionally, the method has the step of modifying a light setting based upon the identified endoscope.
The method may also have the step of comparing image data characteristics of the endoscope with saved image characteristics of an identified endoscope type. The method may also have the step of encoding endoscope information in the field stop. The step of encoding endoscope information in the field stop may include manufacturing holes, slots or notches of varying widths in the field stop and the step of analyzing the image data to identify the endoscope may include scanning the image data for contrast of pixels in a known image location.
In a further exemplary embodiment, the present invention is directed to an endoscope imaging system having an endoscope with a field stop encoded with endoscope information. The system may also have a camera head with an image sensor generating image data including the endoscope information from the field stop. The system may also have a camera control unit receiving image data from the camera head, the camera control unit being configured to analyze the image data to identify the endoscope. Optionally, the endoscope information is encoded as a plurality of holes, slots or notches in the field stop. Optionally, the endoscope information is encoded in a plurality of locations in the field stop. Optionally, the encoded information is digitally removed from an image presented to a user.
These and other features are described below.
The features, aspects and advantages of the present invention will become better understood with regard to the following description, appended claims and accompanying figures wherein:
In the following description of the preferred embodiments, reference is made to the accompanying drawings which show by way of illustration specific embodiments in which the invention may be practiced. Wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts. It is to be understood that other embodiments may be utilized and structural and functional changes may be made without departing from the scope of the present invention.
With reference to
As shown in
As shown in
In an embodiment, the camera control unit 16 also has at least one network interface 42 which may be a wired interface such as Ethernet, or a wireless network connection that allows for the camera control unit to access a network. Optionally, the network interface 42 allows the camera control unit 16 to access the Internet and servers for storing and processing information.
In an embodiment, the external storage device 40 may be, for example, a flash memory storage device or a hard disk storage device, and may be connected to the camera control unit 14 through a USB connection 44 or firewire connection (not shown). In an embodiment, a program application for the camera control unit, or data relevant to a particular patient, surgeon or endoscope, is stored on the external storage device 40 and may be used to quickly configure the camera control unit for future sessions. Preferably, the camera control unit 16 can save images and video in different formats and to different places (e.g. internal memory, an external memory, or to a remote location over the Internet). Additionally, the external storage device 40 may be a server and may be connected to the camera control unit the network interface 42.
With reference to
In an embodiment of the present invention, information features 104 for specific endoscopes are placed in an inner edge of the field stop. The information features 104 may be recognized by an image processing algorithm that scans an image received from a camera head to identify the aperture of the field stop, using the contrast of the pixels and their approximate known location.
A ring of pixels just outside the field stop aperture (in the mask) is extracted. The ring of pixels is then analyzed to locate the orientation notch as well as the information features 104. Optionally, in an embodiment, the information features 104 have start features 106 and end features 108 that are the same for each endoscope so that an algorithm can determine where endoscope data 110 begins and ends.
In an embodiment, the endoscope data 110 is encoded between the start and stop features 106, 108. The endoscope data may be encoded using, for example, a narrow notch or a wide notch constituting “bits” of data. A “bit” may also be, for example, a short notch or a long notch. Although the present specification refers to use of notches, it will be understood that slots, holes or other detectable shapes may be used to encode information into the field stop. In additional embodiments, the encoding may be done with other optically detectable means, such as, for example, a colored marking or a fluorescent phosphor.
The bits are interpreted as a 1 or a 0 as effectively a binary code. In an embodiment, the endoscope data 110 is between about 4 and about 24 “bits” between the start and stop features 106, 108. In a more preferred embodiment, the endoscope data 110 is about 8 “bits” between the start and stop feature 106, 108. Although described herein as a binary system, it is also possible that shapes may be used to encode more than two values. For example, a notch may have three or four different lengths or widths and may thereby be used to encode four or five different values. This allows fewer notches to encode more information.
In a preferred embodiment, two or more copies of the information features 104 are provided at different positions on the field stop 100 so that endoscope information may still be obtained even if one set of information features 104 are oriented off an edge of a video image. Moreover, if the locations of the multiple copies of the information features 104 is known, then that information may be used to infer the location of the orientation notch 102.
A field stop aperture according to a first exemplary embodiment of the present invention is shown in
A field stop aperture according to a second exemplary embodiment of the present invention is shown in
A field stop aperture according to a third exemplary embodiment of the present invention is shown in
A field stop aperture according to a fourth exemplary embodiment of the present invention is shown in
A field stop aperture according to a fifth exemplary embodiment of the present invention is shown in
As shown in
Optionally, the captured image is converted from RGB to gray scale. Removing color simplifies later processing. Optionally, a smoothing filter is applied to remove noise. The smoothing filter cutoff is chosen so as to avoid damaging detail of the identification information while removing unwanted noise. A Gaussian filter can be implemented as a separable low pass with sharp cutoff to remove unwanted noise. Image contrast may be increased to accentuate the black and white differences in the image.
A Hough Circle Transform is performed to locate the center (xcenter, ycenter) and radius (r) of the field stop circle, box 216. A one dimensional line of pixels is extracted from the circumference of the concentric circle with radius r+x where x is chosen so that the line fully intersects the code features in the field stop image, box 218. The one dimensional line of pixels is then processed using a standard linear barcode decoding algorithm, box 220. Standard barcode decoding algorithms may be found in, for example, T. Pavlidis, J. Swartz, and Y. P. Wang, “Fundamentals of bar code information theory,” IEEE Computer, pp. 74-86, April 1990, E. Joseph and T. Pavlidis, “Waveform recognition with applications to bar codes,” in Proc. IEEE SMC 1991 Con&, 1991, pp. 129-134, R. Muniz, L. Junco and A. Otero, “A robust software barcode reader using the Hough transform,” International Conference on Information Intelligence and Systems, Bethesda, Md., USA, pp. 313-319, November 1999, the entire contents of each of which are incorporated herein by reference in their entirety. The endoscope information is then decoded, box 222. The endoscope information may include, for example, a model number, a serial number, a field of view and an angle of view.
Once endoscope information is obtained, the endoscope information can be used to reference image processing, light settings, or display settings in a memory in the camera control unit or in a memory accessible by the camera control unit. The image processing, light settings, or display settings may be automatically implemented by the camera control unit. Also, for example, the image for a particular endoscope may be automatically magnified to a predetermined magnification. Also, for example, the operational settings of the camera head may be automatically adjusted for the specific endoscope. Also, for example, optimal light source settings may be automatically selected for the particular endoscope. Endoscope information can also be used to prevent the use of incompatible endoscopes. For example, if a scope is used that is not validated for use with a given camera system, the camera control unit 16 may cause a warning to be displayed to a user on the display 20.
Additionally, the camera system can record and present usage statistics on which endoscopes are used, with which surgeries, and for how long. This data may be stored in a storage accessible to different camera control units so that the data may be updated when a particular endoscope is used in different locations with different camera heads and camera control units. This data can help surgeons keep records of which endoscopes were used in each procedure. Reliability of endoscopes can be more accurately tracked. Manufacturing, sales, and design may benefit from the data to make the most often used products, rather than relying on sales history.
Endoscope usage information is advantageous because usage time may affect the reliability and health of an endoscope. Once an endoscope is identified, target calibration can be used to gauge the health of the endoscope. For example, optimal light and exposure settings for a given type of endoscope may be compared with light and exposure settings used with the actual endoscope to determine whether the endoscope is functioning properly. The data may also be used to determine the health of different camera heads. For example, if the same endoscope is used with different camera heads and the usage requires very different light and exposure settings, then investigation of the camera heads may be warranted.
The information features 104 may be digitally removed from the image prior to presenting the image to a user and prior to archiving of the image. Removal of the encoded information features 104 from the image may be advantageous to avoid confusing a user.
The present invention is advantageous, because while the method requires physical modification of the field stop, the invention does not require the addition of any components nor direct communication of identification information between an endoscope and a camera head or camera control unit. Moreover, each endoscope may be encoded with a specific identifier so that information about that specific endoscope may be tracked.
There is disclosed in the above description and the drawings, a surgical imaging system and method for processing surgical images that fully and effectively overcomes the disadvantages associated with the prior art. However, it will be apparent that variations and modifications of the disclosed embodiments may be made without departing from the principles of the invention. The presentation of the preferred embodiments herein is offered by way of example only and not limitation, with a true scope and spirit of the invention being indicated by the following claims.
Any element in a claim that does not explicitly state “means” for performing a specified function or “step” for performing a specified function, should not be interpreted as a “means” or “step” clause as specified in 35 U.S.C. § 112.
Number | Name | Date | Kind |
---|---|---|---|
5233416 | Inoue | Aug 1993 | A |
5392067 | Konno | Feb 1995 | A |
5896166 | D'Alfonso et al. | Apr 1999 | A |
6092722 | Heinrichs et al. | Jul 2000 | A |
6293911 | Imaizumi | Sep 2001 | B1 |
6638212 | Oshima | Oct 2003 | B1 |
6712756 | Kura et al. | Mar 2004 | B1 |
6770027 | Banik et al. | Aug 2004 | B2 |
7282026 | Ogawa | Oct 2007 | B2 |
7289139 | Amling et al. | Oct 2007 | B2 |
8266129 | Shimizu et al. | Sep 2012 | B2 |
8373748 | Pang et al. | Feb 2013 | B2 |
8537210 | Omori | Sep 2013 | B2 |
9060674 | Amling | Jun 2015 | B2 |
9438897 | Barreto et al. | Sep 2016 | B2 |
9603512 | Hoeg et al. | Mar 2017 | B2 |
10350001 | Collins | Jul 2019 | B2 |
20040030221 | Ogawa | Feb 2004 | A1 |
20040186351 | Imaizumi | Sep 2004 | A1 |
20090124855 | Urakawa | May 2009 | A1 |
20100141744 | Amling et al. | Jun 2010 | A1 |
20150309284 | Kagawa | Oct 2015 | A1 |
20170007095 | Kutsuma et al. | Jan 2017 | A1 |
20180242828 | Shiga | Aug 2018 | A1 |
20190053693 | Koiso et al. | Feb 2019 | A1 |
Number | Date | Country |
---|---|---|
S56158632 | Dec 1981 | JP |
6161841 | Jul 2017 | JP |
2016040131 | Mar 2016 | WO |
2017072950 | May 2017 | WO |
Entry |
---|
“Fundamentals of Bar Code Information Theory”, by T. Pavlidis, J. Swartz and Y.P. Wang, IEEE Computer, Apr. 1990, pp. 74-86. |
“Waveform Recognition With Application to Bar Codes”, by E. Joseph and T. Pavlidis, Proc. IEEE SMC Aug.1991, pp. 129-134. |
“A Robust Software Barcode Reader Using the Hough Transform”, by Ruben Muniz, et al., IEEE, Sep. 1999, pp. 313-319. |
Annex to form PCT/ISA/206 Communication Relating to the Results of the Partial International Search—for PCT Application No. PCT/US2018/048322, dated Dec. 12, 2018, 13 pages. |
International Search Report and Written Opinion, dated Apr. 10, 2019. |
Number | Date | Country | |
---|---|---|---|
20190069760 A1 | Mar 2019 | US |