The present subject matter relates to display systems. More particularly, the present subject matter relates to stereoscopic display systems and methods for displaying medical or surgical data and information in a surgical microscope or other display device designed for interactive use during patient care.
First generation intrasurgical optical coherence tomography (OCT) systems displayed OCT data onto a separate computer monitor, requiring surgeons to look away from the surgical microscope. In order to provide real-time OCT feedback without requiring surgeons to look away during surgeries, recent prototype research and commercial intrasurgical OCT systems have integrated heads-up display (HUD) systems into the surgical microscopes to allow the surgeons to access the OCT data and the surgical field through the oculars concurrently. However, current intrasurgical OCT systems with a HUD are only capable of displaying through one ocular or eyepiece, thus limiting the surgeon's depth perception of OCT volumes. Stereoscopy is an effective technology to dramatically increase depth perception by presenting an image from slightly different angles to each eye. Conventional stereoscopic HUD use a pair of micro displays which require bulky optics. Several approaches for HUDs are reported to use only one micro display at the expense of image brightness or increased footprint. However, these techniques for HUD are not suitable to be integrated into microscopes. Other display devices intended for medical, industrial, or entertainment use may also benefit from a more compact and efficient optical design. For example, head-mounted displays, enhanced/augmented reality displays, and other immersive display technologies require compact and light weight designs.
For at least the aforementioned reasons, there is a need for improved display systems for use in medical and other applications.
Disclosed herein are stereoscopic display systems and methods for displaying surgical data and information in a surgical microscope. In accordance with embodiments of the present disclosure, a compact, microscope-integrated stereoscopic HUD is disclosed which uses spatial multiplexing to project stereo views into both oculars simultaneously using a single micro-display. The HUD is supported by real-time GPU-enabled stereoscopic OCT image processing. In an example, a stereoscopic HUD having one micro display and only three optical elements is provided for a microscope-integrated OCT system.
According to an aspect, a stereoscopic display system includes first and second eyepieces. The system includes a display having first and second display portions, configured to display first images in the first display portion, and configured to display second images in the second display portion. The first image and the second image are projected along a first pathway and a second pathway leading to the first and second eyepiece, respectively. The system includes a first beamsplitter or other optical element positioned so that the first eyepiece only receives the first images, and positioned to receive and reflect both first and second light into the first microscope optical path but the received second light is rejected by the aperture inside the first eyepiece or any suitable aperture. The system includes a second beamsplitter or other optical element positioned so that the second eyepiece only receives the second images, and positioned to receive and reflect both first and second light into the second microscope optical path but the received first light is rejected by the aperture inside the second eyepiece or any suitable aperture.
The foregoing aspects and other features of the present subject matter are explained in the following description, taken in connection with the accompanying drawings, wherein:
For the purposes of promoting an understanding of the principles of the present disclosure, reference will now be made to various embodiments and specific language will be used to describe the same. It will nevertheless be understood that no limitation of the scope of the disclosure is thereby intended, such alteration and further modifications of the disclosure as illustrated herein, being contemplated as would normally occur to one skilled in the art to which the disclosure relates.
Articles “a” and “an” are used herein to refer to one or to more than one (i.e. at least one) of the grammatical object of the article. By way of example, “an element” means at least one element and can include more than one element.
In this disclosure, “comprises,” “comprising,” “containing” and “having” and the like can have the meaning ascribed to them in U.S. Patent law and can mean “includes,” “including,” and the like; “consisting essentially of” or “consists essentially” likewise has the meaning ascribed in U.S. Patent law and the term is open-ended, allowing for the presence of more than that which is recited so long as basic or novel characteristics of that which is recited is not changed by the presence of more than that which is recited, but excludes prior art embodiments.
Ranges provided herein are understood to be shorthand for all of the values within the range. For example, a range of 1 to 50 is understood to include any number, combination of numbers, or sub-range from the group consisting 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, or 50.
Unless specifically stated or obvious from context, as used herein, the term “about” is understood as within a range of normal tolerance in the art, for example within 2 standard deviations of the mean. The term “about” can be understood as within 10%, 9%, 8%, 7%, 6%, 5%, 4%, 3%, 2%, 1%, 0.5%, 0.1%, 0.05%, or 0.01% of the stated value. Unless otherwise clear from context, all numerical values provided herein are modified by the term “about.”
Unless otherwise defined, all technical terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs.
In accordance with embodiments of the present disclosure, a compact, stereoscopic display system is provided for displaying stereoscopic images utilizing a single micro display for Microscope Integrated OCT (MIOCT). In an example, a MIOCT sample arm may couple the OCT and surgical optical axis and may enable concurrent imaging with both systems. Further, it has been demonstrated that high speed Swept-Source MIOCT (SS-MIOCT) may be used for real-time volumetric imaging of dynamic surgical maneuvers.
The controller 116 can include one or more processors 124 and memory 126 for implementing the function of controlling the display 118 in accordance with embodiments disclosed herein. In an example, the controller 116 may be part of a computing device, such as a desktop computer, a laptop computer, a tablet computer, smartphone, or the like configured with equipment for interfacing with the display 118. The controller 116 may be operably connected to a user interface such that a user may input commands and be presented with data about operation of the system 100, including operation of the display 118. Alternative to the processor(s) 124 and the memory 126, the controller 116 may include any suitable hardware, software, firmware, or combinations thereof for implementing the functionality described herein.
The surgical microscope 106 may be any suitable type of microscope. In this example, the surgical microscope 106 is configured to operate together with an objective lens 130 and contract lens 132 for viewing of a patient's eye 134 during surgery.
In the example of
The HUD 104 may be interfaced with an SS-MIOCT system via USB, HDMI, or the like. The HUD 104 can enable real-time acquisition, processing, and stereoscopic HUD display of volumetric images in real time (up to 10 volumes/sec). GPU-based software may be run by the controller 116 to implement functionality disclosed herein. For example, the controller 116 can enable simultaneous rendering of two separate views of each volume which can be displayed in the separate display portions of the display 116. The two renderings can be rotated relative to each other to create the stereoscopic effect. The volume projected to the left eye can displayed in the right ½ of the display 118 which can only be seen by the left eye. The OCT volume intended for the right eye can be rotated 10° (or any other suitable angle) from that intended for the left eye and displayed in the left ½ of the display 118 which can only be seen by the right eye. The positions of the two images can be calibrated experimentally so that they may be fused by the observer. Furthermore, the overall direction of view of both stereoscopic OCT volumes may be rotatable by a control device (e.g., a foot trackball available from Infogrip, Inc.; Ventura, Calif. or other common computer input device) by the surgeon or other user while operating in real time.
In experiments, the resolution of the microscope with and without the HUD were measured by imaging USAF 1951 resolution bars. Fields of view were measured by imaging a grid paper. The measured resolution and field of view of the surgical field with and without the HUD were both 12.40 μm and 270 mm2 indicating that the HUD did not affect the original field of view of the surgical microscope. Stereoscopic volumetric renders of simulated retinal surgery were acquired by SS-MIOCT and displayed via the HUD. The images used to create the stereoscopic effect inside the HUD oculars are shown in
In accordance with embodiments of the present disclosure, any suitable type of display may be used. For example, a micro display (SXGA060, resolution: 1280×1024, viewing area: 11.941×9.560 mm, pixel pitch: 9.3 μm and contrast ratio: >10000:1, Yunnan OLiGHTEK Opto-Electronic Technology Co., Ltd.; China) may be used.
An objective lens 136 may be used to focus the display 118. The lens may be any suitable type of lens, such as a triplet lens. The triple lens 136 may have a focal length of 50 mm so that the image is magnified enough to show half of the screen to each eye. Distortion and stereoscopic effect may also be characterized carefully. The rotation angle between two volumes intended to different eyes can be suitably optimized.
In accordance with embodiments of the present disclosure, a camera connected with a camera lens may be attached to a surgical microscope's eyepiece with a 3D printed mount for capturing images on the microscope's surgical plane and the HUD's display plane. The camera may be any suitable camera, such as a camera provided by Point Grey Research, http://www.ptgrey.com. The camera lens (e.g., 8.5 mm) may be any suitable lens, such as a lens provided by Edmund Optics Inc., of Barrington, N.J. As a further example,
In order to characterize the effects of the addition of the HUD on the optical path of the operating microscope, resolution, field of view (FOV), and distortion of the surgical microscope with and without the HUD was measured with a USAF resolution test target, a caliper, and a grid chart respectively in the surgical field at 10× total magnification.
Stereopsis is one of the strongest depth cues where human brains reconstruct distance according to lateral displacement between two points perceived by left and right eyes. As two objects like square 700 and square 702 in
Some objects on the display can only be seen by the left eye while other objects can only be seen by the right eye. Through vergence, both eyes can be fixated at point A or point B which appear to be above or beneath the display.
The grid charts placed in the surgical field of the microscope with or without the HUD mounted were imaged through microscope oculars (see
The distortion of the HUD's optical system was also characterized by the same grid chart pattern (see
Table 1 below shows a comparison of resolution, FOV, and distortion between microscope with HUD and without HUD.
Stereoscopic, 4D Microscope Integrated OCT (4DMIOCT) volumetric renders of human retinal surgery have been acquired and displayed via the HUD (see
In accordance with embodiments of the present disclosure, a compact stereoscopic display system may also be incorporated into a head-mounted display unit or any suitable 3D viewing devices intended for medical, industrial, or entertainment use. Whether used with a microscope-integraded HUD, head-mounted display, or any suitable 3D viewing devices, the display system may also include user head movement tracking equipment configured to track movement of a user. The system may include a controller operably connected to the display and configured to alter images displayed by one display portion and another display portion based on the tracked movement. For example,
With continuing reference to
The various techniques described herein may be implemented with hardware or software or, where appropriate, with a combination of both. Thus, the methods and apparatus of the disclosed embodiments, or certain aspects or portions thereof, may take the form of program code (i.e., instructions) embodied in tangible media, such as floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the presently disclosed subject matter. In the case of program code execution on programmable computers, the computer will generally include a processor, a storage medium readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device and at least one output device. One or more programs may be implemented in a high level procedural or object oriented programming language to communicate with a computer system. However, the program(s) can be implemented in assembly or machine language, if desired. In any case, the language may be a compiled or interpreted language, and combined with hardware implementations.
The described methods and apparatus may also be embodied in the form of program code that is transmitted over some transmission medium, such as over electrical wiring or cabling, through fiber optics, or via any other form of transmission, wherein, when the program code is received and loaded into and executed by a machine, such as an EPROM, a gate array, a programmable logic device (PLD), a client computer, a video recorder or the like, the machine becomes an apparatus for practicing the presently disclosed subject matter. When implemented on a general-purpose processor, the program code combines with the processor to provide a unique apparatus that operates to perform the processing of the presently disclosed subject matter.
Features from one embodiment or aspect may be combined with features from any other embodiment or aspect in any appropriate combination. For example, any individual or collective features of method aspects or embodiments may be applied to apparatus, system, product, or component aspects of embodiments and vice versa.
While the embodiments have been described in connection with the various embodiments of the various figures, it is to be understood that other similar embodiments may be used or modifications and additions may be made to the described embodiment for performing the same function without deviating therefrom. Therefore, the disclosed embodiments should not be limited to any single embodiment, but rather should be construed in breadth and scope in accordance with the appended claims. One skilled in the art will readily appreciate that the present subject matter is well adapted to carry out the objects and obtain the ends and advantages mentioned, as well as those inherent therein. The present examples along with the methods described herein are presently representative of various embodiments, are exemplary, and are not intended as limitations on the scope of the present subject matter. Changes therein and other uses will occur to those skilled in the art which are encompassed within the spirit of the present subject matter as defined by the scope of the claims.
This is a 371 national stage patent application, which claims priority to PCT International Patent Application No. PCT/US2016/016830, filed Feb. 5, 2016, and titled STEREOSCOPIC DISPLAY SYSTEMS AND METHODS FOR DISPLAYING SURGICAL DATA AND INFORMATION IN A SURGICAL MICROSCOPE, which claims the benefit of U.S. Provisional Patent Application No. 62/112,698, filed Feb. 6, 2015 and titled STEREOSCOPIC HEADS-UP DISPLAYING SURGICAL DATA IN A SURGICAL MICROSCOPE AND METHODS OF USE, the disclosures of which are incorporated herein by reference in their entireties.
The technology disclosed herein was made in part with government support under Federal Grant No. EY023039 awarded by the National Institutes of Health (NIH). The United States government has certain rights in the technology.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2016/016830 | 2/5/2016 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2016/127088 | 8/11/2016 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5040229 | Lee et al. | Aug 1991 | A |
5483364 | Ishimoto | Jan 1996 | A |
5493109 | Wei et al. | Feb 1996 | A |
5715081 | Chastang et al. | Feb 1998 | A |
5963301 | Volk | Oct 1999 | A |
6671581 | Niemeyer et al. | Dec 2003 | B2 |
7791794 | Reimer et al. | Sep 2010 | B2 |
7839494 | Reimer et al. | Nov 2010 | B2 |
7907166 | Lamprecht et al. | Mar 2011 | B2 |
9207638 | Dubois et al. | Dec 2015 | B2 |
20010031078 | Doane | Oct 2001 | A1 |
20020099290 | Haddad | Jul 2002 | A1 |
20050270486 | Teiwes et al. | Dec 2005 | A1 |
20060050991 | Jerebko et al. | Mar 2006 | A1 |
20070086647 | Grady | Apr 2007 | A1 |
20070299309 | Seibel et al. | Dec 2007 | A1 |
20080002183 | Yatagai et al. | Jan 2008 | A1 |
20080019587 | Wilensky et al. | Jan 2008 | A1 |
20080030497 | Hu et al. | Feb 2008 | A1 |
20080058704 | Hee et al. | Mar 2008 | A1 |
20090018393 | Dick et al. | Jan 2009 | A1 |
20090060332 | Knapp | Mar 2009 | A1 |
20090131921 | Kurtz et al. | May 2009 | A1 |
20090192523 | Larkin et al. | Jul 2009 | A1 |
20090225407 | Nakayama et al. | Sep 2009 | A1 |
20090244485 | Walsh et al. | Oct 2009 | A1 |
20090257065 | Hauger et al. | Oct 2009 | A1 |
20090287223 | Pua et al. | Nov 2009 | A1 |
20100202677 | Imamura et al. | Aug 2010 | A1 |
20100228123 | Brennan et al. | Sep 2010 | A1 |
20100331858 | Simaan et al. | Dec 2010 | A1 |
20110032533 | Izatt et al. | Feb 2011 | A1 |
20110043757 | Everett et al. | Feb 2011 | A1 |
20110122487 | Perelman et al. | May 2011 | A1 |
20120092615 | Izatt et al. | Apr 2012 | A1 |
20120184846 | Izatt et al. | Jul 2012 | A1 |
20120307205 | Zhou et al. | Dec 2012 | A1 |
20130010259 | Carnevale | Jan 2013 | A1 |
20130016319 | Vohnsen et al. | Jan 2013 | A1 |
20130135584 | Alasaarela et al. | May 2013 | A1 |
20130188140 | Bagherinia et al. | Jul 2013 | A1 |
20130245375 | DiMaio et al. | Sep 2013 | A1 |
20130293838 | Makihira et al. | Nov 2013 | A1 |
20140009741 | Levien et al. | Jan 2014 | A1 |
20140139916 | Doi | May 2014 | A1 |
20140142591 | Alvarez et al. | May 2014 | A1 |
20140247425 | Hammer et al. | Sep 2014 | A1 |
20140275760 | Lee et al. | Sep 2014 | A1 |
20140307315 | Bohn | Oct 2014 | A1 |
20140368907 | Minami | Dec 2014 | A1 |
20150173846 | Schneider | Jun 2015 | A1 |
Number | Date | Country |
---|---|---|
1647370 | Apr 2006 | EP |
1020130000023 | Jan 2013 | KR |
2005106786 | Nov 2005 | WO |
2012100030 | Jul 2012 | WO |
2012109301 | Aug 2012 | WO |
2013008033 | Jan 2013 | WO |
2013180773 | Dec 2013 | WO |
2014068058 | May 2014 | WO |
Entry |
---|
Notice of Allowance received in U.S. Appl. No. 14/337,215 dated Dec. 23, 2015. |
Notice of Allowance received in U.S. Appl. No. 15/049,103 dated Oct. 10, 2016. |
Office Action received in U.S. Appl. No. 14/337,215 dated May 12, 2015. |
Office Action received in U.S. Appl. No. 14/337,215 dated Nov. 5, 2014. |
Office Action received in U.S. Appl. No. 15/049,103 dated Jul. 5, 2016. |
Otsu, A Threshold Selection Method from Gray-Level Histograms, IEEE Transactions on Systems, Man, and Cybernetics, vol. SMC-9, No. 1, Jan. 1979, U.S. |
PCT International Search Report for PCT International Application No. PCT/US15/13870. |
PCT International Written Opinion for PCT International Application No. PCT/US15/13870. |
PCT Preliminary Report on Patentability and Written Opinion dated May 8, 2014 for related PCT International Application No. PCT/US2014/013243. |
PCT Search Report and Written Opinion dated Jul. 13, 2012 for related application PCT/US2012/021839; iFD Jan. 19, 2012. |
Perstein, M., Algorithms, Control Data Corp., Palo Alto, California. |
Pircher, Michael et al., Simultaneous SLO/OCT Imaging of the Human Retina with Axial Eye Motion Correction, Optics Express, vol. 15, No. 25, Dec. 4, 2007. |
Related application PCT/US2012/021839 filed Jan. 19, 2012 entitled System Enhancements for Ophthalmic Surgical Microscope Mounted Optical Coherence Tomography, not yet published. |
Schulze, Jürgen P. et al.; “Visualization of Three-Dimensional Ultra-High Resolution OCT in Virtual Reality” Ophthalmology Department, Lariboisière Hospital, APHP, Paris, France. |
SDI/BIOM: Still the Standard in Wide-Angle Viewing for All Microscope Models!, Insight Instruments, Inc., Stuart, Florida. |
Shen, Liangbo et al.; “Novel Microscope-Integrated Stereoscopic Heads-up Display for Intrasurgical OCT in Ophthalmic Surgery”, The Association for Research in Vision and Ophthalmology; Jun. 2015, vol. 56, 3514. |
Shi et al., Normalized Cuts and Image Segmentation, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, No. 8, Aug. 2000, U.S. |
Shi, Minyan et al.; “A Stereo-Fluoroscopic Image-Guided Robotic Biopsy Scheme” IEEE Transactions on Control Systems Technology, vol. 10, No. 3, May 2002, pp. 309-317. |
Stephanie J. Chiu, Cynthia A. Toth, Catherine Bowes Rickman, Joseph A. Izatt, and Sina Farsiu, “Automatic Segmentation of Cloaed-contour Features in Ophthalmic Images Using Graph Theory and Dynamic Programming”, Published Apr. 26, 2012, Optical Society of America. |
Takeda et al., Kernel Regression for Image Processing and Reconstruction, IEEE Transactions on Image Processing, vol. 16, No. 2, Feb. 2007, U.S. |
The Age-Related Eye Disease Study Research Group, The Age-Related Eye Disease Study System for Classifying Age-Related Macular Degeneration From Stereoscopic Color Fundus Photographs: The Age-Related Eye Disease Study Report No. 6, Elsevier Service Inc., vol. 132, No. 5, 2001, U.S. |
Tolliver et al., Unassisted Segmentation of Multiple Retinal Layers via Spectral Rounding, Presented in ARVO 2008 Annual Meeting, Fort Lauderdale, Florida, U.S., Apr. 2008. |
U.S. Final Office Action for U.S. Appl. No. 14/337,215, dated May 12, 2015. |
U.S. Non-Final Office Action for U.S. Appl. No. 13/010,448, dated Jan. 2, 2014. |
U.S. Non-Final Office Action for U.S. Appl. No. 14/337,215, dated Nov. 5, 2014. |
U.S. Non-Final Office Action for U.S. Appl. No. 15/049,103, dated Jul. 5, 2016. |
U.S. Notice of Allowance for U.S. Appl. No. 1301448, dated May 13, 2013. |
U.S. Notice of Allowance for U.S. Appl. No. 14/337,215, dated Jan. 11, 2016. |
U.S. Notice of Allowance for U.S. Appl. No. 15/049,103, dated Oct. 24, 2016. |
U.S. Office Action Response to Non-Final Office Action for U.S. Appl. No. 14/337,215, dated Apr. 6, 2015. |
U.S. Office Action Response to Non-Final Office Action for U.S. Appl. No. 15/049,103, dated Oct. 6, 2016. |
U.S. Response to Non-Final Office Action for U.S. Appl. No. 13/010,448, dated Mar. 31, 2014. |
Viehland, Christian et al.; “Enhanced volumetric visualization for real time 4D intraoperative ophthalmic swept-source OCT” Biomedical Optics Express 1815, May 1, 2016 | vol. 7, No. 5 |. |
Warshall, A Theorem on Boolean Matrices, Computer Associates, Inc., Woburn, Massachusetts, U.S. |
Wieser, Wolfgang et al., Multi-Megahertz OCT: High Quality 3D Imaging at 20 Million A-Scans and 4_5 GVoxels Per Second, Optics Express, vol. 18, No. 14, Jun. 30, 2010. |
Witte, S., Plaw;;ka, A., Ridder, M. C., van Berge, L., Mansvelder, H. D., & Groot, M. L. (2012). Short-coherence off-axis holographic phase microscopy of live cell dynamics. Biomedical Optics Express, 3(9), 2184-2189. http://doi.org/10.1364/BOE.3.002184. |
Yazdanpanah et al., Segmentation of Intra-Retinal Layers from Optical Coherence Tomography Images Using an Active Contour Approach, IEEE, 2010, U.S. |
International Preliminary Report on Patentability issued in PCT Application No. PCT/US2016/016900 dated Aug. 8, 2017. |
“American National Standard for Safe Use of Lasers” American National Standards Institute, Inc. Mar. 16, 2007. |
Aschke et a l., “Augmented Reality in Operating Microscopes for Neurosurgical Interventions.” IEEE, Mar. 22, 2003, pp. 652-654 (Mar. 22, 2003), p. 652, col. 2; p. 653, col. 2—p. 653, col. 1; Fig 3 [online]. |
Goncharov, Alexander V. et al.; “Wide-field schematic eye models with gradient-index lens” J. Opt. Soc. Am. A., vol. 24, No. 8/Aug. 2007, pp. 2157-2174. |
Bellman, On a Routing Problem, Richard Bellman, vol. XVI, No. 1, The RAND Corporation, pp. 87-90, 1957, U.S. |
Bichlmeier, Christoph et al.; “The Tangible Virtual Mirror: New Visualization Paradigm for Navigated Surgery” Chair for Computer Aided Medical Procedures (CAMP), TU Munich, Germany. |
Dabov, Kostadin et al., Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering, IEEE Transactions on Image Processing, vol. 16, No. 8, Aug. 2007. |
Dhalla, Al-Hafeez et al., Complex Conjugate Resolved Heterodyne Swept Source Optical Coherence Tomograph) Using Coherenece Revival, Biomedical Optics Express, vol. 3, No. 3, Feb. 24, 2012. |
Dijkstra, A Note on Two Problems in Connexion with Graphs, Cambridge University Press, 1897, vol. 13, p. 26-8, U.K. |
International Search Report and Written Opinion dated Aug. 16, 2016 from International Application No. PCT/US16/28862. |
International Search Report and Written Opinion dated Aug. 12, 2016 from International Application No. PCT/US16/3105. |
Elias, P., et al., A Note on the Maximum Flow Through a Network, IRE Transactions on Information Theory, 1956, pp. 117-119. |
Fabritius et al., Automated Segmentation of the Macular by Optical Coherence Tomography, Optics Express, vol. 17, No. 18, Aug. 31, 2009, US. |
Farsiu et al., Fast Detection and Segmentation of Drusen in Retinal Optical Coherence Tomography Images, Ophthalmic Technologies XVIII, Proc. of SPIE vol. 6844, 2008, U.S. |
Ferguson, R. Deniel et aL, Tracking Optical Coherence Tomography, Optics Letters, vol. 29, No. 18, Sep. 15, 2004. |
Fernandez et al., Automated Detection of Retinal Layer Structures on Optical Coherence Tomography Images, Optics Express, vol. 13, No. 25, Dec. 12, 2005, U.S. |
Final Rejection received in U.S. Appl. No. 13/353,612 dated May 10, 2017 (eleven (11) pages). |
Garvin et al., Automated 3-D Intraretinal Layer Segmentation of Macular Spectral-Domain Optical Coherence Tomography Images, IEEE Transactions on Medical Imaging, vol. 28, No. 9, Sep. 2009, U.S. |
Garvin, M.K., et al., Intraretinal Layer Segmentation-Search, IEEE, 0278-0062, pp. 1495-1505, 2008. |
Graph cuts segmentation—medical images, Jacquot et al., I EEE,978-0-7695-3122, 2008, pp. 631-635. |
Haeker et al., Automated Segmentation of Intraretinal Layers from Macular Optical Coherence Tomography Images, Society of Photo-Optical Instrumentation Engineers, 2007, U.S. |
Hendargo, Hansford C. et al., Automated Non-Rigid Registration and Mosaicing for Robust Imaging of Distinct Retinal Capillary Beds Using Speckle Variance Optical Coherence Tomography, Biomedical Optics Express, vol. 4, No. 6, May 7, 2013. |
International Search Report and Written Opinion issued in PCT Application No. PCT/US2016/016900 dated May 5, 2016. |
International Preliminary Report on Patentability dated Jul. 23, 2013 for corresponding application PCT/US2012/021839 (filed Jan. 19, 2012). |
Jung, Woonggyu et al.; “Handheld Optical Coherence Tomography Scanner for Primary Care Diagnostics” IEEE Transactions on Biomedical Engineering, vol. 58, No. 3, Mar. 2011. pp. 741-744. |
International Search Report and Written Opinion dated Jan. 5, 2017 from International Application No. PCT/US16/51360. |
International Search Report and Written Opinion dated May 19, 2016 from related International Application No. PCT/US16/16830. |
Intra-retinal segmentation—images, Mishra et al., Optic Express 23719, Published Dec. 11, 2009, pp. 1-10. |
Ishikawa et al., Macular Segmentation with Optical Coherence Tomography, Investigative Ophthalmology & Visual Science, vol. 46, No. 6, Association for Research in Vision and Ophthalmology, Jun. 2005, U.S. |
Jaquot, J.Z., et al.,Graph Cuts Segmentation—Medical Images, IEEE, pp. 631-635. 2008. |
Ji, Na et al., Adaptive Optics Via Pupil Segmentation for High-Resolution Imaging in Biological Tissues, Nature Methods, vol. 7, No. 2, Feb. 2010. |
Kavraki, Lydia E. et al.; “Probabilistic Roadmaps for Path Planning in High-Dimensional Configuration Spaces” IEEE Transactions on Robotics and Automation, vol. 12, No. 4, Aug. 1996. |
Koreeda, Y. et al.; “Development and testing of an endoscopic pseudo-viewpoint alternating system” Int J CARS, Jun. 21, 2014. |
Kozak, Igor et al.; “Virtual reality simulator for vitreoretinal surgery using integrated OCT data” Clinical Ophthalmology 2014:8 pp. 669-672. |
LaRocca et al., “Optimization of confocal scanning laser opthalmoscope design.” Journal of Biomedical Optics. Jul. 2013 {Jul. 20, 2013). pp. 076015-1-076015-2, 076015-8 [online]. |
LaRocca, Francesco et al.; “Handheld simultaneous scanning laser ophthalmoscopy and optical coherence tomography system” Biomedical Optics Express, Nov. 1, 2013 | vol. 4, No. 11, pp. 2307-2321. |
LaValle, Steven M. et al.; “Rapidly-Exploring Random Trees: A New Tool for Path Planning” Department of Computer Science, Iowa State University. |
Lee, K., et al. Segmentation of the Optic Disc in 3-D OCT Scans of the Optic Nerve Head. IEEE Transactions on Imaging, vol. 29(1): pp. 159-168, Jan. 2010. |
Liao, Wen-Hung et al., Robust Pupil Detection for Gaze-Based User Interface, EGIHMI, Feb. 7, 2010. |
Kelly, John P. et al.; “Imaging a Child's Fundus Without Dilation Using a Handheld Confocal Scanning Laser Ophthalmoscope” Arch Ophthalmol/vol. 121, Mar. 2003, pp. 391-396. |
Lu et al., Automated Layer Segmentation of Optical Coherence Tomography Images, IEEE Transactions on Biomedical Engineering, vol. 57, No. 10, Oct. 2010, U.S. |
Lu, Chen D.; “Handheld ultrahigh speed swept source optical coherence tomography instrument using a MEMS scanning mirror” Biomedical Optics Express, Jan. 1, 2014 | vol. 5, No. 1, pp. 293-311. |
Lujan, Brandon J., et al., Revealing Henle's Fiber Layer Using Spectral Domain Optical Coherence Tomography, Investigative Ophthalmology & Visual Science, Mar. 2011, vol. 52, No. 3, 2011. |
Martinez-Conde, Susan et al., The Role of Fixational Eye Movements in Visual Perception, Nature Reviews, Neuroscience, vol. 5, pp. 229-240, Mar. 2004. |
McNabb, Ryan P_ et al., Distributed Scanning Volumetric SDOCT for Motion Corrected Corneal Biometry, Biomedical Optics Express, vol. 3, No. 9, Aug. 10, 2012. |
Scott, Adrienne W. et al.; “Imaging the Infant Retina with a Hand-held Spectral-Domain Optical Coherence Tomography Device” Infant Retina Imaging by Hand-Held SD OCT, vol. 147, No. 2, pp. 364-373. |
Mishra, A., et al., Intra-Retinal Segmentatioin-Images, Optic Express 23719, pp. 1-10, Dec. 11, 2009. |
Thevenaz, Philippe et al.; “User-Friendly Semiautomated Assembly of Accurate Image Mosaics in Microscopy” Microscopy Research and Technique vol. 70: pp. 135-146 (2007). |
Niemeijer et al., Vessel Segmentation in 3D Spectral OCT Scans of the Retina, Medical Imaging 2008, Image Processing, Proc. of SPIE, vol. 6914, 2008, U.S. |
Non-Final Office Action received in U.S. Appl. No. 13/010,448 dated Jan. 2, 2014. |
Notice of Allowance received in U.S. Appl. No. 13/010,448 dated May 13, 2014. |
Aschke et a l., “Augmented Reality in Operating Microscopes for Neurosurgical Interventions.” IEEE, Mar. 22, 2003, pp. 652-654 (Mar. 22, 2003), p. 652, col. 2; p. 653, col. 2—p. 653, col. 1; Fig 3 [online] <URL: http://wwwipr.ipr.uni-karlsruhe.de/en/publications/download/id/288/d/article288.pdf>. |
International Search Report and Written Opinion (PCT/ISA/220 & PCT/ISA/210 & PCT/ISA/237) issued in PCT Application No. PCT/US2016/016830 dated May 19, 2016 (nine (9) pages). |
Number | Date | Country | |
---|---|---|---|
20180008140 A1 | Jan 2018 | US |
Number | Date | Country | |
---|---|---|---|
62112698 | Feb 2015 | US |