This disclosure relates generally to analysis of light field images and three-dimensional images of the eardrum.
The plenoptic otoscope or light field otoscope (LFO) is a type of medical otoscope that enables non-invasive three-dimensional (3D) imaging of the middle ear. Unlike traditional digital otoscopes that capture two-dimensional (2D) images, the LFO captures four-dimensional (4D) light field data from which 3D information can be recovered using digital processing techniques. Past studies have shown that the 3D shape of the eardrum is one of the most important factors for distinguishing Acute Otitis Media (AOM) which is a bacterial infection of the middle ear and requires antimicrobial treatment, from Otitis Media with Effusion (OME) which is a sterile effusion and tends to resolve on its own, and from No Effusion (NOE) which is a normal eardrum condition. Therefore, the 3D reconstruction of the eardrum (also called the tympanic membrane or TM) obtained using the LFO, has significant potential for automated diagnosis of otitis media.
However, automated classification algorithms typically require a registration step or an object detection step before passing the information to the classifier. Since the LFO is a relatively new device, we are not aware of any existing methods for automatic eardrum registration from light field data. Thus, there is a need for automated registration of 3D eardrum images obtained from 4D light field data.
The present disclosure overcomes the limitations of the prior art by providing automatic registration of a three-dimensional image of the eardrum. In one aspect, a three-dimensional image (e.g., depth map) of an eardrum is produced from four-dimensional light field data captured by a light field otoscope. The three-dimensional image is registered as defined by a predefined standard form. The eardrum is then classified based on the registered three-dimensional image. Registration may include compensation for out-of-plane rotation (tilt), for in-plane rotation, for center location, for translation, and/or for scaling.
Other aspects include components, devices, systems, improvements, methods, processes, applications, computer readable mediums, and other technologies related to any of the above.
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
Embodiments of the disclosure have other advantages and features which will be more readily apparent from the following detailed description and the appended claims, when taken in conjunction with the accompanying drawings, in which:
The figures depict various embodiments for purposes of illustration only. One skilled in the art will readily recognize from the following discussion that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles described herein.
The figures and the following description relate to preferred embodiments by way of illustration only. It should be noted that from the following discussion, alternative embodiments of the structures and methods disclosed herein will be readily recognized as viable alternatives that may be employed without departing from the principles of what is claimed.
The imaging optics 260 form a conventional image within the otoscope instrument at an intermediate image plane I2. Rather than a conventional sensor array capturing this image, a plenoptic sensor 270 captures the image. The microlens array 272 is positioned at the intermediate image plane I2 and a sensor array 274 is positioned at a conjugate P3 of the pupil plane P1. In addition, a filter module (not shown in
In the example design of
Other examples are described in U.S. application Ser. No. 13/896,924 “Plenoptic Otoscope,” Ser. No. 15/063,362 “Optical Design of a Light Field Otoscope,” and Ser. No. 14/323,949 “Otoscope Illumination,” which are incorporated by reference herein.
The 4D light field data can be processed 220 in various ways to produce the 3D image of the eardrum. Various approaches are described in U.S. Pat. No. 9,460,515 “Processing of Light Fields by Transforming to Scale and Depth Space” and U.S. Pat. No. 9,569,853 “Processing of Light Fields by Transforming to Scale and Depth Space” (collectively, the “Lisad Patents”) which are incorporated by reference herein.
After registration 230 which will be described in more detail below, the registered 3D image can be classified 240 using different approaches. For example, a depth map of the TM can produce information regarding its shape—whether it is bulging or retracting, and the estimated curvature. Spectral information can include an amber or yellow image, which is especially useful to diagnose conditions of the tympanic membrane. Table 1 lists some features distinguishing the conditions of acute otitis media (AOM), otitis media with effusion (OME), and otitis media with no effusion (NOE). Additional examples are described in Ser. No. 14/318,578 “Use of Plenoptic Otoscope Data for Aiding Medical Diagnosis,” which is incorporated by reference herein.
The example registration 230 in
The following provides a more detailed example of eardrum registration 230. In this example, the input 3D image is obtained by 3D estimation based on the 4D light field data. This can be either a sparse or a dense depth point cloud representation. For a sparse point cloud representation, we can calculate the normalized first derivative Ray-Gaussian light field scale and depth space and/or the Hessian keypoints for the light field scale and depth space, for example as described in the Lisad Patents (as identified above), which are incorporated by reference herein. Lisad stands for light field scale and depth. We thus get a set of points V∈R2 on the eardrum with specific depth. To get the depth graph J={V, E, W} we then connect each point to its neighbors. We use Gaussian weights based on either the 3D or 2D distance between the points and we set small ones to zero, resulting in a sparse graph.
For dense depth point cloud representation of 3D eardrum data, we can use the procedure described in the Lisad Patents to get a dense depth mesh M={V, O} of the eardrum, where V is the point cloud and O is the mesh connectivity (obtained using Delaunay triangulation, for example). This is the representation that we have used in the following example.
In
Returning to
To find the zero orientation from the tilt plane, we use the plane elevation vector. We assume that the plane elevation, defined as the projection of the z-axis (as defined by the 3D image before tilt removal) on the tilt plane, is indicative of the top of the eardrum. In
Since the TMs of the right and left eardrums are mirror-symmetric, we select one of the two as the standard and flip the other so that it is registered correctly relative to the standard form. For example, if the left ear is selected as standard, then we flip the images of the right eardrum. Alternatively, we can flip the left eardrum images if the right ear is chosen as the standard.
The eardrum depth representation, after the tilt removal and rotation to the zero orientation, is an updated mesh represented by MR={VR, OR}. At this point, we have registered the 3D TM data with respect to both out-of-plane rotation (tilt) and in-plane rotation (orientation). The next step in registration is to find the center 236 of the TM.
We use the 3D data to find the center of the TM. In one approach, we designate the location of the umbo as the center of the TM. The umbo is the most depressed part of the TM after tilt removal (i.e., point with most negative depth), as can be seen from
Using only the negative depth locations VRRN, we usually end up with many neighboring vertices that belong to the same area. We use an additional clustering step for the vertices in VRRN. During the clustering, we would like to take into account the fact that the umbo is usually located in inner areas of the eardrum, i.e., areas that are physically separated from the border of the eardrum. From our observations, we have noticed that such areas often exhibit closed depth isocontours, which are the contours of equal depth.
There are cases where due to some irregularity in the shape of the eardrum, closed isocontours that overlap with the locations of the vertices in VRRN are not successfully located. Such an example is shown in
Returning to
Although the detailed description contains many specifics, these should not be construed as limiting the scope of the invention but merely as illustrating different examples and aspects of the invention. It should be appreciated that the scope of the invention includes other embodiments not discussed in detail above. Various other modifications, changes and variations which will be apparent to those skilled in the art may be made in the arrangement, operation and details of the method and apparatus of the present invention disclosed herein without departing from the spirit and scope of the invention as defined in the appended claims. Therefore, the scope of the invention should be determined by the appended claims and their legal equivalents.
Alternate embodiments are implemented in computer hardware, firmware, software, and/or combinations thereof. Implementations can be implemented in a computer program product tangibly embodied in a machine-readable storage device for execution by a programmable processor; and method steps can be performed by a programmable processor executing a program of instructions to perform functions by operating on input data and generating output. Embodiments can be implemented advantageously in one or more computer programs that are executable on a programmable system including at least one programmable processor coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device. Each computer program can be implemented in a high-level procedural or object-oriented programming language, or in assembly or machine language if desired; and in any case, the language can be a compiled or interpreted language. Suitable processors include, by way of example, both general and special purpose microprocessors. Generally, a processor will receive instructions and data from a read-only memory and/or a random access memory. Generally, a computer will include one or more mass storage devices for storing data files; such devices include magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and optical disks. Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM disks. Any of the foregoing can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits) and other forms of hardware.
The term “module” is not meant to be limited to a specific physical form. Depending on the specific application, modules can be implemented as hardware, firmware, software, and/or combinations of these. Furthermore, different modules can share common components or even be implemented by the same components. There may or may not be a clear boundary between different modules.
Number | Name | Date | Kind |
---|---|---|---|
6110106 | MacKinnon et al. | Aug 2000 | A |
6483535 | Tamburrino et al. | Nov 2002 | B1 |
7058441 | Shahar et al. | Jun 2006 | B2 |
7399275 | Goldfain et al. | Jul 2008 | B2 |
7399375 | Leiser et al. | Jul 2008 | B2 |
7433042 | Cavanaugh et al. | Oct 2008 | B1 |
7448753 | Chinnock | Nov 2008 | B1 |
7544163 | MacKinnon et al. | Jun 2009 | B2 |
7723662 | Levoy et al. | May 2010 | B2 |
7901351 | Prescott | Mar 2011 | B2 |
7936392 | Ng et al. | May 2011 | B2 |
7995214 | Forster et al. | Aug 2011 | B2 |
8066634 | Andreassen et al. | Nov 2011 | B2 |
8100826 | MacKinnon et al. | Jan 2012 | B2 |
8107086 | Hart et al. | Jan 2012 | B2 |
8143565 | Berkner et al. | Mar 2012 | B2 |
8617061 | Magalhães Mendes et al. | Dec 2013 | B2 |
8824779 | Smyth | Sep 2014 | B1 |
8944596 | Wood et al. | Feb 2015 | B2 |
8949078 | Berkner et al. | Feb 2015 | B2 |
9001326 | Goldfain | Apr 2015 | B2 |
9392933 | Bedard et al. | Jul 2016 | B2 |
9460515 | Tosic et al. | Oct 2016 | B2 |
9565993 | Okuda et al. | Feb 2017 | B2 |
9569853 | Tosic et al. | Feb 2017 | B2 |
20040147810 | Mizuno | Jul 2004 | A1 |
20050228231 | MacKinnon et al. | Oct 2005 | A1 |
20050283065 | Babayoff | Dec 2005 | A1 |
20080259274 | Chinnock | Oct 2008 | A1 |
20100004513 | MacKinnon et al. | Jan 2010 | A1 |
20110026037 | Forster et al. | Feb 2011 | A1 |
20110058466 | Kwon et al. | Mar 2011 | A1 |
20110152621 | Magalhães Mendes et al. | Jun 2011 | A1 |
20120065473 | Andreassen et al. | Mar 2012 | A1 |
20120182438 | Berkner et al. | Jul 2012 | A1 |
20120226480 | Berkner et al. | Sep 2012 | A1 |
20120320340 | Coleman | Dec 2012 | A1 |
20120327426 | Hart et al. | Dec 2012 | A1 |
20120327427 | Hart et al. | Dec 2012 | A1 |
20130002426 | Hart et al. | Jan 2013 | A1 |
20130002824 | Hart et al. | Jan 2013 | A1 |
20130003078 | Hart et al. | Jan 2013 | A1 |
20130027516 | Hart et al. | Jan 2013 | A1 |
20130128223 | Wood et al. | May 2013 | A1 |
20130237754 | Berglund et al. | Sep 2013 | A1 |
20130289353 | Seth et al. | Oct 2013 | A1 |
20140012141 | Kim et al. | Jan 2014 | A1 |
20140031701 | Berglund | Jan 2014 | A1 |
20140105474 | Lee | Apr 2014 | A1 |
20140192255 | Shroff et al. | Jul 2014 | A1 |
20140206979 | Berkner et al. | Jul 2014 | A1 |
20140316238 | Berkner et al. | Oct 2014 | A1 |
20140350379 | Verdooner | Nov 2014 | A1 |
20150005640 | Davis et al. | Jan 2015 | A1 |
20150005644 | Rhoads | Jan 2015 | A1 |
20150065803 | Douglas | Mar 2015 | A1 |
20150116526 | Meng et al. | Apr 2015 | A1 |
20150117756 | Tosic et al. | Apr 2015 | A1 |
20150126810 | Wood et al. | May 2015 | A1 |
20150250381 | Bedard | Sep 2015 | A1 |
20150373316 | Meng et al. | Dec 2015 | A1 |
20170119237 | Bedard et al. | May 2017 | A1 |
20170119238 | Gao | May 2017 | A1 |
Number | Date | Country |
---|---|---|
1531410 | Sep 2004 | CN |
101305899 | Nov 2008 | CN |
102105780 | Jun 2011 | CN |
102265124 | Nov 2011 | CN |
2000-126116 | May 2000 | JP |
2001-157664 | Jun 2001 | JP |
2002-034916 | Feb 2002 | JP |
2007-004471 | Jan 2007 | JP |
2007-500541 | Jan 2007 | JP |
2009-244429 | Oct 2009 | JP |
2014-138858 | Jul 2014 | JP |
2014-530697 | Nov 2014 | JP |
WO 2012058641 | May 2012 | WO |
WO 2012066741 | May 2012 | WO |
WO 2013138081 | Sep 2013 | WO |
WO 2014021994 | Feb 2014 | WO |
Entry |
---|
Pande et al., “A Mosaicking Approach for In Vivo Thickness Mapping of the Human Tympanic Membrane Using Low Coherence Interferometry,” Journal of the Association for Research in Otolaryngology, 17:403-416. |
Solis et al., “Tympanic membrane contour measurement with two source positions in digital holographic interferometry,” Biomedical Optics Express, Dec. 1, 2012, vol. 3, No. 12. |
Das et al., “A compact structured light based otoscope for three dimensional imaging of the tympanic membrane,” Proc. SPIE 9303, Photonic Therapeutics and Diagnostics XI, 93031F, Feb. 26, 2015. |
Bedard, N. et al., “In Vivo Middle Ear Imaging with a Light Field Otoscope,” Optics in the Life Sciences, OSA Technical Digest, (online), Optical Society of America, Paper BW3A.3, 3 pages. |
Bedard, N. et al., “Light Field Otoscope,” Imaging and Applied Optics 2014, OSA Technical Digest (online), Optical Society of America, 2014, Paper IM3C.6, 4 pages. |
Berkner, K. et al, “Measuring Color and Shape Characteristics of Objects from Light Fields,” Imaging and Applied Optics, 2015, 3 pages. |
Buytaert, J. et al., “Realistic 3D Computer Model of the Gerbil Middle Ear, Featuring Accurate Morphology of Bone and Soft Tissue Structures,” Journal of the Association for Research in Otolaryngology, Jul. 13, 2011, pp. 681-696. |
Chinese Office Action, Chinese Application No. 201410010071.X, May 6, 2015, 7 pages (with concise explanation of relevance). |
Cho, N.H. et al., “Optical Coherence Tomography for the Diagnosis of Human Otitis Media,” Proc. SPIE, 2013, 5 pages, vol. 8879, 88790N. |
European Patent Office, Partial European Search Report, EP Patent Application No. 18159173.6, dated Aug. 3, 2018, 15 pages. |
Hernandez-Montes, M.S. et al., “Optoelectronic Holographic Otoscope for Measurement of Nanodisplacements in Tympanic Membranes,” Journal of Biomedical Optics, Proceedings of the XI th International Congress and Exposition, Society for Experimental Mechanics Inc., Jun. 2-5, 2008, 7 pages, vol. 14, No. 3. |
Japanese Office Action, Japanese Application No. 2014-006668, Sep. 26, 2017, 4 pages (with concise explanation of relevance). |
Japanese Office Action, Japanese Application No. 2016-211050, Sep. 26, 2017, 5 pages (with concise explanation of relevance). |
Kim, C. et al., “Scene Reconstruction from High Spatio-Angular Resolution Light Fields,” Transactions on Graphics (TOG), 2013, 11 pages, vol. 32, No. 4. |
Kubota A. et al.,“View Interpolation using Defocused Multi-View Images,” Proceedings of APSIPA Annual Summit and Conference, Asia-Pacific Signal and Information Processing Association, 2009 Annual Summit and Conference, Sapara, Japan, Oct. 4, 2009, pp. 358-362. |
Kuruvilla, A. et al., “Otitis Media Vocabulary and Grammar,” CMU, ICIP, 2012, 4 pages. |
Levoy, M. “Light Field and Computational Imaging,” IEEE Computer Magazine, 2006, pp. 46-55, vol. 39. |
Levoy, M. et al, “Light Field Microscopy.” ACM Transactions on Graphics 25(3) 2006, pp. 1-11. |
NG, R. et al., “Light Field Photography with a Hand-Held Plenoptic Camera,” Stanford Tech Report, 2005, pp. 1-11. |
Sundberg, M., et al., “Diffuse Reflectance Spectroscopy of the Human Tympanic Membrane in Otitis Media, ”Physiological Measurement, 2004, pp. 1473-93, vol. 25, No. 6. |
Sundberg, M., et al., “Fibre Optic Array for Curvature Assessment—Application in Otitis Media,” Medical & Biological Engineering & Computing, Mar. 2004, pp. 245-252, vol. 42, No. 2. |
Tao, M. et al., “Depth from Combining Defocus and Correspondence Using Light-Field Cameras,” in Proceedings of the International Conference on Computer Vision, 2013, 8 pages. |
United States Office Action, U.S. Appl. No. 14/318,578, filed Dec. 10, 2015, 13 pages. |
United States Office Action, U.S. Appl. No. 13/896,924, Jul. 8, 2015, 15 pages. |
Wanner, S. et al., “Globally Consistent Depth Labeling of 4D Light Fields,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2012, pp. 41-48. |
Yang, T. et al., “High Performance Imaging Through Occlusion via Energy Minimization-Based Optimal Camera Selection,” International Journal of Advanced Robotic Systems, 2013, pp. 19, vol. 10. |
Number | Date | Country | |
---|---|---|---|
20180260611 A1 | Sep 2018 | US |