Technical Field
The present embodiments relate to imaging devices, and in particular, to systems, methods, and apparatus to correct for misalignment of stereoscopic image sensors.
Background
Digital imaging capabilities are currently being integrated into a wide range of devices, including digital cameras and mobile phones. Such devices may include the functionality to capture stereoscopic “3D” images. Device manufacturers have responded to the consumer by introducing devices integrating digital image processing to support capturing stereoscopic images utilizing single or multiple digital imaging sensors. A wide range of electronic devices, including mobile wireless communication devices, personal digital assistants (PDAs), personal music systems, digital cameras, digital recording devices, video conferencing systems, and the like, may make use of stereoscopic imaging capabilities to provide a variety of capabilities and features to their users. These include stereoscopic imaging applications such as 3D photos and videos.
To make stereoscopic data playback comfortable for viewers, it is desirable to provide digital systems wherein the imaging sensors are perfectly aligned, or nearly so. This allows the individual images captured by each imaging sensor to be more perfectly aligned to provide a stereoscopic image that reduces eye strain and other issues with viewing such images. However, this “perfect” image alignment of stereoscopic image pairs may drift over time due to gravity, heat, mechanical assembly, and in-use wear and tear. These sensor alignment imperfections, when present, can lead to capture of unaligned images which may result in visual discomfort to the viewer unless otherwise corrected. In some cases, a drift in yaw causes depth measurement errors. In other cases, autofocus accuracy starts to drift with age. In both cases, correction of the misalignment is desired to provide accurate depth measurements and improved stereoscopic image quality.
The systems, methods and devices of the disclosure each have several innovative aspects, no single one of which is solely responsible for the desirable attributes disclosed herein. Combinations of the innovations, aspects and features described herein can be incorporated in various embodiments of systems, methods, and devices, and such combinations are not limited by the examples of embodiments described herein.
Some of the embodiments may include a system for performing yaw correction of a stereoscopic image sensor pair includes an imaging device comprising a pair of stereoscopic image sensors and a control module. The control module may be configured to capture one or more images of an object with the sensor pair, determine a disparity of the object from the one or more images, estimate the stereoscopic depth of the object from the one or more images, set an autofocus lens position of the stereoscopic image sensor pair, perform autofocus functions on the object to determine a high frequency autofocus position, estimate the autofocus depth from the high frequency autofocus position, and estimate and correct a yaw angle correction using the difference between the autofocus depth and the stereoscopic depth.
Some embodiments may include a method for yaw correction of a stereoscopic image sensor pair using autofocus feedback. In one aspect, the method may include the steps of providing a stereoscopic image sensor pair and capturing one or more images of an object with the sensor pair. This method further includes determining a disparity of the object from the one or more images, estimating the stereoscopic depth of the object from the one or more images, setting an autofocus lens position of the stereoscopic image sensor pair, performing autofocus functions on the object to determine a high frequency autofocus position, estimating the autofocus depth from the high frequency autofocus position, and estimating and correcting a yaw angle correction using the difference between the autofocus depth and the stereoscopic depth.
In another embodiment, a method for autofocus lens position correction of a stereoscopic image sensor pair using autofocus feedback may include the steps of providing a stereoscopic image sensor pair, capturing one or more images of an object with the sensor pair, determining a disparity of the object from the one or more images, estimating the stereoscopic depth of the object from the one or more images, setting an autofocus lens position of the stereoscopic image sensor pair, performing autofocus functions on the object to determine a high frequency autofocus position, estimating the autofocus depth from the high frequency autofocus position, and correcting an autofocus lens position using the difference between the autofocus depth and the stereoscopic depth.
One aspect relates to a system for digitally correcting a physical misalignment of a pair of imaging sensors, including an imaging device comprising a first imaging sensor and a second imaging sensor and a control module. The control module may be configured to capture a first image data of an object with the first imaging sensor, capture a second image data of an object with the second imaging sensor, estimate a first depth of the object using the first image data and the second image data, estimate a second depth of the object from an autofocus lens position of the imaging device, compare the first depth and the second depth, and estimate and correct the misalignment between the first imaging sensor and the second imaging sensor using the difference between the first depth and the second depth. The control module may be further configured to determine a disparity of the object by comparing the first image data and the second image data. In some aspects, estimating the first depth of the object using the first image data and the second image date includes determining a stereoscopic depth estimate by comparing keypoints of the object located in both the first image data and the second image data. In some aspects, the control module may be further configured to perform autofocus functions while the imaging device is focused on the object to determine and set the autofocus lens position of the first imaging sensor and the second imaging sensor. In some aspects, estimating the second depth of the object from an autofocus lens position of the imaging device includes determining a focus position of the first imaging sensor and the second imaging sensor using a high frequency map to set the autofocus lens position of the first imaging sensor and the second imaging sensor and estimating an autofocus depth from the autofocus lens position. In some aspects, the control module may be further configured to warp one of the first image data and the second image data to reduce the disparity of the object if the difference between the first depth and the second depth is not zero.
In another aspect, a method for digitally correcting a physical misalignment of a pair of imaging sensors using autofocus feedback includes the steps of capturing a first image data of an object with a first imaging sensor, capturing a second image data of the object with a second imaging sensor, estimating a stereoscopic depth of the object using the first image data and the second image data, estimating an autofocus depth of the object from an autofocus lens position of the imaging device, comparing the stereoscopic depth and the autofocus depth, and estimating and correcting a misalignment between the first imaging sensor and the second imaging sensor using the difference between the stereoscopic depth and the autofocus depth. In some aspects, the method further includes determining a disparity of the object by comparing the first image data and the second image data. In some aspects, estimating the stereoscopic depth of the object using the first image data and the second image date includes comparing keypoints of the object located in both the first image data and the second image data. In some aspects, the method further includes performing autofocus functions while the imaging device is focused on the object to determine and set the autofocus lens position of the first imaging sensor and the second imaging sensor. In some aspects, estimating the autofocus depth of the object from an autofocus lens position of the imaging device includes determining a focus position of the first imaging sensor and the second imaging sensor using a high frequency map to set the autofocus lens position of the first imaging sensor and the second imaging sensor and estimating the autofocus depth from the autofocus lens position. In some aspects, the method further includes warping one of the first image data and the second image data to reduce the disparity of the object if the difference between the stereoscopic depth and the autofocus depth is not zero.
In yet another aspect, a method for correcting an autofocus lens position of an imaging device having a first imaging sensor and a second imaging sensor using autofocus feedback includes the steps of capturing a first image data of an object with the first imaging sensor, capturing a second image data of the object with the second imaging sensor, estimating a stereoscopic depth of the object using the first image data and the second image data, estimating an autofocus depth of the object from an autofocus lens position of the imaging device, comparing the stereoscopic depth and the autofocus depth, and correcting an autofocus lens position of the imaging device using the difference between the autofocus depth and the stereoscopic depth. In some aspects, estimating a stereoscopic depth of the object further includes determining a disparity of the object from the first image data and the second image data. In some aspects, estimating an autofocus depth of the object from an autofocus lens position further includes performing autofocus functions on the object to determine and set the autofocus lens position of the first imaging sensor and the second imaging sensor. In some aspects, setting the autofocus lens position further includes determining a focus position of the first imaging sensor and the second imaging sensor using a high frequency map. In some aspects, correcting an autofocus lens position of the imaging device further includes correcting a distance estimate of the autofocus lens position if the difference between the autofocus depth and the stereoscopic depth is not zero.
In some aspects, an apparatus for digitally correcting a physical misalignment of a pair of imaging sensors using autofocus feedback, includes means for capturing a first image data of an object with a first imaging sensor, means for capturing a second image data of the object with a second imaging sensor, means for estimating a stereoscopic depth of the object using the first image data and the second image data by comparing keypoints of the object located in both the first image data and the second image data, means for estimating an autofocus depth of the object from an autofocus lens position of the imaging device by determining a focus position of the first imaging sensor and the second imaging sensor using a high frequency map to set the autofocus lens position of the first imaging sensor and the second imaging sensor and estimating the autofocus depth from the autofocus lens position, means for comparing the stereoscopic depth and the autofocus depth, and means for estimating and correct a misalignment between the first imaging sensor and the second imaging sensor using the difference between the stereoscopic depth and the autofocus depth.
In another aspect, an apparatus for correcting an autofocus lens position of an imaging device and correcting a physical misalignment of a pair of imaging sensors using autofocus feedback includes means for capturing a first image data of an object with a first imaging sensor, means for capturing a second image data of the object with a second imaging sensor, means for estimating a stereoscopic depth of the object using the first image data and the second image data by determining a disparity of the object from the first image data and the second image data, means for estimating an autofocus depth of the object from an autofocus lens position of the imaging device by performing autofocus functions on the object to determine and set the autofocus lens position of the first imaging sensor and the second imaging sensor using a high frequency map, means for comparing the stereoscopic depth and the autofocus depth, and means for correcting an autofocus lens position of the imaging device using the difference between the autofocus depth and the stereoscopic depth by correcting a distance estimate of the autofocus lens position if the difference between the autofocus depth and the stereoscopic depth is not zero. In some aspects, the apparatus further includes means for determining a disparity of the object by comparing the first image data and the second image data. In some aspects, the apparatus further includes means for performing autofocus functions while the imaging device is focused on the object to determine and set the autofocus lens position of the first imaging sensor and the second imaging sensor. In some aspects, the apparatus further includes means for warping one of the first image data and the second image data to reduce the disparity of the object if the difference between the stereoscopic depth and the autofocus depth is not zero. In some aspects, the apparatus further includes means for estimating and correcting a misalignment between the first imaging sensor and the second imaging sensor using the difference between the stereoscopic depth and the autofocus depth.
In yet another aspect, a non-transitory computer-readable medium stores instructions that, when executed, cause at least one physical computer processor to perform a method of digitally correcting a physical misalignment of a pair of imaging sensors using autofocus feedback. The method includes the steps of capturing a first image data of an object with a first imaging sensor, capturing a second image data of the object with a second imaging sensor, estimating a stereoscopic depth of the object using the first image data and the second image data, estimating an autofocus depth of the object from an autofocus lens position of the imaging device, comparing the stereoscopic depth and the autofocus depth, and estimating and correct a misalignment between the first imaging sensor and the second imaging sensor using the difference between the stereoscopic depth and the autofocus depth. In some aspects, the method further includes determining a disparity of the object by comparing the first image data and the second image data. In some aspects, estimating the stereoscopic depth of the object using the first image data and the second image date includes comparing keypoints of the object located in both the first image data and the second image data. In some aspect, the method further includes performing autofocus functions while the imaging device is focused on the object to determine and set the autofocus lens position of the first imaging sensor and the second imaging sensor. In some aspects, estimating the autofocus depth of the object from an autofocus lens position of the imaging device includes determining a focus position of the first imaging sensor and the second imaging sensor using a high frequency map to set the autofocus lens position of the first imaging sensor and the second imaging sensor and estimating the autofocus depth from the autofocus lens position. In some aspects, the method further includes warping one of the first image data and the second image data to reduce the disparity of the object if the difference between the stereoscopic depth and the autofocus depth is not zero.
In yet another aspect, a non-transitory computer-readable medium stores instructions that, when executed, cause at least one physical computer processor to perform a method of correcting an autofocus lens position of an imaging device having a first imaging sensor and a second imaging sensor using autofocus feedback. The method includes the steps of capturing a first image data of an object with the first imaging sensor, capturing a second image data of the object with the second imaging sensor, estimating a stereoscopic depth of the object using the first image data and the second image data, estimating an autofocus depth of the object from an autofocus lens position of the imaging device, comparing the stereoscopic depth and the autofocus depth, and correcting an autofocus lens position of the imaging device using the difference between the autofocus depth and the stereoscopic depth. In some aspects, estimating a stereoscopic depth of the object further includes determining a disparity of the object from the first image data and the second image data. In some aspects, estimating an autofocus depth of the object from an autofocus lens position further includes performing autofocus functions on the object to determine and set the autofocus lens position of the first imaging sensor and the second imaging sensor. In some aspects, setting the autofocus lens position further includes determining a focus position of the first imaging sensor and the second imaging sensor using a high frequency map. In some aspects, correcting an autofocus lens position of the imaging device further includes correcting a distance estimate of the autofocus lens position if the difference between the autofocus depth and the stereoscopic depth is not zero.
The disclosed aspects will hereinafter be described in conjunction with the appended drawings, provided to illustrate and not to limit the disclosed aspects, wherein like designations denote like elements.
Certain visual experiments indicate that in order to see a three-dimensional (3D) presentation based on stereoscopic imagery (sometimes referred to herein as “stereo images,” “stereoscopic images” or “stereoscopic image pairs”) with minimal discomfort or strain, the “left” and “right” digital sensors that are used to capture the stereo images should be properly aligned with one another. For example, the optical (or sensing) axis of the two digital sensors may be optically aligned to be parallel, or substantially parallel, e.g., they differ only by known or easily determined horizontal or vertical shifts which are more easily corrected by digital editing of the captured image. For desirable stereoscopic effects and fusibility of images, a horizontal distance between imaging sensors in a stereoscopic sensor pair is, in some examples, around 3.25 cm. In addition, there is preferably only a relatively small horizontal or vertical shift between the sensors in the pair. However, in actual practice, obtaining aligned parallel imaging sensors is often unachievable due to mechanical mounting limitations, alignment measuring devices, gravity and/or heat effects on the sensors. Thus, certain embodiments described herein provide systems and methods for correcting depth measurement errors of stereoscopic imaging sensors due primarily to sensor drift in the yaw direction.
Calibration of stereoscopic images through digital image processing may be required to align the pairs of images after they have been captured. Some methods may digitally process the stereoscopic image pairs to produce aligned images. Aligning stereoscopic images may include, for example, cropping one or both images to correct for horizontal (x axis) or vertical (y axis) shift between the images of a stereoscopic image pair. The two images of a stereoscopic image pair may also be misaligned about a “z” axis, caused when one imaging sensor is slightly closer to a scene being imaged than the other imaging sensor. Cropping may also be required to correct for misalignment due to rotation of the images about an x, y, or z axis. Finally, cropping may also be required to adjust the convergence point of the two images in the stereoscopic image pair.
In addition to the two dimensional x, y and z offsets discussed above, the relative positions of a pair of imaging sensors can also be described by measuring three axes of angular movement and three axes of shift. For purposes of this disclosure, positions on an x, y, and z axis describe relative shift. Angular rotation can be described by rotations about a horizontal (x) axis, also called “pitch,” vertical (y) axis, known as “roll,” and (z) axis or “yaw.”
Some configurations of stereoscopic cameras may have asymmetric sensors wherein one sensor is a low resolution image sensor and the other sensor is a high resolution image sensor. During manufacture, the sensors are calibrated such that the sensor images are parallel. However, for some configurations, such as cameras with high quality autofocus systems, the camera lens positions slowly drift due to gravity, heat, mechanical assembly issues, or wear and tear. Lens position drift in the yaw position in particular can cause depth measurement errors. In some configurations, estimation of the yaw angle difference between the two imaging sensors may be used to correct the depth measurements.
For other configurations, such as cameras with low quality autofocus system, the autofocus accuracy can drift over time due to wear and tear of the components or aging materials used to form the components. In some configurations, a distance or depth measurement of objects in a scene that is estimated using stereo imaging sensors is better than an autofocus depth estimate determined by using the autofocus system in the device. Therefore, in some configurations, the depth estimated using the stereo imaging sensors may be used to correct the autofocus lens position. Special charts or targets may not be needed to correct the depth error measurement and adjust the autofocus lens position of the imaging sensors. In many cases, no user input is needed to perform the correction. Furthermore, image processing of the stereo images such as keypoint detection and matching may not be needed in some embodiments. Additionally, since the autofocus search space is generally small and the yaw drift is generally small, the processes described below are quick and work in real time.
System Overview
Image analysis system 115 may be a stationary device such as a desktop personal computer or it may be a mobile device. A plurality of applications may be available to the user on image analysis system 115. These applications may include traditional photographic applications, high dynamic range imaging, panoramic video, or stereoscopic imaging such as 3D images or 3D video.
Processor 122 may be a general purpose processing unit or a processor specially designed for imaging applications. As shown, the processor 122 is connected to a memory 120 and a working memory 130. In the illustrated embodiment, the memory 120 stores several modules, including an image capture control module 140, a stereo depth calculation module 145, an autofocus control module 155, an autofocus depth calculation module 160, a yaw angle correction module 162, operating system 165, and user interface module 170. These modules may include instructions that configure the processor 122 to perform various image processing and device management tasks. Working memory 130 may be used by processor 122 to store a working set of processor instructions contained in the modules of memory 120. Alternatively, working memory 130 may also be used by processor 122 to store dynamic data created during the operation of image analysis system 115.
Still referring to
The autofocus control module 155 provides instructions that configure the processor 122 to perform an autofocus function using the image sensors 105, 110 to, for example, search for the best high frequency map of the image scene. The autofocus depth calculation module 160 provides instructions that configure the process 122 to calculate a depth of an object in a scene based on one or more characteristics of an autofocus system, for example, the positions of each of the image sensors 105, 110 when the object is determined to be in focus using autofocus functionality. For example, based on the autofocus position of the image sensors 105, 110 during the autofocus function, an equivalent “true” depth or equivalent estimated depth of an object in a scene may be determined. The autofocus characteristic may be, for example, a physical or optical position of one or more components of each of, or one of, the imaging sensors 105, 110 positioned during autofocus operation. The autofocus position may be based on, for example, determining a position of the imaging sensors 105 and 110 to focus on the object using high frequency information (for example, noise) of the object or the scene. Therefore, processor 122, along with autofocus control module 155, autofocus depth calculation module 160, and working memory 130 represent one example of an embodiment of means for performing an autofocus function on the object to determine an autofocus characteristic (for example, position of the imaging sensors) and estimating an autofocus depth of an object from the high frequency autofocus position.
The yaw angle correction module 162 provides instructions that configure the processor 122 to calculate a difference between the depth as measured by the autofocus function and the “depth” of an object in a scene based on the disparity of an object as indicated by the stereoscopic images acquired by the stereoscopic imaging sensors 105 and 110. This difference may used by the yaw angle correction module 162 to estimate and correct the yaw angle of the stereoscopic imaging sensors 105 and 110. Therefore, processor 122, along with yaw angle correction module 162 and working memory 130 represent one example of an embodiment of means for estimating and correcting a yaw angle correction using the difference between the autofocus depth and the stereoscopic depth.
User interface module 170 can include instructions that configure the processor 122 to display information on an electronic display accessible to the user while running the image analysis system 115. Operating system module 165 may configure the processor 122 to manage the memory and processing resources of system 115. For example, operating system module 165 may include device drivers to manage hardware resources such as the electronic display 125 or imaging sensors 105 and 110. In some embodiments, instructions contained in the image processing modules discussed above may not interact with these hardware resources directly, but instead may interact through standard subroutines or APIs located in operating system component 165. Instructions within operating system 165 may then interact directly with these hardware components.
Processor 122 may write data to storage module 130. While storage module 130 is represented graphically as a traditional disk drive, those with skill in the art would understand multiple embodiments could include either a disk-based storage device or one of several other types of storage mediums, including a memory disk, USB drive, flash drive, remotely connected storage medium, virtual disk driver, or the like.
Although
Additionally, although
Method Overview
Embodiments of the invention relate to methods for correcting depth measurement disparity due to misalignment of stereoscopic imaging sensors, for example, imaging sensors 105 and 110, as described above with reference to
The process 300 then transitions to block 306 wherein the focus of one imaging sensor is fixed (such as imaging sensor 105 shown in
One example of a disparity drift due to yaw is shown graphically in
After determining the disparity of the object between the images acquired by each sensor of the imaging sensor pair such as imaging sensors 105, 110, the process 300 transitions to block 312, wherein the stereo depth of the object is estimated using the disparity. In some embodiments, the stereo depth may be calculated as follows:
Depth_stereo=baseline*focal length/disparity,
where baseline is the distance between the two imaging sensors of the stereoscopic imaging sensor pair.
Referring again to
One example of the autofocus statistics is shown in
In the next block 320 of the process 300 illustrated in
In another embodiment that may be particularly useful for imaging devices having poor or low quality autofocus systems, the autofocus accuracy may drift with age. In some embodiments, the estimated depth measurement to an object in an imaged scene is more accurate using stereo depth estimation rather than autofocus depth estimations. In these embodiments, the stereo depth measurement may be used to correct the autofocus lens position of the imaging device 150. One embodiment of a process to correct autofocus lens position using estimated stereo depth measurements is shown in
After determining the disparity of the object between the image data acquired by each sensor of the imaging sensor pair, the process 700 transitions to block 712, wherein the stereo depth of the object is estimated using the disparity. In some embodiments, the stereo depth may be calculated as follows:
Depth_stereo=baseline*focal length/disparity,
where baseline is the distance between the two imaging sensors of the stereoscopic imaging sensor pair.
After determining the stereo depth of the object, the process 700 transitions to block 714, where the autofocus feature of the imaging device maybe activated to determine the autofocus lens position. The process 700 then transitions to block 716, wherein the imaging device may sweep through two or more autofocus positions of the imaging sensors to search for the best high frequency map. The process 700 then transitions to block 718, wherein an autofocus depth of an object, as calculated by the autofocus feature, may be found from the autofocus statistics. As discussed above, one example of the autofocus statistics is shown in
In the next block 720 of the process 700 shown in
Clarifications Regarding Terminology
Unless expressly limited by its context, the term “signal” is used herein to indicate any of its ordinary meanings, including a state of a memory location (or set of memory locations) as expressed on a wire, bus, or other transmission medium. Unless expressly limited by its context, the term “generating” is used herein to indicate any of its ordinary meanings, such as computing or otherwise producing. Unless expressly limited by its context, the term “calculating” is used herein to indicate any of its ordinary meanings, such as computing, evaluating, smoothing, and/or selecting from a plurality of values. Unless expressly limited by its context, the term “obtaining” is used to indicate any of its ordinary meanings, such as calculating, deriving, receiving (e.g., from an external device), and/or retrieving (e.g., from an array of storage elements). Unless expressly limited by its context, the term “selecting” is used to indicate any of its ordinary meanings, such as identifying, indicating, applying, and/or using at least one, and fewer than all, of a set of two or more. Where the term “comprising” is used in the present description and claims, it does not exclude other elements or operations. The term “based on” (as in “A is based on B”) is used to indicate any of its ordinary meanings, including the cases (i) “derived from” (e.g., “B is a precursor of A”), (ii) “based on at least” (e.g., “A is based on at least B”) and, if appropriate in the particular context, (iii) “equal to” (e.g., “A is equal to B”). Similarly, the term “in response to” is used to indicate any of its ordinary meanings, including “in response to at least.”
Unless indicated otherwise, any disclosure of an operation of an apparatus having a particular feature is also expressly intended to disclose a method having an analogous feature (and vice versa), and any disclosure of an operation of an apparatus according to a particular configuration is also expressly intended to disclose a method according to an analogous configuration (and vice versa). The term “configuration” may be used in reference to a method, apparatus, and/or system as indicated by its particular context. The terms “method,” “process,” “procedure,” and “technique” are used generically and interchangeably unless otherwise indicated by the particular context. The terms “apparatus” and “device” are also used generically and interchangeably unless otherwise indicated by the particular context. The terms “element” and “module” are typically used to indicate a portion of a greater configuration. Unless expressly limited by its context, the term “system” is used herein to indicate any of its ordinary meanings, including “a group of elements that interact to serve a common purpose.” Any incorporation by reference of a portion of a document shall also be understood to incorporate definitions of terms or variables that are referenced within the portion, where such definitions appear elsewhere in the document, as well as any figures referenced in the incorporated portion.
Those having skill in the art will further appreciate that the various illustrative logical blocks, modules, circuits, and process steps described in connection with the implementations disclosed herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention. One skilled in the art will recognize that a portion, or a part, may comprise something less than, or equal to, a whole. For example, a portion of a collection of pixels may refer to a sub-collection of those pixels.
The various illustrative logical blocks, modules, and circuits described in connection with the implementations disclosed herein may be implemented or performed with a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
The steps of a method or process described in connection with the implementations disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, a removable disk, a CD-ROM, or any other form of non-transitory storage medium known in the art. An exemplary computer-readable storage medium is coupled to the processor such the processor can read information from, and write information to, the computer-readable storage medium. In the alternative, the storage medium may be integral to the processor. The processor and the storage medium may reside in an ASIC. The ASIC may reside in a user terminal, camera, or other device. In the alternative, the processor and the storage medium may reside as discrete components in a user terminal, camera, or other device.
Headings are included herein for reference and to aid in locating various sections. These headings are not intended to limit the scope of the concepts described with respect thereto. Such concepts may have applicability throughout the entire specification.
The previous description of the disclosed implementations is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these implementations will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other implementations without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the implementations shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
This application claims the benefit of U.S. Provisional Patent Application No. 61/866,950, filed Aug. 16, 2013, titled “STEREO YAW CORRECTION USING AUTOFOCUS FEEDBACK,” the disclosure of which is hereby incorporated herein by reference in its entirety and for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
3698803 | Toshio | Oct 1972 | A |
4114171 | Altman | Sep 1978 | A |
4437745 | Hajnal | Mar 1984 | A |
4639586 | Fender et al. | Jan 1987 | A |
4740780 | Brown et al. | Apr 1988 | A |
4751570 | Robinson | Jun 1988 | A |
4890314 | Judd et al. | Dec 1989 | A |
5012273 | Nakamura et al. | Apr 1991 | A |
5016109 | Gaylord | May 1991 | A |
5063441 | Lipton et al. | Nov 1991 | A |
5142357 | Lipton et al. | Aug 1992 | A |
5194959 | Kaneko et al. | Mar 1993 | A |
5207000 | Chang et al. | May 1993 | A |
5231461 | Silvergate et al. | Jul 1993 | A |
5243413 | Gitlin et al. | Sep 1993 | A |
5313542 | Castonguay | May 1994 | A |
5475617 | Castonguay | Dec 1995 | A |
5506913 | Ibison et al. | Apr 1996 | A |
5539483 | Nalwa | Jul 1996 | A |
5606627 | Kuo | Feb 1997 | A |
5614941 | Hines | Mar 1997 | A |
5640222 | Paul | Jun 1997 | A |
5642299 | Hardin et al. | Jun 1997 | A |
5686960 | Sussman et al. | Nov 1997 | A |
5721585 | Keast et al. | Feb 1998 | A |
5734507 | Harvey | Mar 1998 | A |
5745305 | Nalwa | Apr 1998 | A |
5760846 | Lee | Jun 1998 | A |
5793527 | Nalwa | Aug 1998 | A |
5798791 | Katayama et al. | Aug 1998 | A |
5903306 | Heckendorn et al. | May 1999 | A |
5926411 | Russell | Jul 1999 | A |
5990934 | Nalwa | Nov 1999 | A |
6111702 | Nalwa | Aug 2000 | A |
6115176 | Nalwa | Sep 2000 | A |
6128143 | Nalwa | Oct 2000 | A |
6141034 | McCutchen | Oct 2000 | A |
6141145 | Nalwa | Oct 2000 | A |
6144501 | Nalwa | Nov 2000 | A |
6195204 | Nalwa | Feb 2001 | B1 |
6219090 | Nalwa | Apr 2001 | B1 |
6285365 | Nalwa | Sep 2001 | B1 |
6356397 | Nalwa | Mar 2002 | B1 |
6421185 | Wick et al. | Jul 2002 | B1 |
6445815 | Sato | Sep 2002 | B1 |
6611289 | Yu et al. | Aug 2003 | B1 |
6628897 | Suzuki | Sep 2003 | B2 |
6650774 | Szeliski | Nov 2003 | B1 |
6700711 | Nalwa | Mar 2004 | B2 |
6701081 | Dwyer et al. | Mar 2004 | B1 |
6768509 | Bradski et al. | Jul 2004 | B1 |
6775437 | Kazarinov et al. | Aug 2004 | B2 |
6782137 | Avinash | Aug 2004 | B1 |
6798406 | Jones et al. | Sep 2004 | B1 |
6809887 | Gao et al. | Oct 2004 | B1 |
6850279 | Scherling | Feb 2005 | B1 |
6855111 | Yokoi et al. | Feb 2005 | B2 |
6861633 | Osborn | Mar 2005 | B2 |
6862364 | Berestov | Mar 2005 | B1 |
6987534 | Seta | Jan 2006 | B1 |
6992700 | Sato et al. | Jan 2006 | B1 |
7006123 | Yoshikawa et al. | Feb 2006 | B2 |
7039292 | Breiholz | May 2006 | B1 |
7084904 | Liu et al. | Aug 2006 | B2 |
7116351 | Yoshikawa | Oct 2006 | B2 |
7215479 | Bakin | May 2007 | B1 |
7253394 | Kang | Aug 2007 | B2 |
7271803 | Ejiri et al. | Sep 2007 | B2 |
7336299 | Kostrzewski et al. | Feb 2008 | B2 |
7612953 | Nagai et al. | Nov 2009 | B2 |
7710463 | Foote | May 2010 | B2 |
7805071 | Mitani | Sep 2010 | B2 |
7817354 | Wilson et al. | Oct 2010 | B2 |
7860214 | Haff | Dec 2010 | B1 |
7893957 | Peters et al. | Feb 2011 | B2 |
7961398 | Tocci | Jun 2011 | B2 |
7978222 | Schneider | Jul 2011 | B2 |
8004557 | Pan | Aug 2011 | B2 |
8098276 | Chang et al. | Jan 2012 | B2 |
8115813 | Tang | Feb 2012 | B2 |
8139125 | Scherling | Mar 2012 | B2 |
8228417 | Georgiev et al. | Jul 2012 | B1 |
8267601 | Campbell et al. | Sep 2012 | B2 |
8284263 | Oohara et al. | Oct 2012 | B2 |
8294073 | Vance et al. | Oct 2012 | B1 |
8356035 | Baluja et al. | Jan 2013 | B1 |
8400555 | Georgiev et al. | Mar 2013 | B1 |
8442392 | Ollila et al. | May 2013 | B2 |
8482813 | Kawano et al. | Jul 2013 | B2 |
8791984 | Jones et al. | Jul 2014 | B2 |
8836693 | Katano | Sep 2014 | B2 |
8928988 | Ford et al. | Jan 2015 | B1 |
8988564 | Webster et al. | Mar 2015 | B2 |
9049375 | Wade et al. | Jun 2015 | B2 |
9055208 | Kim | Jun 2015 | B2 |
9185296 | Wade et al. | Nov 2015 | B2 |
9264610 | Duparre | Feb 2016 | B2 |
9294672 | Georgiev et al. | Mar 2016 | B2 |
9316810 | Mercado | Apr 2016 | B2 |
9332188 | Takei et al. | May 2016 | B2 |
9602806 | Stafford | Mar 2017 | B1 |
9609210 | Djordjevic et al. | Mar 2017 | B2 |
9952371 | Ambur et al. | Apr 2018 | B2 |
9973680 | Osborne et al. | May 2018 | B2 |
20010028482 | Nishioka | Oct 2001 | A1 |
20020070365 | Karellas | Jun 2002 | A1 |
20020136150 | Mihara et al. | Sep 2002 | A1 |
20030024987 | Zhu | Feb 2003 | A1 |
20030034395 | Tsikos et al. | Feb 2003 | A1 |
20030038814 | Blume | Feb 2003 | A1 |
20030156751 | Lee | Aug 2003 | A1 |
20030214575 | Yoshikawa | Nov 2003 | A1 |
20040021767 | Endo et al. | Feb 2004 | A1 |
20040051805 | Yoshikawa et al. | Mar 2004 | A1 |
20040066449 | Givon | Apr 2004 | A1 |
20040105025 | Scherling | Jun 2004 | A1 |
20040183907 | Hovanky et al. | Sep 2004 | A1 |
20040195492 | Hsin | Oct 2004 | A1 |
20040201769 | Yoshikawa et al. | Oct 2004 | A1 |
20040246333 | Steuart et al. | Dec 2004 | A1 |
20040263611 | Cutler | Dec 2004 | A1 |
20050053274 | Mayer et al. | Mar 2005 | A1 |
20050057659 | Hasegawa | Mar 2005 | A1 |
20050081629 | Hoshal | Apr 2005 | A1 |
20050111106 | Matsumoto et al. | May 2005 | A1 |
20050185711 | Pfister et al. | Aug 2005 | A1 |
20050218297 | Suda et al. | Oct 2005 | A1 |
20050243175 | Yamada et al. | Nov 2005 | A1 |
20050253951 | Fujimoto et al. | Nov 2005 | A1 |
20060023074 | Cutler | Feb 2006 | A1 |
20060023106 | Yee et al. | Feb 2006 | A1 |
20060023278 | Nishioka | Feb 2006 | A1 |
20060061660 | Brackmann | Mar 2006 | A1 |
20060084852 | Mason et al. | Apr 2006 | A1 |
20060098267 | Togawa | May 2006 | A1 |
20060140446 | Luo et al. | Jun 2006 | A1 |
20060193509 | Criminisi et al. | Aug 2006 | A1 |
20060215054 | Liang et al. | Sep 2006 | A1 |
20060215903 | Nishiyama | Sep 2006 | A1 |
20060238441 | Benjamin et al. | Oct 2006 | A1 |
20070024739 | Konno | Feb 2007 | A1 |
20070058961 | Kobayashi et al. | Mar 2007 | A1 |
20070064142 | Misawa et al. | Mar 2007 | A1 |
20070085903 | Zhang | Apr 2007 | A1 |
20070146530 | Nose | Jun 2007 | A1 |
20070164202 | Wurz et al. | Jul 2007 | A1 |
20070216796 | Lenel et al. | Sep 2007 | A1 |
20070242152 | Chen | Oct 2007 | A1 |
20070263115 | Horidan et al. | Nov 2007 | A1 |
20070268983 | Elam | Nov 2007 | A1 |
20080029708 | Olsen et al. | Feb 2008 | A1 |
20080030573 | Ritchey | Feb 2008 | A1 |
20080030597 | Olsen et al. | Feb 2008 | A1 |
20080058629 | Seibel et al. | Mar 2008 | A1 |
20080088702 | Linsenmaier et al. | Apr 2008 | A1 |
20080117289 | Schowengerdt et al. | May 2008 | A1 |
20080117532 | Shafer | May 2008 | A1 |
20080218612 | Border et al. | Sep 2008 | A1 |
20080259172 | Tamaru | Oct 2008 | A1 |
20080266404 | Sato | Oct 2008 | A1 |
20080290435 | Oliver et al. | Nov 2008 | A1 |
20080291543 | Nomura et al. | Nov 2008 | A1 |
20080297612 | Yoshikawa | Dec 2008 | A1 |
20080316301 | Givon | Dec 2008 | A1 |
20090003646 | Au et al. | Jan 2009 | A1 |
20090005112 | Sorek et al. | Jan 2009 | A1 |
20090015812 | Schultz et al. | Jan 2009 | A1 |
20090051804 | Nomura et al. | Feb 2009 | A1 |
20090080695 | Yang | Mar 2009 | A1 |
20090085846 | Cho et al. | Apr 2009 | A1 |
20090096994 | Smits | Apr 2009 | A1 |
20090153726 | Lim | Jun 2009 | A1 |
20090160931 | Pockett et al. | Jun 2009 | A1 |
20090268210 | Prince | Oct 2009 | A1 |
20090268983 | Stone et al. | Oct 2009 | A1 |
20090268985 | Wong | Oct 2009 | A1 |
20090296984 | Nijim et al. | Dec 2009 | A1 |
20090315808 | Ishii | Dec 2009 | A1 |
20100044555 | Ohara et al. | Feb 2010 | A1 |
20100045774 | Len et al. | Feb 2010 | A1 |
20100066812 | Kajihara et al. | Mar 2010 | A1 |
20100165155 | Chang | Jul 2010 | A1 |
20100202766 | Takizawa et al. | Aug 2010 | A1 |
20100215249 | Heitz et al. | Aug 2010 | A1 |
20100232681 | Fujieda et al. | Sep 2010 | A1 |
20100259655 | Takayama | Oct 2010 | A1 |
20100265313 | Liu et al. | Oct 2010 | A1 |
20100265363 | Kim | Oct 2010 | A1 |
20100278423 | Itoh et al. | Nov 2010 | A1 |
20100289878 | Sato et al. | Nov 2010 | A1 |
20100290703 | Sim et al. | Nov 2010 | A1 |
20100290769 | Nasiri et al. | Nov 2010 | A1 |
20100302396 | Golub et al. | Dec 2010 | A1 |
20100309286 | Chen et al. | Dec 2010 | A1 |
20100309333 | Smith et al. | Dec 2010 | A1 |
20110001789 | Wilson et al. | Jan 2011 | A1 |
20110007135 | Okada et al. | Jan 2011 | A1 |
20110009163 | Fletcher et al. | Jan 2011 | A1 |
20110012998 | Pan | Jan 2011 | A1 |
20110038535 | Wang et al. | Feb 2011 | A1 |
20110043623 | Fukuta et al. | Feb 2011 | A1 |
20110090575 | Mori | Apr 2011 | A1 |
20110096089 | Shenhav et al. | Apr 2011 | A1 |
20110096988 | Suen et al. | Apr 2011 | A1 |
20110128412 | Milnes et al. | Jun 2011 | A1 |
20110181588 | Barenbrug | Jul 2011 | A1 |
20110213664 | Osterhout et al. | Sep 2011 | A1 |
20110235899 | Tanaka | Sep 2011 | A1 |
20110249341 | DiFrancesco et al. | Oct 2011 | A1 |
20110262122 | Minamisawa et al. | Oct 2011 | A1 |
20110304764 | Shigemitsu et al. | Dec 2011 | A1 |
20120008148 | Pryce et al. | Jan 2012 | A1 |
20120033051 | Atanassov et al. | Feb 2012 | A1 |
20120044368 | Lin et al. | Feb 2012 | A1 |
20120056987 | Fedoroff | Mar 2012 | A1 |
20120075168 | Osterhout et al. | Mar 2012 | A1 |
20120127276 | Tsai | May 2012 | A1 |
20120229688 | Tajiri | Sep 2012 | A1 |
20120249536 | Sutou | Oct 2012 | A1 |
20120249750 | Izzat et al. | Oct 2012 | A1 |
20120249815 | Bohn et al. | Oct 2012 | A1 |
20120269400 | Heyward | Oct 2012 | A1 |
20120281072 | Georgiev et al. | Nov 2012 | A1 |
20120293607 | Bhogal et al. | Nov 2012 | A1 |
20120293632 | Yukich | Nov 2012 | A1 |
20120327195 | Cheng | Dec 2012 | A1 |
20130003140 | Keniston et al. | Jan 2013 | A1 |
20130010084 | Hatano | Jan 2013 | A1 |
20130038689 | McDowall | Feb 2013 | A1 |
20130057655 | Su | Mar 2013 | A1 |
20130070055 | Atanassov et al. | Mar 2013 | A1 |
20130076924 | Wade et al. | Mar 2013 | A1 |
20130077945 | Liu et al. | Mar 2013 | A1 |
20130128030 | Georgiev | May 2013 | A1 |
20130141802 | Yang | Jun 2013 | A1 |
20130182325 | Minamisawa et al. | Jul 2013 | A1 |
20130222556 | Shimada | Aug 2013 | A1 |
20130229529 | Lablans | Sep 2013 | A1 |
20130250045 | Ki et al. | Sep 2013 | A1 |
20130250053 | Levy | Sep 2013 | A1 |
20130250123 | Zhang et al. | Sep 2013 | A1 |
20130260823 | Shukla et al. | Oct 2013 | A1 |
20130278785 | Nomura et al. | Oct 2013 | A1 |
20130286451 | Verhaegh | Oct 2013 | A1 |
20130329015 | Pulli | Dec 2013 | A1 |
20130335598 | Gustavsson et al. | Dec 2013 | A1 |
20130335600 | Gustavsson et al. | Dec 2013 | A1 |
20140009631 | Topliss | Jan 2014 | A1 |
20140016832 | Kong et al. | Jan 2014 | A1 |
20140085502 | Lin et al. | Mar 2014 | A1 |
20140104378 | Kauff et al. | Apr 2014 | A1 |
20140111650 | Georgiev et al. | Apr 2014 | A1 |
20140139623 | McCain et al. | May 2014 | A1 |
20140152852 | Ito et al. | Jun 2014 | A1 |
20140184749 | Hilliges et al. | Jul 2014 | A1 |
20140192253 | Laroia | Jul 2014 | A1 |
20140285673 | Hundley et al. | Sep 2014 | A1 |
20140340568 | Sano et al. | Nov 2014 | A1 |
20150043076 | Nakayama | Feb 2015 | A1 |
20150070562 | Nayar et al. | Mar 2015 | A1 |
20150085363 | Liu et al. | Mar 2015 | A1 |
20150125092 | Zhuo et al. | May 2015 | A1 |
20150177524 | Webster et al. | Jun 2015 | A1 |
20150201128 | Dong | Jul 2015 | A1 |
20150244934 | Duparre et al. | Aug 2015 | A1 |
20150253647 | Mercado | Sep 2015 | A1 |
20150286033 | Osborne | Oct 2015 | A1 |
20150288865 | Osborne | Oct 2015 | A1 |
20150370040 | Georgiev | Dec 2015 | A1 |
20150371387 | Atanassov | Dec 2015 | A1 |
20150373252 | Georgiev | Dec 2015 | A1 |
20150373263 | Georgiev | Dec 2015 | A1 |
20150373268 | Osborne | Dec 2015 | A1 |
20150373269 | Osborne | Dec 2015 | A1 |
20150373279 | Osborne | Dec 2015 | A1 |
20160014332 | De et al. | Jan 2016 | A1 |
20160085059 | Mercado | Mar 2016 | A1 |
20160127641 | Gove | May 2016 | A1 |
20160127646 | Osborne | May 2016 | A1 |
20160198087 | Georgiev et al. | Jul 2016 | A1 |
20160269602 | Osborne | Sep 2016 | A1 |
20160286121 | Georgiev et al. | Sep 2016 | A1 |
20160295112 | Georgiev et al. | Oct 2016 | A1 |
20160353008 | Osborne | Dec 2016 | A1 |
20160373263 | Zaidi et al. | Dec 2016 | A1 |
20170026570 | Shepard et al. | Jan 2017 | A1 |
20170038502 | Georgiev | Feb 2017 | A1 |
20170118421 | Georgiev | Apr 2017 | A1 |
20180084193 | Georgiev et al. | Mar 2018 | A1 |
Number | Date | Country |
---|---|---|
101046534 | Oct 2007 | CN |
101065955 | Oct 2007 | CN |
101201459 | Jun 2008 | CN |
101257576 | Sep 2008 | CN |
101571666 | Nov 2009 | CN |
101581828 | Nov 2009 | CN |
101867720 | Oct 2010 | CN |
101902657 | Dec 2010 | CN |
101926171 | Dec 2010 | CN |
101952762 | Jan 2011 | CN |
201917706 | Aug 2011 | CN |
202405984 | Aug 2012 | CN |
103038689 | Apr 2013 | CN |
103376613 | Oct 2013 | CN |
203519911 | Apr 2014 | CN |
203519914 | Apr 2014 | CN |
203551875 | Apr 2014 | CN |
0610605 | Aug 1994 | EP |
0751416 | Jan 1997 | EP |
1176812 | Jan 2002 | EP |
1383342 | Jan 2004 | EP |
1816514 | Aug 2007 | EP |
1832912 | Sep 2007 | EP |
2242252 | Oct 2010 | EP |
2354390 | Mar 2001 | GB |
2354391 | Mar 2001 | GB |
S60213178 | Oct 1985 | JP |
H06217184 | Aug 1994 | JP |
H06251127 | Sep 1994 | JP |
H089424 | Jan 1996 | JP |
H0847001 | Feb 1996 | JP |
H08125835 | May 1996 | JP |
8194274 | Jul 1996 | JP |
H08242453 | Sep 1996 | JP |
H0946729 | Feb 1997 | JP |
H09214992 | Aug 1997 | JP |
H10142490 | May 1998 | JP |
2001194114 | Jul 2001 | JP |
2002158913 | May 2002 | JP |
2003304561 | Oct 2003 | JP |
2004260787 | Sep 2004 | JP |
3791847 | Jun 2006 | JP |
2006279538 | Oct 2006 | JP |
2007147457 | Jun 2007 | JP |
2007323615 | Dec 2007 | JP |
2008009424 | Jan 2008 | JP |
2009122842 | Jun 2009 | JP |
2010041381 | Feb 2010 | JP |
2010067014 | Mar 2010 | JP |
2010128820 | Jun 2010 | JP |
2010524279 | Jul 2010 | JP |
2012085102 | Apr 2012 | JP |
2013117568 | Jun 2013 | JP |
20060049992 | May 2006 | KR |
20080071400 | Aug 2008 | KR |
WO-199321560 | Oct 1993 | WO |
WO-199847291 | Oct 1998 | WO |
WO-2006075528 | Jul 2006 | WO |
WO-2007129147 | Nov 2007 | WO |
WO-2008112054 | Sep 2008 | WO |
WO-2009047681 | Apr 2009 | WO |
WO-2009086330 | Jul 2009 | WO |
WO-2010019757 | Feb 2010 | WO |
2011108276 | Sep 2011 | WO |
WO-2012136388 | Oct 2012 | WO |
WO-2012164339 | Dec 2012 | WO |
WO-2013154433 | Oct 2013 | WO |
WO-2014012603 | Jan 2014 | WO |
WO-2014025588 | Feb 2014 | WO |
Entry |
---|
Arican, et al., “Intermediate View Generation for Perceived Depth Adjustment of Sterio Video”, Mitsubishi Electric Research Laboratories, http://www.merl.com, TR2009-052, Sep. 2009; 12 pages. |
Hoff, et al., “Surfaces from Stereo: Integrating Feature Matching, Disparity Estimation, and Contour Detection”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 11, No. 2, pp. 121-136, Feb. 1989. |
International Search Report and Written Opinion—PCT/US2014/049776—ISA/EPO—dated Nov. 14, 2014. |
Krotkov E., et al., “Active vision for reliable ranging: Cooperating focus, stereo, and vergence”, International Journal of Computer Vision. vol. 11, No. 2, Oct. 1, 1993 (Oct. 1, 1993), pp. 187-203, XP055149875, ISSN: 0920-5691. DOI: 10.1007/BF01469228. |
Murphy M., et al., “Lens Drivers Focus on Performance in High-Resolution Camera Modules,” Analog Dialogue, Nov. 2006, vol. 40, pp. 1-3. |
Narkhede, et al., “Stereoscopic Imaging: A Real-Time, in Depth Look,” IEEE Potentials, Feb./Mar. 2004, vol. 23, Issue 1, pp. 38-42. |
Sun W.S., et al., “Single-Lens Camera Based on a Pyramid Prism Array to Capture Four Images,” Optical Review, 2013, vol. 20 (2), pp. 145-152. |
Han Y., et al., “Removing Illumination from Image Pair for Stereo Matching”, Audio, Language and Image Processing (ICALIP), 2012 International Conference on, IEEE, Jul. 16, 2012, XP032278010, pp. 508-512. |
Hao M., et al., “Object Location Technique for Binocular Stereo Vision Based on Scale Invariant Feature Transform Feature Points”, SIFT, Journal of Harbin Engineering University, Jun. 2009, vol. 30, No. 6 pp. 649-653. |
Kawanishi T., et al., “Generation of High-Resolution Stereo Panoramic Images by Omnidirectional Imaging Sensor Using Hexagonal Pyramidal Mirrors”, Pattern Recognition, 1998, Proceedings, Fourteenth International Conference on Brisbane, QLD., Australia Aug. 16-20, 1998, Los Alamitos, CA, USA,IEEE Comput. Soc, US, Jan. 1, 1998 (Jan. 1, 1998), pp. 485-489, vol. 1, XP031098377, ISBN: 978-0-8186-8512-5. |
Ricoh Imagine Change: “New Ricoh Theta Model, Capturing 360-degree Images in One Shot, is on Sale Soon—Spherical Video Function, API and SDK (Beta Version)”, News Release, Oct. 28, 2014, 3 pages. |
Shuchun Y., et al., “Preprocessing for stereo vision based on LOG filter”, Proceedings of 2011 6th International Forum on Strategic Technology, Aug. 2011, XP055211077, pp. 1074-1077. |
Tan K-H., et al., “Multiview Panoramic Cameras Using a Pyramid”, Omnidirectional Vision, 2002, Proceedings, Third Workshop on Jun. 2, 2002, Piscataway, NJ, USA,IEEE, Jan. 1, 2002 (Jan. 1, 2002), pp. 87-93, XP010611080, ISBN: 978-0-7695-1629-5. |
Hua et al., “Design Analysis of a High-Resolution Panoramic Camera Using Conventional Imagers and a Mirror Pyramid,” IEEE Transactions on Pattern Analysis and Machine Intelligence; Feb. 2007; 29(2): 356-361. |
Meng et al., “Single-shot Specular Surface Reconstruction with Gonio-plenoptic Imaging,” 2015 IEEE International Conference on Computer Vision; pp. 3433-3441. |
Chowdhury A., et al., “Challenges of Megapixel Camera Module Assembly and Test,” Electronic Components and Technology Conference, 2005, pp. 1390-1401. |
Hung K-Y., et al., “Integrated the Back-Side Inclined Exposure Technology to Fabricate the 45 Degree K-Type Prism with Nanometer Roughness,” NEMS 2012, Kyoto, Japan, Mar. 2012, pp. 120-124. |
Wenyi Zhao and N. Nandhakumar, “Effects of Camera Alignment Errors on Stereoscopic Depth Estimates”, 24 pages, Dec. 1996, Pattern Recognition. |
Number | Date | Country | |
---|---|---|---|
20150049172 A1 | Feb 2015 | US |
Number | Date | Country | |
---|---|---|---|
61866950 | Aug 2013 | US |