Some contemporary imaging devices contain multi-component sensors comprising color (R, G, B) and infrared photosites (where R, G and B are sometimes used herein for red, green and blue, respectively, and IR for infrared). Depending on an environment in which the sensor is used, the R, G, and B part of the photosites (referred to as R0, G0, B0 hereinafter where the subscript zero represents the component state as initially captured) often contain a significant amount of IR component, such as when ambient light contains IR, or when IR is projected into a scene for depth sensing or other purposes.
When the captured R0, G0, B0 data contains IR information, the data do not provide true color information. For example, if the data are used directly for demosaicing (to generate per pixel RGB), the resulting colors look washed out.
This Summary is provided to introduce a selection of representative concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used in any way that would limit the scope of the claimed subject matter.
Briefly, one or more of various aspects of the subject matter described herein are directed towards calibrating a color correction (extraction) transform that corrects for infrared light in at least one of red, green or blue parts of a photosite. Ground truth color data is captured as raw image data via a sensor comprised of photosites. The ground truth color data is also captured as long-pass-filtered image data via the sensor capturing through a long pass filter. The long-pass-filtered image data for at least one of the red, green or blue parts of a photosite of the sensor are subtracted from the raw image data for each corresponding part of the photosite to obtain true color data values for the photosite. Data corresponding to the true color data values are used to produce one or more tables or curves that are accessible during online usage to color correct an image.
In one or more aspects, a sensor comprising photosites having infrared, red, green, and blue parts is configured to capture a first image of ground truth data without a filter, capture a second image of the ground truth data with a long pass filter, and capture a third image of the ground truth data with a short pass filter. A processing component obtains true red, green and blue data based upon the first and second images, and obtains true infrared data based upon the first and third images. The processing component outputs data corresponding the true red, green and blue data and true infrared data into one or more tables or curves.
One or more aspects are directed towards (a) selecting a current infrared value, (b) accessing table or curve data to determine predicted red, green and blue values based upon the current infrared value, (c) accessing table or curve data to determine a predicted infrared value based upon the predicted red, green and blue values, (d) setting the current infrared value as the predicted infrared value, (e) returning to step (b) until a stopping criterion is met; and (f) outputting an infrared value and red green and blue values based upon the current infrared value and last predicted red, green and blue values.
Other advantages may become apparent from the following detailed description when taken in conjunction with the drawings.
The present invention is illustrated by way of example and not limited in the accompanying figures in which like reference numerals indicate similar elements and in which:
Various aspects of the technology described herein are generally directed towards extracting true RGB from sensor data. In one aspect, this is facilitated by a calibration process, e.g., using ground truth colors (e.g., a color chart) to determine how a specific camera or the like captures R, G, B and IR values in the presence of IR illumination. Note that as used herein, “true” is an inexact, relative concept, and thus the calibration is based upon whatever is decided as ground truth, subject to varying lighting conditions and the like. Further, curves, tables, mappings and/or other structures and the like described below may use approximations, interpolations and so forth, whereby “true” typically means approximately achieving or approaching the ground truth, e.g., to the extent possible. In practice, significant improvement in image appearance has been obtained by outputting true color after compensating for IR.
It should be understood that any of the examples herein are non-limiting. For instance, while the examples herein are directed towards “true” RGB being approximated using IR component data, “true” IR may be approximated using RGB component data. As such, the present invention is not limited to any particular embodiments, aspects, concepts, structures, functionalities or examples described herein. Rather, any of the embodiments, aspects, concepts, structures, functionalities or examples described herein are non-limiting, and the present invention may be used various ways that provide benefits and advantages in image processing in general.
The color chart may, for example, be any known set of differently-colored patches or the like, but in one implementation was a six column by four-column grid of twenty-four different colors (including black white and gray variations). Each color in the grid corresponded to fixed, known R, G and B values.
An IR light source 228 may be used to project a consistent level of IR onto the color chart 224, generally under controlled lighting conditions. As described below, the IR light source may be variable so that different amounts of IR may be emitted for different types of calibration, or different IR light sources may be used. Note that the IR light source may be centered (e.g., slightly above or below) the camera at a suitable distance so that the color chart is illuminated relatively evenly. Notwithstanding, the amount of IR is determined at a per-pixel level and subtracted out in one implementation, whereby reasonable variations in the IR value across the color chart are not significant.
In reality, infrared (or near infrared, NIR, which is synonymous with IR as used herein) contaminates the R, G and B values detected by a camera sensor, whereby additional processing is needed to get the (approximately) true RGB-IR values at each pixel. The output may be generated using a “demosaicing” process as shown by block 230 in
As can be seen, the photosites 232 and 233 contains the values captured with no filter and with the long pass filter that blocks visible light, respectively. Thus, the non-filtered (raw) photosite 232 comprises IR, and R, G, and B values contaminated with some amount of IR, shown as R+IRR, G+IRG, and B+IRB. The long-pass-filtered photosite 233 contains IR, IRR, IRG, IRB values.
In a processing component 236, subtracting the filtered IR from each of the raw R, G and B parts removes the difference (Δ) that the IR contamination is contributing:
IRGB-IRΔ=IRGB-IRraw−IRGB-IRLPF.
The first part of the above equation corresponds to RΔ, GΔBΔ, which can be linearized through radiometric calibration, for example, which is a known technique in image processing to compensate for sensors' non-linear response to light. The non-linearity may be modeled in any of many known ways, e.g., empirically determined via various images captured in different lighting conditions.
Once linearized, an affine matrix transform may be offline computed for performing true RGB correction, which may then be used “online” in actual usage. The following describes the transform, which may be modeled as lookup tables/curves (e.g., corresponding to blocks 237-239 in
In one alternative, where IR leakage is discounted, the IR component may be ignored, whereby a 3×3 matrix may be used.
In another alternative, (shown in the block diagram of
This calibration configuration in
The boxed “IR” (when the long pass filter is used) in photosite 233, and boxed “R”, “G”, and “B” (when the short pass filter is used) in photosite 334, represent the true signals to recover. As before, subtraction (in block 336 of
Once the signals are linearized, “true” RGB may be used to predict RGBIR. At step 414 a mapping C may be used to map RGB to RGBIR (three dimensions to one dimension). As before, “true” IR may be used to predict IRR, I RG, IRB (three one-to-one mappings: “true” IR to IRR, “true” IR to I RG, “true” IR to IRB). Each mapping can be in a form of lookup tables or fitted parametric curves (referred to as tables/curves QR, QG, QB), shown in
After calibration, the tables/curves are known, and can be stored in the camera as in
IR0=IR+RGBIR,
R0=R+IRR,
G0=G+IRG,
B0=B+IRB.
To get true R, G, B, IR, the following procedure (also shown in example steps in the flow diagram of
Step 512 repeats the process from step 504 until convergence (e.g., the updated IR value does not change over some number of iterations), or for a fixed number of steps. Step 514 outputs the computed true values. As can be seen, this process iteratively hones in on the true values by predicting the IR component to predict a closer true R, G, B and uses those predicted R, G, B values to find a closer IR value, which is used to find even closer R, G, B values and so on, until some convergence is reached or some iteration limit is reached. The process may be repeated for each photosite. The process can be done in the other order, that is, by starting with RGB values and predicting IR, then updating the RGB with the predicted IR and so on.
Note that this is only one example, and that other optimizations may be used to extract (R, G, B, IR) from the original data (R0, G0, B0, IR0).
In another alternative, IR illumination may be changed via a variable or multiple IR light sources; filters may or may not be used. The calibration thus captures the color chart under different IR lighting conditions. The different IR lighting conditions allow effective separation of true color from the IR component.
The linearization may be done via lookup tables generated using a standard technique involving multiple exposures, while the matrix transform M3×4 is extracted through the calibration process involving a color chart. Since there is no ground-truth for IR when no filters are used, the data may be used as-is and directly interpolated.
The equation for mapping the inputs R0, G0, B0, and IR0 to RGB is the same as above, that is:
Each block in the color chart has a unique known color. To deal with the multiple image captures, Let N=P×Q, where P is the number of blocks in the chart and Q the number of image captures, each under different IR illumination conditions. To extract RGB, the following over-determined linear equations are solved:
The above techniques are based upon broad spectrum distribution, which may not be the case. Thus, in any alternative, calibration may be performed under different lighting conditions, with a different affine matrix computed for each lighting condition. For example, one lighting condition may be general room lighting, another may be dark room with some IR lighting, and so forth. It may also be desirable (or needed) to have multiple affine transforms for the same lighting condition, split based on an amount of IR that is present.
After calibration during online operation, a broad spectral distribution (or user input) may be used to estimate the current lighting condition, and the estimated lighting condition used to select the matching calibration parameters to apply. An option is to compute a weighting scheme based on a similarity measure between the current lighting condition and the predefined ones, and weight average the pixel values as output.
Example Operating Environment
The invention is operational with numerous other general purpose or special purpose computing system environments or configurations. Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the invention include, but are not limited to: personal computers, server computers, hand-held or laptop devices, tablet devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
The invention may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, and so forth, which perform particular tasks or implement particular abstract data types. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in local and/or remote computer storage media including memory storage devices.
With reference to
The computer 610 typically includes a variety of computer-readable media. Computer-readable media can be any available media that can be accessed by the computer 610 and includes both volatile and nonvolatile media, and removable and non-removable media. By way of example, and not limitation, computer-readable media may comprise computer storage media and communication media. Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, solid-state device memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by the computer 610. Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above may also be included within the scope of computer-readable media.
The system memory 630 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 631 and random access memory (RAM) 632. A basic input/output system 633 (BIOS), containing the basic routines that help to transfer information between elements within computer 610, such as during start-up, is typically stored in ROM 631. RAM 632 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 620. By way of example, and not limitation,
The computer 610 may also include other removable/non-removable, volatile/nonvolatile computer storage media. By way of example only,
The drives and their associated computer storage media, described above and illustrated in
The computer 610 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 680. The remote computer 680 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 610, although only a memory storage device 681 has been illustrated in
When used in a LAN networking environment, the computer 610 is connected to the LAN 671 through a network interface or adapter 670. When used in a WAN networking environment, the computer 610 typically includes a modem 672 or other means for establishing communications over the WAN 673, such as the Internet. The modem 672, which may be internal or external, may be connected to the system bus 621 via the user input interface 660 or other appropriate mechanism. A wireless networking component such as comprising an interface and antenna may be coupled through a suitable device such as an access point or peer computer to a WAN or LAN. In a networked environment, program modules depicted relative to the computer 610, or portions thereof, may be stored in the remote memory storage device. By way of example, and not limitation,
An auxiliary subsystem 699 (e.g., for auxiliary display of content) may be connected via the user interface 660 to allow data such as program content, system status and event notifications to be provided to the user, even if the main portions of the computer system are in a low power state. The auxiliary subsystem 699 may be connected to the modem 672 and/or network interface 670 to allow communication between these systems while the main processing unit 620 is in a low power state.
While the invention is susceptible to various modifications and alternative constructions, certain illustrated embodiments thereof are shown in the drawings and have been described above in detail. It should be understood, however, that there is no intention to limit the invention to the specific forms disclosed, but on the contrary, the intention is to cover all modifications, alternative constructions, and equivalents falling within the spirit and scope of the invention.
The present application claims priority to U.S. provisional patent application Ser. No. 61/812,232, filed Apr. 15, 2013.
Number | Name | Date | Kind |
---|---|---|---|
3938102 | Morrin et al. | Feb 1976 | A |
5351152 | Kuo et al. | Sep 1994 | A |
5471326 | Hall et al. | Nov 1995 | A |
5586200 | Devaney et al. | Dec 1996 | A |
5739906 | Evans et al. | Apr 1998 | A |
6105139 | Dey et al. | Aug 2000 | A |
6751344 | Grumbine | Jun 2004 | B1 |
7315383 | Abdollahi | Jan 2008 | B1 |
7512262 | Criminisi et al. | Mar 2009 | B2 |
7565003 | Ashizaki et al. | Jul 2009 | B2 |
7634395 | Flandrin et al. | Dec 2009 | B2 |
8077034 | Borlez et al. | Dec 2011 | B2 |
8331654 | Abraham et al. | Dec 2012 | B2 |
8442940 | Faletti et al. | May 2013 | B1 |
8787656 | Park et al. | Jul 2014 | B2 |
8818077 | Hwang | Aug 2014 | B2 |
9508003 | Eguro et al. | Nov 2016 | B2 |
9697424 | Kirk et al. | Jul 2017 | B2 |
9760770 | Eguro et al. | Sep 2017 | B2 |
9922249 | Kang et al. | Mar 2018 | B2 |
9928420 | Kirk et al. | Mar 2018 | B2 |
9959465 | Georgiou et al. | May 2018 | B2 |
20020136444 | Brown et al. | Sep 2002 | A1 |
20030043270 | Rafey et al. | Mar 2003 | A1 |
20030048459 | Gooch | Mar 2003 | A1 |
20030081833 | Tilton | May 2003 | A1 |
20040105580 | Hager et al. | Jun 2004 | A1 |
20040125222 | Bradski et al. | Jul 2004 | A1 |
20040201586 | Marschner et al. | Oct 2004 | A1 |
20050058362 | Kita | Mar 2005 | A1 |
20050234527 | Slatkine | Oct 2005 | A1 |
20050257748 | Kriesel et al. | Nov 2005 | A1 |
20050279172 | Schreier et al. | Dec 2005 | A1 |
20060176306 | Nagaraj et al. | Aug 2006 | A1 |
20060210146 | Gu | Sep 2006 | A1 |
20060238714 | Fox et al. | Oct 2006 | A1 |
20060291020 | Knox et al. | Dec 2006 | A1 |
20070009150 | Suwa et al. | Jan 2007 | A1 |
20070145273 | Chang | Jun 2007 | A1 |
20070146512 | Suzuki et al. | Jun 2007 | A1 |
20070183657 | Kidono et al. | Aug 2007 | A1 |
20070253310 | Ikenaka | Nov 2007 | A1 |
20070263903 | St. hilaire et al. | Nov 2007 | A1 |
20080118143 | Gordon et al. | May 2008 | A1 |
20080130015 | Lu | Jun 2008 | A1 |
20080165357 | Stern et al. | Jul 2008 | A1 |
20080187711 | Alam et al. | Aug 2008 | A1 |
20080205748 | Lee et al. | Aug 2008 | A1 |
20080218612 | Border et al. | Sep 2008 | A1 |
20080283729 | Hosaka | Nov 2008 | A1 |
20090021750 | Korner et al. | Jan 2009 | A1 |
20090080048 | Tsao | Mar 2009 | A1 |
20090096783 | Shpunt et al. | Apr 2009 | A1 |
20090217213 | Meserve | Aug 2009 | A1 |
20090217214 | Meserve | Aug 2009 | A1 |
20090231425 | Zalewski | Sep 2009 | A1 |
20090273679 | Gere | Nov 2009 | A1 |
20100042964 | Meserve | Feb 2010 | A1 |
20100046004 | Lee et al. | Feb 2010 | A1 |
20100074532 | Gordon et al. | Mar 2010 | A1 |
20100177164 | Zalevsky et al. | Jul 2010 | A1 |
20100202725 | Popovich et al. | Aug 2010 | A1 |
20100277571 | Xu et al. | Nov 2010 | A1 |
20100289885 | Lu | Nov 2010 | A1 |
20110063427 | Fengler et al. | Mar 2011 | A1 |
20110078189 | Bonchi et al. | Mar 2011 | A1 |
20110091096 | Morris et al. | Apr 2011 | A1 |
20110103711 | Su et al. | May 2011 | A1 |
20110149031 | Um et al. | Jun 2011 | A1 |
20110222757 | Yeatman et al. | Sep 2011 | A1 |
20110228097 | Motta | Sep 2011 | A1 |
20110310220 | Mceldowney | Dec 2011 | A1 |
20120002045 | Tony et al. | Jan 2012 | A1 |
20120025080 | Liu et al. | Feb 2012 | A1 |
20120038986 | Pesach | Feb 2012 | A1 |
20120056982 | Katz et al. | Mar 2012 | A1 |
20120087572 | Dedeoglu et al. | Apr 2012 | A1 |
20120120494 | Takayama | May 2012 | A1 |
20120154397 | Chernikov et al. | Jun 2012 | A1 |
20120155747 | Hwang | Jun 2012 | A1 |
20120242829 | Shin et al. | Sep 2012 | A1 |
20120253201 | Reinhold | Oct 2012 | A1 |
20120281087 | Kruse | Nov 2012 | A1 |
20120294510 | Zhang et al. | Nov 2012 | A1 |
20120307075 | Margalit | Dec 2012 | A1 |
20130002814 | Park et al. | Jan 2013 | A1 |
20130003069 | Umeda et al. | Jan 2013 | A1 |
20130051657 | Ostermann et al. | Feb 2013 | A1 |
20130083062 | Geisner et al. | Apr 2013 | A1 |
20130095302 | Pettis et al. | Apr 2013 | A1 |
20130100256 | Kirk et al. | Apr 2013 | A1 |
20130100282 | Siercks | Apr 2013 | A1 |
20130141545 | Macchia et al. | Jun 2013 | A1 |
20130141611 | Hirai | Jun 2013 | A1 |
20130215235 | Russell | Aug 2013 | A1 |
20130229396 | Huebner | Sep 2013 | A1 |
20130265623 | Sugiyama et al. | Oct 2013 | A1 |
20130278631 | Border et al. | Oct 2013 | A1 |
20130287291 | Cho | Oct 2013 | A1 |
20130335531 | Lee et al. | Dec 2013 | A1 |
20140055560 | Fu et al. | Feb 2014 | A1 |
20140098342 | Webb | Apr 2014 | A1 |
20140104387 | Klusza et al. | Apr 2014 | A1 |
20140112573 | Francis et al. | Apr 2014 | A1 |
20140120319 | Joseph | May 2014 | A1 |
20140132501 | Choi et al. | May 2014 | A1 |
20140132728 | Verano et al. | May 2014 | A1 |
20140139717 | Short | May 2014 | A1 |
20140168380 | Heidemann et al. | Jun 2014 | A1 |
20140180639 | Cheatham et al. | Jun 2014 | A1 |
20140184584 | Reif et al. | Jul 2014 | A1 |
20140206443 | Sharp et al. | Jul 2014 | A1 |
20140225985 | Klusza et al. | Aug 2014 | A1 |
20140225988 | Poropat | Aug 2014 | A1 |
20140241612 | Rhemann et al. | Aug 2014 | A1 |
20140293011 | Lohry et al. | Oct 2014 | A1 |
20140307047 | Kirk et al. | Oct 2014 | A1 |
20140307055 | Kang et al. | Oct 2014 | A1 |
20140307057 | Kang et al. | Oct 2014 | A1 |
20140307058 | Kirk et al. | Oct 2014 | A1 |
20140307307 | Georgiou et al. | Oct 2014 | A1 |
20140307952 | Sweeney et al. | Oct 2014 | A1 |
20140307953 | Kirk et al. | Oct 2014 | A1 |
20140309764 | Socha-leialoha et al. | Oct 2014 | A1 |
20140310496 | Eguro et al. | Oct 2014 | A1 |
20140320605 | Johnson | Oct 2014 | A1 |
20140354803 | Chida | Dec 2014 | A1 |
20150078672 | Eguro et al. | Mar 2015 | A1 |
20150316368 | Moench et al. | Nov 2015 | A1 |
20180173947 | Kang et al. | Jun 2018 | A1 |
20180218210 | Georgiou et al. | Aug 2018 | A1 |
20180260623 | Kang et al. | Sep 2018 | A1 |
Number | Date | Country |
---|---|---|
1244008 | Feb 2000 | CN |
1414412 | Apr 2003 | CN |
1414420 | Apr 2003 | CN |
1445724 | Oct 2003 | CN |
1541483 | Oct 2004 | CN |
1669051 | Sep 2005 | CN |
1735789 | Feb 2006 | CN |
101061367 | Oct 2007 | CN |
101124514 | Feb 2008 | CN |
101309429 | Nov 2008 | CN |
101443809 | May 2009 | CN |
101501442 | Aug 2009 | CN |
101509764 | Aug 2009 | CN |
101711354 | May 2010 | CN |
101878409 | Nov 2010 | CN |
102027434 | Apr 2011 | CN |
102036599 | Apr 2011 | CN |
102231037 | Nov 2011 | CN |
102362150 | Feb 2012 | CN |
102385237 | Mar 2012 | CN |
102572485 | Jul 2012 | CN |
102638692 | Aug 2012 | CN |
102760234 | Oct 2012 | CN |
102803894 | Nov 2012 | CN |
102831380 | Dec 2012 | CN |
103308517 | Sep 2013 | CN |
0085210 | Aug 1983 | EP |
2295932 | Mar 2011 | EP |
2400261 | Dec 2011 | EP |
2481459 | Dec 2011 | GB |
2000341721 | Dec 2000 | JP |
2003058911 | Feb 2003 | JP |
2004135209 | Apr 2004 | JP |
2005-341470 | Dec 2005 | JP |
2006229802 | Aug 2006 | JP |
2009014501 | Jan 2009 | JP |
2010011223 | Jan 2010 | JP |
2010145186 | Jul 2010 | JP |
2011514232 | May 2011 | JP |
2013544449 | Dec 2013 | JP |
20110132260 | Dec 2011 | KR |
2237284 | Sep 2004 | RU |
2006016303 | Feb 2006 | WO |
2007132399 | Nov 2007 | WO |
2009046268 | Apr 2009 | WO |
2012137434 | Oct 2012 | WO |
Entry |
---|
“International Preliminary Report on Patentability Issued in PCT Patent Application No. PCT/US2014/033909”, dated Jun. 29, 2015, 17 Pages. |
“International Search Report & Written Opinion Issued in PCT Patent Application No. PCT/US2014/033909”, dated Jul. 25, 2014, 16 Pages. |
“Second Written Opinion Issued in PCT Patent Application No. PCT/US2014/033909”, dated Mar. 25, 2015, 14 Pages. |
“First Office Action and Search Report Issued in Chinese Patent Application No. 201480021958.7”, dated Oct. 18, 2017, 19 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/915,626”, dated Jul. 11, 2016, 13 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/915,626”, dated Mar. 28, 2017, 17 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/915,626”, dated Dec. 9, 2016, 16 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/915,626”, dated Jan. 29, 2016, 17 Pages |
“Final Office Action Issued in U.S. Appl. No. 13/918,892”, dated Dec. 21, 2016, 6 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/918,892”, dated Nov. 19, 2015, 7 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/918,892”, dated Mar. 28, 2016, 6 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/918,892”, dated May 11, 2015, 6 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/923,135”, dated Jul. 27, 2015, 21 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/923,135”, dated Oct. 3, 2016, 15 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/923,135”, dated May 5, 2017, 17 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/923,135”, dated Mar. 31, 2016, 14 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/923,135”, dated Dec. 19, 2014, 13 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/924,464”, dated Dec. 17, 2015, 25 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/924,464”, dated Sep. 15, 2016, 25 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/924,464”, dated Sep. 27, 2017, 37 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/924,464”, dated May 5, 2017, 38 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/924,464”, dated May 17, 2016, 26 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/924,464”, dated Aug. 10, 2015, 20 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/924,475”, dated May 6, 2016, 13 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/924,475”, dated Sep. 8, 2016, 15 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/924,475”, dated Mar. 24, 2015, 17 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/924,475”, dated Oct. 7, 2015, 14 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/924,485”, dated Sep. 13, 2016, 16 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/924,485”, dated Jul. 20, 2017, 16 Pages. |
“Non Final office Action issued in U.S. Appl. No. 13/924,485”, dated Dec. 15, 2015, 15 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 13/924,485”, dated Mar. 7, 2017, 16 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/925,762”, dated May 11, 2016, 22 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/925,762”, dated May 9, 2017, 26 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/925,762”, dated Oct. 3, 2016, 28 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/925,762”, dated Nov. 5, 2015, 13 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 14/088,408”, dated Sep. 29, 2016, 16 Pages. |
“Final Office Action Issued in U.S. Appl. No. 14/253,696”, dated Feb. 25, 2016, 23 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 14/253,696”, dated Sep. 10, 2015, 18 Pages. |
“Notice of Allowance Issued in U.S. Appl. No. 14/253,696”, dated Jul. 21, 2016, 17 Pages. |
“Office Action Issued in European Patent Application No. 14723271.4”, dated Sep. 19, 2016, 9 Pages. |
“Office Action Issued in European Patent Application No. 14723271.4”, dated Jun. 29, 2017, 12 Pages. |
“Oral Hearing Issued in European Patent Application No. 14723271.4”, Mailed Date: Mar. 29, 2018, 16 Pages. |
“Office Action Issued in Australian Patent Application No. 2014254219”, dated Apr. 19, 2017, 3 Pages. |
“First Office Action and Search Report Issued in Chinese Patent Application No. 201480021199.4”, dated May 31, 2017, 17 Pages. |
“Second Office Action Issued in Chinese Patent Application No. 201480021199.4”, dated Jan. 17, 2018, 7 Pages. |
“Third Office Action Issued in Chinese Patent Application No. 201480021199.4”, dated Jun. 19, 2018, 13 Pages. |
“Office Action Issued in Chinese Patent Application No. 201480021422.5”, dated Jul. 5, 2018, 10 Pages. |
“First Office Action & Search Report Issued in Chinese Patent Application No. 201480021422.5”, dated Mar. 21, 2017, 16 Pages. |
“Second Office Action Issued in Chinese Patent Application No. 201480021422.5”, dated Dec. 11, 2017, 14 Pages. |
“First Office Action and Search Report Issued in Chinese Patent Application No. 201480021460.0”, dated Mar. 28, 2017, 17 Pages. |
“Second Office Action Issued in Chinese Patent Application No. 201480021460.0”, dated Dec. 11, 2017, 6 Pages. |
“Third Office Action Issued in Chinese Patent Application No. 201480021460.0”, dated Jul. 2, 2018, 7 Pages. |
“Office Action Issued in Chinese Patent Application No. 201480021487.X”, dated Jun. 20, 2018, 7 Pages. |
“First Office Action Issued in Chinese Application No. 201480021487.X”, dated Apr. 21, 2017, 13 Pages. |
“Second Office Action Issued in Chinese Patent Application No. 201480021487.X”, dated Dec. 15, 2017, 10 Pages. |
“Office Action Issued in Chinese Patent Application No. 201480021493.5”, dated Dec. 19, 2017, 12 Pages. |
“Second Office Action Issued in Chinese Patent Application No. 201480021493.5”, dated Aug. 15, 2018, 10 Pages. |
“Office Action Issued in Chinese Patent Application No. 201480021519.6”, dated Sep. 4, 2018, 10 Pages. |
“First Office Action and Search Report Issued in Chinese Patent Application No. 201480021519.6”, dated Aug. 30, 2016, 18 Pages. |
“Fourth Office Action Issued in Chinese Patent Application No. 201480021519.6”, dated Mar. 8, 2018, 15 Pages. |
“Office Action Issued in Chinese Patent Application No. 201480021519.6”, dated Sep. 19, 2017, 14 Pages. |
“Second Office Action and Search Report Issued in Chinese Patent Application No. 201480021519.6”, dated Mar. 30, 2017, 17 Pages. |
“First Office Action and Search Report Issued in Chinese Patent Application No. 201480021528.5”, dated Dec. 11, 2017, 14 Pages. |
“Second Office Action Issued in Chinese Patent Application No. 201480021528.5”, dated Aug. 2, 2018, 8 Pages. |
“Second Office Action Issued in Chinese Patent Application No. 201480021958.7”, dated Jun. 13, 2018, 16 Pages. |
“Office Action Issued in Russian Patent Application No. 2015143654”, dated Jan. 31, 2018, 8 Pages. |
“Office Action Issued in Japanese Patent Application No. 2016-508993”, dated Feb. 6, 2018, 5 Pages. |
“Office Action Issued in Japanese Patent Application No. 2016-508993”, dated Aug. 7, 2018, 5 Pages. |
Anderson, et al., “The IBM System/360 Model 91: Machine Philosophy and Instruction-Handling”, Published in IBM Journal of Research and Development, vol. 11, Issue 1, Jan. 1, 1967, pp. 8-24. |
Chan, et al., “Regular Stereo Matching Improvement System Based on Kinect-Supporting Mechanism”, In Open Journal of Applied Sciences, vol. 3, Issue 1, Mar. 30, 2013, 5 Pages. |
Chihoub, et al., “A Band Processing Imaging Library for a Tricore-Based Digital Still Camera”, In Research Article of Real-Time Imaging, vol. 7, Issue 4, Aug. 1, 2001, pp. 327-337. |
Gao, et al., “Stereo Matching Algorithm Based on Illumination Normal Similarity and Adaptive Support Weight”, In Optical Engineering, vol. 52, Issue 2, Feb. 2013, 8 Pages. |
Gu, et al., “Trinocular Disparity Estimation With Adaptive Support Weight and Disparity Calibration”, In Optical Engineering, vol. 47, Issue 3, Mar. 1, 2008, 5 Pages. |
Hariyama, et al., “Optimal Periodical Memory Allocation for Logic-In-Memory Image Processors”, In Proceedings of the Emerging VLSI Technologies and Architectures (ISVLSI'06), Mar. 2, 2006, pp. 193-198. |
Hosni, et al., “Near Real-Time Stereo With Adaptive Support Weight Approaches”, In Proceedings of International Symposium 3D Data Processing, Visualization and Transmission, May 17, 2010, 8 Pages. |
Kanade, et al., “Development of a Video-Rate Stereo Machine”, In Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems. Human Robot Interaction and Cooperative Robots, Aug. 5, 1995, pp. 95-100. |
Kuon, et al., “FPGA Architecture: Survey and Challenges”, In Foundations and Trends in Electronic Design Automation vol. 2, Issue 2, Apr. 18, 2008, pp. 135-253. |
Langmann, et al., “Depth Camera Technology Comparison and Performance Evaluation”, In Proceedings of the 1st International Conference on Pattern Recognition Applications and Methods, Feb. 2012, 7 Pages. |
Mcllroy, et al., “Kinectrack: Agile 6-DoF Tracking Using a Projected Dot Pattern”, Published in IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Nov. 5, 2012, 7 Pages. |
“Office Action Issued in Mexican Patent Application No. Mx/a/2015/014577”, dated Nov. 10, 2017, 3 Pages. |
“International Preliminary Report on Patentability Issued in PCT Application No. PCT/US2014/033910”, dated Jul. 24, 2015, 8 Pages. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2014/033910”, dated Jul. 25, 2014, 9 Pages. |
“Second Written Opinion Received for PCT Patent Application No. PCT/US2014/033910”, dated Dec. 18, 2014, 5 Pages. |
“International Preliminary Report on Patentability Issued in PCT Application No. PCT/US2014/033911”, dated Jul. 13, 2015, 6 Pages. |
“International Search Report & Written Opinion Issued in PCT Application No. PCT/US2014/033911”, dated Aug. 29, 2014, 8 Pages. |
“Second Written Opinion Issued in PCT Application No. PCT/US2014/033911”, dated Apr. 2, 2015, 5 Pages. |
“International Preliminary Report on Patentability Issued in PCT Application No. PCT/US2014/033915”, dated Apr. 7, 2015, 10 Pages. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2014/033915”, dated Jul. 16, 2014, 11 Pages. |
“Second Written Opinion Issued in PCT Application No. PCT/US2014/033915”, dated Jan. 8, 2015, 9 Pages. |
“International Preliminary Report on Patentability Issued in PCT Application No. PCT/US2014/033916”, dated Jul. 13, 2015, 7 Pages. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2014/033916”, dated Jul. 18, 2014, 8 Pages. |
“Second Written Opinion Issued in PCT Application No. PCT/US2014/033916”, dated Mar. 27, 2015, 6 Pages. |
“International Preliminary Report on Patentability Issued in PCT Patent Application No. PCT/US2014/033917”, dated Jul. 20, 2015, 8 Pages. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2014/033917”, dated Jul. 18, 2014, 10 Pages. |
“International Preliminary Report on Patentability Issued in PCT Patent Application No. PCT/US2014/033919”, dated Jul. 13, 2015, 7 Pages. |
“International Search Report & Written Opinion Issued in PCT Application No. PCT/US2014/033919”, dated Jul. 17, 2014, 8 Pages. |
“Second Written Opinion Issued in PCT Application No. PCT/US2014/033919”, dated Mar. 27, 2015, 6 Pages. |
“International Search Report Issued in PCT Application No. PCT/US2014/033996”, dated Jul. 31, 2014, 3 Pages. |
Yamaguchi, et al., “Interleaved Pixel Lookup for Embedded Computer Vision”, In Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Jun. 23, 2008, pp. 1-8. |
Yang, et al., “Spatial-Depth Super Resolution for Range Images”, In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Jun. 17, 2007, 8 Pages. |
Yoon, et al., “Locally Adaptive Support-Weight Approach for Visual Correspondence Search”, In Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 20, 2005, 8 Pages. |
“Connected component (Graph Theory)”, Retrieved From: https://en.wikipedia.org/w/index.php?title=Connected_component_(graph_theory)&oldid=841872370, May 18, 2018, 3 Pages. |
“Third Office Action Issued in Chinese Patent Application No. 201480021958.7”, dated Dec. 3, 2018, 10 Pages. |
Number | Date | Country | |
---|---|---|---|
20140307098 A1 | Oct 2014 | US |
Number | Date | Country | |
---|---|---|---|
61812232 | Apr 2013 | US |