The present application claims the benefit of priority of Indian patent application number 3723/CHE/2011, entitled “MULTI-SPECTRAL IP CAMERA”, filed Oct. 31, 2011, and Indian patent application number 3724/CHE/2011, entitled “MULTI-SENSOR IP CAMERA WITH EDGE ANALYTICS”, filed Oct. 31, 2011, the entirety of each of which is hereby incorporated herein for all purposes.
The number of sensors used for security applications is increasing rapidly, leading to a requirement for intelligent ways to present information to the operator without information overload, while reducing the power consumption, weight and size of systems. Security systems for military and paramilitary applications can include sensors sensitive to multiple wavebands including color visible, intensified visible, near infrared, thermal infrared and tera hertz imagers.
Typically, these systems have a single display that is only capable of showing data from one camera at a time, so the operator must choose which image to concentrate on, or must cycle through the different sensor outputs. Sensor fusion techniques allow for merging data from multiple sensors. Traditional systems employing sensor fusion operate at the server end, assimilating data from multiple sensors into one processing system and performing data or decision fusion.
Present day camera systems that support multi-sensor options may typically provide two ways of visualizing data from the sensors. One method is to toggle between the sensors based on user input. The other method is to provide a “Picture in Picture” view of the sensor imagery. Toggling can provide a view of only one sensor at any given time. “Picture in Picture” forces the operator to look at two images within a frame and interpret them.
It may be desirable to have means of providing a unified method of visualizing data from multiple sensors in real time. It may be desirable to have such a means within a compact, light-weight package.
Various embodiments allow for real-time fusion of multi-band imagery sources in one tiny, light-weight package, thus offering a real-time multi-sensor camera. Various embodiments maximize scene detail and contrast in the fused output, and may thereby provide superior image quality with maximum information content.
Various embodiments include a camera system that can improve the quality of long-wave infrared (LWIR) and electro-optical (EO) image sensors. Various embodiments include a camera system that can fuse the signals from the LWIR and EO sensors. Various embodiments include a camera system that can fuse such signals intelligently to image simultaneously in zero light and bright daylight conditions. Various embodiments include a camera system that can package the fused information in a form that is suitable for a security camera application.
The following are incorporated by reference herein for all purposes:
U.S. Pat. No. 7,535,002, entitled “Camera with visible light and infrared image blending”, to Johson, et al., filed Jan. 19, 2007; U.S. Pat. No. 7,538,326, entitled “Visible light and IR combined image camera with a laser pointer”, to Johson, et al., filed Dec. 5, 2005; United States Patent Application No. 20100045809, entitled “INFRARED AND VISIBLE-LIGHT IMAGE REGISTRATION”, to Corey D. Packard, filed Aug. 22, 2008; United States Patent Application No. 20110001809, entitled “THERMOGRAPHY METHODS”, to Thomas J. McManus et al, filed Jul. 1, 2010.
The following is incorporated by reference herein for all purposes: Kirk Johnson, Tom McManus and Roger Schmidt, “Commercial fusion camera”, Proc. SPIE 6205, 62050H (2006); doi:10.1117/12.668933
Various embodiments include a multi-resolution image fusion system in the form of a standalone camera system. In various embodiments, the multi-resolution fusion technology integrates features available from all available sensors into one camera package. In various embodiments, the multi-resolution fusion technology integrates features available from all available sensors into one light-weight camera package. In various embodiments, the multi-resolution fusion technology integrates the best features available from all available sensors into one light-weight camera package.
Various embodiments enhance the video feed from each of the input sensors. Various embodiments fuse the complementary features. Various embodiments encode the resultant video feed. Various embodiments encode the resultant video feed into an H.264 video stream. Various embodiments transmit the video feed over a network. Various embodiments transmit the video feed over an IP network.
In various embodiments, the multi-resolution fusion technology integrates the best features available from all available sensors into one light-weight camera package, enhances the video feed from each of the input sensors, fuses the complementary features, encodes the resultant video feed into a H.264 video stream and transmits it over an IP network.
In various embodiments, sensor image feeds are enhanced in real-time to get maximum quality before fusion. In various embodiments, sensor fusion is done at a pixel level to avoid loss of contrast and introduction of artifacts.
In various embodiments, the resultant fused feed is available as a regular IP stream that can be integrated with existing security cameras.
A multi-sensor camera according to some embodiments overcomes the limitations of a single sensor vision system by combining the images from imagery in two spectrums to form a composite image.
A camera according to various embodiments may benefit from an extended range of operation. Multiple sensors that operate under different operating conditions can be deployed to extend the effective range of operation.
A camera according to various embodiments may benefit from extended spatial and temporal coverage. In various embodiments, joint information from sensors that differ in spatial resolution can increase the spatial coverage.
A camera according to various embodiments may benefit from reduced uncertainty. In various embodiments, joint information from multiple sensors can reduce the uncertainty associated with the sensing or decision process.
A camera according to various embodiments may benefit from increased reliability. In various embodiments, the fusion of multiple measurements can reduce noise and therefore improve the reliability of the measured quantity.
A camera according to various embodiments may benefit from robust system performance. In various embodiments, redundancy in multiple measurements can help in systems robustness. In the event that one or more sensors fail or the performance of a particular sensor deteriorates, the system can depend on the other sensors.
A camera according to various embodiments may benefit from compact representation of information. In various embodiments, fusion leads to compact representations. Instead of storing imagery from several spectral bands, it is comparatively more efficient to store the fused information.
Various embodiments include a camera system capable of real-time pixel level fusion of long wave IR and visible light imagery.
Various embodiments include a single camera unit that performs sensor data acquisition, fusion and video encoding.
Various embodiments include a single camera capable of multi-sensor, depth of focus and dynamic range fusion.
Referring to
A first input signal may be received at the LWIR sensor 104, and may include an incident LWIR signal. The first input signal may represent an image captured in the LWIR spectrum. The sensor 104 may register and/or record the signal in digital format, such as an array of bits or an array of bytes. As will be appreciated, there are many ways by which the input signal may be recorded. In some embodiments, the input signal may be registered and/or recorded in analog forms. The signal may then be passed to image enhancement circuitry 108, which may perform one or more operations or transformations to enhance the incident signal.
On a parallel track, a second input signal may be received at the EO sensor 112. The second input signal may include an incident signal in the visible light spectrum. The second input signal may represent an image captured in the visible light spectrum. The sensor 112 may register and/or record the signal in digital format, such as an array of bits or an array of bytes. As will be appreciated, there are many ways by which the input signal may be recorded. In some embodiments, the input signal may be registered and/or recorded in analog forms. The signal may then be passed to image enhancement circuitry 116, which may perform one or more operations or transformations to enhance the incident signal.
It will be appreciated that, whereas a given stage (e.g., LWIR sensor, EO sensor 112, Image Enhancement Circuitry 108, Image Enhancement 116) may operate on a single image at a given instant of time, such sensors may perform their operations repeatedly in rapid succession, thereby processing a rapid sequence of images, and thereby effectively operating on a video.
Image enhancement circuitry 108, and image enhancement circuitry 116 may, in turn, pass their respective output signals to circuitry 120, for the process of video alignment, video fusion, and H.264 encoding.
LWIR sensor 104 may take various forms, as will be appreciated. An exemplary LWIR sensor may include an uncooled microbolometer based on an ASi substrate manufactured by ULIS.
EO sensor 112 may take various forms, as will be appreciated. EO sensor may include a charge-coupled device (CCD), a complementary metal-oxide semiconductor (CMOS) active pixel sensor, or any other image sensor. EO sensor may include a lens, shutter, illumination source (e.g., a flash), a sun shade or light shade, mechanisms and/or circuitry for focusing on a target, mechanisms and/or circuitry for automatically focusing on a target, mechanisms and/or circuitry for zooming, mechanisms and/or circuitry for panning, and/or any other suitable component. An exemplary EO sensor may include a CMOS sensor manufactured by Omnivision.
Image enhancement circuitry 108 may include one or more special purpose processor, such as digital signal processors (DSPs) or graphics processing units. Image enhancement circuitry 108 may include general purpose processors. Image enhancement circuitry 108 may include custom integrated circuits, field programmable gate arrays, or any other suitable circuitry. In various embodiments, image enhancement circuitry 108 is specifically programmed and/or designed for performing image enhancement algorithms quickly and efficiently. Image enhancement circuitry 116 may, in various embodiments, include circuitry similar to that of circuitry 108.
Circuitry 120 may receive input signals from the outputs of image enhancement circuitry 108 and image enhancement circuitry 116. The signals may comprise image signals and/or video signals. The signals may be transmitted to circuitry 120 via any suitable connector or conductor, as will be appreciated. Circuitry 120 may then perform one or more algorithms, processes, operations and/or transformations on the input signals.
Processes performed may include video alignment, which may ensure that features present in the respective input signals are properly aligned for combination. As will be appreciated, signals originating from LWIR sensor 104 and from EO sensor 112 may both represent captured images and/or videos of the same scene. It may thus be desirable that these two images and/or videos are aligned, so that information about a given feature in the scene can be reinforced from the combination of the two signals.
In some embodiments, as the LWIR sensor 104 and EO sensor 112 may be at differing physical positions, the scene captured by each will be from slightly differing vantage points, and may thus introduce parallax error. The process of video alignment may seek to minimize and/or correct this parallax error, in some embodiments.
Circuitry 120 may also be responsible for video fusion, which may include combining the two signals originating from the respective sensors into a single, combined signal. In various embodiments, the combined signals may contain more information about the captured scene than do one or either of the original signals.
Circuitry 120 may also be responsible for video encoding, which may include converting the combined video signal into a common or recognized video format, such as the H.264 video format.
Circuitry 120 may output one or more video signals, which may include a video signal in common format, such as an H.264 video signal. In some embodiments, circuitry 120 may include a port or interface for linking to an internet protocol (IP) network. The circuitry 120 may be operable to output a video signal over an IP network.
In various embodiments, camera 100 may include one or more additional components, such as a view finder, viewing panel (e.g., a liquid crystal display panel for showing an image or a fused image of the camera), power source, power connector, memory card, solid state drive card, hard drive, electrical interface, universal serial bus connector, sun shade, illumination source, flash, and any other suitable component. Components of camera 100 may be enclosed within, and/or attached to a suitable housing, in various embodiments. Whereas various components have been described as separate or discrete components, it will be appreciated that, in various embodiments, such components may be physically combined, attached to the same circuit board, part of the same integrated circuit, utilize common components (e.g., common processors; e.g., common signal busses), or otherwise coincide. For example, in various embodiments, image enhancement circuitry 108 and image enhancement circuitry 116 may be one and the same, and may be capable of simultaneously or alternately operating on input signals from both the LWIR sensor 104 and from the EO sensor 112.
It will be appreciated that certain components that have been described as singular may, in various embodiments, be broken into multiple components. For example, in some embodiments, circuitry 120 may be instantiated over two or more separate circuit boards, utilize two or more integrated circuits or processors, and so on. Where there are multiple components, such components may be near or far apart in various embodiments.
Whereas various embodiments have described LWIR and EO sensors, it will be appreciated that other types of sensors may be used, and that sensors for other portions of the electromagnetic spectrum may be used, in various embodiments.
Referring to
Various embodiments utilize hardware on an FPGA system with DSP coprocessors. In some embodiments, the multi-sensor camera performs algorithms on a Texas Instruments DaVinci chip.
In various embodiments, a hardware implementation allows for an advantageously light camera. In various embodiments, a camera weighs in the vicinity of 1.2 kg. The camera may minimize weight by utilizing a light-weight LWIR sensor, and/or by utilizing a light-weight DSP board that performs both video capture and processing on a single board.
Referring to
Initially, input signals may come from sensor 304, and from sensor 308. These may correspond respectively to LWIR sensor 104, and to EO sensor 116. However, as will be appreciated, other types of sensors may be used, in various embodiments (e.g., sensors for different portions of the spectrum). In various embodiments, input signals may be derived from other sources. For example, input signals may be derived over a network or from an electronic storage medium. For example, the input signals may represent raw, pre-recorded video signals.
In various embodiments, there may be more than two input signals. For example, there may be three or more input signals, each stemming from a different sensor. In some embodiments, input sensors may include a short wave infrared (SWIR) sensor, a LWIR sensor, and a visible light sensor.
At step 312, a process of image enhancement may be performed. Image enhancement may include altering or increasing sharpness, brightness, contrast, color balance, or any other aspect of the image. Image enhancement may be performed via digital manipulation, e.g., via manipulation of pixel data. In some embodiments, image enhancement may occur via manipulation of analog image data. In some embodiments, image enhancement may include the application of one or more filters to an image. In various embodiments, image enhancement may include the application of any algorithm or transformation to the input image signal. As will be appreciated, image enhancement, when applied to frames of a video signal, may include video enhancement.
At step 316, a process of image alignment may occur. Image alignment may operate on image signals originating, respectively, from image enhancement circuitry 108, and from image enhancement circuitry 116. In the process of image alignment, two separate images may be compared. Common signals, features, colors, textures, regions, patterns, or other characteristics may be sought between the two images. A transformation may then be determined which would be necessary to bring such common signals, features, etc., into alignment. For example, it may be determined that shifting a first image a certain number of pixels along a notional x-axis and y-axis may be sufficient to align the first image with a second image that is also presumed to fall within the same coordinate system. As will be appreciated, in various embodiments, other transformations may be utilized in the process of image alignment. For example, transformations may include shifting, rotating, or scaling.
At step 320, video fusion may be performed. Video fusion may include combining images from each of two input video streams. Such input video streams may consist of images that have been aligned at step 316. Video fusion may be performed in various ways, according to various embodiments. In some embodiments, data from two input images may be combined into a single image. The single image may contain a better representation of a given scene than do one or both of the input images. For example, the single image may contain less noise, finer detail, better contrast, etc. The process of video fusion may include determining the relative importance of the input images, and determining an appropriate weighting for the contribution of the respective input images. For example, if a first input image contains more detail than does a second input image, then more information may be used from the first image than from the second image in creating the fused image.
In various embodiments, a weighting determination may be made on more localized basis than on an entire image. For example, a certain region of a first image may be deemed more important than an analogous region of a second image. However, another region of the first image may be deemed less important than its analogous region in the second image. Thus, different regions of a given image may be given different weightings with respect to their contribution to a fused image. In some embodiments, weightings may go down to the pixel level. In some embodiments, weightings may be applied to images in some transform domain (e.g., in a frequency domain). In such cases, relative contributions of the two images may differ by frequency (or other metric) in the transform domain.
In various embodiments, other methods may be used for combining or fusing images and/or videos.
In various embodiments a fusion algorithm may be used for different wavelengths, different depths of field and/or different fields of view.
In various embodiments, a determination may be made as to whether or not a sensor is functional, and/or whether or not the sensor is functioning properly. If the sensor is not functioning properly, or not functioning at all, then video input from that sensor may be disregarded. For example, video input from the sensor may be omitted in the fusion process, and the fusion process may only utilize input from remaining sensors.
In various embodiments, an image quality metric is derived in order to determine if input from a given sensor is of good visual quality. In various embodiments, the image quality metric is a derivative of the singular value decomposition of local image gradient matrix, and provides a quantitative measure of true image content (i.e., sharpness and contrast as manifested in visually salient geometric features such as edges,) in the presence of noise and other disturbances. This measure may have various advantages in various embodiments. Advantages may include that the image quality metric 1) is easy to compute, 2) reacts reasonably to both blur and random noise, and 3) works well even when the noise is not Gaussian.
In various embodiments, the image quality metric may be used to determine whether or not input from a given sensor should be used in a fused video signal.
At step 324, video encoding may be performed. Video encoding may be used to compress a video signal, prepare the video signal for efficient transmission, and/or to convert the signal into a common, standard, or recognized format that can be replayed by another device. The process of video encoding may convert the fused video signal into any one or more known video formats, such as MPEG-4 or H.264. Following the encoding process, an output signal may be generated that is available for transmission, such as for transmission over an IP network.
In various embodiments, some portion or segment of fused video data may be stored prior to transmission, such as transmission over an IP network. In some embodiments, fused video data is transmitted immediately, and little or no data may be stored. In various embodiments, some portion or segment of encoded video data may be stored prior to transmission, such as transmission over an IP network. In some embodiments, encoded video data is transmitted immediately, and little or no data may be stored.
Whereas
In various embodiments, more or fewer steps may be performed than are shown in
While fusing data from different sources, it may be desirable to preserve the more significant detail from each of the video streams on a pixel by pixel basis. An easy combination of the video streams is to perform an averaging function of the two video streams. However, contrast is reduced significantly and sometimes detail from one stream cancels detail from the other stream. The Laplacian pyramid fusion on the other hand may provide excellent automatic selection of the important image detail for every pixel from both images at multiple image resolutions. By performing this selection in the multiresolution representation, the reconstructed—fused—image may provide a natural-looking scene.
In addition, the Laplacian pyramid fusion algorithm allows for additional enhancement of the video. It can provide multi-frequency sharpening, contrast enhancement, and selective de-emphasis of image detail in either video source.
Laplacian pyramid fusion is a pattern selective fusion method that is based on selecting detail from each image on a pixel by pixel basis over a range of spatial frequencies. This is accomplished in three basic steps (assuming the source images have already been aligned). First, each image is transformed into a multiresolution, bandpass representation, such as the Laplacian pyramid. Second, the transformed images are combined in the transform domain—i.e. combine the Laplacian pyramids on a pixel by pixel basis. Finally, the fused image is recovered from the transform domain through an inverse transform—i.e. Laplacian pyramid reconstruction.
The Laplacian pyramid is derived from a Gaussian pyramid. The Gaussian pyramid is obtained by sequence of filter and subsample steps. First a low pass filter is applied to the original image G0. The filtered image is then subsampled by a factor of two providing level 1 of the Gaussian pyramid, G1. The subsampling can be applied since the spatial frequencies have been limited to half the sample frequency. This process is repeated for N levels computing G2 . . . GN.
The Laplacian pyramid is obtained by taking the difference between each of the Gaussian pyramid levels. These are often referred to as DoG (difference of Gaussians). So Laplacian level 0 is the difference between G0 and G1. Laplacian level 1 is the difference between G1 and G2. The result is a set of bandpass images where L0 represents the upper half of the spatial frequencies (all the fine texture detail), L1 represents the frequencies between ¼ and ½ the full bandwidth, L2 represents the frequencies between ⅛ and ¼ the full bandwidth, etc.
This recursive computation of the Laplacian pyramid is a very efficient method for computing effectively very large filters with one small filter kernel.
Further, the Laplacian pyramid plus the lowest level of the Gaussian pyramid, represent all the information of the original image. So an inverse transform that combines the lowest level of the Gaussian pyramid with the Laplacian pyramid images, can reconstruct the original image exactly.
When using the Laplacian pyramid representation as described above, certain dynamic artifacts in video scenes will be noticeable. This often manifests itself as “flicker” around areas with reverse contrast between the image. This effect is magnified by aliasing that has occurred during the subsampling of the images.
Double density Laplacian pyramids are computed using double the sampling density of the standard Laplacian pyramid. This requires larger filter kernels, but can still be efficiently implemented using the proposed hardware implementation in the camera. This representation is essential in reducing the image flicker in the fused video.
Most video sources are represented as an interlaced sequence of fields. RS170/NTSC video has a 30 Hz frame rate, where each frame consists of 2 fields that are captured and displayed 1/60 sec. apart. So the field rate is 60 Hz. The fusion function can operate either on each field independently, or operate on full frames. By operating on fields there is vertical aliasing present in the images, which will reduce vertical resolution and increase image flicker in the fused video output. By operating the fusion on full frames, the flicker is much reduced, but there may be some temporal artifacts visible in areas with significant image motion.
If the images are the same size, the images are reduced at step 512. The images may be reduced by sub-sampling of the images. In some embodiments, a filtering step is performed on the images before sub-sampling (e.g., a low pass filter is applied to the image before sub-sampling). The reduced images are then expanded at step 514. The resultant images will represent the earlier images but with less detail, as the sub-sampling will have removed some information.
At step 516, pyramid coefficients of the actual level for both images are calculated. Pyramid coefficients may represent possible weightings for each of the respective images in the fusion process. Pyramid coefficients may be calculated in various ways, as will be appreciated. For example, in some embodiments, coefficients may be calculated based on a measure of spatial frequency detail and/or based on a level of noise.
At step 518, maximum coefficients are chosen, which then results in fused level L.
At step 520, it is determined whether or not consistency is on. Consistency may be a user selectable or otherwise configurable setting, in some. In some embodiments, applying consistency may include ensuring that there is consistency among chosen coefficients at different iterations of process flow 500. Thus, for example, in various embodiments, applying consistency may include altering the coefficients determined at step 518. If consistency is on, then flow proceeds to step 522, where consistency is applied. Otherwise, step 522 is skipped.
At step 524, a counter is decreased. The counter may represent the level of recursion that will be carried out in the fusion process. For example, the counter may represent the number of levels of a Laplacian or Gaussian pyramid that will be employed. If, at 526, the counter has not yet reached zero, then the algorithm may run anew on reduced image 1528, and reduced image 2530, which may become image 1502, and image 2504, for the next iteration. At the same time, the fused level L may be added to the overall fused image 536 at step 534. If, on the other hand, the counter has reached zero at step 526, then flow proceeds to step 532, where the fused level becomes the average of the reduced images. This average is in turn combined with the overall fused image 530.
Ultimately, upon completion of all levels of recursion of the algorithm, the fused image 530 will represent the separately weighted contributions of multiple different pyramid levels stemming from original image 1 and original image 2.
Whereas
It will be appreciated that, whereas certain algorithms are described herein, other algorithms are also possible and are contemplated. For example, in various embodiments other algorithms may be used for one or more of image enhancement and fusion.
As will be appreciated, in various embodiments, alternative hardware implementations and components are possible. In various embodiments, certain components may be combined, or partially combined. In various embodiments, certain components may be separated into multiple components, which may divide up the pertinent functionalities.
Because the fusion function operates in the Laplacian pyramid transform domain, several significant image enhancement techniques may be readily performed, in various embodiments.
Various embodiments may employ a technique to make video look sharper by boosting the high spatial frequencies. This may be accomplished by adding a gain factor to Laplacian level 0. This “sharpens” the edges and fine texture detail in the image.
Since the Laplacian pyramid consists of several frequency bands, various embodiments contemplate boosting the lower spatial frequencies, which effectively boosts the image contrast. Note that peaking often results in boosting noise also. So the Laplacian pyramid provides the opportunity to boost level 1 instead of level 0, which often boosts the important detail in the image, without boosting the noise as much.
In various embodiments, the video from each of the sensors (e.g., sensors 104 and 112) is enhanced before it is presented to the fusion module. The fusion system accepts the enhanced feeds and then fuses the video.
In various embodiments, the input feeds may be fused first and then the resultant video may be enhanced.
In various embodiments, the fusion process combines the video data on each of the Laplacian pyramid levels independently. This provides the opportunity to control the contribution of each of the video sources for each of the Laplacian levels.
For example, if the IR image does not have much high spatial frequency detail, but has a lot of noise, then it is effective to reduce the contribution at L0 from the IR image. It is also possible that very dark regions of one video source reduce the visibility of details from the other video source. This can be compensated for by changing the contribution of the lowest Gaussian level.
The following are incorporated by reference herein for all purposes:
U.S. Pat. No. 5,912,993, entitled “Signal encoding and reconstruction using pixons”, to Puetter, et al., filed Jun. 8, 1993; U.S. Pat. No. 6,993,204, entitled “High speed signal enhancement using pixons”, to Yahil, et al., filed Jan. 4, 2002; United States Patent Application No. 20090110321, entitled “Determining a Pixon Map for Image Reconstruction”, to Vija, et al., filed Oct. 31, 2007
The following are incorporated by reference herein for all purposes:
The following are incorporated by reference herein for all purposes:
The following are incorporated by reference herein for all purposes:
The following are embodiments, not claims:
A. A camera comprising:
In various embodiments, data is interlaced, so there may be two ways the fusion could happen. One is to separately fuse each field, and the other is to fuse based on the full frame, in various embodiments
A.y The camera of embodiment A in which the first aligned data comprises a first field and a second field that are interlaced, and in which the second aligned data comprises a third field and a fourth field that are interlaced.
A.y.1 The camera of embodiment A.y in which, in performing video fusion, the circuitry is operable to fuse the first field and the third field, and to separately fuse the second field and the fourth field.
A.y.2 The camera of embodiment A.y in which, in performing video fusion, the circuitry is operable to fuse the full frames of the first aligned data and the second aligned data.
In various embodiments, the image may be sharpened.
A.11 The camera of embodiment A in which, in performing video fusion, the circuitry is operable to apply a sharpening algorithm to result in increased sharpness in the fused data.
A.11.1 The camera of embodiment A, in which the sharpening algorithm includes boosting high spatial frequencies in the first enhanced data and in the second enhanced data.
A.11.2 The camera of embodiment A, in which the sharpening algorithm includes performing a Laplacian pyramid fusion algorithm and adding a gain factor to Laplacian level 0.
In various embodiments, contrast may be enhanced.
A.12 The camera of embodiment A in which, in performing video fusion, the circuitry is operable to apply a contrast enhancing algorithm to result in increased contrast in the fused data.
A.12.1 The camera of embodiment A, in which the contrast enhancing algorithm includes performing a Laplacian pyramid fusion algorithm and adding a gain factor to Laplacian level 1.
In various embodiments, there may be selective contribution of the first enhanced data and the second enhanced data.
A.13 The camera of embodiment A in which, in performing video fusion, the circuitry is operable to weight the contributions of the first enhanced data and the second enhanced data to the fused data.
In various embodiments, it is determined how to weight the contribution of the first enhanced data based on some detail.
A.13.1 The camera of embodiment A in which, in performing video fusion, the circuitry is further operable to determine a level of detail in the first enhanced data, in which the contribution of the first enhanced data is weighted based on the level of detail.
In various embodiments, it is determined how to weight the contribution of the first enhanced data based on spatial frequency detail.
A.13.2 The camera of embodiment A in which, in performing video fusion, the circuitry is further operable to determine a level of spatial frequency detail in the first enhanced data, in which the contribution of the first enhanced data is weighted based on the level of spatial frequency detail.
In various embodiments, it is determined how to weight the contribution of the first enhanced data based on noise.
A.13.3 The camera of embodiment A in which, in performing video fusion, the circuitry is further operable to determine a level of noise in the first enhanced data, in which the contribution of the first enhanced data is weighted based on the level of noise.
In various embodiments, it is determined how to weight the contribution of the first enhanced data based on the presence of dark regions.
A.13.4 The camera of embodiment A in which, in performing video fusion, the circuitry is further operable to determine an existence of dark regions in the first enhanced data, in which the contribution of the first enhanced data is weighted based on the existence of the dark regions.
A.5 The camera of embodiment A in which, in generating the encoded data, the circuitry is operable to generate the encoded data using the discrete cosine transform algorithm.
A.5 The camera of embodiment A in which, in generating the encoded data, the circuitry is operable to generate an H.264 encoded internet protocol stream.
In various embodiments, the camera can enhance data in real time.
A.6 The camera of embodiment A, in which the circuitry is operable to generate the first enhanced data, the second enhanced data, the first aligned data, the second aligned data, the fused data, and the encoded data, each in real time.
In various embodiments, the camera can enhance data at a rate of 30 frames per second.
A.7 The camera of embodiment A, in which the circuitry is operable to generate the first enhanced data, the second enhanced data, the first aligned data, the second aligned data, the fused data, and the encoded data, each at a rate of at least 30 frames per second.
In various embodiments, the camera can enhance data at a rate of 60 frames per second.
A.8 The camera of embodiment A, in which the circuitry is operable to generate the first enhanced data, the second enhanced data, the first aligned data, the second aligned data, the fused date, and the encoded data, each at a rate of at least 60 frames per second.
A.z The camera of embodiment A in which the circuitry comprises a field programmable gate array system with digital signal processing coprocessors.
A.q The camera of embodiment in which the circuitry comprises a Texas Instruments DaVinci chip.
In various embodiments, there may be multiple stages of circuitry, each with separate functions.
A.w The camera of embodiment A in which the circuitry comprises:
In various embodiments, where one sensor fails, another may be used.
B. A camera comprising:
Number | Date | Country | Kind |
---|---|---|---|
3723/CHE/2011 | Oct 2011 | IN | national |
3724/CHE/2011 | Oct 2011 | IN | national |