This application claims the benefit under Title 35 United States Code §119(e) of U.S. Provisional Patent Application Ser. No. 62/107,326; Filed: Jan. 23, 2015, the full disclosure of which is incorporated herein by reference.
Not applicable
Not applicable
Not applicable
The present invention generally relates to a system and method for image/videos processing, and more specifically to an image/videos quality (including image/videos features) improvement.
This invention also relates generally to a system and method of image/video enhancement/assessment to be used in the following as examples and not limited to 1) transportation systems, 2) medical imaging systems, 3) thermal imaging, 4) security systems, and 5) aerospace applications etc.
Without limiting the scope of the disclosed systems and methods, the background is described in connection with a system and method for image/videos processing.
There are several iterative image enhancement techniques represented in the literature. The prior art attempts to enhance an image in each iteration until satisfying a termination condition or reaching a certain number of iteration. Patent U.S. Pat. No. 8,270,760, modify object space data at each iteration and U.S. Pat. No. 8,111,943 attempts to enhance the sharpening and lightness of the image in each iteration. U.S. Pat. No. 7,162,100 performing a plurality of Richardson and Lucy (RL) iterations of each of the frequency zones to obtain a succession of intermediary enhanced images, wherein a first intermediary enhanced image is obtained prior to a second intermediary enhanced image; WO, 2012142624, determines the pixon map from a variable that is used to update the image in the iteration, i.e., an “updating variable”, and smoothes this updating variable during the iteration. The updated image is usually also further smoothed at the end of the iteration, using the pixon map determined during the iteration. By contrast, the existing pixon methods determine the pixon map from the image after it has been updated and proceed to smooth the image with that pixon map. WO2011120588 A1 For each iteration, the soft-saliency matte value at the current tile position can be used to control at least one of the parameters controlling the selection of tile shape; the parameters of the geometric tile transform; and the relative opacity of the blend of the source image, any auxiliary image, and the current pixel in the output image. US20120155728 the desired image is not reconstructed directly in the iterative process. Instead, an image is reconstructed that yields the desired image when filtered by the de-convolution filter. WO2002005212, iteratively using the image result from a former iteration as the starting point image for a subsequent iteration. A preferred realization involves holding the non-fixed parameter from the first iteration to completion.
Numerous image/videos processing procedures are available for the enhancement of digital image/videos that either embolden or thin features in a digital image/videos. However, there currently doesn't exit any techniques that provide image/videos quality improvement by using feedback methodologies. In general the current invention as feedback image enhancement attempt to enhance the quality of the image in aspect of contrast and color at each iteration. However, the main difference between the current invention and the existing iterative image enhancement is using the information provided in the current loop/iteration for the next image in the next loop/iteration based on the combination of the previous image and distance information between the mentioned image and desired image quality measurement. The desired image quality measurement is the value assigned for any image based on the category of that image, for example dark, light, indoor, outdoor, foggy, hazy, rainy, snowy, and etc.
The innovative method addresses feedback image enhancement systems for depth, color and gray images and videos. The inherent characteristics of the innovative functions provides superior performance over the existing systems and methods known in the art. Embodiments of the invention provide image quality enhancement/assessments by relying on information of image enhancements in each iteration and using that information for the next one, a differentiator from existing iterative enhancement methods.
The features and advantages described in the specification are not all inclusive and, in particular, many additional features and advantages will be apparent to one of ordinary skill in the art in view of the drawings, specification, and claims. Furthermore, the language used in the specification has been principally selected for readability and instructional purposes, and may not have been selected to delineate or circumscribe the inventive subject matter. Illustrative examples show the effectiveness of both in color and gray removal haze in comparison with the existing methods.
Methods and systems are disclosed herein for generating shape detection using an image control system. Also disclosed are methods and systems for generating smart image quality assessments using a swarm intelligence methods or algorithms.
For a more complete understanding of the features and advantages of the present invention, reference is now made to the detailed description of the invention along with the accompanying figures in which:
Described herein is a method and system for image processing. The numerous innovative teachings of the present invention will be described with particular reference to several embodiments (by way of example, and not of limitation).
The main component of the system in embodiments is a computer or computing device configured to perform the steps discussed herein for automatic image enhancement for feedback control. The computing device being comprised of a storage device, memory, and processor.
The storage device is any non-transitory computer-readable storage medium, such as a hard drive, compact disk read-only memory (CD-ROM), DVD, or a solid-state memory device. The memory holds instructions and data used by the processor. The pointing device may be a mouse, track ball, or other type of pointing device, and is used in combination with the keyboard to input data into the computer system or computing device. The graphics adapter displays images and other information on the display. The network adapter couples the computer system to a local or wide area network.
As is known in the art, a computer can have different and/or other components than those described above. In addition, the computer can lack certain components described above. In one embodiment, a computer acting as an image processing server lacks a keyboard, pointing device, graphics adapter, and/or display. Moreover, the storage device can be local and/or remote from the computer (such as embodied within a storage area network (SAN)). As is known in the art, the computer is adapted to execute computer program modules for providing functionality previously described herein. In one embodiment, program modules are stored on the storage device, loaded into the memory, and executed by the processor.
An embodiment proposes an innovation for image enhancement by implementing control theory concepts. Combination of feedback and feed forward loops and making such kind of iterative learning control structure is introduced as a novel image enhancement method. The following block diagram presents a feedback image enhancement is depicted in
In the mentioned image enhancement technique, feed forward loop imparts a local enhancement. While the repetitive loop make the enhancement in global form. The enhanced image at pixel (i,j) is constructed by the following relation:
X(i,j)new=X(i,j)old+U(i,j) Equation 1
The result obtained from the feed forward loop after a run time iteration is considered as the new input signal for the next iteration. The iteration process would be terminated till the error signal converges to zero asymptotically.
The innovative controller enhances images locally and globally. The controller consists of two parts: The first is a nonlinear part, which provides enhancement locally by using equation 8. The mentioned part works as a feed forward strategy to collect the information from the current version of image and attempts to make decision based on achieved data from a pixel and all neighborhoods considered in a block.
The second part makes global enhancement using feedback systems. The quality of enhanced image is calculated by an enhancement measure in iterations and compared with a desired quality defined for the image. The difference used as feedback signal which adjust the mechanism in a way that the error asymptotically approaching zero.
In general the controller signal could be defined as follow:
U(i, j)=f(E(i, j))+{r−EMEE(X)} Equation2
E(i, j)=Σβ=j−wj+wΣα=i−wi+wγα,βX(i+α, j+β) Equation 3
In the above relation it is supposed to consider a (2w+1)×(2w+1) block, f is a nonlinear function; r is a set point and γα,β are constants. For the color image a cube should be considered around the pixel. There are many choice to define as error signals around the center pixel (i,j) as sketched in
The controller formulation for this case is updated as follows:
In the above relation size of cube is considered as (2w+1)×(2w+1))×3
To show the effectiveness of the proposed algorithm, illustrative examples would be considered as follows:
The examples expressed the above considered feedback system image enhancement. Consequently, this section will provide some examples of the results of applying the represented algorithms to the different input images.
To describe the controller design, consider an 3×3 block with the pixel (i, j) set as the center (in general, the block size could be changed and larger block could be assigned). Regards to the mentioned expression, the error signals is defined as follows:
E
1(i, j)=X(i+1, j)−X(i−1, j+1)
E
2(i, j)=X(i, j+1)−X(i+1, j−1)
E(i, j)=E1(i, j)+E2(i, j) Equation 5
where in the above relation the coefficient is expressed as follows:
γ1,0=γ0,1=1 and γ−1,1=γ1,−1=−1
The controller signal is defined as follows:
U(i, j)=E(i, j)+α3Ue
U
e
=r−EMEE(X(i, j))
Where α1, α2 and α3 are arbitrary coefficients that could be optimized by GA. In the above relation, “r” is a set point and EMEE is a measure for enhancement, which could be found in table 2. The results of applying the innovative controller are illuminated in
Various kind of errors could be defined according to the block size considered around the pixel (i,j). The following errors are defined for this example and the results are depicted in
E
1(i, j)=X(i+1, j)−X(i−1, j+1)
E
2(i, j)=X(i, j+1)−X(i+1, j−1)
E
3(i, j)=X(i−1, j)−X(i+1, j−1)
E
4(i, j)=X(i−1, j−1)−X(i+1, j+1) Equation 6
Where in the above relation the coefficient is expressed as follows:
γ1,0=γ0,1=γ−1,0=γ−1,−1=1, γ−1,1=γ1,1=−1 and γ1,−1=−2
In this case a 3×3 block is considered and the error definition is similar to what stated in example 2. To show the effectiveness of the proposed scheme the nonlinear function, f is expressed as follow:
The proposed innovative algorithm could be used for color images. The errors which used for this case are as follows:
E
1(i, j, k)=X(i+1, j, k)−X(i−1, j+1, k)
E
2(i, j, k)=X(i, j+1, k)−X(i+1, j−1, k)
E
3(i, j, k)=X(i−1, j, k)−X(i+1, j−1, k)
E
4(i, j, k)=X(i−1, j−1, k)−X(i+1, j+1, k)
E
5(i, j, k)=α1E1(i, j, k)+α2E2(i, j, k)+α3E3(i, j, k)+α4E4(i, j, k)+α5Ue
In the proposed embodiment, control theory concepts are implemented for image processing application. Combination of sliding mode control and repetitive control structure is introduced as a novel method for image enhancement. The following block diagram presents a feedback image enhancement is illustrated in
X(i, j)new=X(i, j)old+U(i, j)
U(i, j)=α+k×sign(Δε)
Δε=EMEE(Xp
In the above diagram, X(i, j)old represents of the image intensity value for the pixel (i, j), U(i,j) is controller and X(i, j)new is the result of applying controller on pixel(i, j). Xp stands for the image created in the previous iteration. Various enhancement measures which already have been developed (See table 3) or newly developed measures could be used in the proposed algorithm. α and k are two arbitrary constants which could be adjusted based on image category.
The proposed algorithm is applicable for color image. At first the color image should be transformed from RGB to HSV and the represented method applied on “V” channel. Another transformation should be implemented to return the image back in the RGB mode.
The fifth embodiment of the present invention proposed new algorithms for image enhancement by intelligent methods such as fuzzy logic. Intelligent techniques could be combined with the classic image enhancement systems for getting better performances. In the following, a represented intelligent model used for image enhancement has two inputs and one output as illustrated in
X
new(i, j)=Xold(i, j)+Uf
U
f=fuzzy(Eold,Enew)
E
old=Setpoint−Enhance Measure(Xpold)
E
new=Setpoint−Enhance Measure(Xpnew) Equation 9
Xp stands for the image created in the previous iteration. Please note that the above mentioned block has been depicted for one run iteration and it should be repeated until the errors converge to zero.
Among intelligent methods, fuzzy logic systems are very powerful in modeling the error systems, demonstrated in
The rule map and membership functions regards to the fuzzy controller are considered as
The mentioned method is a global image enhancement and is applicable for color and gray scale images.
Corollary 1: Type-II fuzzy could also be used in the proposed algorithm. It needs to replace the membership function with the type-II illustrated in
Fuzzy controller is introduced in this embodiment as a technique for image enhancement locally. The fuzzy inputs are the error signals which are introduced in the third embodiment, equation 9, as E1(i,j), E2(i,j), . . . , En(i,j). The block diagram of the represented algorithm is depicted in
X(i, j)new=X(i, j)old+Ul(i, j)
U
l(i, j)=fuzzy(E1, E2, En) Equation 10
Corollary2: The local fuzzy enhancement algorithm could be used for the color images by the following relations:
X(i, j, k)new=X(i, j, k)old+Ul(i, j, k)
U
l(i, j, k)=fuzzy(E1,E2, . . . , En)
The current embodiment is a process according to the depth image enhancement. A depth map (sometimes referred to as a shadow map) is a texture that holds the depth of each pixel rendered from the perspective of the light source. The proposed method could be used to reconstruct the depth map. Reconstructed Depth Map (RDM) could be achieved as depicted in
X(i, j)new=X(i, j)old+U(i, j)
U(i, j)=f(RDMold(i, j))+{r−EMEE(Xold)} Equation 11
RDM(i, j)=Σβ=j−wi+wγα,βXold(i±α, j±β) Equation 12
In the above diagram, X(i, j)old represents of the image intensity value for the pixel (i, j), U(i, j) is controller and X(i, j)new is the result of applying controller on pixel (i, j).
Where γα, β are arbitrary constants that could be adjusted by GA algorithm with some measure of enhancement as cost function like EME or EMEE.
Photometric stereo is recovering an object's three-dimensional structure from a series of images where the camera location does not change but the lighting location does change. The general idea is that, based on the highlights and shadows of the object, you can recover its shape, similar to how humans perceive shape. In the proposed innovation, it is assumed that there is a photometric stereo image and the goal is to reconstruct the original image.
The RDM block considered for the present example is:
RDM(i, j)=5[X(i+1, j)−X(i−1, j+1)X(i, j+1)−X(i+1, j−1)]
Using the proposed algorithm, there is possibility to have face detection for depth images with more detail. There are many techniques that could detect face from depth images. According to the proposed enhancement not only is the face detection easier but recognizing more details in face is also achieved.
Shape Detection (SD) from depth Images is proposed in the present embodiment. The represented algorithm is different from traditional schemes for shape detection like as shape form shading. The traditional methods attempts to construct 3D image from 2D. The main difference between the proposed and existing schemes is the former make 2D image but there is a kind of depth appears as achievements, (see
X(i, j)new=X(i, j)old+U(i, j)
U(i, j)=f(SDold(i, j))+{r−EMEE(Xold)} Equation 13
SD(i, j)=Σβ=j−wj+wΣα=i−wi+wγα,βXold(i±α, j±β) Equation 14
Observation and recognition of objects in infrared and thermal images are difficult. According to the
Fingerprint image processing is very important for security area. Regards to the
Empirical Mode Decomposition with combination of feedback system introduced in the first embodiment is proposed in the present embodiment. In the
To show the effectiveness of the proposed contribution, thermal images have been examined to enhance by the new method.
A new measure for image enhancement is the process considered in the ninth embodiment. In the current measure, the image is divided to some blocks with k1 rows and k2 columns. kik2 elements make a set of intensity values as:
{x1, . . . , xk1k2} where xi=intensity value of ith pixel
The following relations are defined for the considered intensity set:
A new matrix is expressed for the mentioned block as:
Where wij is the weight between pixel i and j obtained by the following fuzzy system:
The membership function for the inputs and output fuzzy system demonstrated in
By introducing the fuzzy system, new measure for image enhancement is defined as:
Throughout this application, the term “about” is used to indicate that a value includes the standard deviation of error for the device or method being employed to determine the value.
The disclosed system and method of use is generally described, with examples incorporated as particular embodiments of the invention and to demonstrate the practice and advantages thereof. It is understood that the examples are given by way of illustration and are not intended to limit the specification or the claims in any manner.
To facilitate the understanding of this invention, a number of terms may be defined below. Terms defined herein have meanings as commonly understood by a person of ordinary skill in the areas relevant to the present invention.
Terms such as “a”, “an”, and “the” are not intended to refer to only a singular entity, but include the general class of which a specific example may be used for illustration. The terminology herein is used to describe specific embodiments of the invention, but their usage does not delimit the disclosed device or method, except as may be outlined in the claims. Consequently, any embodiments comprising a one component or a multi-component system having the structures as herein disclosed with similar function shall fall into the coverage of claims of the present invention and shall lack the novelty and inventive step criteria.
It will be understood that particular embodiments described herein are shown by way of illustration and not as limitations of the invention. The principal features of this invention can be employed in various embodiments without departing from the scope of the invention. Those skilled in the art will recognize, or be able to ascertain using no more than routine experimentation, numerous equivalents to the specific device and method of use described herein. Such equivalents are considered to be within the scope of this invention and are covered by the claims.
All publications, references, patents, and patent applications mentioned in the specification are indicative of the level of those skilled in the art to which this invention pertains. All publications, references, patents, and patent application are herein incorporated by reference to the same extent as if each individual publication, reference, patent, or patent application was specifically and individually indicated to be incorporated by reference.
In the claims, all transitional phrases such as “comprising,” “including,” “carrying,” “having,” “containing,” “involving,” and the like are to be understood to be open-ended, i.e., to mean including but not limited to. Only the transitional phrases “consisting of” and “consisting essentially of,” respectively, shall be closed or semi-closed transitional phrases.
The system and/or methods disclosed and claimed herein can be made and executed without undue experimentation in light of the present disclosure. While the system and methods of this invention have been described in terms of preferred embodiments, it will be apparent to those skilled in the art that variations may be applied to the system and/or methods and in the steps or in the sequence of steps of the method described herein without departing from the concept, spirit, and scope of the invention.
More specifically, it will be apparent that certain components, which are both shape and material related, may be substituted for the components described herein while the same or similar results would be achieved. All such similar substitutes and modifications apparent to those skilled in the art are deemed to be within the spirit, scope, and concept of the invention as defined by the appended claims.
This application claims the benefit under Title 35 United States Code §119(e) of U.S. Provisional Patent Application Ser. No. 62/107,326; Filed: Jan. 23, 2015, the full disclosure of which is incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
62107326 | Jan 2015 | US |