This application is related to U.S. application Ser. No. 10/826,108, entitled “AUTOMATIC RED EYE REMOVAL,” filed Apr. 16, 2004, which is commonly assigned and incorporated by reference in its entirety.
This invention relates to a method for red eye removal in photographs and images.
The pupil is an opening that lets light into the eye. Since most of the light entering the eye does not escape, the pupil appears black. In dim light, the pupil expands to allow more light to enter the eye. In bright light, the pupil contracts to allow less light to enter the eye.
“Red eye” is a phenomenon where a person's pupils appear red in a photograph taken with a flash. Red eye comes from light reflecting off of the blood vessels in the retinas (on the back interior of the eyeballs).
Some cameras have a “red eye reduction” feature. In these cameras, the flash goes off twice—once right before the picture is taken, and then again to actually take the picture. The first flash causes people's pupils to contract, thereby reducing red eye significantly.
Some photography software have a “red eye removal” feature. These software require the user to identify the red eye, usually by dragging a rectangular box around the red eye with a mouse, and then remove red eye from the specified area. Others software, such as those available from Pixology of Guildford, England, and FotoNation of San Francisco, Calif., require little or no user intervention and the software identifies the red eye automatically.
In addition, there are literatures that describe methods for red eye removal. Of particular interest is U.S. Patent Application Publication 2002/0176623, filed Mar. 29, 2002 (“Steinberg”). Steinberg describes a 3-step process: (1) search candidate areas by color constraints; (2) eliminate candidates with shape and other criteria; and (3) output the results for user interactive verification of the red eye candidates. However, this process has been widely used and is processing common sense.
For example, in an article by de la Escalera et al., a process for detecting road traffic signs is disclosed. de la Escalera et al., “Road Traffic Sign Detection and Classification,” IEEE Trans. on Industrial Electronics, Vol. 44, No. 6, December 1997. de la Escalera et al. discloses two steps: (1) localize the sign in the image depending on the color and the form; and (2) recognize the sign through a neural network.
In another example, in an article by Yang et al., a process for face detection is disclosed. “Detecting Faces in Images; A Survey,” Yang et al., IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol. 24, No. 1, January 2002. Yang et al. states, “Most of [software] utilize global features such as skin color, size, and shape to find face candidates, and then verify these candidates using local, detailed features such as eye brows, nose, and hair. A typical approach begins with the detection of skin-like regions. Next, skin-like pixels are grouped together using connected component analysis or clustering algorithms. If the shape of a connected region has an elliptic or oval shape, it becomes a face candidate. Finally, local features are used for verification.” Id. at p. 40.
Use of the same reference numbers in different figures indicates similar or identical elements.
In one embodiment of the invention, a method for removing red eyes includes (1) identifying potential red eye pixels in a low resolution image generated from a high resolution image (hereafter “LR red eye pixels”), (2) grouping contiguous LR red eye pixels into red eye regions (herafter “LR red eye regions”), (3) determining working areas around the LR red eye regions (hereafter “LR working areas”), (4) determining classifiers from the LR working areas, (5) retrieving working areas from the high resolution image that correspond to the LR working areas (hereafter “HR working areas”), (6) applying the classifiers to the HR working areas to identify potential red eye pixels (hereafter “HR red eye pixels”), (7) grouping contiguous HR red eye pixels into red eye regions (hereafter “HR red eye regions”), (8) determining some of the HR red eye regions as the red eyes, and (9) replacing the HR red eye pixels in the red eyes with black pixels. By doing some of the processing involved in red eye removal on the low resolution image, the number of computational steps is reduced and the required memory used in these computational steps is also reduced.
In step 102, the software determines a weighted purple-red value of each pixel in image 10. The weighted purple-red value represents the likelihood that a pixel is a purple-red pixel that forms part of a red eye. In one embodiment for the rgb (red, green, and blue) color space, the weighted purple-red value of a pixel is determined from its red, green, and blue color values, and its luminance as follows:
where ƒ1 is the weighted purple-red value, r is the red color value, g is the green color value, b is the blue color value, c1(1) is a weight given to the red color value, c2(1) is a weight given to the green color value, c3(1) is a weight given to the blue color value, and Y is the luminance calculated as Y=a1r+a2g+a3b. Note that the weighted red value is independent of luminance change. In one embodiment, c1(1) is 0.5, c2(1) is 0.5, c3(1) is −1, and Y=0.299r+0.587g+0.114b. In other words, equation (1) can be rewritten as:
In one embodiment for the YCrCb (luminance, red chrominance, and blue chrominance) color space, the weighted purple-red value of a pixel is determined as follows:
where Cr is the red chrominance, Cb is the blue chrominance, and Y is the luminance.
In step 104, the software selects the pixels in image 10 that have weight purple-red values greater than a threshold as red eye pixels. In one embodiment, this threshold is 0.5.
In step 106, the software determines a weighted orange-red value of each pixel in image 10. The weighted orange-red value represents the likelihood that a pixel is an orange-red pixel that forms part of a red eye. In one embodiment for the rgb color space, the weighted orange-red value for a pixel is determined from its red, green, and blue color values as follows:
where ƒ2 is the weighted orange-red value, c1(2) is a weight given to the red color value, c2(2) is a weight given to the green color value, and c3(2) is a weight given to the blue color value. As noted above, the weighted red value is independent of luminance change. In one embodiment, c1(2) is 0.6667, c2(2) is 0.3333, c3(2) is −1.0. In other words, equation (4) can be rewritten as:
In one embodiment for the YCrCb color space, the weighted orange-red value of a pixel is determined as follows:
In step 108, the software selects the pixels in image 10 that have weighted orange-red values greater than another threshold as red eye pixels. In one embodiment, this threshold is 0.5.
In step 110, the software groups contiguous red eye pixels into red eye regions. Red eye pixels hereafter refer to the purple-red and orange-red pixels selected in steps 104 and 108. A red eye region can contain purple-red pixels, orange-red pixels, or a combination thereof.
In step 112, the software determines if each red eye region is a substantially round pupil. In step 114, the software rejects the red eye pixels in each red eye region that is not a substantially round pupil. Steps 112 and 114 are illustrated in reference to red eye region 204 in
First, the software determines the geometric center 206 of red eye region 204. The software then determines the distance Dmax from geometric center 206 to the farthest red eye pixel 202F. Distance Dmax is used to set a range of radii of circles where the weighted red values of corresponding pixels located on two adjacent circles are compared. This is illustrated in reference to
The software first determines the sum of the weighted red values of the pixels located at a first circle having a radius Ri as follows:
where S(Ri) is the sum of the first circle, Max(ƒ1,ƒ2) is a function that outputs the greater of a weighted purple-red value ƒ1 and a weighted orange-red value ƒ2, and Ri and θ are the polar coordinates of a pixel.
The software then determines the sum of the weight red values of the pixels located at an adjacent second circle having a radius Ri+1.
where S(Ri+1) is the sum of the second adjacent circle. In one embodiment, the radius Ri+1 is radius Ri incremented by 1 pixel, and angles θ consists 0 to 360° increased at 4° increments (e. 0, 4, 8 . . . 356).
The software then determines the difference between the two sums: S(Ri)−S(Ri+1). If the absolute value of the difference is small, then there has not been a change in the red color between the pixels on the adjacent circles, which indicates that the image has not transitioned from a pupil having red eye to the eye (e.g., from pupil 14 to eye 16). If the difference is positive, then there has been a decrease in the red color between the pixels on the adjacent circles, which may indicate that the image has transitioned from a pupil having red eye to the eye.
In one embodiment, this process is repeated for radii ranging from 0.5*Dmax to 1.5*Dmax where the radius is incremented at 1 pixel between adjacent circles. After the process is completed for these radii, the software selects the radius of the circle that generates the largest difference with its adjacent circle as the radius of a pupil (hereafter “Rpupil”) having red eye region 204.
Referring to
where R1 is the first ratio, NR
The software also determines a second ratio between (1) the number of red eye pixels located within a ring 224 having an inner radius of Rpupil and an outer radius of 1.5* Dmax and (2) the area of ring 224 in pixels as follows:
where R2 is the second ratio, NR
The software then determines a difference between the first ratio and the second ratio: R1-R2. A large difference indicates that red eye region 204 is probably a pupil with red eye. If the difference is less than a threshold, the software then rejects the red eye pixels in red eye region 204 as candidates for red eye removal. In one embodiment, the threshold is 30% (i.e., 0.3).
In step 116, the software determines if each remaining red eye region is too close to another red eye region. In step 118, the software rejects red eye pixels in each red eye region that is too close to another red eye region. Steps 116 and 118 are illustrated in reference to red eye regions 204 and 242 in
The software determines a distance Dpupil between the geometric centers of red eye regions 204 and 242. The software then determines if distance Dpupil is within a range that makes red eye regions 204 and 242 too close. In one embodiment, red eye regions are too close if they are within 10*Rpupil to 14*Rpupil of each other. The software thus rejects the red eye pixels in red eye regions 204 and 242 as candidates for red eye removal if they are too close to each other.
In step 120, the software determines if each remaining red eye region is proximate to a facial region (e.g., face 20 in
The software first generates a histogram 262 of the pixels located in a ring 264 about geometric center 206. The function of histogram 262 is simply the number of pixels that has a particular color (e.g., a particular combination of Y,Cr,Cb color values). In one embodiment, ring 264 has an inner radius of 4*Rpupil and an outer radius of 9*Rpupil. The software then selects the most common color 266 in histogram 262 and compares it to a range of threshold skin color values. If color 266 is within the range of threshold skin color values, then red eye region 204 is probably a pupil with red eye that is proximate to a facial region. In one embodiment, the threshold skin color values are expressed in HSV (hue, saturation, value) color space as —80<H<50, 5<S<80, and 20<V<80. Thus, the software first converts the most common color 266 into HSV color space and then compares it with the threshold skin color values. If color 266 is not within the range of threshold skin color values, the software rejects the pixels in red eye region 204 as candidates for red eye removal. In order to remove the luminance change, the luminance (Y) of the image within the circle having radius 9*Rpupil will be normalized to [0,255] before the software generates histogram 262 and compares color 266 to the range of threshold skin color values. This is because without the normalization, any luminance change will introduce a color cast (i.e., unwanted color effect) into the HSV color space.
In step 124, the software determines if each remaining red eye region is proximate to a sclera of an eye (e.g., sclera 18 in
The software generates a luminance histogram 282 of the pixels located in a ring 284 about geometric center 206. In one embodiment, ring 284 has an inner radius of 2*Rpupil and an outer radius of 5*Rpupil . From histogram 282, the software determines a ratio between (1) the number of pixels having the brightest color 288 in ring 284 and (2) the number of red eye pixels in a circle 286 having a radius of 2* Rpupil as follows:
where Rsclera is the ratio, Nbrightest is the number of pixels having the brightest color in ring 284, and N2*R
If the ratio is greater than a sclera threshold, then red eye region 204 is probably a pupil with red eye proximate to the sclera. If the ratio is less than the threshold, then the software rejects the red eye pixels in red eye region 204. In one embodiment, the sclera threshold is 82% (e.g., 0.82).
In step 128, the software replaces the remaining red eye pixels with black pixels to remove the red eye.
In step 1102, the software generates a low resolution image 10A (
In step 1104, the software performs steps 102 to 110, 116, and 118 described above to low resolution image 10A instead of image 10 at full resolution. This saves the computational steps that would otherwise be performed for all the pixels in image 10. Furthermore, this saves the RAM that would otherwise be required for performing the computational steps for all the pixels in image 10.
Specifically, in steps 102 and 104, the software selects pixels in low resolution image 10A that have weighted purple-red values greater than a threshold as red eye pixels (also referred to as “LR red eye pixels”). In steps 106 and 108, the software selects pixels in image 10A that have weighted orange-red values greater than another threshold as red eye pixels (also referred to as “LR red eye pixels”). The thresholds for steps 104 and 108 are selected so they are less discriminatory in order to capture all red eye pixels and some non-red eye pixels. At step 110, the software groups contiguous red eye pixels in image 10A into red eye regions (also referred to as “LR red eye regions”).
In step 1106, the software determines working areas around the red eye regions in low resolution image 10A (also referred to as “LR working area”). Each LR working area is centered at the corresponding red eye region. Each LR working area has a width of twice the maximum width (Wmax) of the corresponding red eye region, and a height of twice the maximum height (Hmax) of the corresponding red eye region. For simplicity, only one LR working area 1202 and one red eye region 1204 has been labeled in
The software then uses the selected red eye pixels and the non-red eye pixels in each LR working area to train a Bayesian classifier for that LR working area. The Bayesian classifier is later used to determine red eye pixels in a corresponding working area in image 10 (also referred to as “HR working area”). The Bayesian classifier for a LR working area is defined as follows:
P(x, y)=P(x|y)P(y)=P(y|x)P(x), (12)
wherein P(x) is the probability of a pixel taking a weighted red value of x, P(y) is the probability of a pixel being a red eye pixel (i.e., P(y)=1) or a non-red eye pixel (i.e., P(y)=0), P(x,y) is the joint probability that a pixel taking a weighted red value of x and a red-eye pixel value of y, P(x|y) is the conditional probability that a pixel taking a weighted red value of x for a given a red-eye pixel value of y, and P(y|x) is the conditional probability that a pixel taking a weighted red value of y for a given red-eye pixel value of x. In one embodiment, the weighted red value of a pixel is the greater of the purple-red value and the orange-red value of that pixel.
From Equation 12, the probability for a pixel taking a weighted red value of x when that pixel is a red eye pixel can be calculated as follows:
P(x,y=1)=P(x|y=1)P(y=1)=P(y=1|x)P(x) (13)
which can be rewritten as:
From Equation 12, the probability for a pixel taking a weighted red value of x when that pixel is a non-red eye pixel x can be calculated as follows:
P(x,y=0)=P(x|y=0)P(0)=P(y=0|x)P(x), (15)
which can be rewritten as:
In one embodiment of the invention, a pixel is assumed to be a red eye pixel when P(y=1|x)>P(y=0|x). This can be rewritten from Equations 14 and 16 as follows:
which can be rewritten as:
P(x|y=1)P(y=1)>P(x|y=0)P(y=0). (18)
Thus, the software trains the Bayesian classifiers for the LR working areas by determining probabilities P(x|y=1), P(y=1), P(x|y=0), and P(y=0) for each LR working area.
Probability P(x|y=1) can be determined from a histogram of the red eye pixels in the LR working area as shown in
Probability P(y=1) is the total number of red eye pixels divided by the total number of pixels in the LR working area. Similarly, probability P(y=0) is the total number of non-red eye pixels divided by the total number of pixels in the LR working area.
In step 1108, the software retrieves HR working areas in image 10 that correspond to LR working areas in low resolution image 10A.
In step 1110, the software determines the weighted red values of the pixels in each of the HR working area. As discussed above, the weighted red value is the greater of the purple-red value and the orange-red value of that pixel in one embodiment of the invention. From the weighted red values and the determined probabilities for each LR working area, the software applies Equation 18 to identify (e.g., flag) the red eye pixels in the corresponding HR working area (also referred to as “HR red eye pixels”). The software then groups contiguous red eye pixels in the HR working areas search area into red eye regions.
In step 1112, the software performs steps 112, 114, and 120 to 128 described above. Specifically, in steps 112 and 114, the software rejects red eye regions that are not round. In steps 120 and 122, the software rejects red eye regions that are not proximate to a facial region within their HR working areas. In steps 124 and 126, the software rejects red eye regions that are not proximate to a sclera of an eye within their working areas. In step 128, the software replaces the red eye pixels in the remaining red eye regions with black pixels.
Various other adaptations and combinations of features of the embodiments disclosed are within the scope of the invention. Numerous embodiments are encompassed by the following claims.
Number | Name | Date | Kind |
---|---|---|---|
6009209 | Acker et al. | Dec 1999 | A |
6204858 | Gupta | Mar 2001 | B1 |
6407777 | DeLuca | Jun 2002 | B1 |
6714665 | Hanna et al. | Mar 2004 | B1 |
6718051 | Eschbach | Apr 2004 | B1 |
6728401 | Hardeberg | Apr 2004 | B1 |
6798903 | Takaoka | Sep 2004 | B2 |
6873743 | Steinberg | Mar 2005 | B2 |
6895112 | Chen et al. | May 2005 | B2 |
7027662 | Baron | Apr 2006 | B2 |
7035461 | Luo et al. | Apr 2006 | B2 |
7035462 | White et al. | Apr 2006 | B2 |
7042505 | DeLuca | May 2006 | B1 |
7062086 | Chen et al. | Jun 2006 | B2 |
7088855 | Vide | Aug 2006 | B1 |
7116820 | Luo et al. | Oct 2006 | B2 |
7155058 | Gaubatz et al. | Dec 2006 | B2 |
7171044 | Chen et al. | Jan 2007 | B2 |
7224850 | Zhang et al. | May 2007 | B2 |
20020176623 | Steinberg | Nov 2002 | A1 |
20040240747 | Jarman et al. | Dec 2004 | A1 |
20050047656 | Luo et al. | Mar 2005 | A1 |
20050196067 | Gallagher et al. | Sep 2005 | A1 |
20050232481 | Donghui Wu | Oct 2005 | A1 |
Number | Date | Country | |
---|---|---|---|
20060072815 A1 | Apr 2006 | US |