The invention relates generally to the area of digital photography, and more specifically to filtering “red-eye” artefacts from a flash-induced digital camera image.
“Red-eye” is a phenomenon in flash photography where a flash is reflected within a subject's eye and appears in a photograph as a red dot where the black pupil of the subject's eye would normally appear. The unnatural glowing red of an eye is due to internal reflections from the vascular membrane behind the retina, which is rich in blood vessels. This objectionable phenomenon is well understood to be caused in part by a small angle between the flash of the camera and the lens of the camera. This angle has decreased with the miniaturization of cameras with integral flash capabilities. Additional contributors include the relative closeness of the subject to the camera and ambient light levels.
The red-eye phenomenon can be minimized by causing the iris to reduce the opening of the pupil. This is typically done with a “pre-flash”, a flash or illumination of light shortly before a flash photograph is taken. This causes the iris to close. Unfortunately, the pre-flash is an objectionable 0.2 to 0.6 seconds prior to the flash photograph. This delay is readily discernible and easily within the reaction time of a human subject. Consequently the subject may believe the pre-flash is the actual photograph and be in a less than desirable position at the time of the actual photograph. Alternately, the subject must be informed of the pre-flash, typically loosing any spontaneity of the subject captured in the photograph.
Those familiar with the art have developed complex analysis processes operating within a camera prior to invoking a pre-flash. Various conditions are monitored prior to the photograph before the pre-flash is generated; the conditions include the ambient light level and the distance of the subject from the camera. Such a system is described in U.S. Pat. No. 5,070,355, which is hereby incorporated by reference. Although that invention minimizes the occurrences where a pre-flash is used, it does not eliminate the need for a pre-flash.
Digital cameras are becoming more popular and smaller in size. Digital cameras have several advantages over film cameras. Digital cameras eliminate the need for film as the image is digitally captured and stored in a memory array for display on a display screen on the camera itself. This allows photographs to be viewed and enjoyed virtually instantaneously as opposed to waiting for film processing. Furthermore, the digitally captured image may be downloaded to another display device such as a personal computer or color printer for further enhanced viewing. Digital cameras include microprocessors for image processing and compression and camera systems control. It is possible to exploit the computation capabilities of such microprocessors for performing operations to improve the red-eye detection and elimination. Thus, what is needed is a method of better tools for eliminating red-eye phenomenon within, for example, a digital camera having a flash unit without the distraction of a pre-flash.
U.S. Patent Application 2002/0150306 (Baron), which is hereby incorporated by reference, describes a method for the removal of flash artefacts by capturing two digital images of a subject, one with flash and one without flash, and subtracting one image from the other to provide an artefact image which is then thresholded and subtracted from the flash image. However, the technique is directed to flash artefacts in general, and not specifically to red-eye removal. There is no attempt to identify red-eye regions as compared to any other flash-induced artefacts. Indeed, there is no attempt to identify particular regions at all, since the technique is simply one of subtraction and thresholding.
A system in accordance with the present invention there is provided a digital image acquisition system having no photographic film, comprising a portable apparatus for capturing digital images, a flash unit for providing illumination during image capture, and a red-eye filter for detecting a region within a captured image indicative of a red-eye phenomenon, said detection being based upon a comparison of said captured image and at least one reference eye color characteristic.
According to one embodiment of the invention, the at least one reference eye color characteristic includes a reference image of nominally the same scene taken without flash.
According to this embodiment, the reference image may be a preview image of lower pixel resolution than the captured image, the filter may include programming code for matching the pixel resolutions of the captured and reference images by up-sampling the preview image and/or sub-sampling the captured image.
To allow for inadvertent movement in the subject between taking the two images, preferably the filter may further include programming code for aligning at least portions of the captured image and reference image prior to said comparison.
In the embodiment the filter detects said region indicative of a red-eye phenomenon by identifying a region in the captured image at least having a color indicative of a red-eye phenomenon and comparing said identified region with the corresponding region in the reference image. The filter may further designate the region as indicative of a red-eye phenomenon if said corresponding region does not have a color indicative of a red-eye phenomenon. The decision as to whether a region has a color indicative of a red-eye phenomenon may be determined on a statistical basis as a global operation on the entire region.
According to a further embodiment, the filter includes a shape analyser so that the filter is configured to identify a region in the captured image having both a shape and color indicative of a red-eye phenomenon for subsequent comparison with a reference eye shape characteristic and the reference eye color characteristic, respectively.
In another embodiment, a pixel modifier is included for modifying the color of the pixels within a region indicative of a red-eye phenomenon.
A method in accordance with the present invention is also provided for detecting a red-eye phenomenon within a digital image acquired by a digital image acquisition device having no photographic film. The device includes a portable apparatus for capturing digital images, and a flash unit for providing illumination during image capture. The method includes identifying a region within a captured image indicative of a red-eye phenomenon including comparing the captured image and at least one reference eye color characteristic.
According to one embodiment, the at least one reference eye color characteristic includes an eye color characteristic of a reference image of nominally the same scene taken without flash.
The reference image may be a preview image of lower pixel resolution than the captured image. The method may further include matching the pixel resolutions of the captured and reference images including up-sampling the preview image or sub-sampling the captured image, or a combination thereof.
The method may further include aligning at least portions of the captured image and reference image prior to said comparison.
The method may further include analysing a shape so that the identifying comprises identifying a region in the captured image having both a shape and color indicative of a red-eye phenomenon for subsequent comparison with the corresponding region in the reference image. A shape may be analysed to determine subsequent to the comparison whether a region designated as indicative of a red-eye phenomenon has a shape indicative of a red-eye phenomenon.
The method may also include detecting a region indicative of a red-eye phenomenon by identifying a region in the captured image at least having a color indicative of a red-eye phenomenon and comparing the identified region with the corresponding region in the reference image, and designating the region as indicative of a red-eye phenomenon if the corresponding region does not have a color indicative of a red-eye phenomenon.
The method may also include deciding whether a region has a color indicative of a red-eye phenomenon by determining on a statistical basis as a global operation on the entire region.
The method may also include modifying the color of the pixels within a region indicative of a red-eye phenomenon.
FIGS. 3-a to 3-d illustrates the pixelation process of an image in different resolutions.
FIG. 3-e is a enlargement of a hypothetical digitization of an eye in an image.
FIGS. 5-a to 5-d and 6-a and 6-b illustrate the detailed workflow of the red-eye filter according to the embodiment, and alternatives.
The red-eye filter 90 can be integral to the camera 20 or part of an external processing device 10 such as a desktop computer, a hand held device, a cell phone handset or a server. In this embodiment, the filter receives the captured image from the full resolution image storage 80 as well as one or a plurality of preview images from the temporary storage 82. The filter 90 analyzes the stored image for characteristics of red-eye and, if found, modifies the image and removes the red-eye phenomenon from the image as will be describe in more detail. The red-eye filter includes a pixel locator 92 for locating pixels having a color indicative of red-eye; a shape analyzer 94 for determining if a grouping of at least a portion of the pixels located by the pixel locator comprise a shape indicative of red-eye; an falsing analyzer 96 for processing the image around the grouping for details indicative of an image of an eye; and a pixel modifier 98 for modifying the color of pixels within the grouping. The modified image may be either displayed on image display 100, saved on a persistent storage 112 which can be internal or a removable storage such as CF card, SD card or the like, or downloaded to another device, such as a personal computer, server or printer via image output means 110 which can be tethered or wireless.
In a further embodiment where the red-eye filter 90 is located on an external application in a separate device, 10, such as a desktop computer, the final captured image stored in block 80 along with a representation of the preview image as temporarily stored in 82, may be stored prior to modification on the storage device 112, or transferred together via the image output means 110 onto the external device 10, later to be processed by the red-eye filter 90.
In a simple embodiment, if the test criteria are met, or if no test criteria exist, the system will constantly replace the previous saved preview image with the new preview image, 230. Alternatively, where multiple preview images can be saved, 240, the new image will be placed on a chronological FIFO stack, namely First In First Out, where the system continuously captures and saves new preview images 244 while each time clearing the oldest image 242 from the stack, until the user takes the final picture. The reason for storing multiple preview images is based on the fact that the last image, or any single image, may not be the best reference image for comparison with the final full resolution image in the red-eye correction process. By storing multiple images, a better reference image can be achieved, and a closer alignment between the preview and the final captured image can be achieved. This concept will be further discussed in
The test criteria 222 may involve the actual analysis of the preview image content before deciding whether the new preview image should replace a previously saved image. Such criteria may be based on image analysis such as the existence of faces in the image, detection of eyes or metadata analysis such as the exposure condition, whether a flash is going to happen, the distance to the subjects, etc.
As part of the red-eye filter 90 the full resolution image 292 and the preview image or images 294 will be loaded into working memory, 292 and 294, unless they are already in memory in which they will just be accessed through a pointer.
Referring to
FIG. 3-b depicts the optical projection of a face 310 onto the sensor 60. Each of the cells 302 records the average light information it receives for the image. This is the process of digitization and quantization.
The degree of details is determined by the resolution of the sensor as depicted in FIG. 3-c. In this illustration a much smaller sensor is used, and in this case the same face 310 is digitized into a smaller number of pixels, or alternatively subsampled from the full resolution sensor data into a smaller number of pixel representations.
FIG. 3-d describes the inverse process where the subsampled image of FIG. 3-c is upsampled to the same size as the original. When comparing the resolution, naturally the some of the details are lost in this process. For illustrative example, while in FIG. 3-b the face was roughly 25×25=625 pixels, in FIG. 3-d the face is made of only 5×5=25 pixels.
Of course, the above are only for illustration purposes. In practice, due to the larger resolution of the sensors than in this illustration, a normal eye will be depicted by a much larger pixel count to be noticeable. FIG. 3-e displays such a digitized eye. In this figure, an eye 350, as imaged on a sensor 60, will consist of roughly 25 pixels wide, 352. In particular interest for this invention the inner portion, the iris 360, in this case will be roughly 8 pixels in diameter, as illustrated in 462.
According to a preferred embodiment of this invention, the preview image and the final image, or portions of them, need to be aligned as depicted in
The flat regions should display no significant differences due to resolution changes. The main difference will be caused be two reasons namely edge regions where changes in value occur such as in blocks 410. However, there is another cause for difference which is of interest to this invention and displayed in 430. In these pixels, the difference is caused by the actual change in the color of the eye from normal eyes to red-eyes. Not only is there a change in the pixel value but the change is also more specifically reflected as change to a red or light color from the normal color of the iris or form the black color of the pupil.
FIGS. 5-a to 5-d illustrate the workflow of the red-eye filter 90 of this embodiment, as well as variations thereof.
Referring first to FIG. 5-a, there are two input images into the filter, namely a full resolution image 510, I(x,y) which is the one that was captured by full depression of the shutter button and needs to be analyzed for red-eye artefacts, and a preview image 520, P(x,y) which is used as a reference image and is nominally the same scene as the image I(x,y) but taken without the flash. The preview image may be a result of some image processing taking into account multiple preview images and creating a single image, 522. Methods of improving image quality based on multiple images are familiar to those versed in the art of image processing. The resulting output from the analysis process of 522 is a single preview image.
The preview image 520 is normally, but not necessarily, of lower resolution than the full resolution image 510, typically being generated by clocking out a subset of the sensor cells or by averaging the raw sensor data. Therefore, the two images, or alternatively the relevant regions in the images (i.e. the regions containing or suspected to contain eyes, which can be determined by image processing techniques known in the art), need to be matched in pixel resolution, 530. In the present context “pixel resolution” means the size of the image, or relevant region, in terms of the number of pixels constituting the image or region concerned. Such a process may be done by either upsampling the preview image, 534, downsampling the acquired image, 532, or a combination thereof. Those familiar in the art will be aware of several techniques best used for such sampling methods. The result of step 530 is a pair of images I′(x,y) and P′(x,y) corresponding to the original images I(x,y) and P(x,y), or relevant regions thereof, with matching pixel resolution. The system and method of the preferred embodiment involves the detection and removal of red-eye artefacts. The actual removal of the red-eye will eventually be performed on the full resolution image. However, all or portions of the detection of red-eye candidate pixel groupings, the subsequent testing of said pixel groupings for determining false red-eye groupings, and the initial step of the removal, where the image is presented to the user for user confirmation of the correction, can be performed on the entire image, the subsampled image, or a subset of regions of the entire image or the subsampled image.
Although nominally of the same scene, the preview image and the finally acquired full resolution image may differ spatially due to the temporal lag between capturing the two images. Therefore, the two images, or relevant regions thereof, may need to be aligned, 540, especially in respect of regions of the images containing or suspected to contain eyes. Essentially, alignment means transforming at least one of the images, and in this embodiment the preview image P′(x,y), to obtain maximum correlation between the images, or relevant regions thereof, based on measurable characteristics such as color, texture, edge analysis. Those familiar in the art are aware of several algorithms to achieve such alignment; see, for example, U.S. Pat. No. 6,295,367 which is hereby incorporated by reference and which describes alignment of images due to object and camera movement, and U.S. Pat. No. 5,933,546, which is also hereby incorporated by reference and which addresses the use of multi-resolution data for pattern matching.
Further discussion on the alignment is presented in FIG. 5-c. In this Figure, the inputs are the two images I′(x,y) and P′(x,y) as defined in FIG. 5-a. The alignment may be global for the entire image or local for specific regions. For example, a simple linear alignment, such as a shift in the horizontal direction by H pixels, and/or in the vertical direction by V pixels, or a combination of the two. Mathematically, the shifted image, P″(x,y), can be described as:
P″(x,y)=P′(x−H,y−V)
However, simple translation operation may not suffice in the need to align the image. Therefore, there may be a need for X-Y shearing, which is a symmetrical shift of the object's points in the direction of the axis to correct for perspective changes; X-Y tapering where the object is pinched by shifting its coordinates towards the axis, the greater the magnitude of the coordinate the further the shift; or rotation around an arbitrary point.
In general, the alignment process may involve an affine transformation, defined as a special class of projective transformations that do not move any objects from the affine space to the plane at infinity or conversely, or any transformation that preserves collinearity (i.e. all points lying on a line initially still lie on a line after transformation) and ratios of distances (e.g., the midpoint of a line segment remains the midpoint after transformation). Geometric contraction, expansion, dilation, reflection, rotation, shear, similarity transformations, spiral similarities and translation are all affine transformations, as are their combinations. In general, the alignment 540 may be achieved via an affine transformation which is a composition of rotations, translations, dilations, and shears, all well-known to one familiar in the art of image processing.
If it is determined through a correlation process that a global transformation suffices, as determined in block 542=YES, one of the images, and for simplicity the preview image, will undergo an affine transformation, 544, to align itself with the final full resolution image. Mathematically, this transformation can be depicted as:
P″=AP′+q
where A is a linear transformation and q is a translation.
However, in some cases a global transformation may not work well, in particular for cases where the subject matter moved, as could happen when photographing animated objects. In such case, in particular in images with multiple human subjects, and when the subjects move in independent fashion, the process of alignment 540 may be broken down, 546, to numerous local regions each with its own affine transformation. What is important is to align the eyes between the images. Therefore, according to this alternative, one or multiple local alignments may be performed, 548, for regions in the vicinity surrounding the eyes, such as faces.
Only after the images are aligned can one compare the potential red-eye colors.
In the preferred embodiment of FIG. 5-a, the preview image information is used as part of the falsing stage 96. Blocks 92, 94 and 98 correspond to the same blocks in
Referring to block 96, for each region of the image I′(x,y) suspected as red-eye, step 596-2, as identified by steps 92 and 94, the suspected region is tested by comparing the pixel values of the region with the pixel values of the corresponding region in the aligned preview image P″(x,y), 596-6. However, prior to doing so, the regions need to be prepared and modified for such comparison, 596-4.
Due to the fact that the regions may not match exactly, a pixel-by-pixel comparison may not suffice. The reason for the mismatch may occur due to the original size discrepancy. For example, in edges this phenomenon is graphically illustrated in
This process as illustrated in block 596-4 will be further described in FIG. 5-d. The underlying concept behind step 596-4 is to distinguish between differences that are caused due to the acquisition process and the differences that are caused due to the existence of red-eye in the image. This problem is well known to one familiar in the art of statistical pattern matching and scene analysis and image recognition. An example of such an application taking into account differences due to resolution is described in U.S. Pat. No. 5,933,546, which is hereby incorporated by reference.
If a region in the aligned preview image P″(x,y) was red and the equivalent region is red in the image I′(x,y), 596-6, that region will be eliminated from I′(x,y) as a red-eye artefact, 596-9, and the corresponding region will be eliminated as a red-eye artefact from the original full resolution image I(x,y). Otherwise, the region will continue to remain suspected as red-eye, 596-8. The process will continue, 596-3, for all suspected regions.
The comparison of the regions for a color value is done as a global operation on the entire region, and the answer to the question of whether a region is red or not is made statistically for the entire region and not pixel by pixel, i.e. it does not depend on the value of any particular individual pixel. Such approach will account for inconsistencies on the pixel level that may be resolved statistically when analyzing a larger collection of pixels consisting of a region. For example, some regions of the eye may not be fully red, or display other artefacts such as a glint of high luminance. Other example for the need of a global statistical operation is the presence of noise in the image. Techniques are known in the art for such global comparison.
Based on the information above, the regions finally identified as red-eye artefacts can be modified, 98, to eliminate the red-eye from the original full resolution image I(x,y). The modification can be done using any one of numerous available techniques such as luminance reduction, chrominance reduction, or subtraction of the artefact, as described in US Published Patent Application 2002/0150306 (Baron), which is hereby incorporated by reference.
FIG. 5-d describes the preparation of regions suspected of red-eye for comparison as described in FIG. 5-a, block 596-4. As discussed above, a simple pixel level comparison may not be enough to determine whether the region is not of red-eye nature. The process of preparation may include a combination of several components such as creating color balance between the regions of the preview image and the final image, 1510, analyzing the texture, or differences in high frequency patterns between the two regions that may have occurred due to the change in resolution, 1520, and comparing the edges between the two regions, 1530, where the differences may have occurred due to change in exposure, color balance, resolution or alignment, and in particular sub pixel alignment. The color balance step 1510 comprises marking each red-eye region in I′(x,y) and the corresponding region in P″(x,y), steps 1512 and 1514, determining the difference in color balance between the region in I′(x,y) surrounding, but not including, the suspected red-eye region and the corresponding region of P″(x,y), step 1516, and transforming the entire region, including the suspected red-eye region, based on the color balance difference so determined, step 1518.
As an alternative embodiment of this invention, the preview image can be used as part of the pixel locator stage 92, as illustrated in FIG. 5-b, rather than as part of the falsing analyzer 96. In FIG. 5-b, blocks 510, 520, 522, 530, 532, 534, 540, 94 and 98 are identical to those in FIG. 5-a. According to this embodiment, the use of the preview image in order to detect red-eye artefacts is implemented as part of the red-eye identification process, otherwise described as the pixel locator 92 in
After the suspected red-eye regions are identified, the process continues via the shape analysis 94, false detection elimination 96 and correction 98 as described in
According to this embodiment, after the alignment step 540 the following steps 592-1a and 592-1b analyse both images I′(x,y) and P″(x,y) for the presence of pixels having a color indicative of red-eye (592-1a), for example in the manner of DeLuca, and then identifies clusters of contiguous red pixels so detected (592-1b). This is known as segmentation and is more fully described in US Pat. Appn. 2002/0176623, which is hereby incorporated by reference.
Now, each region (cluster) with red content in the acquired image I′(x,y), step 592-2, is compared with the corresponding region in the aligned preview image P″(x,y). The regions will need to be prepared, 592-4, as previously described in relation to block 596-4 of FIG. 5-a. If the regions are red in both cases, 592-6=YES, the region will not be marked as red-eye, no action will be taken and the process will continue to the next suspected region, 592-3. If the region is red in the acquired image I′(x,y) while the corresponding region is not red in the preview image P″(x,y), 592-6=NO, then the region will be marked as suspected red-eye, 592-8.
FIG. 6-a shows a modification of the embodiment of FIG. 5-b in which Step 540 (Align Images) has been divided into two steps, Step 541 (If Possible Globally Align Images) and Step 592-3 (If Required Locally Align Images). Step 541 corresponds to Steps 542 and 544 of FIG. 5-c. However, if a global alignment is not possible or practical, the local alignment is deferred until after red pixel identification and clustering has been performed, since the presence of such clusters in the two images I′(x,y) and P′(x,y) will assist in the local alignment. FIG. 6-b shows a similar modification applied to FIG. 5-a.
In the embodiments of the invention, in the comparison stages, 592-6 and 596-6 the pixel values do not necessarily have to be compared with red but may alternatively or additionally be compared with other values such as yellow, white, pink, brown or other color indicative of a red-eye phenomenon, or to a range of values, to accommodate other flash related eye artefacts that are not purely red. Due to the fact that the eye surface is retro-reflective (due to the smoothness created by the tears, and the spherical shape of the eyeball), the technique as described in this specification can assist in the detection of the eyes in an image. Such existence of an eye can be found by comparison of the spectral reflection of the flash in the eye with the same region where no flash was used, and thus without spectral reflection. This comparison may assist in locating eyes in general and not just eyes with red-eye artefacts. This process may be implemented by finding the change of small specular reflections that occur in the eye region when flash illumination is used such as described in WO 03/026278 (Jarman). The specular reflections may be used as another indication of suspected regions as defined in blocks 592-2 and 596-2 by comparing the specular reflection of the flash image with no specular reflection of the preview image.
Alternatively to a binary decision of adding or eliminating a region, 596-8 and 596-9, in the case of a continuous probability for each region, the process will be revised from a binary decision to changing a probability decision. The quantitative determination of such change in probability may be decided based on analysis of the confidence level of the comparison 592-4 and 596-4.
The preferred embodiments described above may be modified by adding or changing operations, steps and/or components in many ways to produce advantageous alternative embodiments. For example, the reference image can be a post-view image rather than a preview image, i.e. an image taken without flash immediately after the flash picture is taken.
A red-eye correction procedure may begin as described by block 92 with detecting a human face in a digital image and, based on this detection, finding the eyes in the face (see, e.g., U.S. Pat. No. 6,252,976, U.S. Publ. Pat. App. No. 2003/0044070 and U.S. Pat. No. 6,278,491, which are hereby incorporated by reference). This procedure may also be used for creating the regional alignment 546 and color balance 1510.
A range of alternative techniques may be employed to detect and verify the existence of red-eye defects in an image (see, e.g., U.S. Publ. Pat. Apps. No. 2003/0044177 and 2003/0044178, which are hereby incorporated by reference). These techniques may be incorporated into the pixel locator, shape analyzer, falsing analyzer and pixel modifier corresponding to blocks 92, 94, 96 and 98. A camera may include software or firmware for automatically detecting a red-eye image using a variety of image characteristics such as image brightness, contrast, the presence of human skin and related colors. The analysis of these image characteristics may be utilized, based on certain pre-determined statistical thresholds, to decide if red-eye defects exist and if a flash was used to take the original image.
The preferred embodiments described herein may involve expanded digital acquisition technology that inherently involves digital cameras, but that may be integrated with other devices such as cell-phones equipped with an acquisition component, toy cameras etc. The digital camera or other image acquisition device of the preferred embodiment has the capability to record not only image data, but also additional data referred to as meta-data. The file header of an image file, such as JPEG, TIFF, JPEG-2000, etc., may include capture information including the preview image, for processing and red-eye detection at a later post processing stage, which may be performed in the acquisition device or in a separate device such as a personal computer. The preferred embodiments described herein serve to improve the detection of red-eyes in images, while eliminating or reducing the occurrence of false positives, and to improve the correction of the detected artefacts.
While an exemplary drawing and specific embodiments of the present invention have been described and illustrated, it is to be understood that that the scope of the present invention is not to be limited to the particular embodiments discussed. Thus, the embodiments shall be regarded as illustrative rather than restrictive, and it should be understood that variations may be made in those embodiments by workers skilled in the arts without departing from the scope of the present invention, as set forth in the claims below and structural and functional equivalents thereof.
In addition, in methods that may be performed according to preferred embodiments herein and that may have been described above, the operations have been described in selected typographical sequences. However, the sequences have been selected and so ordered for typographical convenience and are not intended to imply any particular order for performing the operations, unless expressly set forth or understood by those skilled in the art being necessary.
Thus, the preferred embodiments described herein provide an improved method and apparatus for detecting red-eye phenomenon within images taken by a digital camera having a flash while eliminating or reducing the occurrence of false positives by using preview information.
In addition to all of the references cited above that have been incorporated by reference, the sections entitled BACKGROUND, SUMMARY OF THE INVENTION, ABSTRACT and BRIEF DESCRIPTION OF THE DRAWINGS, are hereby incorporated by reference into the DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT.
This application is a continuation of U.S. patent application Ser. No. 10/919,226, filed Aug. 16, 2004, published as US2005/0041121, which is a continuation-in-part (CIP) application which claims the benefit of priority to U.S. patent application Ser. No. 10/772,767, filed Feb. 4, 2004, now U.S. Pat. No. 7,352,394 issued Apr. 1, 2008, which is a CIP that claims the benefit of priority to U.S. patent application Ser. No. 10/635,862, filed Aug. 5, 2003, published as US2004/0223063, now U.S. Pat. No. 7,630,006, which is a CIP that claims the benefit of priority to U.S. patent application Ser. No. 10/170,511, filed Jun. 12, 2002, now U.S. Pat. No. 7,042,505 issued May 9, 2006, which is a continuation of U.S. patent application Ser. No. 08/947,603, filed Oct. 9, 1997, now U.S. Pat. No. 6,407,777, issued Jun. 18, 2002, which are each hereby incorporated by reference. This application is also related to U.S. patent application Ser. Nos. 10/635,918, filed Aug. 5, 2003, published as US2005/0031224, 10/773,092, filed Feb. 4, 2004, published as US2005/0140801, 11/379,346 filed Apr. 19, 2006, 11/690,834 filed Mar. 25, 2007, published as US2007/0263104, 11/769,206 filed Jun. 27, 2007 published as US2008/0002060, 12/035,416 filed Feb. 21, 2008, PCT application no. PCT/EP2005/005907 filed Jun. 1, 2005 published as WO2006/018056.
Number | Name | Date | Kind |
---|---|---|---|
4285588 | Mir | Aug 1981 | A |
4577219 | Klie et al. | Mar 1986 | A |
4646134 | Komatsu et al. | Feb 1987 | A |
4777620 | Shimoni et al. | Oct 1988 | A |
4881067 | Watanabe et al. | Nov 1989 | A |
4978989 | Nakano et al. | Dec 1990 | A |
5016107 | Sasson et al. | May 1991 | A |
5070355 | Inoue et al. | Dec 1991 | A |
5130789 | Dobbs et al. | Jul 1992 | A |
5164831 | Kuchta et al. | Nov 1992 | A |
5164833 | Aoki | Nov 1992 | A |
5202720 | Fujino et al. | Apr 1993 | A |
5249053 | Jain | Sep 1993 | A |
5274457 | Kobayashi et al. | Dec 1993 | A |
5301026 | Lee | Apr 1994 | A |
5303049 | Ejima et al. | Apr 1994 | A |
5335072 | Tanaka et al. | Aug 1994 | A |
5384601 | Yamashita et al. | Jan 1995 | A |
5400113 | Sosa et al. | Mar 1995 | A |
5432863 | Benati et al. | Jul 1995 | A |
5432866 | Sakamoto | Jul 1995 | A |
5452048 | Edgar | Sep 1995 | A |
5455606 | Keeling et al. | Oct 1995 | A |
5537516 | Sherman et al. | Jul 1996 | A |
5568187 | Okino | Oct 1996 | A |
5568194 | Abe | Oct 1996 | A |
5671013 | Nakao | Sep 1997 | A |
5694926 | DeVries et al. | Dec 1997 | A |
5719639 | Imamura | Feb 1998 | A |
5719951 | Shackleton et al. | Feb 1998 | A |
5724456 | Boyack et al. | Mar 1998 | A |
5734425 | Takizawa et al. | Mar 1998 | A |
5748764 | Benati et al. | May 1998 | A |
5748784 | Sugiyama | May 1998 | A |
5751836 | Wildes et al. | May 1998 | A |
5761550 | Kancigor | Jun 1998 | A |
5781650 | Lobo et al. | Jul 1998 | A |
5805720 | Suenaga et al. | Sep 1998 | A |
5805727 | Nakano | Sep 1998 | A |
5805745 | Graf | Sep 1998 | A |
5815749 | Tsukahara et al. | Sep 1998 | A |
5818975 | Goodwin et al. | Oct 1998 | A |
5847714 | Naqvi et al. | Dec 1998 | A |
5850470 | Kung et al. | Dec 1998 | A |
5862217 | Steinberg et al. | Jan 1999 | A |
5862218 | Steinberg | Jan 1999 | A |
5892837 | Luo et al. | Apr 1999 | A |
5949904 | Delp | Sep 1999 | A |
5974189 | Nicponski | Oct 1999 | A |
5990973 | Sakamoto | Nov 1999 | A |
5991456 | Rahman et al. | Nov 1999 | A |
5991549 | Tsuchida | Nov 1999 | A |
5991594 | Froeber et al. | Nov 1999 | A |
5999160 | Kitamura et al. | Dec 1999 | A |
6006039 | Steinberg et al. | Dec 1999 | A |
6009209 | Acker et al. | Dec 1999 | A |
6011547 | Shiota et al. | Jan 2000 | A |
6016354 | Lin et al. | Jan 2000 | A |
6028611 | Anderson et al. | Feb 2000 | A |
6035072 | Read | Mar 2000 | A |
6035074 | Fujimoto et al. | Mar 2000 | A |
6036072 | Lee | Mar 2000 | A |
6101271 | Yamashita et al. | Aug 2000 | A |
6104839 | Cok et al. | Aug 2000 | A |
6118485 | Hinoue et al. | Sep 2000 | A |
6134339 | Luo | Oct 2000 | A |
6151403 | Luo | Nov 2000 | A |
6172706 | Tatsumi | Jan 2001 | B1 |
6192149 | Eschbach et al. | Feb 2001 | B1 |
6195127 | Sugimoto | Feb 2001 | B1 |
6201571 | Ota | Mar 2001 | B1 |
6204858 | Gupta | Mar 2001 | B1 |
6233364 | Krainiouk et al. | May 2001 | B1 |
6249315 | Holm | Jun 2001 | B1 |
6252976 | Schildkraut et al. | Jun 2001 | B1 |
6266054 | Lawton et al. | Jul 2001 | B1 |
6268939 | Klassen et al. | Jul 2001 | B1 |
6275614 | Krishnamurthy et al. | Aug 2001 | B1 |
6278491 | Wang et al. | Aug 2001 | B1 |
6285410 | Marni | Sep 2001 | B1 |
6292574 | Schildkraut et al. | Sep 2001 | B1 |
6295378 | Kitakado et al. | Sep 2001 | B1 |
6298166 | Ratnakar et al. | Oct 2001 | B1 |
6300935 | Sobel et al. | Oct 2001 | B1 |
6381345 | Swain | Apr 2002 | B1 |
6393148 | Bhaskar | May 2002 | B1 |
6396963 | Shaffer et al. | May 2002 | B2 |
6407777 | DeLuca | Jun 2002 | B1 |
6421468 | Ratnakar et al. | Jul 2002 | B1 |
6426775 | Kurokawa | Jul 2002 | B1 |
6429924 | Milch | Aug 2002 | B1 |
6433818 | Steinberg et al. | Aug 2002 | B1 |
6438264 | Gallagher et al. | Aug 2002 | B1 |
6441854 | Fellegara et al. | Aug 2002 | B2 |
6459436 | Kumada et al. | Oct 2002 | B1 |
6473199 | Gilman et al. | Oct 2002 | B1 |
6496655 | Malloy Desormeaux | Dec 2002 | B1 |
6501911 | Malloy Desormeaux | Dec 2002 | B1 |
6505003 | Malloy Desormeaux | Jan 2003 | B1 |
6510520 | Steinberg | Jan 2003 | B1 |
6516154 | Parulski et al. | Feb 2003 | B1 |
6614471 | Ott | Sep 2003 | B1 |
6614995 | Tseng | Sep 2003 | B2 |
6621867 | Sazzad et al. | Sep 2003 | B1 |
6628833 | Horie | Sep 2003 | B1 |
6700614 | Hata | Mar 2004 | B1 |
6707950 | Burns et al. | Mar 2004 | B1 |
6714665 | Hanna et al. | Mar 2004 | B1 |
6718051 | Eschbach | Apr 2004 | B1 |
6724941 | Aoyama | Apr 2004 | B1 |
6728401 | Hardeberg | Apr 2004 | B1 |
6765686 | Maruoka | Jul 2004 | B2 |
6786655 | Cook et al. | Sep 2004 | B2 |
6792161 | Imaizumi et al. | Sep 2004 | B1 |
6798913 | Toriyama | Sep 2004 | B2 |
6859565 | Baron | Feb 2005 | B2 |
6873743 | Steinberg | Mar 2005 | B2 |
6885766 | Held et al. | Apr 2005 | B2 |
6895112 | Chen et al. | May 2005 | B2 |
6900882 | Iida | May 2005 | B2 |
6912298 | Wilensky | Jun 2005 | B1 |
6937997 | Parulski | Aug 2005 | B1 |
6967680 | Kagle et al. | Nov 2005 | B1 |
6980691 | Nesterov et al. | Dec 2005 | B2 |
6984039 | Agostinelli | Jan 2006 | B2 |
7024051 | Miller et al. | Apr 2006 | B2 |
7027662 | Baron | Apr 2006 | B2 |
7030927 | Sasaki | Apr 2006 | B2 |
7035461 | Luo et al. | Apr 2006 | B2 |
7035462 | White et al. | Apr 2006 | B2 |
7042501 | Matama | May 2006 | B1 |
7042505 | DeLuca | May 2006 | B1 |
7062086 | Chen et al. | Jun 2006 | B2 |
7116820 | Luo et al. | Oct 2006 | B2 |
7133070 | Wheeler et al. | Nov 2006 | B2 |
7155058 | Gaubatz et al. | Dec 2006 | B2 |
7216289 | Kagle et al. | May 2007 | B2 |
7224850 | Zhang et al. | May 2007 | B2 |
7269292 | Steinberg | Sep 2007 | B2 |
7289664 | Enomoto | Oct 2007 | B2 |
7295233 | Steinberg et al. | Nov 2007 | B2 |
7310443 | Kris et al. | Dec 2007 | B1 |
7315631 | Corcoran et al. | Jan 2008 | B1 |
7336821 | Ciuc et al. | Feb 2008 | B2 |
7352394 | DeLuca et al. | Apr 2008 | B1 |
7362368 | Steinberg et al. | Apr 2008 | B2 |
7369712 | Steinberg et al. | May 2008 | B2 |
7403643 | Ianculescu et al. | Jul 2008 | B2 |
7436998 | Steinberg et al. | Oct 2008 | B2 |
7454040 | Luo et al. | Nov 2008 | B2 |
7515740 | Corcoran et al. | Apr 2009 | B2 |
20010015760 | Fellegara et al. | Aug 2001 | A1 |
20010031142 | Whiteside | Oct 2001 | A1 |
20010052937 | Suzuki | Dec 2001 | A1 |
20020019859 | Watanabe | Feb 2002 | A1 |
20020041329 | Steinberg | Apr 2002 | A1 |
20020051571 | Jackway et al. | May 2002 | A1 |
20020054224 | Wasula et al. | May 2002 | A1 |
20020085088 | Eubanks | Jul 2002 | A1 |
20020090133 | Kim et al. | Jul 2002 | A1 |
20020093577 | Kitawaki et al. | Jul 2002 | A1 |
20020093633 | Milch | Jul 2002 | A1 |
20020105662 | Patton et al. | Aug 2002 | A1 |
20020114513 | Hirao | Aug 2002 | A1 |
20020126893 | Held et al. | Sep 2002 | A1 |
20020131770 | Meier et al. | Sep 2002 | A1 |
20020136450 | Chen et al. | Sep 2002 | A1 |
20020141661 | Steinberg | Oct 2002 | A1 |
20020150292 | O'Callaghan | Oct 2002 | A1 |
20020150306 | Baron | Oct 2002 | A1 |
20020159630 | Buzuloiu et al. | Oct 2002 | A1 |
20020172419 | Lin et al. | Nov 2002 | A1 |
20020176623 | Steinberg | Nov 2002 | A1 |
20030007687 | Nesterov et al. | Jan 2003 | A1 |
20030021478 | Yoshida | Jan 2003 | A1 |
20030025808 | Parulski et al. | Feb 2003 | A1 |
20030025811 | Keelan et al. | Feb 2003 | A1 |
20030044063 | Meckes et al. | Mar 2003 | A1 |
20030044070 | Fuersich et al. | Mar 2003 | A1 |
20030044176 | Saitoh | Mar 2003 | A1 |
20030044177 | Oberhardt et al. | Mar 2003 | A1 |
20030044178 | Oberhardt et al. | Mar 2003 | A1 |
20030052991 | Stavely et al. | Mar 2003 | A1 |
20030058343 | Katayama | Mar 2003 | A1 |
20030058349 | Takemoto | Mar 2003 | A1 |
20030095197 | Wheeler et al. | May 2003 | A1 |
20030107649 | Flickner et al. | Jun 2003 | A1 |
20030113035 | Cahill et al. | Jun 2003 | A1 |
20030118216 | Goldberg | Jun 2003 | A1 |
20030137597 | Sakamoto et al. | Jul 2003 | A1 |
20030142285 | Enomoto | Jul 2003 | A1 |
20030161506 | Velazquez et al. | Aug 2003 | A1 |
20030190072 | Adkins et al. | Oct 2003 | A1 |
20030194143 | Iida | Oct 2003 | A1 |
20030202715 | Kinjo | Oct 2003 | A1 |
20040017481 | Takasumi et al. | Jan 2004 | A1 |
20040027593 | Wilkins | Feb 2004 | A1 |
20040032512 | Silverbrook | Feb 2004 | A1 |
20040032526 | Silverbrook | Feb 2004 | A1 |
20040033071 | Kubo | Feb 2004 | A1 |
20040037460 | Luo et al. | Feb 2004 | A1 |
20040041924 | White et al. | Mar 2004 | A1 |
20040046878 | Jarman | Mar 2004 | A1 |
20040047491 | Rydbeck | Mar 2004 | A1 |
20040056975 | Hata | Mar 2004 | A1 |
20040057623 | Schuhrke et al. | Mar 2004 | A1 |
20040057705 | Kohno | Mar 2004 | A1 |
20040057715 | Tsuchida et al. | Mar 2004 | A1 |
20040090461 | Adams | May 2004 | A1 |
20040093432 | Luo et al. | May 2004 | A1 |
20040114796 | Kaku | Jun 2004 | A1 |
20040114797 | Meckes | Jun 2004 | A1 |
20040114829 | LeFeuvre et al. | Jun 2004 | A1 |
20040114904 | Sun et al. | Jun 2004 | A1 |
20040119851 | Kaku | Jun 2004 | A1 |
20040120598 | Feng | Jun 2004 | A1 |
20040125387 | Nagao et al. | Jul 2004 | A1 |
20040126086 | Nakamura et al. | Jul 2004 | A1 |
20040141657 | Jarman | Jul 2004 | A1 |
20040150743 | Schinner | Aug 2004 | A1 |
20040160517 | Iida | Aug 2004 | A1 |
20040165215 | Raguet et al. | Aug 2004 | A1 |
20040184044 | Kolb et al. | Sep 2004 | A1 |
20040184670 | Jarman et al. | Sep 2004 | A1 |
20040196292 | Okamura | Oct 2004 | A1 |
20040196503 | Kurtenbach et al. | Oct 2004 | A1 |
20040213476 | Luo et al. | Oct 2004 | A1 |
20040223063 | DeLuca et al. | Nov 2004 | A1 |
20040227978 | Enomoto | Nov 2004 | A1 |
20040228542 | Zhang et al. | Nov 2004 | A1 |
20040233299 | Ioffe et al. | Nov 2004 | A1 |
20040233301 | Nakata et al. | Nov 2004 | A1 |
20040234156 | Watanabe et al. | Nov 2004 | A1 |
20040239779 | Washisu | Dec 2004 | A1 |
20040240747 | Jarman et al. | Dec 2004 | A1 |
20040258308 | Sadovsky et al. | Dec 2004 | A1 |
20050001024 | Kusaka et al. | Jan 2005 | A1 |
20050013602 | Ogawa | Jan 2005 | A1 |
20050013603 | Ichimasa | Jan 2005 | A1 |
20050024498 | Iida et al. | Feb 2005 | A1 |
20050031224 | Prilutsky et al. | Feb 2005 | A1 |
20050041121 | Steinberg et al. | Feb 2005 | A1 |
20050047655 | Luo et al. | Mar 2005 | A1 |
20050047656 | Luo et al. | Mar 2005 | A1 |
20050053279 | Chen et al. | Mar 2005 | A1 |
20050058340 | Chen et al. | Mar 2005 | A1 |
20050058342 | Chen et al. | Mar 2005 | A1 |
20050062856 | Matsushita | Mar 2005 | A1 |
20050063083 | Dart et al. | Mar 2005 | A1 |
20050068452 | Steinberg et al. | Mar 2005 | A1 |
20050074164 | Yonaha | Apr 2005 | A1 |
20050074179 | Wilensky | Apr 2005 | A1 |
20050078191 | Battles | Apr 2005 | A1 |
20050117132 | Agostinelli | Jun 2005 | A1 |
20050129331 | Kakiuchi et al. | Jun 2005 | A1 |
20050140801 | Prilutsky et al. | Jun 2005 | A1 |
20050147278 | Rui et al. | Jul 2005 | A1 |
20050151943 | Iida | Jul 2005 | A1 |
20050163498 | Battles et al. | Jul 2005 | A1 |
20050168965 | Yoshida | Aug 2005 | A1 |
20050196067 | Gallagher et al. | Sep 2005 | A1 |
20050200736 | Ito | Sep 2005 | A1 |
20050207649 | Enomoto et al. | Sep 2005 | A1 |
20050212955 | Craig et al. | Sep 2005 | A1 |
20050219385 | Terakawa | Oct 2005 | A1 |
20050219608 | Wada | Oct 2005 | A1 |
20050220346 | Akahori | Oct 2005 | A1 |
20050220347 | Enomoto et al. | Oct 2005 | A1 |
20050226499 | Terakawa | Oct 2005 | A1 |
20050232490 | Itagaki et al. | Oct 2005 | A1 |
20050238230 | Yoshida | Oct 2005 | A1 |
20050243348 | Yonaha | Nov 2005 | A1 |
20050275734 | Ikeda | Dec 2005 | A1 |
20050276481 | Enomoto | Dec 2005 | A1 |
20050280717 | Sugimoto | Dec 2005 | A1 |
20050286766 | Ferman | Dec 2005 | A1 |
20060008171 | Petschnigg et al. | Jan 2006 | A1 |
20060017825 | Thakur | Jan 2006 | A1 |
20060038916 | Knoedgen et al. | Feb 2006 | A1 |
20060039690 | Steinberg et al. | Feb 2006 | A1 |
20060045352 | Gallagher | Mar 2006 | A1 |
20060050300 | Mitani et al. | Mar 2006 | A1 |
20060066628 | Brodie et al. | Mar 2006 | A1 |
20060082847 | Sugimoto | Apr 2006 | A1 |
20060093212 | Steinberg et al. | May 2006 | A1 |
20060093213 | Steinberg et al. | May 2006 | A1 |
20060093238 | Steinberg et al. | May 2006 | A1 |
20060098867 | Gallagher | May 2006 | A1 |
20060098875 | Sugimoto | May 2006 | A1 |
20060119832 | Iida | Jun 2006 | A1 |
20060120599 | Steinberg et al. | Jun 2006 | A1 |
20060140455 | Costache et al. | Jun 2006 | A1 |
20060150089 | Jensen et al. | Jul 2006 | A1 |
20060203108 | Steinberg et al. | Sep 2006 | A1 |
20060204052 | Yokouchi | Sep 2006 | A1 |
20060204110 | Steinberg et al. | Sep 2006 | A1 |
20060221408 | Fukuda | Oct 2006 | A1 |
20060285754 | Steinberg et al. | Dec 2006 | A1 |
20070110305 | Corcoran et al. | May 2007 | A1 |
20070116379 | Corcoran et al. | May 2007 | A1 |
20070116380 | Ciuc et al. | May 2007 | A1 |
20070133863 | Sakai et al. | Jun 2007 | A1 |
20070154189 | Harradine et al. | Jul 2007 | A1 |
20070201724 | Steinberg et al. | Aug 2007 | A1 |
20070263104 | DeLuca et al. | Nov 2007 | A1 |
20080002060 | DeLuca et al. | Jan 2008 | A1 |
20080013798 | Ionita et al. | Jan 2008 | A1 |
20080043121 | Prilutsky et al. | Feb 2008 | A1 |
20080112599 | Nanu et al. | May 2008 | A1 |
20080144965 | Steinberg et al. | Jun 2008 | A1 |
20080186389 | DeLuca et al. | Aug 2008 | A1 |
20080232711 | Prilutsky et al. | Sep 2008 | A1 |
20080240555 | Nanu et al. | Oct 2008 | A1 |
Number | Date | Country |
---|---|---|
0884694 | Dec 1998 | EP |
0911759 | Apr 1999 | EP |
911759 | Jun 2000 | EP |
1199672 | Apr 2002 | EP |
1229486 | Aug 2002 | EP |
1288858 | Mar 2003 | EP |
1288859 | Mar 2003 | EP |
1288860 | Mar 2003 | EP |
1293933 | Mar 2003 | EP |
1296510 | Mar 2003 | EP |
1429290 | Jun 2004 | EP |
1429290 | Jun 2004 | EP |
1478169 | Nov 2004 | EP |
1528509 | May 2005 | EP |
979487 | Mar 2006 | EP |
841609 | Jul 1960 | GB |
4-192681 | Jul 1992 | JP |
5224271 | Sep 1993 | JP |
9214839 | Aug 1997 | JP |
2000 134486 | May 2000 | JP |
22247596 | Aug 2002 | JP |
2002 271808 | Sep 2002 | JP |
2003-030647 | Jan 2003 | JP |
WO-9802844 | Jan 1998 | WO |
WO 9917254 | Apr 1999 | WO |
WO-9933684 | Jul 1999 | WO |
WO 0171421 | Sep 2001 | WO |
WO-0192614 | Dec 2001 | WO |
WO 0245003 | Jun 2002 | WO |
WO 03026278 | Mar 2003 | WO |
WO 03071484 | Aug 2003 | WO |
WO-2004034696 | Apr 2004 | WO |
WO 2005015896 | Feb 2005 | WO |
WO 2005041558 | May 2005 | WO |
WO 2005076217 | Aug 2005 | WO |
WO 2005076217 | Aug 2005 | WO |
WO-2005087994 | Sep 2005 | WO |
WO-2005109853 | Nov 2005 | WO |
WO 2006018056 | Feb 2006 | WO |
WO-2006011635 | Feb 2006 | WO |
WO 2006045441 | May 2006 | WO |
WO-2007057063 | May 2007 | WO |
WO-2007057064 | May 2007 | WO |
WO-2007093199 | Aug 2007 | WO |
WO-2007093199 | Aug 2007 | WO |
WO-2007095553 | Aug 2007 | WO |
WO-2007095553 | Aug 2007 | WO |
WO-2007142621 | Dec 2007 | WO |
WO-2008023280 | Feb 2008 | WO |
WO-2008109644 | Sep 2008 | WO |
WO-2008109644 | Sep 2008 | WO |
Number | Date | Country | |
---|---|---|---|
20080211937 A1 | Sep 2008 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10919226 | Aug 2004 | US |
Child | 12119614 | US | |
Parent | 08947603 | Oct 1997 | US |
Child | 10170511 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10772767 | Feb 2004 | US |
Child | 10919226 | US | |
Parent | 10635862 | Aug 2003 | US |
Child | 10772767 | US | |
Parent | 10170511 | Jun 2002 | US |
Child | 10635862 | US |