This application claims the benefit of priority under 35 USC 119 of Japanese application no. 2008-079232, filed on Mar. 25, 2008, which is incorporated herein by reference.
1. Technical Field
The present invention relates to detecting a face area in an image.
2. Related Art
Techniques of detecting a face area corresponding to an image of a face are disclosed, for example, in JP-A-2007-94633. In this disclosure, sections of an image represented by image data are successively extracted and it is then determined whether each extracted section is an image of a face.
The face area is preferably detected at a high speed.
The present invention provides a technique of detecting a face area in an image at a high speed, and may be embodied in one of the following aspects.
In accordance with one aspect of the invention, an image processing apparatus includes an application identification information acquisition unit that acquires application identification information that identifies an application of detection results of a face area corresponding to an image of a face in a target image, a condition setting unit that sets a condition, relating to an angle of the face to be represented in the face area, based on the application identification information, and a face area detection unit that detects the face area corresponding to the image of the face satisfying the set condition, based on image data representing the target image.
The image processing apparatus acquires the application identification information that identifies an application of the detection results of the face area corresponding to the image of the face in the target image, sets the condition, relating to the angle of the face to be represented in the face area, based on the application identification information, and detects the face area corresponding to the image of the face satisfying the set condition, based on image data representing the target image. A face area detection process is thus performed at a high speed.
The condition may relate to at least one of an angle of rotation of the face to be represented in the face area about an axis substantially perpendicular to a plane of the image and an angle of rotation of the face about an axis extending substantially in parallel with the plane of the image.
The image processing apparatus detects the face area corresponding to the image of the face satisfying the condition of at least one of an angle of rotation of the face to be represented in the face area about an axis substantially perpendicular to the plane of the image and an angle of rotation of the face about an axis extending in substantially parallel with the plane of the image. The face area detection process is thus performed at a high speed.
In the image processing apparatus, the face area detection unit may include a determination target setter that sets a determination target image area as an image area on the target image, a storage that stores a plurality of pieces of evaluation data, the plurality of pieces of evaluation data being predetermined in response to a value set as the condition and being used to calculate an evaluation score that represents a likelihood that the determination target image area is an image area corresponding to the image of the face satisfying the set condition, an evaluation score calculator that calculates the evaluation score, based on the evaluation data corresponding to the set condition and image data corresponding to the determination target image area, and an area setting unit that sets the face area based on the evaluation score and a position and a size of the determination target image area.
In the image processing apparatus, the determination target image area is set as the image area on the target image, and a plurality of pieces of evaluation data are stored. The plurality of pieces of evaluation data are predetermined in response to the value set as the condition and are used to calculate the evaluation score that represents a likelihood that the determination target image area is the image area corresponding to the image of the face satisfying the set condition. The evaluation score is calculated, based on the evaluation data corresponding to the set condition and image data corresponding to the determination target image area. The face area is set based on the evaluation score and the position and the size of the determination target image area. The face area detection process is thus performed in the target image at a high speed.
The area setting unit may determine based on the evaluation score whether the determination target image area is the image area corresponding to the image of the face satisfying the set condition, and set the face area based on the position and size of the determination target image area that has been determined as the image area corresponding to the image of the face satisfying the set condition.
In the image processing apparatus, it is determined based on the evaluation score whether the determination target image area is the image area corresponding to the image of the face satisfying the set condition. The face area is set based on the position and size of the determination target image area that has been determined as the image area corresponding to the image of the face satisfying the set condition. The face area detection process is thus performed in the target image at a high speed.
The application may include a predetermined image processing process that processes the image area set in accordance with the detected face area.
The image processing apparatus sets the condition in accordance with the predetermined image processing process that processes the image area set in accordance with the detected face area, and detects the face area corresponding to the image of the face satisfying the condition. The face area detection process is thus performed in the image at a high speed.
The predetermined image processing process may include at least one of a skin complexion correction operation, a metamorphosis operation, a red-eye correction operation, a determination operation determining an expression of the image of the face, and a detection operation detecting the image of the face having a particular tilt and a particular size.
The image processing apparatus sets the condition in response to at least one of the skin complexion correction operation, the metamorphosis operation, the red-eye correction operation, the determination operation determining the expression of the image of the face, and the detection operation detecting the image of the face having the particular tilt and the particular size, and detects the face area of the image of the face satisfying the condition. The face area detection process is thus performed in the target image at a high speed.
The image processing apparatus may further include an image generator that generates the image data by capturing an image of a subject. The application sets a timing of capturing the image of the subject.
The face area detection process is used to set the timing of image capturing. The image processing apparatus can thus perform the face area detection process at a high speed.
The invention may be embodied in a variety of modes. For example, the invention may be embodied as an image processing method, an image processing apparatus, a face area detection method, a face area detection apparatus, a computer program for performing these methods and functions of these apparatuses, a recording medium storing the computer program, or the like.
The invention will be described with reference to the accompanying drawings, wherein like numbers reference like elements.
The printer engine 160 is a print mechanism for performing a print operation in response to print data. The card interface 170 exchanges data with a memory card MC inserted into a card slot 172. In the first embodiment, the memory card MC stores a image file containing image data.
The internal memory 120 is a computer-readable medium including an image processor 200, a display processor 310, and a printer processor 320. Under the control of a predetermined operating system, the image processor 200 is a computer program embodied in memory 120 that executes an image processing process to be discussed later. The display processor 310 as a display driver controls the display unit 150, thereby displaying on the display unit 150 a process menu, a message, an image, etc. The printer processor 320 as a computer program embodied in memory 120 controls the printer engine 160, thereby printing an image responsive to the print data. The CPU 110 reads from the internal memory 120 these programs and executes the read programs in order to perform the function of each element.
The image processor 200, as a program module, includes an area detector 210, a processing mode setter 220, and a condition setter 230. The area detector 210 detects a face area of an image of a subject of a predetermined type (an image of a face and an image of a part of the face) in a target image represented by target image data. The area detector 210 includes a determination target setter 211, an evaluation score calculator 212, a determiner 213, and an area setter 214. The functions of these elements will be described later in the discussion of the image processing process. The area detector 210 functions as a face area detection unit that detects a face area of an image of a face. The determiner 213 and the area setter 214 function as an area setting unit.
The processing mode setter 220 sets a mode of the image processing process to be executed. The processing mode setter 220 includes a designation acquisition unit 222 that acquires a mode of the image processing process designated by the user. The condition setter 230 sets a condition related to an angle or the like of a face displayed in a face area detected in the image processing process.
The internal memory 120 stores a plurality of pieces of preset face learning data FLD and a plurality of pieces of preset part learning data OLD. The face learning data FLD and the part learning data OLD are used when the area detector 210 detects a predetermined image area.
The content of the face learning data FLD is described later in the discussion of the image processing process. The face learning data FLD is set in association with a face tilt and a face orientation of a face. The face tilt means the tilt of the face (angle of rotation) within the plane of the image (in-plane tilt). More specifically, the face tilt is an angle of rotation about an axis substantially perpendicular to the plane of the image. In accordance with the first embodiment of the invention, a tilt of each of an area of a target image, a subject, and other feature is represented by an angle of clockwise rotation from a reference state (with the tilt being zero). The reference state is a state in which an upward looking direction of each of the area of the target image, the subject, and the other feature matches an upward looking direction of the target image. More specifically, the face tilt is represented by an angle of clockwise rotation of the face from the reference state. The reference state (with the face tilt=O) is a state in which the face is positioned along a substantially vertical direction of the target image (with the top of the head looking upward and the jaw looking downward).
The face orientation means an out-plane orientation of the face (a turnaround angle of the face). The turnaround angle of the face refers to an angle at which the face looks about an axis of the neck having a substantially circular cylindrical shape. More specifically, the face orientation is an angle of rotation of the face about an axis extending substantially in parallel with the plane of the image. In accordance with the first embodiment of the invention, the orientation of the face looking toward an imaging plane of an image generating apparatus such as a digital still camera is referred to as a “full-faced orientation,” the orientation of the face looking rightward if viewed from the imaging plane is referred to as a “rightward looking orientation” (with the image of the face looking leftward if viewed from an observer of the image) and the orientation of the face looking leftward if viewed from the imaging place is referred to as a “leftward looking orientation” (with the image of the face looking leftward if viewed from the observer of the image).
The internal memory 120 stores four pieces of face learning data FLD as illustrated in
As will be described later, the face learning data FLD responsive to a given face tilt is so set through learning that an image of a face having a face tilt falling with a range of from −15° to +15° with respect to the given face tilt is detected. The human face is generally symmetrical. If two pieces of data at the full-faced face orientation, one for the face learning data FLD at a face tilt of 0° (
The part learning data OLD is set in association with a combination of a type of a part of the face and a part tilt. In accordance with the first embodiment of the invention, the face parts include the eyes (the right and left eyes) and the mouth. The part tilt means a tilt of a part of the face (angle of rotation) within the image (in-plane tilt). More specifically, the part tilt refers to an angle of rotation of the part of the face about an axis substantially perpendicular to the plane of the image. As the face tilt, the part tilt is represented by a clockwise angle of rotation of the part of the face from the reference state (with the part tilt=0). The reference state is a state in which the part of the face is positioned in a substantially vertical direction of the target image.
The internal memory 120 stores four pieces of part learning data OLD as illustrated in
As the face learning data FLD, the part learning data OLD responsive to a given part tilt is so set through learning that an image of a part having a part tilt falling with a range of from −15° to +15° with respect to the given pat tilt is detected. The human eyes and mouth are generally symmetrical. If two pieces of data of the eye, one for the part learning data OLD at a part tilt of 0° (
The internal memory 120 (
In step S110 of the image processing process, the processing mode setter 220 sets the mode of the image processing process. More specifically, the processing mode setter 220 controls the display processor 310, thereby causing a user interface to be displayed on the display unit 150 for image processing mode setting.
Skin complexion correction is an image processing mode for correcting the skin complexion of a person to a favorite color. Face metamorphosis is an image processing mode for metamorphosing an image in a face area or an image in an image area containing the image of the face set in the face area. Red-eye correction is an image processing mode for correcting the color of the eye suffering for the red-eye effect to a natural color. Smiling face detection is an image processing mode for detecting an image of a face of a smiling person. Identity photograph detection is an image processing mode for detecting an image appropriate for an identity photograph.
When the user selects one of the image processing modes using the operation unit 140, the designation acquisition unit 222 acquires information identifying the image processing mode selected and designated (hereinafter referred to as “image processing mode identification information”). The processing mode setter 220 sets the image processing mode, identified by the image processing mode identification information, as the image processing mode to be executed. In the image processing mode, a predetermined process is performed on a face area (or an image area set based on the face area) detected in a face area detection process (step S140 in
In step S120, the condition setter 230 sets a condition related to an angle or the like of a face to be displayed in the face area detected in the face area detection process, based on the image processing mode identification information and the condition table CT. On the basis of the set condition, the condition setter 230 sets the learning data and a window size to be used.
The detection tilt specified in the condition table CT is a tilt of a face to be displayed in the image area detected as a face area in an image (a face detection image FDImg (
In identity photograph detection, the detection tilt is defined as a range of from −15° to +15° with respect to each of 90° and 270° (i.e., from 75° to 105° and from 255° to 285°). The detection tilt is defined because the identity photograph is typically in a portrait frame with the frame vertical sides in substantially parallel with vertical sides of the face. If the image processing mode is identity photograph detection, the condition setter 230 sets the detection tilt within a range of from −15° to +15° with respect to each of 90° and 270°.
The detection orientation specified in the condition table CT is a face orientation of a face displayed in the image area to be detected as the face area in the image (face detection image FDImg) handled as a target of the face area detection process. In skin complexion correction and red-eye correction, the detection orientations include all of the full-faced face, rightward looking and leftward looking orientations. This is because each of skin complexion correction and red-eye correction is performed regardless of the orientation of the face. If one of the image processing modes is one of skin complexion correction and red-eye correction, the condition setter 230 sets all the orientations as the detection orientation.
Only the full-faced face orientation is specified as the detection orientation in each of face metamorphosis, smiling face detection and identity photograph detection. This is because if the metamorphosis process is performed on the face that is in a rightward looking position or a leftward looking position (other than a full-faced position), the process results are likely to become unnatural. Smiling face detection and identity photograph detection are based on the premise that the full-faced face is handled as a detection target. If the image processing mode is one of face metamorphosis, smiling face detection, and identity photograph detection, the condition setter 230 sets the full-faced face orientation as the detection orientation.
The condition setter 230 sets the face learning data FLD for use in the face area detection process based on the combination of the detection tilt and the detection orientation specified in the condition table CT. More specifically, the condition setter 230 sets as the face learning data FLD in use the face learning data FLD responsive to the detection tilt and the detection orientation so that the face area responsive to the image of the face having the detection tilt and the detection orientation specified in the condition table CT is detected. More specifically, if the image processing mode is one of skin complexion correction and red-eye correction, face learning data FLD responsive to combinations of all face tilts and all face orientations is set as the learning data in use. If the image processing mode is one of face metamorphosis and smiling face detection, face learning data FLD responsive to combinations of all face tilts and the full-faced face orientation is set as the learning data in use. If the image processing mode is identity photograph detection, face learning data FLD responsive to combinations of face tilts of 90° and 270° and the full-faced face orientation is set as the learning data in use.
The detection part specified in the condition table CT is a type of the part of the face to be detected in a part area detection process (step S180 in
The condition setter 230 sets the detection part specified in the condition table CT as the type of a part of a face to be detected in the part area detection process. The condition setter 230 sets the part learning data OLD to be used in the part area detection process, in accordance with the detection part specified in the condition table CT. More specifically, the condition setter 230 sets as the part learning data OLD the part learning data OLD responsive to the detection part so that a part area responsive to the detection part specified in the condition table CT is detected in accordance with the set image processing mode. More specifically, if the image processing modes are face metamorphosis and identity photograph detection, the part learning data OLD for the eye and the part learning data OLD for the mouth is set as the learning data to be used. If image processing mode is red-eye correction, the part learning data OLD for the eye is set as the learning data to be used. If image processing mode is smiling face detection, the part learning data OLD for the mouth is set as the learning data to be used.
The detection size specified in the condition table CT is a size of an image area to be detected as the face area in the image (face detection image FDImg) serving as a target of the face area detection process (step S140 in
Referring to
The condition setter 230 sets, as the size of the image area to be detected as the face area in the face area detection process, the detection size specified in the condition table CT. The condition setter 230 also sets an operative window size so that the face area matching the detection size specified in the condition table CT is detected. As will be described later, a square window SW is placed on the face detection image FDImg with the size and position thereof modified in the face area detection process. The face area is detected by determining whether a determination target image area JIA, as an image area on the face detection image FDImg, defined by the placed window SW, corresponds to an image area corresponding to the image of the face (see
The condition setter 230 sets as the operative window size a standard size falling within the range of the detection size specified in the condition table CT out of the standard sizes of the fifteen windows. The detection size used in skin complexion correction and red-eye correction ranges from 20-240 pixels, and all the standard sizes of the 15 windows SW are set as the operative window sizes.
The detection size in face metamorphosis ranges from 60-180 pixels as illustrated in
In step S130 of the image processing process, the image processor 200 acquires image data representing an image as a target of the image processing process. The display unit 150 displays a thumbnail image of an image file stored on the memory card MC inserted in the card slot 172. The user selects at least one image to be processed using the operation unit 140 while monitoring the displayed thumbnail image. The image processor 200 acquires from the memory card MC an image file containing image data responsive to at least one image selected, and stores the image file onto a predetermined area of the internal memory 120. The acquired image data is referred to as original image data, and an image represented by the original image data is referred to as an original image OImg.
In step S140, the area detector 210 performs the face area detection process. In the face area detection process, the image area corresponding to the image of the face is detected as a face area FA. The operative learning data and the operative window size are used in the face area detection process.
In step S310 of the face area detection process (
In step S320, the determination target setter 211 sets an initial value of the size of the window SW for use in setting the determination target image area JIA (to be discussed later). In step S330, the determination target setter 211 places the window SW at an initial position on the face detection image FDImg. In step S340, the determination target setter 211 sets as the determination target image area JIA the image area defined by the window SW placed on the face detection image FDImg. The determination target image area JIA serves as a target area, i.e., a determination is made as to whether the determination target image area JIA is the face area corresponding to the image of the face. As illustrated in the middle portion of
In step S350, the evaluation score calculator 212 calculates a cumulative evaluation score Tv for use in the face determination in the determination target image area JIA on the basis of the image data of the determination target image area JIA. The face determination is performed based on each combination responsive to the image processing mode to be performed, out of the predetermined combinations of particular face tilts and particular face orientations, using the operative learning data set in step S120. More specifically, it is determined for each combination responsive to the image processing process to be performed whether the determination target image area JIA is the face area responsive to the image of the face having the particular face tilt and the particular face orientation. For this reason, the cumulative evaluation score Tv is also calculated for each of the combinations of particular face tilt and particular face orientations responsive to the image processing process to be performed. The particular face tilt means a specific face tilt. In accordance with the first embodiment, there are twelve particular face tilts including a reference face tilt (face tilt=0 degrees), and eleven other face tilts increasing in steps of 30 degrees (30°, 60°, . . . , 330°). The predetermined face orientation is a specific face orientation. In accordance with the first embodiment of the invention, there are three predetermined face orientations, i.e., the full-faced face orientation, the rightward looking orientation, and the leftward looking orientation.
The calculated evaluation score vX is compared with a threshold value thX set for each evaluation score vX (i.e., th1-thN). If the evaluation score vX is equal to or higher than the threshold value thX, it is determined that the determination target image area JIA is the image area corresponding to the image of the face with respect to the filter X, and a value “1” is set for the output value for the filter X. If the evaluation score vX is lower than the threshold value thX, it is determined that the determination target image area JIA is not the image area corresponding to the image of the face with respect to the filter X, and a value “0” is set for the output value for the filter X. Weight coefficients Wex (i.e., We1-WeN) are set respectively for the filters X. The sum of products of the outputs of each filter X and the respective weights Wex is calculated as the cumulative evaluation score Tv.
The specifications of each filter X, the threshold value thX, the weight coefficient Wex, and a threshold value TH are specified in the face learning data FLD. For example, the specifications of each filter X, the threshold value thX, the weight coefficient Wex, and the threshold value TH, specified in the face learning data FLD responsive to a combination of the full-faced face orientation and a face tilt of 0° (see
The face learning data FLD is set through a learning process of a sample image.
The face sample image group includes images of faces having twelve particular face tilts as shown in
The face sample image group responsive to each particular face tilt includes a plurality of face sample images, each face sample image having within a predetermined range a ratio of a size of the image of the face to the image size and the tilt of the image of the face equal to a particular face tilt (hereinafter referred to as a “reference face sample image FIo”). The face sample image group includes at least an image that is obtained by scale-expanding or scale-contracting the reference face sample image FIo within a range of 1.2-0.8 times (such as one of images FIa and FIb in
The learning process using the sample image is performed using a method of a neural network, a method of boosting (such as AdaBoosting), or a method of using a support vector machine. If the learning process is performed using a neural network, the evaluation score vX (vX1-vN) is calculated with respect to each filter 1-N (
An initial value is set for the weight coefficient Wex (We1-WeN) for the filter X, and the cumulative evaluation score Tv is calculated for one sample image selected from the face sample image group and the non-face sample image group. As will be discussed later, the image is determined as the image corresponding to the image of the face if it is determined in the face determination that the cumulative evaluation score Tv calculated for the image is equal to or higher than the predetermined threshold value TH. The weight coefficient Wex set for each filter X is modified in response to the threshold value determination results of the cumulative evaluation score Tv calculated for the selected sample image (one of the face sample image and the non-face sample image). The selection of the sample image, the threshold value determination based on the cumulative evaluation score Tv calculated for the selected sample image, and the modification of the weight coefficient Wex based on the threshold value determination are repeated for all the sample images contained in the face and non-face sample image groups. Through this process, the face learning data FLD responsive to the combination of the full-faced face orientation and the particular face tilt is set.
Similarly, the face learning data FLD responsive to the other particular face orientation (the rightward looking face and the leftward looking face) is also set through the learning process. The learning process uses a face sample image group composed of a plurality of face sample images known beforehand as images responsive to the other particular face orientation (the rightward looking orientation and the leftward looking orientation) and a non-face sample image group composed a plurality of non-face sample images known beforehand as images not responsive to the rightward looking face (or the leftward looking face).
If the image processing mode is one of skin complexion correction and red-eye correction, face learning data FLD responsive to the combinations of all the particular face orientations and all the particular face tilts is set as operative learning data. The cumulative evaluation score Tv based on the face learning data FLD is thus calculated. Similarly, if the image processing mode is one of face metamorphosis and smiling face detection, face learning data FLD responsive to the combinations of all the particular face orientations and all the particular face tilts is used to calculate the cumulative evaluation score Tv. If the image processing mode is identity photograph detection, the face learning data FLD responsive to the combinations of the particular face tilts of 90° and 270° and the full-faced face orientation is used to calculate the cumulative evaluation score Tv.
When the cumulative evaluation score Tv is calculated for each combination of a particular face orientation and tilt in the determination target image area JIA (step S350 in
In step S380, the area detector 210 determines whether the entire face detection image FDImg is scanned with currently set window SW. If the entire face detection image FDImg is not scanned with currently set window SW, the determination target setter 211 moves the window SW by a predetermined distance in a predetermined direction (step S390).
If it is determined in step S380 (
If all the operative window sizes set in step S120 have been used, the area setter 214 executes a face area setting process (step S420).
A plurality of windows SW mutually overlapping each other at a given particular face tilt may be stored in step S370. In such a case, the area setter 214 sets a new window SW having as the center of gravity thereof the average of the coordinates of predetermined points (centers of gravity) of the windows SW and the average size of the sizes of the windows SW. Such a new window SW is referred to as an “average window AW.” If four windows SW1-SW4 mutually overlapping each other are stored as illustrated in
A single window SW having no overlapping portion with another window SW, as illustrated in
In accordance with the first embodiment, the face sample image group (see
If the face area FA is not detected in the face area detection process (step S140 of
If the image processing mode set in step S110 is other than skin complexion correction, a detection part is specified in the condition table CT. It is thus determined in step S160 that the part area detection process is to be performed (YES in step S160). Processing proceeds to step S170 where the area detector 210 selects one detected face area FA.
In step S180, the area detector 210 performs the part area detection process. In the part area detection process, the image area corresponding to the image of a part of the face in the selected face area FA is detected as a part area. In response to the image processing mode set in step S110, the area detector 210 detects a part corresponding to the detected part associated with image processing mode in the condition table CT. For example, if the image processing mode is red-eye correction, the area detector 210 detects a right-eye area EA(r) for the image of the right eye and a left-eye area EA(l) for the image of the left eye. If the image processing mode is smiling face detection, the area detector 210 detects a mouth area MA for the image of the mouth.
In step S510 of the part area detection process (
The detection of the part area from the part detection image ODImg may be performed in the same manner as the detection of the face area FA from the face detection image FDImg. More specifically, as illustrated in
The cumulative evaluation score Tv is calculated at each particular face tilt in the part area detection process (step S140 of
If the cumulative evaluation score Tv calculated at each detection part is equal to or higher than the threshold value TH, the determination target image area JIA is the image area corresponding to the image of the part. The position of the determination target image area JIA, i.e., the coordinates of the currently set window SW are stored (step S570 in
In step S190 (
In step S200, the image processor 200 executes the image processing mode set in step S110. More specifically, if the image processing mode is skin complexion correction, the skin complexion of a person in the face area FA or in the image area containing the image of the face set in accordance with the face area FA is corrected to a preferable complexion. If the image processing mode is face metamorphosis, the face area FA is adjusted based on the positional relationship of the detected part area (the right-eye area EA(r), the left-eye area EA(l) and the mouth area MA), and the image within the adjusted face area FA or the image within the image area containing the image of the face set in accordance with the adjusted face area FA is metamorphosed. If the image processing mode is red-eye correction, a red-eye image is detected from the part area (the right-eye area EA(r) and the left-eye area EA(l)) detected in the face area FA and is then adjusted to be closer to a natural eye color. If the image processing mode is smiling face detection, an outline of the detected face area FA or an outline of the detected part area (mouth area MA) is detected. A smiling face determination of whether the image within the face area FA is the image of a smiling face is performed by evaluating the opening of the corner of the mouth. Techniques on smiling face determination are disclosed in JP-A-2004-178593 and the paper entitled “A Study For Moving Object Tracking In Scene Changing Environment” authored by Yoshitaka SOEJIMA, Feb. 15, 1998, Japan Advanced Institute of Science and Technology (JAIST). If the image processing mode is identity photograph detection, the face area FA is adjusted based on the positional relationship of the detected part area (the right-eye area EA(r), the left-eye area EA(l) and the mouth area MA), and it is then determined whether the target image is an identity photograph.
In the image processing process of the printer 100, information identifying the image processing mode as an application of the face area detection is acquired. The condition related to the angle of the face to be displayed in the face area FA is set based on the acquired information. The face area FA corresponding to the image of the face matching the set condition is detected in accordance with the image data representing the target image. In the face area detection process (step S140 in
The face area sorter 240 sorts the face areas FA into matching face areas FAf and unmatching face areas FAn. The matching face area FAf corresponds to the image of the face satisfying the condition of the detection size, the detection tilt, and the detection orientation defined according to the image processing mode to be performed in the condition table CT. The unmatching face area FAn corresponds to the image of the face not satisfying the condition.
The process content in step S110 of the image processing process (
The process content in steps S130-S190 in the image processing process (
In step S192 of the image processing process (
In the displaying of the face area detection results, the face area sorter 240 (
As illustrated in
As in the first embodiment, the image processing mode set in step S110 is performed subsequent to the displaying of the face area detection results (step S192 in
In the image processing process of the printer 100a as described above, the face area FA is detected from the target image. The face areas FA are sorted into the matching face area FAf and the unmatching face area FAn. The face areas FA are then displayed on the display unit 150 in a manner such that the matching face area FAf is distinctively identified from the unmatching face area FAn. In the image processing process, the printer 100a of the second embodiment causes the detected face area FA to be determined as to whether the face area FA matches the application of the face area FA.
The invention is not limited to the above-described embodiments. Changes and modifications are possible without departing from the scope of the invention. For example, the following modifications may be performed.
The image processing modes (
The detection size, the detection tilt, the detection orientation, and the detection part defined in accordance with the image processing mode in the condition table CT (
The image processing mode to be performed is set by the user. Alternatively, the image processing mode may be automatically set. The detection size, tilt, orientation and part in the face area and part area detection processes are set in accordance with the image processing mode. Alternatively, the detection size, tilt, orientation and part may be set directly by the user or automatically.
The image processing apparatus is the printer 100 in the above-described embodiments. The invention is also applicable to an imaging apparatus such as a digital still camera or a digital video camera. When the invention is applied to an imaging apparatus such as a digital still camera or a digital video camera, the application of the face area detection results may be the setting of an image capturing timing. In other words, the condition relating to at least one of the detection size, tilt, orientation and part is set as in the condition table CT. If a face area satisfying the set condition is detected in a preparation image generated at the moment an image capturing button is pressed, image capturing is automatically performed, or an image capturing instruction is issued by pressing fully the image capturing button.
In the above-described embodiments, the part detection image ODImg is generated and the determination target image area JIA is set up with the window SW placed on the generated part detection image ODImg. The part detection image ODImg is not necessarily generated, and the determination target image area JIA is set up with the window SW placed on the face detection image FDImg. In such a case, the cumulative evaluation score Tv may be calculated using the part learning data OLD responsive to the same tilt as the tilt of detected face area FA. Conversely, it the part detection image ODImg is generated, the cumulative evaluation score Tv can be calculated using only the part learning data OLD responsive to a tilt of 0°. It suffices if only the part learning data OLD responsive to a tilt of 0° is stored on the internal memory 120.
In accordance with the first embodiment of the invention, the condition (see
The display format of the matching face area FAf and the unmatching face area FAn on the display unit 150 in accordance with the second embodiment are illustrated for exemplary purposes only. A different display format is perfectly acceptable as long as the matching face area FAf is clearly distinguished from the unmatching face area FAn. For example, the matching face area FAf and the unmatching face area FAn may be displayed within regions or within outlines of the regions, with the regions and the outlines of the regions distinguished by color difference. One of the matching face area FAf and the unmatching face area FAn may be flashed in display for distinction. Either or both of the matching face area FAf and the unmatching face area FAn may be labeled with characters or symbols for distinction.
In the above-described embodiments, the face area detection process (
In accordance with the above-described embodiments, the face and part determinations are performed by comparing the cumulative evaluation score Tv with the threshold value TH (see
In the above-described embodiments, twelve particular face tilts are set in steps of 30°. Face tilts of more than or less than twelve may be set. The particular face tilts are not necessarily set. The face determination may be performed at a face tilt of 0°. In the above-described embodiments, the face sample image group includes image obtained by scale-contracting and scale-expanding the reference face sample image FIo and image obtained by rotating the reference face sample image FIo. The face sample image group does not necessarily include such images.
In the above-described embodiments, the determination target image area JIA may be determined as corresponding to the image of the face (the image of the part of the face) in the face determination (the part determination) in the determination target image area JIA specified by the window SW of a given size. When the window SW having a size smaller than a predetermined rate than the given size is placed, the determination target image area JIA image area determined as corresponding to the image of the face may be avoided. In this way, the process speed is increased.
In the above-described embodiments, image data stored on the memory card MC is set as the original image data. The original image data is not limited to image data stored on the memory card MC. For example, the original image data may be received via a network.
In the above-described embodiments, the right eye and the left eye are set, and the right-eye area EA(r), the left-eye area EA(l), and the mouth area MA are then detected. The type of part of the face to be set is modifiable. For example, either of the right and left eyes or both of the right and left eyes may be set as the type of part of the face. In addition to the right eye, the left eye, and the mouth, or instead of at least one of the right eye, the left eye, and the mouth, another type of part of the face (such as the nose or the eyebrow) may be set, and an area corresponding to such a part may be detected as the part area.
In the above-described embodiments, the face area FA and the part area are rectangular. Alternatively, the face area FA and the part area may have another shape other than a rectangular shape.
The above-described embodiments are related to the image processing process performed by the printer 100 as the image processing apparatus. Part or all of the image processing process may be performed by another type of image processing apparatus, such as a personal computer, a digital still camera, or a digital video camera. The printer 100 is not limited to an ink jet printer. The printer 100 may be another type of printer such as a laser printer or a sublimation printer.
In the above-described embodiments, part of the embodiment implemented using hardware may be implemented using software. Conversely, part of the embodiment implemented using software may be implemented using hardware.
If part or all of the functions of the invention are implemented using software (a computer program), the software may be supplied on a computer readable recording medium. A “computer readable recording medium” in the invention is not limited to a removable recording medium such as a flexible disk or a CD-ROM, but also includes an internal recording device such as a RAM, a ROM, and an external recording device fixed to a computer, such as a hard disk.
Number | Date | Country | Kind |
---|---|---|---|
2008-079232 | Mar 2008 | JP | national |