Claims
- 1. A computer software system for multi-modal human gender classification, comprising:
a first-mode classifier classifying first-mode data pertaining to male and female subjects according to gender and rendering a first-mode gender-decision for each male and female subject; a second-mode classifier classifying second-mode data pertaining to male and female subjects according to gender and rendering a second-mode gender-decision for each male and female subject; and a fusion classifier integrating the individual gender decisions obtained from said first-mode classifier and said second-mode classifier and outputting a joint gender decision for each of said male and female subjects.
- 2. A computer software system as set forth in claim 1, wherein said first mode classifier is a vision-based classifier; and
wherein said second mode classifier is a speech-based classifier.
- 3. A computer software system as set forth in claim 2, wherein said speech-based classifier comprises a support vector machine.
- 4. A computer software system as set forth in claim 2, wherein said first-mode classifier, second-mode classifier, and fusion classifier each comprise a support vector machine.
- 5. A computer software system for multi-modal human gender classification, comprising:
means for storing a database comprising a plurality of male and female facial images to be classified according to gender; means for classifying the male and female facial images according to gender; means for storing a database comprising a plurality of male and female utterances to be classified according to gender; means for classifying the male and female utterances according to gender; means for integrating the individual gender decisions obtained from the vision and speech based classification means to obtain a joint gender decision, said multi-modal gender classification having a higher performance measurement than the vision or speech based means individually.
- 6. A multi-modal method for human gender classification, comprising the following steps, executed by a computer:
generating a database comprising a plurality of male and female facial images to be classified; extracting a thumbnail face image from said database; training a support vector machine classifier to differentiate between a male and a female facial image, comprising determining an appropriate polynomial kernel and the bounds on Lagrange multiplier; generating a database comprising a plurality of male and female utterances to be classified; extracting a Cepstrum feature from said database; training a support vector machine classifier to differentiate between a male and a female utterance, comprising determining an appropriate Radial Basis Function and the bounds on Lagrange multiplier; integrating the individual gender decisions obtained from the speech and vision based support vector machine classifiers, using a semantic fusion method, to obtain a joint gender decision, said multi-modal gender classification having a higher performance measurement that the speech or vision based modules individually.
- 7. The method of claim 6 wherein the performance of the support vector machine classifier is further augmented, comprising the steps of:
testing the support vector machine classifier by employing a plurality of refinement male and female facial images to be classified by the support vector machine classifier according to gender; and using the refinement facial images for which gender was improperly detected to augment and reinforce the support vector machine learning process.
- 8. The method of claim 7 wherein the performance of the support vector machine classifier is further augmented, comprising the steps of:
testing the support vector machine classifier by employing a plurality of refinement male and female utterances to be classified by the support vector machine classifier according to gender; and using the refinement utterances for which gender was improperly detected to augment and reinforce the support vector machine learning process.
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application is based on and claims priority to U.S. Provisional Application No. 60/330,492, filed Oct. 16, 2001, which is fully incorporated herein by reference.
STATEMENT REGARDING FEDERALLY-SPONSORED RESEARCH
[0002] This development was supported in part by the NSF Career Grant IIS-97-33644 and NSF Grant IIS-0081935. The government may have certain rights in this invention.
Provisional Applications (1)
|
Number |
Date |
Country |
|
60330492 |
Oct 2001 |
US |