Claims
- 1. A method for generating a photorealistic talking head, comprising:receiving an input stimulus; reading data from a first library comprising one or more parameters associated with mouth shape images of sequences of at least three concatenated phonemes which correspond to the input stimulus; reading, based on the data read from the first library, corresponding data from a second library comprising images of a talking subject; and generating, using the data read from the second library, an animated sequence of a talking head tracking the input stimulus.
- 2. The method of claim 1, further comprising the steps of:reading acoustic data from the second library associated with the corresponding image data read from the second library; converting the acoustic data into sound; and outputting the sound in synchrony with the animated sequence of the talking head.
- 3. The method of claim 2, wherein the data read from the first library comprises one or more equations characterizing mouth shapes.
- 4. The method of claim 2, wherein said converting step is performed using a data-to-voice converter.
- 5. The method of claim 2, wherein the data read from the second library comprises segments of sampled images of a talking subject.
- 6. The method of claim 5, wherein said first library comprises a coarticulation library, and wherein said second library comprises an animation library.
- 7. The method of claim 5, wherein said generating step is performed by overlaying the segments onto a common interface to create frames comprising the animated sequence.
- 8. The method of claim 2, wherein the data read from the first library comprises mouth parameters characterizing degree of lip opening.
- 9. The method of claim 2, wherein said receiving, said generating, said converting, and all said reading steps are performed on a personal computer.
- 10. The method of claim 2, wherein said first and second libraries reside in a memory device on a computer.
- 11. The method of claim 1, wherein the data read from the first library comprises one or more equations characterizing mouth shapes.
- 12. A method for generating a photorealistic talking entity, comprising:receiving an input stimulus; reading, first data from a library comprising one or more parameters associated with mouth shape images of sequences of two concatenated phonemes and images of commonly-used sequences of at least three concatenated phonemes which correspond to the input stimulus; reading, based on the first data, corresponding second data comprising stored images; and generating, using the second data, an animated sequence of a talking entity tracking the input stimulus.
- 13. A method for generating a photorealistic talking entity, comprising:receiving an input stimulus; reading, based on at least one diphone, first data comprising one or more parameters associated with mouth shape images of sequences of concatenated phonemes which correspond to the input stimulus, the first data stored in a library comprising images of sequences associated with diphones and the most common images associated with triphones; reading, based on the first data, corresponding second data comprising stored images; and generating, using the second data, an animated sequence of a talking entity tracking the input stimulus.
- 14. The method of claim 13, wherein reading first data is based on at least one triphone.
- 15. The method of claim 13, wherein reading first data is based on at least one quadriphone.
Parent Case Info
This is a Continuation of application Ser. No. 08/965,702 filed Nov. 7, 1997, now U.S. Pat. No. 6,112,177. The entire disclosure of the prior application is hereby incorporated by reference herein in its entirety.
US Referenced Citations (13)
Continuations (1)
|
Number |
Date |
Country |
Parent |
08/965702 |
Nov 1997 |
US |
Child |
09/390704 |
|
US |