research onVisual Speech attractsmore and more attention. It has become a very active research field of the Human-Machine Interface. The chief information relative to speech is lip motion
face
and even the whole head
which is called “Talking Head”. To synthesis a lifelike three-dimension (3D) talking head model
a novel method is proposed in this paper
which is based on an individual independent 3D-model and photos of human face. At first
the features of face shape and the position of facial organs are extracted from a front-face and a side-face phototo revise the 3D-model and make it adaptthe real person. Then
the textures of the skin and hair are picked from the photos and pasted on the revised 3D-model to make it looks like the person. This method integrates the techniques of 3D-model based modeling and photo lib based modeling
and has both of their advantages: the model has strong flexibility of synthesizing lip motions and expressions
can be rotated freely
can be synthesized in real-time
and can achieve a highly natural
lifelike 3D talking head visual effect. Then
the model is applied in a visualText-to-Speech (TTS) talking head synthesis system