PowerPoint Presentation - Research | Department...
-
Upload
hoangthuan -
Category
Documents
-
view
214 -
download
0
Transcript of PowerPoint Presentation - Research | Department...
Subtle Facial Expression Synthesis using Motion ManifoldEmbedding and Nonlinear Decomposable Generative Models
Chan-Su Lee1, Yang Wang2, Zhiguo Li1, Atul Kanaujia1, Ahmed Elgammal1, Dimitris Samaras2, Dimitris Metaxas1, Xiangfeng Gu2, and Peisen Huang3
1Computer Science, Rutgers University 2Computer Science, SUNY Stony Brook 3Mechanical Emgineering, SUNY Stony Brook{chansu,zhiguo,kanaujia,elgmmal, dnm}@cs.rutgers edu, {yangwang, samaras,gu}@cs.sunysb.edu, [email protected]
MotivationsModeling high resolution facial
expressions–Previous Work: Modeling person
specific facial motion style using high resolution tracking from range data [EG04]
–Extension 1: Model multidimensional variations like expression type across people in addition to person style
–Extension 2: Control and synthesis of high resolution facial expressions from single-view video sequences
Estimation of Parameters– To synthesize new facial expression, we need to know states of
facial expressions ( configuration, expression type, person style)
– Estimation of configuration for the known style and expression factors is a nonlinear 1-dimensional search problem
– Obtain style(expression type) conditional class probability by assuming a Gaussian density around the mean of the style classes
Conclusions and Future worksModeling dynamic facial expressions – Modeling nonlinear facial deformations in kernel map using conceptual
embedding– Decomposition of multiple factors (dynamics, person style, and expression type)
in facial expressions
Control of the high resolution facial expression model from video– Transferring style and expression weight estimated from video sequence based on
corresponding 3D and 2D training sequence and decomposition models– We plan to develop better models that couple 3D range data tracking and 2D tracking
Tracking of facial expressions from video–Single camera and no active
lighting –Tracking with coarse 3D
generic models
Components of high resolution facial expression synthesis system
Ndes NNN ⋅×× : 'B
b’se
Order-three coefficient tensor
= S’ F’E’'Z ×
1
×2 ×3
Core tensor Style basisExpression Basis
Mapping coefficient space basis'
3'
2'
1 FES ×××= 'Ζ'Bss NN ×
ee NN ×
dNdN NN ×
Ndes NNN ⋅×× : 'Z
)(321 tse
t xesy ψ×××=CGenerative Model
Ndes NNN ⋅×× : B
bse
Order-three coefficient tensor
= S FEZ ×1
×2 ×3
Core tensor Style basisExpression Basis
Mapping coefficient space basisFES 321 ×××= ΖB
ss NN ×ee NN ×
dNdN NN ×
Ndes NNN ⋅×× : Z
)(),,( ttt xseCysexE Ψ×××−=
∑=
eK
k
kke
1
β ∑=
sK
k
kk s
1
α
)),((),,|( ∑×××≈kskk xseCNesxyp ψ
),,|( exysp kk ∝α ),,|( sxyep k
k ∝β
Parameter estimation and transfer from Video sequences
Synthesis of Facial expressions with subtle differences
Extension 1: Decomposition of expression type across people–Conceptual manifold embedding based on motion parameters for uniform
representation of multiple motions in low dimensional space
–Kernel mapping to high dimensional space and then projection to motion parameter space
–Decomposition of projections to person style and expression type across peopleExtension 2: Control and synthesis of high resolution expressions from video–Correspondence of 3D range data and 2D video in training sequences
– Learning each decomposable model (style and expression) from training seq.
–Control of 3D model by varying style and expression parameters estimated from video
Contributions
Direct ControlFor Synthesis
Data Acquisition For Modeling
Learning Video Sequence:Coarse Facial Expression Decomposition
Estimation of Control Parameters
from Video Input
Learning 3D:Dense Facial Expression Decomposition
Learning Decomposed 3D
Facial Expression Model with Dense
Tracking Data
Learning Decomposed 3D
Facial Expression Model with Coarse
Tracking Data
Decomposition of Style & Expression
Decomposition of Style & Expression
Estimate Style & Expression
Factors for 3D
Estimate Style and Expression Factors
From Video Sequence
Synthesis of Dense Stylized
Facial Expression
Actor’s facial expressionSequence
Actor’s facial expressionSequence
3D RangeData
3D RangeData
Video SequenceVideo Sequence
Data Acquisition For Synthesis
Actor’s facial expression
Performance
Video Sequence
Animator’sExpression Style, Expression Type
Weighting
Key Observation:Although these facial expression bases associated to high resolution facial motions (from 3D range data) and low resolution facial motions (from 2D video) are different, an important observation is that a new expression vector estimated from different resolution tracking shares a similar distance to each expression basis