Figure-centric averages Antonio Torralba & Aude Oliva (2002) Averages: Hundreds of images containing...
-
Upload
maude-hardy -
Category
Documents
-
view
213 -
download
0
Transcript of Figure-centric averages Antonio Torralba & Aude Oliva (2002) Averages: Hundreds of images containing...
Figure-centric averages
Antonio Torralba & Aude Oliva (2002)Averages: Hundreds of images containing a person are averaged to reveal regularities
in the intensity patterns across all the images.
More by Jason Salavon
More at: http://www.salavon.com/
Computing Means
Two Requirements:• Alignment of objects• Objects must span a subspace
Useful concepts:• Subpopulation means• Deviations from the mean
How to align faces?
http://www2.imm.dtu.dk/~aam/datasets/datasets.html
Average Face
1. Warp to mean shape2. Average pixels
http://graphics.cs.cmu.edu/courses/15-463/2004_fall/www/handins/brh/final/
Subpopulation meansExamples:
• Happy faces
• Young faces
• Asian faces
• Etc.
• Sunny days
• Rainy days
• Etc.
• Etc.
Average male
Average female
Manipulating Facial Appearance through Shape and Color
Duncan A. Rowland and David I. Perrett
St Andrews University
IEEE CG&A, September 1995
Face ModelingCompute average faces
(color and shape)
Compute deviations between male and female (vector and color differences)
Changing gender
Deform shape and/or color of an input face in the direction of “more female” original shape
color both
Linear Subspace: convex combinations
m
iiiXaX
1
Any new image X can beobtained as weighted sum of stored “basis” images.
Our old friend, change of basis!What are the new coordinates of X?
The Morphable Face Model
The actual structure of a face is captured in the shape vector S = (x1, y1, x2, …, yn)T, containing the (x, y) coordinates of the n vertices of a face, and the appearance (texture) vector T = (R1, G1, B1, R2, …, Gn, Bn)T, containing the color values of the mean-warped face image.
Shape S
Appearance T
The Morphable face model
Again, assuming that we have m such vector pairs in full correspondence, we can form new shapes Smodel and new appearances Tmodel as:
If number of basis faces m is large enough to span the face subspace then:
Any new face can be represented as a pair of vectors
(1, 2m)T and (1, 2m)T !
m
iiimodel a
1
SS
m
iiimodel b
1
TT
27Learned-Miller
Five Applications
Image factorizations• For transfer learning, learning from one example
Alignment for Data Pooling• 3D MR registration• EEG registration
Artifact removal• Magnetic resonance bias removal
Improvements to recognition algorithms• Alignment before recognition
Defining anchor points for registration• Find highly repeatable regions for future registrations
28Learned-Miller
Congealing
Process of joint “alignment” of sets of arrays (samples of continuous fields).
3 ingredients• A set of arrays in some class• A parameterized family of continuous transformations• A criterion of joint alignment
29Learned-Miller
Congealing Binary Digits
3 ingredients• A set of arrays in some class:
• Binary images• A parameterized family of continuous transformations:
• Affine transforms• A criterion of joint alignment:
• Entropy minimization
30Learned-Miller
Criterion of Joint Alignment Minimize sum of pixel stack
entropies by transforming each image.
A pixel stack
31Learned-Miller
ObservedImage
“Latent Image”
Transform
(Previous work by Grenander,, Frey and Jojic.)
An Image Factorization
32Learned-Miller
QuickTime™ and aTIFF (LZW) decompressor
are needed to see this picture. A pixel stack
33Learned-Miller
The Independent Pixel Assumption Model assumes independent pixels A poor generative model:
• True image probabilities don’t match model probabilities.• Reason: heavy dependence of neighboring pixels.
However! This model is great for alignment and separation of causes!
• Why? • Relative probabilities of “better aligned” and “worse aligned” are
usually correct. Once components are separated, a more accurate (and
computationally expensive) model can be used to model each component.
34Learned-Miller
Before After
Each pair implicitly creates a sample of the transform T.
Congealing
A transform
35Learned-Miller
Character Models
LatentImages
Transforms
Image KernelDensity Estimator
(or other estimator)
Transform KernelDensity Estimator
(CVPR 2003)
Latent ImageProbability Density
for Zeroes
P(IL)
TransformProbability Density
for Zeroes
P(T)Con
geal
ing
36Learned-Miller
How do we line up a new image?
Sequence of successively “sharper” models
…
step 0 step 1 step N
…
Take one gradient step with respect to each model.