Surrogate loss functions, divergences and...

62
Surrogate loss functions, divergences and decentralized detection XuanLong Nguyen Department of Electrical Engineering and Computer Science U.C. Berkeley Advisors: Michael Jordan & Martin Wainwright 1

Transcript of Surrogate loss functions, divergences and...

Page 1: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Surrogate loss functions, divergences and

decentralized detection

XuanLong Nguyen

Department of Electrical Engineering and Computer Science

U.C. Berkeley

Advisors: Michael Jordan & Martin Wainwright

1

Page 2: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Talk outline

• nonparametric decentralized detection algorithm

– use of surrogate loss functions and marginalized kernels

– use of convex analysis

• study of surrogate loss functions and divergence functionals

– correspondence of losses and divergences

– M-estimator of divergences (e.g., Kullback-Leibler divergence)

2

Page 3: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Decentralized decision-making problem

learning both classifier and experiment

• covariate vector X and hypothesis (label) Y = ±1

• we do not have access directly to X in order to determine Y

• learn jointly the mapping (Q, γ)

XQ−→ Z

γ−→ Y

3

Page 4: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Decentralized decision-making problem

learning both classifier and experiment

• covariate vector X and hypothesis (label) Y = ±1

• we do not have access directly to X in order to determine Y

• learn jointly the mapping (Q, γ)

XQ−→ Z

γ−→ Y

• roles of “experiment” Q:

– due to data collection constraints (e.g., decentralization)

– data transmission constraints

– choice of variates (feature selection)

– dimensionality reduction scheme

3-a

Page 5: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

A decentralized detection system

. . . Sensors

Communication channel

Fusion center

Decision rule

• Decentralized setting: Communication constraints between sen-

sors and fusion center (e.g., bit constraints)

• Goal: Design decision rules for sensors and fusion center

• Criterion: Minimize probability of incorrect detection

4

Page 6: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Concrete example – wireless sensor network

...

...

...

...

Light source

sensors

Set-up:

• wireless network of tiny sensor motes, each equiped with light/ humidity/

temperature sensing capabilities

• measurement of signal strength ([0–1024] in magnitude, or 10 bits)

Goal: is there a forest fire in a certain region?

5

Page 7: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Related work

• Classical work on classification/detection:

– completely centralized

– no consideration of communication-theoretic infrastructure

6

Page 8: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Related work

• Classical work on classification/detection:

– completely centralized

– no consideration of communication-theoretic infrastructure

• Decentralized detection in signal processing (e.g., Tsitsiklis, 1993)

– joint distribution assumed to be known

– locally-optimal rules under conditional independence assumptions

(i.e., Naive Bayes)

6-a

Page 9: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Overview of our approach

• Treat as a nonparametric estimation (learning) problem

– under constraints from a distributed system

• Use kernel methods and convex surrogate loss functions

– tools from convex optimization to derive an efficient algorithm

7

Page 10: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Problem set-up

. . .

. . .

. . .

Y ∈ {±1}

X1 X2 X3 XS

Z1 Z2 Z3 ZS

γ1 γ2 γ3 γS

γ(Z1, . . . , ZS)

X ∈ {1, . . . , M}S

Z ∈ {1, . . . , L}S ; L ¿ M

Problem: Given training data (xi, yi)ni=1, find the decision rules

(γ1, . . . , γs; γ) so as to minimize the detection error probability:

P (Y 6= γ(Z1, . . . , Zs)).

8

Page 11: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Kernel methods for classification

• Classification: Learn γ(z) that predicts label y

• K(z, z′) is a symmetric positive semidefinite kernel function

• feature space H in which K acts as an inner product, i.e., K(z, z′) =

〈Ψ(z), Ψ(z′)〉

• Kernel-based algorithm finds linear function in H, i.e.

γ(z) = 〈w, Ψ(z)〉 =

n∑

i=1

αiK(zi, z)

• Advantages:

– kernel function classes are sufficiently rich for many applications

– optimizing over kernel function classes is computionally efficient

9

Page 12: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Convex surrogate loss function φ to 0-1 loss

−1 −0.5 0 0.5 1 1.5 20

0.5

1

1.5

2

2.5

3

Margin value

Sur

roga

te lo

ss

Zero−oneHingeLogisticExponential

• minimizing (regularized) empirical φ-risk Eφ(Y γ(Z)):

minγ∈H

n∑

i=1

φ(yiγ(zi)) +λ

2‖γ‖2,

• (zi, yi)ni=1 are training data in Z × {±1}

• φ is a convex loss function (surrogate to non-convex 0-1 loss)

10

Page 13: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Stochastic decision rules at each sensor

. . .

. . .

. . .

Y

X1 X2 X3 XS

Z1 Z2 Z3 ZS

γ(Z)

Q(Z|X)

Kz(z, z′) = 〈Ψ(z), Ψ(z′)〉γ(z) = 〈w, Ψ(z)〉

• Approximate deterministic sensor decisions by stochastic rules Q(Z|X)

• Sensors do not communicate directly =⇒ factorization:

Q(Z|X) =∏S

t=1 Qt(Zt|Xt)

• The overall decision rule is represented by

Q =∏

Qt,

γ(z) = 〈w, Ψ(z)〉

11

Page 14: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

High-level strategy:

Joint optimization

• Minimize over (Q, γ) an empirical version of Eφ(Y γ(Z))

• Joint minimization:

– fix Q, optimize over γ: A simple convex problem

– fix γ, perform a gradient update for Q, sensor by sensor

12

Page 15: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Approximating empirical φ-risk

• The regularized empirical φ-risk Eφ(Y γ(Z)) has the form:

G0 =∑

z

n∑

i=1

φ(yiγ(z))Q(z|xi) +λ

2||w||2

• Challenge: even evaluating G0 at a single point is intractable

Requires summing over LS possible values for z

• Idea:

– approximate G0 by another objective function G

– G is φ-risk of a “marginalized” feature space

– G0 ≡ G for deterministic Q

13

Page 16: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

“Marginalizing” over feature space

x1

x2

z1

z2 γ(z) = 〈w, Ψ(z)〉 fQ(x) = 〈w, ΨQ(x)〉

Q(z |x)

Original: X−spaceQuantized: Z−space

Stochastic decision rule Q(z |x):

• maps between X and Z

• induces marginalized feature map ΨQ from base map Ψ (or marginalized

kernel KQ from base kernel K)

14

Page 17: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Marginalized feature space {ΨQ(x)}

15

Page 18: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Marginalized feature space {ΨQ(x)}

• Define a new feature space ΨQ(x) and a linear function over ΨQ(x):

ΨQ(x) =∑

z Q(z|x)Ψ(z) ⇐= Marginalization over z

fQ(x) = 〈w, ΨQ(x)〉

15-a

Page 19: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Marginalized feature space {ΨQ(x)}

• Define a new feature space ΨQ(x) and a linear function over ΨQ(x):

ΨQ(x) =∑

z Q(z|x)Ψ(z) ⇐= Marginalization over z

fQ(x) = 〈w, ΨQ(x)〉

• The alternative objective function G is the φ-risk for fQ:

G =

n∑

i=1

φ(yifQ(xi)) +λ

2‖w‖2

15-b

Page 20: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Marginalized feature space {ΨQ(x)}

• Define a new feature space ΨQ(x) and a linear function over ΨQ(x):

ΨQ(x) =∑

z Q(z|x)Ψ(z) ⇐= Marginalization over z

fQ(x) = 〈w, ΨQ(x)〉

• The alternative objective function G is the φ-risk for fQ:

G =

n∑

i=1

φ(yifQ(xi)) +λ

2‖w‖2

• ΨQ(x) induces a marginalized kernel over X :

KQ(x, x′) := 〈ΨQ(x), ΨQ(x′)〉 =∑

z,z′

Q(z|x)Q(z′|x′) Kz(z, z′)

⇒ Marginalization taken over message z conditioned on sensor signal x

15-c

Page 21: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Marginalized kernels

• Have been used to derive kernel functions from generative models

(e.g. Tsuda, 2002)

• Marginalized kernel KQ(x, x′) is defined as:

KQ(x, x′) :=∑

z,z′

Q(z|x)Q(z′|x′)︸ ︷︷ ︸

Factorized distributions

Kz(z, z′)︸ ︷︷ ︸

Base kernel

,

• If Kz(z, z′) is decomposed into smaller components of z and z′, then

KQ(x, x′) can be computed efficiently (in polynomial-time)

16

Page 22: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Centralized and decentralized function

• Centralized decision function obtained by minimizing φ-risk:

fQ(x) = 〈w, ΨQ(x)〉

– fQ has direct access to sensor signal x

17

Page 23: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Centralized and decentralized function

• Centralized decision function obtained by minimizing φ-risk:

fQ(x) = 〈w, ΨQ(x)〉

– fQ has direct access to sensor signal x

• Optimal w also define decentralized decision function:

γ(z) = 〈w, Ψ(z)〉

– γ has access only to quantized version z

17-a

Page 24: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Centralized and decentralized function

• Centralized decision function obtained by minimizing φ-risk:

fQ(x) = 〈w, ΨQ(x)〉

– fQ has direct access to sensor signal x

• Optimal w also define decentralized decision function:

γ(z) = 〈w, Ψ(z)〉

– γ has access only to quantized version z

• Decentralized γ behaves on average like the centralized fQ:

fQ(x) = E[γ(Z)|x]

17-b

Page 25: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Optimization algorithm

Goal: Solve the problem:

infw;Q

G(w; Q) :=∑

i

φ

(

yi〈w,∑

z

Q(z|xi)Ψ(z)〉

)

2||w||2

• Finding optimal weight vector:

– G is convex in w with Q fixed

– solve dual problem (quadratic convex program) to obtain optimal

w(Q)

• Finding optimal decision rules:

– G is convex in Qt with w and all other {Qr, r 6= t} fixed

– efficient computation of subgradient for G at optimal (w(Q), Q)

Overall: Efficient joint minimization by blockwise coordinate descent

18

Page 26: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Simulated sensor networks

. . .

Y

X1

X2

X3

X4

X10

. . .

Y

X1

X2

X3

X4

X10

X1 X2 X3

X4 X5 X6

X7 X8 X9

Y

Naive Bayes net Chain-structured network Spatially-dependent network

19

Page 27: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Kernel Quantization vs. Decentralized LRT

0 0.1 0.2 0.30

0.1

0.2

0.3

Kernel quantization (KQ)

Dec

entr

aliz

ed L

RT

Naive Bayes networks

0 0.1 0.2 0.3 0.4 0.50

0.1

0.2

0.3

0.4

0.5

Kernel quantization (KQ)

Dec

entr

aliz

ed L

RT

Chain−structured (1st order)

0 0.1 0.2 0.3 0.4 0.50

0.1

0.2

0.3

0.4

0.5

Kernel quantization (KQ)

Dec

entr

aliz

ed L

RT

Chain−structured (2nd order)

0 0.1 0.2 0.3 0.4 0.50

0.1

0.2

0.3

0.4

0.5

Kernel quantization (KQ)

Dec

entr

aliz

ed L

RT

Fully connected networks

20

Page 28: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Wireless network with tiny Berkeley motes

...

...

...

...

Light source

sensors

• 5 × 5 = 25 tiny sensor motes, each equipped with a light receiver

• Light signal strength requires 10-bit ([0–1024] in magnitude)

• Perform classification with respect to different regions

• Each problem has 25 training positions, 81 test positions

(Data collection courtesy Bruno Sinopoli)

21

Page 29: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

0 5 10 15 20 250

0.1

0.2

0.3

0.4

0.5

0.6

0.7

Classification problem instances

Tes

t err

orClassification with Mica sensor motes

centralized SVM (10−bit sig)centralized NB classifierdecentralized KQ (1−bit)decentralized KQ (2−bit)

22

Page 30: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Outline

• nonparametric decentralized detection algorithm

– use of surrogate loss functions and marginalized kernels

• study of surrogate loss functions and divergence functionals

– correspondence of losses and divergences

– M-estimator of divergences (e.g., Kullback-Leibler divergence)

23

Page 31: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Consistency question

• recall that our decentralized algorithm essentially solves

minγ,Q

Eφ(Y, γ(Z))

• does this also imply optimality in the sense of 0-1 loss?

P (Y 6= γ(Z))

24

Page 32: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Consistency question

• recall that our decentralized algorithm essentially solves

minγ,Q

Eφ(Y, γ(Z))

• does this also imply optimality in the sense of 0-1 loss?

P (Y 6= γ(Z))

• answers:

– hinge loss yields consistent estimates

– all losses corresponding to variational distance yield consistency

and we can identify all of them

– exponential loss, logistic loss do not

• the gist lies in the correspondence between loss functions and

divergence functionals

24-a

Page 33: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Divergence between two distributions

The f-divergence between two densities µ and π is given by

If (µ, π) :=

z

π(z)f

(µ(z)

π(z)

)

dν.

where f : [0, +∞) → R ∪ {+∞} is a continuous convex function

25

Page 34: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Divergence between two distributions

The f-divergence between two densities µ and π is given by

If (µ, π) :=

z

π(z)f

(µ(z)

π(z)

)

dν.

where f : [0, +∞) → R ∪ {+∞} is a continuous convex function

• Kullback-Leibler divergence: f(u) = u log u.

If (µ, π) =z

µ(z) logµ(z)

π(z).

• variational distance: f(u) = |u − 1|.

If (µ, π) :=z|µ(z) − π(z)|.

• Hellinger distance: f(u) = 1

2(√

u − 1)2.

If (µ, π) :=z∈Z

( � µ(z) − � π(z))2.

25-a

Page 35: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Surrogate loss and f-divergence

Map Q induces measures on Z:

µ(z) := P (Y = 1, z); π(z) := P (Y = −1, z)

Theorem: Fixing Q, the optimal risk for each φ loss is an f -divergence for

some convex f , and vice versa:

Rφ(Q) = −If (µ, π), where Rφ(Q) := minγ

Eφ(Y, γ(Z))

φ1

φ2

φ3

f1

f2

f3

Class of loss functions Class of f -divergences

26

Page 36: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

“Unrolling” divergences by convex duality

• Legendre-Fenchel convex duality: f(u) = supv∈R uv − f∗(v),

where f∗ is the convex conjugate of f

If (µ, π) =

πf

π

)

27

Page 37: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

“Unrolling” divergences by convex duality

• Legendre-Fenchel convex duality: f(u) = supv∈R uv − f∗(v),

where f∗ is the convex conjugate of f

If (µ, π) =

πf

π

)

=

πsupγ

(γµ/π − f∗(γ)) dν

27-a

Page 38: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

“Unrolling” divergences by convex duality

• Legendre-Fenchel convex duality: f(u) = supv∈R uv − f∗(v),

where f∗ is the convex conjugate of f

If (µ, π) =

πf

π

)

=

πsupγ

(γµ/π − f∗(γ)) dν

= supγ

γµ − f∗(γ)π dν

27-b

Page 39: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

“Unrolling” divergences by convex duality

• Legendre-Fenchel convex duality: f(u) = supv∈R uv − f∗(v),

where f∗ is the convex conjugate of f

If (µ, π) =

πf

π

)

=

πsupγ

(γµ/π − f∗(γ)) dν

= supγ

γµ − f∗(γ)π dν

= − infγ

f∗(γ)π − γµ dν

27-c

Page 40: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

“Unrolling” divergences by convex duality

• Legendre-Fenchel convex duality: f(u) = supv∈R uv − f∗(v),

where f∗ is the convex conjugate of f

If (µ, π) =

πf

π

)

=

πsupγ

(γµ/π − f∗(γ)) dν

= supγ

γµ − f∗(γ)π dν

= − infγ

f∗(γ)π − γµ dν

• The last quantity can be viewed as a risk functional with respect to

loss functions f∗(γ) and −γ

27-d

Page 41: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Examples

• 0-1 loss:

Rbayes(Q) = 1

2− 1

2 z∈Z|µ(z) − π(z)| ⇒ variational distance

• hinge loss:

Rhinge(Q) = 2Rbayes(Q) ⇒ variational distance

• exponential loss:

Rexp(Q) = 1 −z∈Z

(µ(z)1/2 − π(z)1/2)2 ⇒ Hellinger distance

• logistic loss:

Rlog(Q) = log 2 − KL(µ||µ+π2

) − KL(π||µ+π2

) ⇒ capacitory dis. distance

28

Page 42: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Examples

Equivalent surrogage losses corresponding to Hellinger distance (left)

and variational distance (right)

−1 0 1 20

1

2

3

4

margin

φ l

oss

g = exp(u−1)g = u

g = u2

−1 0 1 20

1

2

3

4

margin value

φ lo

ss

g = eu−1

g = u

g = u2

the part after 0 is a fixed map of the part before 0!

29

Page 43: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Comparison of loss functions

φ1

φ2

φ3

f1

f2f3

Class of loss functions Class of f -divergences

• two loss functions φ1 and φ2, corresponding to f -divergence induced

by f1 and f2

• φ1 and φ2 are universally equivalent, denoted by

φ1U≈ φ2 (or, equivalently) f1

U≈ f2

if for any P (X, Y ) and quantization rules QA, QB , there holds:

Rφ1(QA) ≤ Rφ1

(QB) ⇔ Rφ2(QA) ≤ Rφ2

(QB).

30

Page 44: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

An equivalence theorem

Theorem:

φ1U≈ φ2 (or, equivalently) f1

U≈ f2

if and only if

f1(u) = cf2(u) + au + b

for constants a, b ∈ R and c > 0

• in particular, surrogate losses universally equivalent to 0−1 loss are those

whose induced f divergence has the form:

f(u) = c min{u, 1} + au + b

31

Page 45: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Empirical risk minimization procedure

• let φ be a convex surrogate equivalent to 0 − 1 loss

• (Cn,Dn) is a sequence of increasing function classes for(γ, Q)

(C1,D1) ⊆ (C2,D2) ⊆ . . . ⊆ (Γ,Q)

• our procedure learns:

(γ∗n, Q∗

n) := argmin(γ,Q)∈(Cn,Dn)Eφ(Y γ(Z))

• let R∗bayes := inf(γ,Q)∈(Γ,Q) P (Y 6= γ(Z)) ⇐ optimal Bayes error

• our procedure is consistent if

Rbayes(γ∗n, Q∗

n) − R∗bayes → 0

32

Page 46: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Consistency result

Theorem: If

• ∪∞n=1(Cn,Dn) is dense in the space of pairs of classifier and quantizer

(γ, Q) ∈ (Γ,Q)

• sequence (Cn,Dn) increases in size sufficiently slowly

then our procedure is consistent, i.e.,

limn→∞

Rbayes(γ∗n, Q∗

n) − R∗bayes = 0 in probability.

• proof exploits the equivalence of φ loss and 0 − 1 loss

• decomposition of φ risk into approximation error and estimation error

33

Page 47: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Outline

• nonparametric decentralized detection algorithm

– use of surrogate loss functions and marginalized kernels

• study of surrogate loss functions and divergence functionals

– correspondence of losses and divergences

– M-estimator of divergences (e.g., Kullback-Leibler divergence)

34

Page 48: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Estimating divergence and likelihood ratio

• given i.i.d {x1, . . . , xn} ∼ Q, {y1, . . . , yn} ∼ P

• want to estimate two quantities

– KL divergence functional

DK(P, Q) =

p0 logp0

q0dµ

– likelihood ratio function

g0(.) = p0(.)/q0(.)

35

Page 49: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Variational characterization

• recall the correspondence:

minγ

Eφ(Y, γ(Z)) = −If (µ, π)

• f -divergence can be estimated by minimizing over some associated

φ-risk functional

36

Page 50: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Variational characterization

• recall the correspondence:

minγ

Eφ(Y, γ(Z)) = −If (µ, π)

• f -divergence can be estimated by minimizing over some associated

φ-risk functional

• for the Kullback-Leibler divergence:

DK(P, Q) = supg>0

log g dP −

gdQ + 1.

• furthermore, the supremum is attained at g = p0/q0.

36-a

Page 51: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

M-estimation procedure

• let G be a function class of X → R+

•∫

dPn and∫

dQn denote the expectation under empirical measures

Pn and Qn, respectively

• our estimator has the following form:

DK = supg∈G

log g dPn −

gdQn + 1.

• supremum is attained at gn, which estimates the likelihood ratio

p0/q0

37

Page 52: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Convex empirical risk with penalty

• in practice, control the size of the function class G by using penalty

• let I(g) be a measure of complexity for g

• decompose G as follows:

G = ∪1≤M≤∞GM ,

where GM is restricted to g for which I(g) ≤ M .

• the estimation procedure involves solving:

gn = argming∈G

gdQn −

log g dPn +λn

2I2(g).

38

Page 53: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Convergence analysis

• for KL divergence estimation, we study

|DK − DK(P, Q)|

• for the likelihood ratio estimation, we use Hellinger distance

h2Q(g, g0) :=

1

2

(g1/2 − g1/20 )2 dQ.

39

Page 54: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Assumptions for convergence analysis

• true likelihood ratio g0 is bounded from below by some positive con-

stant:

g0 ≥ η0 > 0.

Note: we don’t assume that G is bounded away from 0 (not yet)!

• uniform norm of GM is Lipchitz with respect to the penalty measure

I(g): for any M ≥ 1:

supg∈GM

|g|∞ ≤ cM.

• on the entropy of G: For some 0 < γ < 2,

HBδ (GM , L2(Q)) = O(M/δ)γ .

40

Page 55: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Convergence rates

• when λn vanishes sufficiently slowly:

λ−1n = OP (n2/(2+γ))(1 + I(g0)),

• then under P:

hQ(g0, gn) = OP (λ1/2n )(1 + I(g0))

I(gn) = OP (1 + I(g0)).

41

Page 56: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Convergence rates

• when λn vanishes sufficiently slowly:

λ−1n = OP (n2/(2+γ))(1 + I(g0)),

• then under P:

hQ(g0, gn) = OP (λ1/2n )(1 + I(g0))

I(gn) = OP (1 + I(g0)).

• if G is bounded away from 0:

|DK − DK | = OP (λ1/2n )(1 + I(g0)).

41-a

Page 57: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

G is RKHS function class

• {xi} ∼ Q, {yj} ∼ P

• G is a RKHS with Mercer kernel k(x, x′) = 〈Φ(x), Φ(x′)〉

• I(g) = ‖g‖H

gn = argming∈G

gdQn −

log g dPn +λn

2‖g‖2

H

42

Page 58: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

G is RKHS function class

• {xi} ∼ Q, {yj} ∼ P

• G is a RKHS with Mercer kernel k(x, x′) = 〈Φ(x), Φ(x′)〉

• I(g) = ‖g‖H

gn = argming∈G

gdQn −

log g dPn +λn

2‖g‖2

H

• Convex dual formulation:

α := argmax1

n

n∑

j=1

log αj −1

2λn‖

n∑

j=1

αjΦ(yj) −1

n

n∑

i=1

Φ(xi)‖2

DK(P, Q) := −1

n

n∑

j=1

log αj − log n

42-a

Page 59: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

log G is RKHS function class

• {xi} ∼ Q, {yj} ∼ P

• log G is a RKHS with Mercer kernel k(x, x′) = 〈Φ(x), Φ(x′)〉

• I(g) = ‖ log g‖H

gn = argming∈G

gdQn −

log g dPn +λn

2‖ log g‖2

H

• Convex dual formulation:

α := argmax1

n

n∑

i=1

(

αi log αi+αi logn

e

)

−1

2λn‖

n∑

i=1

αiΦ(xi)−1

n

n∑

j=1

Φ(yj)‖2

DK(P, Q) := 1 +

n∑

i=1

αi log αi + αi logn

e

43

Page 60: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

−0.1

0

0.1

0.2

0.3

0.4

100 200 500 1000 2000 5000 10000 20000 50000

Estimate of KL(Beta(1,2),Unif[0,1])

0.1931M1, σ = .1, λ = 1/nM2, σ = .1, λ = .1/n

WKV, s = n1/2

WKV, s = n1/3

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

100 200 500 1000 2000 5000 10000

Estimate of KL(1/2 Nt(0,1)+ 1/2 N

t(1,1),Unif[−5,5])

0.414624M1, σ = .1, λ = 1/nM2, σ = 1, λ = .1/n

WKV, s = n1/3

WKV, s = n1/2

WKV, s = n2/3

0

0.5

1

1.5

2

2.5

100 200 500 1000 2000 5000 10000

Estimate of KL(Nt(0,1),N

t(4,2))

1.9492M1, σ = .1, λ = 1/nM2, σ = .1, λ = .1/n

WKV, s = n1/3

WKV, s = n1/2

WKV, s = n2/3

−2

−1

0

1

2

3

4

5

6

100 200 500 1000 2000 5000 10000

Estimate of KL(Nt(4,2),N

t(0,1))

4.72006M1, σ = 1, λ = .1/nM2, σ = 1, λ = .1/n

WKV, s = n1/4

WKV, s = n1/3

WKV, s = n1/2

44

Page 61: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

0

0.5

1

1.5

100 200 500 1000 2000 5000 10000

Estimate of KL(Nt(0,I

2),N

t(1,I

2))

0.959316M1, σ = .5, λ = .1/nM2, σ = .5, λ = .1/n

WKV, n1/3

WKV, n1/2

0

0.5

1

1.5

100 200 500 1000 2000 5000 10000

Estimate of KL(Nt(0,I

2),Unif[−3,3]2)

0.777712M1, σ = .5, λ = .1/nM2, σ = .5, λ = .1/n

WKV, n1/3

WKV, n1/2

−0.2

0

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

1.8

100 200 500 1000 2000 5000 10000

Estimate of KL(Nt(0,I

3),N

t(1,I

3))

1.43897M1, σ = 1, λ = .1/nM2, σ = 1, λ = .1/n

WKV, n1/2

WKV, n1/30

0.5

1

1.5

2

100 200 500 1000 2000 5000 10000

Estimate of KL(Nt(0,I

3),Unif[−3,3]3)

1.16657

M1 σ = 1, λ = .1/n1/2

M2, σ = 1, λ = .1/n

M2, σ = 1, λ = .1/n2/3

WKV, n1/3

WKV, n1/2

45

Page 62: Surrogate loss functions, divergences and …dept.stat.lsa.umich.edu/~xuanlong/Talks/thesis_talk.pdfFusion center Decision rule • Decentralized setting: Communication constraints

Conclusion

• nonparametric decentralized detection algorithm

– use of surrogate loss functions and marginalized kernels

• study of surrogate loss functions and divergence functionals

– correspondence of losses and divergences

– M-estimator of divergences (e.g., Kullback-Leibler divergence)

46