Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks...

43
1/42 Overview Approximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Feature Extraction in Deep Learning and Image Processing Yiran Li Applied Mathematics, Statistics, and Scientific Computation Norbert Wiener Center Department of Mathematics University of Maryland, College Park

Transcript of Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks...

Page 1: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

1/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Feature Extraction in Deep Learning and Image Processing

Yiran Li

Applied Mathematics, Statistics, and Scientific ComputationNorbert Wiener Center

Department of MathematicsUniversity of Maryland, College Park

Page 2: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

2/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Outline

1 Overview

2 Approximation Properties of Neural Networks

3 Gabor Invariant Representation in Quantum Energy Regression

Page 3: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

3/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Outline

1 Overview

2 Approximation Properties of Neural Networks

3 Gabor Invariant Representation in Quantum Energy Regression

Page 4: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

4/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Overview

My dissertation consists of the following topics:

Approximation Properties of Neural Networks

Maximal Function Pooling in Convolutional Sparse Coding

Quantum Energy Regression Using Gabor Transform

Detection of Epithelial versus Mesenchymal Regions in 2D Images ofTumor Biopsies Using Shearlets

Due to the time constraint, I will discuss the topics in bold in this presentation.

Page 5: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

5/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Outline

1 Overview

2 Approximation Properties of Neural Networks

3 Gabor Invariant Representation in Quantum Energy Regression

Page 6: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

6/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Approximation Properties of Neural Networks

Deep Neural Networks (DNNs) and deep learning algorithms have achievedsuccessful results in many areas of machine learning, and there has beengrowing interest in the theoretical study of DNNs.Some important topics in the theoretical analysis of neural networks include:

1 Specification of the network topology to obtain certain approximationproperties of functions;

2 The stability analysis of the network;

3 Study of the training algorithms to obtain desired convergence rate.

Page 7: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

7/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Previous Works

1 Chui, C. K., Li, X., & Mhaskar, H. N. (1994). Neural networks for localizedapproximation. Mathematics of Computation, 63(208), 607-623.

2 Shaham, U., Cloninger, A., & Coifman, R. R. (2016). Provableapproximation properties for deep neural networks. Applied andComputational Harmonic Analysis.

3 Bolcskei, H., Grohs, P., Kutyniok, G., & Petersen, P. (2017). Optimalapproximation with sparsely connected deep neural networks. preprintarXiv:1705.01714.

4 Balan, R., Singh, M., & Zou, D. (2017). Lipschitz properties for deepconvolutional networks. arXiv preprint arXiv:1701.05217.

5 Jacobs, R. A. (1988). Increased rates of convergence through learningrate adaptation. Neural networks, 1(4), 295-307.

Page 8: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

8/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Universal Approximation Theorem

The most well-known early result is by Cybenko in 1989 states that:

Any continuous function can be uniformly approximated by acontinuous neural network having only one internal hidden layer andwith arbitrary continuous sigmoidal nonlinearity.

Theorem (Cybenko, 1989)

Let σ be any continuous discriminatory sigmoidal function. Then the finitesums

G(x) =n∑

k=1

ckσ(wk · x + bk ), (1)

are dense in C(Id ), where Id is the unit cube in Rd .

Here σ is the sigmoidal activation function, defined as σ(u) withlimu→−∞ σ(u) = 0 and limu→∞ σ(u) = 1.

Page 9: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

9/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Fourier Approximation

The number of neurons and number of layers required to yield anapproximation rate of a given quantity is not addressed.The first work to address this problem is by Barron in 1991:

Theorem (Fourier Approximation, Barron, 1991)

Given a function f : Rm → R with

Cf =

∫Rm|ω||f (ω)|dω <∞, (2)

there exsists a single layer artificial neural network (ANN) of N sigmoid units,s.t. the output of the network fN satisfies

‖f − fN‖2≤cf√N, (3)

with cf proportional to Cf .

Page 10: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

10/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Approximation Properties with Wavelets

Smooth functions defined on low-dimensional subspace can be highdimensional in ambient space. The goal is to find approximation rate relatedto the dimension of the manifold, not the ambient space.

Theorem (Cloninger, Coifman, Shaham, 2015)

Let Γ ⊂ Rm be a smooth d-dimensional manifold, f ∈ L2(Γ) and let ε > 0 bean approximation level. Then if a network has at least 4 layers, there exists asparsely-connected neural network with N total units whereN = CΓm + C′ΓdNf ,ε, computing function fN such that

‖f − fN‖22< ε, (4)

where Nf ,ε depends on the complexity of f in terms of its local waveletrepresentation, and CΓ on the curvature and dimension of the manifold Γ.

If f ∈ C2(Γ) and has bounded Hessian, then

‖f − fN‖∞= O(N−2d ). (5)

Page 11: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

11/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Construction of Wavelet Frame

A wavelet like frame of Rd is constructed in which the frame elementsare built using rectified linear units. A rectified linear unit (ReLU) isdefined as

rect(x) = max{0, x}. (6)

Define a trapezoid-shaped function t : R→ R by

t(x) = rect(x + 3)− rect(x + 1)− rect(x − 1) + rect(x − 3). (7)

Define the scaling function φ : Rd → R by

φ(x) = rect

d∑j=1

t(xj )− 2(d − 1)

, (8)

and normalize it so that the integral of φ is 1.

Let Sk (x , b) = 2kφ(2kd (x − b)). Define the mother wavelet as

Dk (x , b) = Sk (x , b)− Sk−1(x , b). The wavelets are defined asψk,b(x) = 2−

kd Dk (x , b).

Page 12: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

12/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Construction of Wavelet Frame

Figure: Wavelets constructed from ReLUs (Source: U. Shaham, A. Cloninger, R. R.Coifman, 2015.)

Page 13: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

13/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Approximation of Functions on Manifold

Given a d-dimensional manifold Γ ⊂ Rm, cover Γ by set of pairs{(Ui , φi )}CΓ

i=1. Here φi is the orthogonal projection from Ui onto Hi , whereHi is the hyperplane tangent to Γ at xi .

Use the corresponding partition of unity {ηi} to define

fi (x) = f (x)ηi (x). (9)

Define f ∈ Rd as

fi (x) =

{fi (φ−1

i (x)), x ∈ φi (Ui ),0, otherwise.

(10)

For all x ∈ Γ, we have ∑i:x∈Ui

fi (φi (x)) = f (x). (11)

Assuming fi ∈ L2(Rd ), it can be expanded using wavelet frame.

Page 14: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

14/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Gabor System in Machine Learning

Current study of the approximation properties of neural networks ismainly in terms of affine transformations (e.g., wavelet transform).

We extend the study of approximation properties of neural networks tofunctions in the modulation space.

Gabor transform, or the short-time Fourier transform, arises naturally inanalysis of 1D data such as speech and music.

There have been algorithms developed for voiced-unvoiced speechdiscrimination in noise, where short segments of speech are modeled asa sum of basis functions from a Gabor dictionary.

Gabor filters and Gabor wavelets are widely used as convolutionalkernels for neural networks for 2D image processing.

Page 15: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

15/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Motivation for Gabor System in Neural Networks

There are cortical receptive fields that best respond to signals withorientation. They also capture spatial frequency information.Two-dimensional spatial linear filters are constrained by generaluncertainty relations. The theoretical lower limit for the uncertainty isachieved by Gabor functions.Gabor filters have been used as units of a neural network to model theprofile of cortical receptive fields (Daugman, 1988).

Figure: Illustration of experimentally measured 2D receptive-field profiles of threesimple cells in cat striate cortex (top row). Each plot shows the excitatory or inhibitoryeffect of a small flashing light or dark spot on the firing rate of the cell, as a function ofthe (x, y) location of the stimulus. Best fit using Gabor functions (second row).

Page 16: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

16/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Approximation Properties of Neural Networks

We design a novel type of neural network and prove its theoreticalapproximation rate to functions f based on the network topology. Informally,we show that

Theorem (Informal)

Let f ∈ L2(R), and let δ > 0 be an approximation level. There exists a 4-layersparsely-connected neural network with N units where N = N(f , δ),computing fN with

‖f − fN‖∞≤ δ.

We demonstrate a method to build a Gabor frame of L2(R) based on atype of activation function in neural networks: rectified linear units.

We construct a 4-layer neural network based on the Gabor frame anddemonstrate its approximation properties.

Page 17: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

17/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Gabor System

We first introduce the notion of Gabor system. Let the time shift T of afunction g ∈ L2(Rd ) by x ∈ Rd be defined by

Tx g(t) = g(t − x),

and let the modulation of g by ω ∈ Rd be defined by

Mωg(t) = e2πiω·tg(t).

Definition

A Gabor system G(g, α, β) is the set of time-frequency shifts of a non-zerowindow function g ∈ L2(Rd ) with lattice parameters α, β > 0:

{Tαk Mβng : k , n ∈ Zd}.

Page 18: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

18/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Frame

We introduce the notion of a frame. A frame can be thought of as ageneralization of a basis that may be linearly dependent.

Definition

A sequence {ej , j ∈ J} in a separable Hilbert space H is called a frame ifthere exist positive constants A,B > 0 such that for all f ∈ H

A‖f‖2≤∑j∈J

|〈f , ej〉|2≤ B‖f‖2.

Any two constants A,B where 0 < A ≤ B <∞ satisfying the above statementare called frame bounds. If A = B, then {ej : j ∈ J} is called a tight frame.

A frame provides a redundant way of representing a signal.

Page 19: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

19/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Construction of a Gabor Frame Using ReLUs

We build the window function g using rectified linear units.Rectified linear unit, or ReLU, is commonly used as activation function of theneuron of deep neural networks. A rectified linear unit is defined as:

rect(x) = max{0, x}.

We define the window function g as a triangular-shaped window function:

g(x) = rect(12

x + 1)− rect(x) + rect(12

x − 1). (12)

We take g as the window function of a Gabor system G(g, α, β).

-4 -3 -2 -1 0 1 2 3 4

-1

-0.5

0

0.5

1

1.5

2window function g(x)

Figure: Window function g definied in (12).

Page 20: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

20/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Construction of a Gabor Frame Using ReLUs

It can be shown that G(g, α, β) is a Gabor frame with specific choices of αand β.

Lemma

Given window function g(x) = rect( 12 x + 1)− rect(x) + rect( 1

2 x − 1), theGabor system G(g, α, β) is a Gabor frame for L2(R) with values of α, βsatisfying α = 1 and β ≤ 1

6 .

Page 21: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

21/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Correlation Functions and Wiener Space

We introduce the following definitions for the proof.

Definition

Given g, γ ∈ L2(Rd ) and α, β > 0, the correlation functions of the pair (g, γ)are defined to be

Gn(x) = G(α,β)n (x) =

∑k∈Zd

g(x − nβ− αk)γ(x − αk) (13)

for n ∈ Zd .

Denote the cube [0, α]d by Qα and write Q = Q1 = [0, 1]d for the unit cube.

Definition

A function g ∈ L∞(Rd ) belongs to the Wiener space W = W (Rd ) if

‖g‖W =∑n∈Zd

ess supx∈Q

|g(x + n)|<∞. (14)

Page 22: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

22/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Existence of Gabor Frames

We introduce the following Theorem on conditions of existence of Gaborframe.

Theorem (Walnut, 1992)

Suppose that g ∈ W (Rd ) and that α > 0 is chosen such that for constantsa, b > 0

a ≤∑k∈Zd

|g(x − αk)|2≤ b <∞ x − a.e. (15)

Then there exists value β0 = β0(α) > 0, such that G(g, α, β) is a Gaborframe for all β ≤ β0. Specifically, if β0 > 0 is chosen such that∑

n∈Zd ,n 6=0

‖G(α,β0)n ‖∞< ess inf

x∈Rd|G0(x)|, (16)

then G(g, α, β) is a frame for all β ≤ β0.

Page 23: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

23/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Proof of Lemma

By Theorem (Walnut, 1992), we need to show that g ∈ W (R) and that gsatisfies

a ≤∑k∈Zd

|g(x − αk)|2≤ b <∞ x − a.e. (17)

for some a, b > 0.We know that supp g = [−2, 2], and that sup|g|= 1 by construction of g.Since x ∈ Q1 = [0, 1], and n ∈ Z, we have

‖g‖W =∑n∈R

ess supx∈Q|g(x + n)|≤ 4 sup|g|= 4 <∞. (18)

We can choose α = 1 so that the infinite sum in (18) has only four non-zeroterms for all x ∈ R. Given any x ∈ R, we have∑

k∈Z

|g(x − k)|2≤ 4 sup|g|2= 4.

Thus the upper bound b is b = 4.

Page 24: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

24/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Proof of Lemma

Note that the window function g can be expressed as a piecewise linearfunction:

g(x) =

{12 x + 1, −2 ≤ x ≤ 0;

− 12 + 1, 0 < x ≤ 2.

(19)

Hence in order to find the lower bound a, we simplify the sum in (17) for somex ∈ [−2,−1], and rewrite the equation as∑

k∈Z

|g(x − k)|2=|g(x)|2+|g(x + 1)|2+|g(x + 2)|2+|g(x + 3)|2

=(x + 1)2 +52.

(20)

Therefore, the minimum is reached when x = −1 and a = 52 .

Given α = 1, we can choose β ≤ β0 = 16 so that the the condition for β0 listed

in Theorem (Walnut, 1992) is satisfied.

Page 25: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

25/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Approximation Property

Now that we have introduced the Gabor frame, we will build a 4-layer neuralnetwork that can be used to approximate functions, and we show that

Lemma (Approximation Property, Czaja, Li, 2017)

Let f ∈ L2(R) be s times continuously differentiable, and let ‖f (s)‖1<∞. Thenfor every x ∈ R, there exists a construction fN using Gabor coefficients ofmodulations up to scale N such that:

|f − fN |= O(1

Ns−1 ), (21)

where |·| denotes the point-wise absolute value.

Page 26: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

26/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Construction of the Neural Network

We construct the neural network with specified number of nodes and layersas the following.

The input layer: x ∈ R.

The first layer: all the shifts {x − αk} of x for k ∈ [−K ,K ].

The second layer: shifted x ′s are activated by modulated ReLUs of threetypes: rect( 1

2 x + 1), −rect(x), rect( 12 x − 1), with each of them

modulated by Mβn for n ∈ [−N,N].

The third layer: outputs from different ReLUs of the same modulationterm are added together to obtain Tαk Mβng for all k ∈ [−K ,K ] andn ∈ [−N,N].

The output layer: outputs from the third layer are added to produce thefinal output function:

fK ,N =∑|k|≤K

∑|n|≤N

wk,nTαk Mβng. (22)

Page 27: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

27/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Construction of the Neural Network

Figure: Illustration of the neural network

Page 28: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

28/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Approximation Rate of Neural Networks

Theorem (Czaja, Li, 2017)

Let f ∈ L2(R). If f is at s times continuously differentiable for s ≥ 2, then fcan be approximated on the order of O( 1

Ns−1 ) using a 4-layer network with(2K + 1)(4(2N + 1) + 1) units. There are 2K + 1 linear units in the first layer;(2K + 1)× 3× (2N + 1) units in the second layer; (2K + 1)(2N + 1) linearunits in the third layer and a single linear unit in the fourth layer.Here K is the number of translations and N is the number of modulations inthe Gabor system used to construct the neural network.

Page 29: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

29/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Proof of Theorem

The output of the neural network can be written as

fK ,N =∑|k|≤K

∑|n|≤N

wk,nTαk Mβng (23)

with weights wk,n. It remains to prove the Lemma (approximation property).We have shown that G(g, α, β) is a Gabor frame for L2(R) with α = 1 andβ ≤ 1

6 .

Proposition (Grochenig)

If G(g, α, β) is a frame for L2(Rd ), then there exists a dual windowγ ∈ L2(Rd ), such that the dual frame of G(g, α, β) is G(γ, α, β).Consequently, every f ∈ L2(Rd ) possesses the expansions

f =∑

k,n∈Zd

∑〈f ,Tαk Mβng〉Tαk Mβnγ

=∑

k,n∈Zd

∑〈f ,Tαk Mβnγ〉Tαk Mβng

(24)

with unconditional convergence in L2(Rd ).

Page 30: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

30/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Proof of Theorem

Let fK ,N be the approximation obtained by the first (2K + 1)(2N + 1) terms inthe expansion:

fK ,N =∑|k|≤K

∑|n|≤N

〈f ,Tαk Mβnγ〉Tαk Mβng. (25)

Then for any x ∈ R,

|f (x)− fK ,N(x)|=

∣∣∣∣∣∣∑|k|>K

∑|n|>N

〈f ,Tαk Mβnγ〉Tαk Mβng(x)

∣∣∣∣∣∣≤∑|k|>K

∑|n|>N

|〈f ,Tαk Mβnγ〉|·|e2πiβn·(x−αk)|·|g(x − αk)|.

≤∑|k|>K

∑|n|>N

|〈f ,Tαk Mβnγ〉||g(x − αk)|.

(26)

Page 31: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

31/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Proof of Theorem

Note that we can consider the Gabor coefficients as

〈f ,Tαk Mβnγ〉 = Hα,β,k (n), where Hα,β,k (t0) = f (1β

t0 + αk)γ(1β

t0), (27)

for t0 ∈ R. We need to discuss the properties of the dual window function γ.

In fact, from the work by Christensen, Kim, Kim on regularity of dual Gaborwindows, we obtain the following Lemma:

Lemma (Construction of Smooth Dual Window)

Given window function g = rect( 12 x + 1)− rect(x) + rect( 1

2 x − 1), thereexists dual window function γ such that γ is smooth with finite support.

Page 32: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

32/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Proof of Theorem

Since f ∈ Cs(R), γ ∈ C∞(R) and γ has finite support, we have

|Hα,β,k (n)|< Cs

ns , (28)

and ∑|n|>N

|Hα,β,k (n)|<∑|n|>N

Cs

ns < s∫ ∞

N

2Cs

ns dn =2sCs

Ns−1 , (29)

where Cs is a constant proportional to the L1 norm of the sth derivative ofHα,β,k . we plug in the bound in (29) back into (26) and obtain

|f − fK ,N |<∑|k|>K

2sCs

Ns−1 |g(x − αk)|. (30)

Note that g is compactly supported on [−2, 2]. Then, for any x , there are onlyfinitely many k ’s (d 4

αe) with |k |> K such that g(x − αk) 6= 0. Therefore,

|f − fK ,N |<∑|k|>K

2sCs

Ns−1 |g(x − αk)|<d 4αe2sCs

Ns−1 . (31)

Page 33: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

33/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Representation of Network Dictionary

We illustrate with figures the neurons of our neural network in the third layerand the output of our neural network with random weights.

-4 -3 -2 -1 0 1 2 3 4

-1

-0.5

0

0.5

1

1.5

2

Re(Tk*alpha

Mn*beta

g(x)), k=1, alpha=1, n=6, beta=1/6

-10 -8 -6 -4 -2 0 2 4 6 8 10

-6

-4

-2

0

2

4

6

8

10

12

Figure: Translated and modulated ReLUs (left); Output of the neural network withrandom weights when K = 8, and N = 8 (right).

Page 34: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

34/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Outline

1 Overview

2 Approximation Properties of Neural Networks

3 Gabor Invariant Representation in Quantum Energy Regression

Page 35: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

35/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Background on Quantum Energy Regression

Computation of energy of a single chemical molecule has become anessential topic in computational chemistry.A chemical molecule is represented by its state x = {rk , zk}k , whererk ∈ R3 is the position of the k th nuclei and zk > 0 is the charge of k thnuclei.The molecular energy E can be written as a functional of the electrondensity ρ(u) ≥ 0 at every position u ∈ R3.The ground state energy f (x) which is unique for every molecule x , canbe obtained by minimizing energy E over a set of electronic densities ρ:

f (x) = E(ρx ) = infρ

E(ρ).

atomic density valence density core density

Figure: Approximated Electron Density ρx

Page 36: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

36/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Invariant Properties of Chemical Molecules

The quantum energy f (x) of molecule x must satisfy the following invariantproperties:

Permutation invariance The energy functional f (x) is invariant underpermutation of indices {k = 1, ...,K} in x = {rk , zk}k .

Isometric invariance The energy functional f (x) is invariant underglobal translations, rotations, and symmetries of atomic positions rk .

In machine learning, one way to avoid direct computation of f (x) is to buildset of dictionaries of functions Φ(x) = {φi (x)}i such that the energy f (x) canbe approximated by f (x), where

f (x) = 〈w ,Φ(x)〉 =∑

i

wiφi (x).

The weights {wi} are computed such that the error∑n

j=1

∣∣∣f (xj )− f (xj )∣∣∣2 on

the training data set is minimized.We intend to build a set of dictionaries from the electronic density ofmolecules with desired invariant properties.

Page 37: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

37/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Gabor Transform on Electronic Density of Molecules

We take the Gabor transform Gρ(t , γ) for the electronic density ρ by windowfunction g:

Gρ(t , γ) =

∫ρ(x)g(x − t)e−2πixγdx ,

where t is the center location of the window.

2

4

6

8

10-6

10

20

30

0.02

0.04

0.06

0.1

0.2

0.3

1000

2000

3000

0.5

1

1.5

5

10

15

10-5

0.01

0.02

0.03

1

2

3

4

5

10-8

Figure: Gabor transform of electron density at different translation locations

Page 38: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

38/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Translation Invariance

For translation invariance, we take the modulus of Gρ(t , γ) and integrate overall t : Gρ(γ) =

∫R3 |Gρ(t , γ)|dt .

Gabor coefficients integrated over translations

Figure: Gabor coefficients integrated over translations

Page 39: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

39/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Rotation Invariance

In order to obtain rotation invariance in the representation, we take averageof the coefficients across locations of the same distance to the center.

0 50 100 150 200 250 300

distance to center

0

2

4

6

8

10

12

1410

4 L1 average over radii

0 50 100 150 200 250 300

distance to center

0

2

4

6

8

10

12

14

16

1810

9 L2 average over radii

Figure: Rotation and translation invariant representation

To capture information of f of different widths at t , we adopt two differentGaussian functions g1 and g2. The Gabor invariant dictionary is defined as:

Φρ = {‖ρ‖1, ‖G1ρ,kε‖1, ‖G1

ρ,kε‖22, ‖G2

ρ,kε‖1, ‖G2ρ,kε‖2

2}0≤k≤ε−2 . (32)

Page 40: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

40/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Quantum Energy Regression

We use sparse orthogonal least squares regression in dictionariesΦ(x) = {φk (x)}k . We compare our results with state-of-the-art methods:

M RMSE MAE

Coulomb Matrix 6.7 ±2.8 14.8 ± 12.2

Fourier 73±27 6.7±0.7 8.5±0.9

Wavelet 38±13 6.9±0.6 9.1±0.8

Scattering 16 74 6.9 9.0

Gabor 71±31 5.3±0.3 7.0±0.6

Scattering 17 107±41 3.2±0.1 4.5±0.2

Table: Average Error ± Standard Deviation over the five folds in kcal/mol

The Gabor invariant representation is promising for its extend-ability to 3D.

Page 41: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

41/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

Thank You!

Page 42: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

42/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

References I

Helmut Bolcskei, Philipp Grohs, Gitta Kutyniok, and Philipp Petersen,Optimal approximation with sparsely connected deep neural networks,arXiv preprint arXiv:1705.01714 (2017).

Radu Balan, Maneesh Singh, and Dongmian Zou, Lipschitz propertiesfor deep convolutional networks, arXiv preprint arXiv:1701.05217 (2017).

Ole Christensen, Hong Oh Kim, and Rae Young Kim, Regularity of dualgabor windows, Abstract and Applied Analysis, vol. 2013, Hindawi, 2013.

CK Chui, Xin Li, and HN Mhaskar, Neural networks for localizedapproximation, Mathematics of Computation 63 (1994), no. 208,607–623.

George Cybenko, Approximation by superpositions of a sigmoidalfunction, Mathematics of control, signals and systems 2 (1989), no. 4,303–314.

John G Daugman, Complete discrete 2-d gabor transforms by neuralnetworks for image analysis and compression, IEEE Transactions onacoustics, speech, and signal processing 36 (1988), no. 7, 1169–1179.

Page 43: Feature Extraction in Deep Learning and Image ProcessingApproximation Properties of Neural Networks Gabor Invariant Representation in Quantum Energy Regression Approximation Properties

43/42

OverviewApproximation Properties of Neural Networks

Gabor Invariant Representation in Quantum Energy Regression

References II

Karlheinz Grochenig, Foundations of time-frequency analysis, SpringerScience & Business Media, 2013.

Matthew Hirn, Stephane Mallat, and Nicolas Poilvert, Wavelet scatteringregression of quantum chemical energies, Multiscale Modeling &Simulation 15 (2017), no. 2, 827–863.

Matthew Hirn, Nicolas Poilvert, and Stephane Mallat, Quantum energyregression using scattering transforms, arXiv preprint arXiv:1502.02077(2015).

Arthur P Lobo and Philipos C Loizou, Voiced/unvoiced speechdiscrimination in noise using gabor atomic decomposition, Acoustics,Speech, and Signal Processing, 2003. Proceedings.(ICASSP’03). 2003IEEE International Conference on, vol. 1, IEEE, 2003, pp. I–I.

Uri Shaham, Alexander Cloninger, and Ronald R Coifman, Provableapproximation properties for deep neural networks, Applied andComputational Harmonic Analysis (2016).