Discrete time processes - ETH Zürich · Stochastic Systems, ... Brownian Motion is a discrete-time...

44
Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013

Transcript of Discrete time processes - ETH Zürich · Stochastic Systems, ... Brownian Motion is a discrete-time...

Discrete time processes

Predictions are difficult. Especially about the future

Mark Twain.

Florian Herzog

2013

Modeling observed data

When we model observed (realized) data, we encounter usually the following situations:

• Time-series data of a single time series (Dt), e.g. economic data.

• Online observations also single time series (Dt)

• Ensemble data from measurements (experiments) or seasonal data (Djt )

Modeling ensemble data is often easier than single time series, because we have the

same stochastic information of a given point in time t with multiple realizations.

By Dt t = 1, ..., T we denote the data observation at time in the case that we have

only one single time series observation. By Djt t = 1, ..., T j = 1, ..., l we denote

the time series observation at time t and measurement j.

Data Dt is recorded in discrete intervals (sampled data) and assumed to be generated

by an SDE.

Stochastic Systems, 2013 2

Modeling real-world data

Stochastic differential equations are can be viewed as stochastic process from two point

of view:

• Stochastic process with a probability distribution p(t|θ) across-time, where θ

denotes the parameters

• A stochastic process, where the changes in the resulting time series is the stochastic

process, i.e. p(Dt −Dt−1|θ) or p(Dt−Dt−1Dt−1

|θ)

The first interpretation is help full to describe ensemble data and the second to analyze

single time series.

In order to deal with discrete data, all SDEs need to be discretized.

Stochastic Systems, 2013 3

Discrete-time white noise

When we assume a discrete white noise process εt which identically and independently

distributed (i.i.d) with zero expectation and σ2 variance. Then following statistical

properties apply:

E[εt] = 0 (1)

E[ε2t ] = σ

2(2)

E[εtετ ] = 0 for t ̸= τ (3)

The last property is derived from the assumption of independence. A discretized

Brownian Motion is a discrete-time white noise process with εt ∼ N (0,∆t) where

∆t is the time discretization.

In discrete-time models, a white noise process can be normally distributed (Gaussian

white noise) but can be distributed by any other distribution as long as the i.i.d.

assumption is valid.

Stochastic Systems, 2013 4

Statistical Theory

Stationarity: A process is covariance-stationary if and only if (iff):

E[Dt] = µ ∀t (4)

E[(Dt − µ)(Dt−j − µ)] = νj ∀t ∀j . (5)

This form of stationarity is often also called weak stationarity. For example the process:

Dt = µ + εt where εt is white noise process with zero mean and σ2 variance. In this

case Dt is stationary since:

E[Dt] = µ

E[(Dt − µ)(Dt−j − µ)] =

{σt j = 0

0 j ̸= 0

Stochastic Systems, 2013 5

Statistical Theory

Stationarity: A process is strictly stationary iff for any value of j1, j2, j3, ... the joint

distribution of Dt, Dt+j1, Dt+j2

, Dt+j3, .. depend only on the intervals separating the

different values of j and not the time t.

Ergodicity: A process is called ergodic for the mean iff:

limt→∞

E[Dt] = µ . (6)

A process is said to be ergodic for the second moment iff:

limT→∞

1

T − jE[(Dt − µ)(Dt−j − µ)] →p

νj (7)

where →p denotes convergence in probability. From an ergodic process, we can estimate

all necessary data without possing ensemble data. If an process is not ergodic, we have

only one observation, we can not identify its parameters.

Stochastic Systems, 2013 6

Statistical Theory

It can be shown that if the autocovariance of a covariance stationary process satisfies

the following:

∞∑j=0

|νj| < ∞ (8)

then the process is ergodic for the mean.

Stochastic Systems, 2013 7

First Order Moving Average Process

When εk be a Gaussian white noise processes with mean zeros and variance σ2 and

consider the process:

Yk = µ+ εk + θεk−1 , (9)

where µ and θ are nay real valued constant. The process is called first order moving

average process and ussualy denotes as MA(1). The expectation of Yk is given as:

E[Yk] = E[µ+ εk + θεk−1]

= µ+ E[εk] + θE[εk−1] = µ

The constant µ is thus the mean of the processes.

Stochastic Systems, 2013 8

First Order Moving Average Process

The variance of the MA(1) is calculated as

Var[Yk] = E[(Yk − µ)2] = E[(εk + θεk−1)

2]

= E[ε2k + θ

2ε2k−1 + 2θεkεk−1]

= σ2+ 0 + θ

2= (1 + θ

2)σ

2

The first autocovariance is computed as

E[(Yk − µ)(Yk−1 − µ)] = E[(εk + θεk−1)(εk−1 + θεk−2)]

= E[εkεk−1 + θε2k−1 + θεk−1εkθ

2εk−1εk−2]

= 0 + θσ2+ 0 + 0 = θσ

2

Stochastic Systems, 2013 9

First Order Moving Average Process

The higher order autocovariances are all zero:

E[(Yk − µ)(Yk−j − µ)] = E[(εk + θεk−1)(εk−j + θεk−j−1)] = 0 .

Since the mean and autocovariance are not function of times, an MA(1) process is

covariance stationary regardless of the value of θ, since

∞∑j=0

|νj| = (1 + θ2)σ

2+ |θσ2| . (10)

When εk is a Gaussian white noise process, the MA(1) is ergodic for all moments.

Stochastic Systems, 2013 10

First Order Moving Average Process

Autocorrelation is defined as ρ1 = Cov[YkYk−1]/√V ar[Yk]V ar[Yk−1] and

V ar[Yk] = V ar[Yk+1] = (1 + θ2)σ2. Therefore we get

ρ1 =θσ2

(1 + θ2)σ2=

θ

1 + θ2. (11)

All higher order autocorrelations are zero. Positive values of θ induce positive auto-

correlation, negative values induces negative autocorrelation. The autocorrelation of an

MA(1) process can be between -0.5 and 0.5. For any value of ρ1, there exists two

values of θ, see

ρ1 =(1/θ)

1 + (1/θ)2=

θ2(1/θ)

θ2(1 + (1/θ)2)=

θ

1 + θ2. (12)

Stochastic Systems, 2013 11

Q-th Order Moving Average Process

A q-th order moving average process, denotes as MA(q) is given by

Yk = µ+ εk + θ1εk−1 + θ2εk−2 + ......+ θqεk−q , (13)

where µ and θ1, ...θ1 are real valued constant. The expectation of Yk is given as:

E[Yk] = µ+ E[εk] + θ1E[εk−1] + ......+ θqE[εk−q] = µ

The constant µ is thus the mean of the processes. The variance is computed as The

variance of the MA(1) is calculated as

Var[Yk] = E[(Yk − µ)2] = E[(εk + θ1εk−1 + θ2εk−2 + ......+ θqεk−q)

2] .

Since all ε’s are uncorrelated, all cross terms εjεk are zero.

Stochastic Systems, 2013 12

Q-th Order Moving Average Process

The variance is therefore

Var[Yk] = σ2+ θ

21σ

2+ ......+ θ

2qσ

2

= σ2(1 +

q∑j=1

θ2j) .

The autocovariance is for the i-th lag is computed as

νi = E[(εk +

q∑j=1

θjεk−j)(εk−i +

q∑j=1

θjεk−j−i)]

= E[θiε2k−i + θi+1θ1ε

2k−i−1 + θi+2θ2εk−i−2 + ......+ θqθq−iε

2k−q]

Stochastic Systems, 2013 13

Q-th Order Moving Average Process

The ε’s at different dates have been drop since their product is zero. For i > q there

are no ε’s with common dates in the definition of of νi. We get for the autocovariance

thus

νi =

{(θi + θi+1θ1 + θi+2θ2 + ...+ θqθq−j)σ

2 i = 1, 2, ..., q

0 i > q

The autocorrelation is given as

νi =

θi+θi+1θ1+θi+2θ2+...+θqθq−j

(1+∑qj=1

θ2j)

i = 1, 2, ..., q

0 i > q

Stochastic Systems, 2013 14

Autoregressive first order Process

We analyze an so-called first order autoregressive (AR(1)):

Yk+1 = c+ ΦYk + εk

Long-term mean: We take the mean on both sides of the equation

E[Yk+1] = E[c+ ΦYk + εk]

E[Yk+1] = c+ ΦE[Yk]

µ = = c+ Φµ

µ =c

1 − Φ

Stochastic Systems, 2013 15

Autoregressive first order Process

µ =c

1 − Φ

From this formula, we can see that Φ must be smaller than 1, otherwise the mean

would be negative when c ≥ 0. Also we have assumed that E[Yk] = E[Yk+1] = µ

and exists. Second moment:

Yk+1 = µ(1 − Φ) + ΦYk + εk

Yk+1 − µ = Φ(Yk − µ) + εk

Stochastic Systems, 2013 16

Autoregressive first order Process

E[(Yk+1 − µ)2] = E[(Φ(Yk − µ) + εk)

2]

= Φ2E[(Yk − µ)] + 2ΦE[(Yk − µ)εk] + E[ε

2k]

E[(Yk+1 − µ)2] = Φ

2E[(Yk − µ)] + σ

2

where E[(Yk−µ)εk] = 0. Assuming covariance stationarity, i.e. E[(Yk+1−µ)2] = ν

and E[(Yk − µ)2] = ν. We get:

ν = Φ2ν + σ

2

ν =σ2

1 − Φ2

Stochastic Systems, 2013 17

Autoregressive first order Process

Autocorrelation:

E[(Yk+1 − µ)(Yk − µ)] = E[(Φ(Yk − µ) + εk)(Yk − µ)]

E[(Yk+1 − µ)(Yk − µ)] = ΦE[((Yk − µ)2] + E[(εk)(Yk − µ)] .

Again E[(εk)(Yk − µ)] = 0 and E[((Yk − µ)2] = ν we get:

E[(Yk+1 − µ)(Yk − µ)] = Φν = Φσ2

1 − Φ2.

Autocorrelation is defined as Cov[Yk+1Yk]/√V ar[Yk+1]V ar[Yk] and V ar[Yk] =

V ar[Yk+1] = ν. Therefore we get

acf(Yk+1Yk) = ϕ (14)

Stochastic Systems, 2013 18

Autoregressive first order Process

AR(1) process with Yk+1 = ΦYk + c+ εk hast the following properties:

Mean : µ =c

1 − Φ

V ariance : ν =σ2

1 − Φ2

Autocorr. : acf(Yk+jYk) = ϕj

A covariance stationary AR(1) process is only well defined when |Φ| ≤ 1, otherwise

the variance is infinite and mean makes no sense. We can check if a process is a AR(1)

process by calculating the mean and variance over-time and the autocorrelation. From

these three calculations the parameters can be estimated. It is important to notice that

the autocorrelation is exactly the coefficient of the autoregression.

Stochastic Systems, 2013 19

Example: GBM and mean reversion

The two examples have a discrete-time structure as an first-order autoregressive process

with:

(I) Φ = 1 + µ∆t

(II) Φ = 1 − b∆t

When we assume that µ > 0, the first model is instable and we can not use the

techniques from auto-regressions to estimate the parameters. We have to deal different-

ly in this case, by using p = ln(x), since then we get an SDE with constant coefficients.

When we assume that b > 0, the process is a true mean reverting process that can be

estimated by AR(1) methods. In this case, the discrete-time model is also stable.

Stochastic Systems, 2013 20

First order difference equation

A first order difference equation give by

yk+1 = Φyk + wk

is of the same type as an AR(1) equation. The can be solved by recursive substitution

since y1 = Φy0 + w0, y2 = Φy1 + w1, y3 = Φy2 + w2 ... . When we plug in the

results of y1 into the equation for y2, we get y2 = Φ2y0 + Φw0 + w1, which in turn

we plug in the equation for y3 and get y3 = Φ3y0 + Φ2w0 + Φw1 + w2. Expanding

this further on, we get the general solution

yk+1 = Φk+1y0 +

k∑j=0

Φk−jwj

Stochastic Systems, 2013 21

p-th order difference equation

A p-th order difference equation give by

yk+1 = Φ1yk + Φ2yk−1 + ......+ Φpyk−p+1 + wk

We write this in a vectorized form with:

ξk =

ykyk−1

yk−2...

yk−p−1

F =

Φ1 Φ2 Φ3 · · · Φp−1 Φp

1 0 0 · · · 0 0

0 1 0 · · · 0 0... ... ... · · · ... ...

0 0 0 · · · 1 0

vk =

wk0

0...

0

and get

ξk+1 = Fξk + vk .

Stochastic Systems, 2013 22

p-th order difference equation

The solution is given as:

ξk = Fkξ0 + F

k−1v0 + F

k−2v1 + F

k−3v2 + ......+ Fvk−1 + vk

The p-th order difference equation is stable (or stationary) when the eigenvalues of F

are smaller than one. The eigenvalue polynom is given as:

λp − Φ1λ

p−1 − Φ2λp−2 − ......− Φp−1λ− Φp = 0

Stochastic Systems, 2013 23

Lag operators

A highly useful operator for time series models is the lag operator. Suppose that we

start with an sequence of data xk where k = −∞...∞ and generate a new sequence

yk where the value of y for date k is equal to the value x took at time k − 1

yk = xk−1 .

This is described as applying the Lag operator on xk and is represented by symbol L:

Lxk = xk−1 .

Consider applying the lag operator twice to a series:

L(Lxk) = L(xk−1) = xk−2 ,

which is denotes as L2 or L2xk = xk−2.

Stochastic Systems, 2013 24

Lag operators

In general for any integer l

Llxk = xk−l ,

The lag operator is linear and can be treated as the multiplation operator and thus

L(βxk) = βLxk = βxk−1

L(xk + wk) = Lxk + Lwk

yk = (a+ bL)Lxk = (aL+ bL2)xk = axk−1 + bxk−2

An expression such as (aL+ bL2) is referred as a polynomial lag operators.

Notice if xk = c then Lxk = xk−1 = c

Stochastic Systems, 2013 25

First order difference equation and lag operators

A first order difference equation give by

yk+1 = Φyk + wk

can be written as

yk+1 = ΦLyk+1 + wk

yk+1 − ΦLyk+1 = wkyk+1(1 − L) = wk

when we multiply both sides with (1+ΦL+Φ2L2 +Φ3L3 + ......+ΦkLk) we get

yk+1(1 + ΦL+ Φ2L

2+ Φ

3L

3+ ......+ Φ

kLk)(1 − L)

= (1 + ΦL+ Φ2L

2+ Φ

3L

3+ ......+ Φ

kLk)wk

Stochastic Systems, 2013 26

First order difference equation and lag operators

yk+1(1 − Φk+1Lk+1

) = (1 + ΦL+ Φ2L

2+ Φ

3L

3+ ......+ Φ

kLk)wk

yk+1(1) = Φk+1Lk+1yk+1 + (1 + ΦL+ Φ

2L

2+ Φ

3L

3+ ......+ Φ

kLk)wk

yk+1 = Φk+1y0 + wk + Φwk−1 + Φ

2wk−2 + Φ

3wk−3 + ......+ Φ

kw0

which results exactly in the same solution as by the usage of the recursive substitution.

Stochastic Systems, 2013 27

P-th order difference equation and lag operators

The p-th order difference equation given by

yk+1 = Φ1yk + Φ2yk−1 + ......+ Φpyt−p+1 + wk ,

can be written using the lag operator as:

yk+1(1 − Φ1L− Φ2L2 − Φ3L

3 − ....− ΦpLp) = wk ,

which results in the same polynomial as in the case of the eigenvalues of the matrix F .

Stochastic Systems, 2013 28

P-th order difference equation and lag operators

Since

(1 − Φ1z − Φ2z2 − Φ3z

3 − ....− Φpzp) = (1 − λ1z)(1 − λ2z)....(1 − λpz) ,

and by defining z−1 = λ and multiply both sides with z−p and we get

(λp − Φ1λ

p−1 − Φ2λp−2 − ....− Φp) = (λ− λ1)(λ− λ2z)....(λ− λpz) ,

which is nothing but the root search of the left side polynomial:

λp − Φ1λ

p−1 − Φ2λp−2 − ......− Φp−1λ− Φp = 0

Stochastic Systems, 2013 29

Second order Autoregression

A second order autoregressive process (denotes as AR(2)) has the form:

Yk+1 = c+ Φ1Yk + Φ2Yk−1 + εk

written in lag notation:

Yk+1(1 − Φ1L− Φ2L2) = c+ εk

The AR(2) is stable if the roots of (1 − Φ1z − Φ2z2) = 0 lie outside the unit circle.

The process has the mean value:

E[Yk+1] = c+ Φ1E[Yk] + Φ2E[Yk−1] + E[εk]

µ = c+ Φ1µ+ Φ2µ

µ =c

1 − Φ1 − Φ2

Stochastic Systems, 2013 30

Second order Autoregression

To find the second moments of the AR(2), we write

Yk+1 = µ(1 − Φ1 − Φ2) + Φ1Yk + Φ2Yk−1 + εk

Yk+1 − µ = Φ1(Yk − µ) + Φ2(Yk−1 − µ) + εk .

multiplying both sides with Yk+1−j − µ we get and taking the expectation yields:

νj = Φ1νj−1 + Φ2νj−2 j = 1, 2, 3..

The auto-covariances follow the same second order difference equation as the AR(2)

process. The autocorrelation is given by dividing the last equation by ν0 we get

ρj = Φ1ρj−1 + Φ2ρj−2 j = 1, 2, 3..

Stochastic Systems, 2013 31

Second order Autoregression

For j = 1 we get for the autocorrelation:

ρ1 = Φ1 + Φ2ρ1

ρ1 =Φ1

1 − Φ2

,

since ρ0 = 1 and ρ−1 = ρ1. For j=2 we get

ρ2 = Φ1ρ1 + Φ2

ρ2 =Φ2

1

1 − Φ2

+ Φ2 .

The covariance is computed by

E[(Yk+1 − µ)(Yk+1 − µ)] = Φ1E[(Yk − µ)(Yk+1 − µ)] + Φ2E[(Yk−1 − µ)(Yk+1 − µ)]

+ E[εk(Yk+1 − µ)] ,

Stochastic Systems, 2013 32

Second order Autoregression

The covariance equation can be written as

ν0 = Φ1ν1 + Φ2ν2 + σ2j = 1, 2, 3..

The last term from noticing

E[εk(Yk+1 − µ)] = E[(εk)(Φ1(Yk − µ)

+Φ2(Yk−1 − µ) + εk)]

= Φ10 + Φ20 + σ2

The covariance equation can be written as

ν0 = Φ1ρ1ν0 + Φ2ρ2ν0 + σ2,

ν0 =(1 − Φ2)σ

2

(1 + Φ2)[(1 − Φ22) − Φ2

1]

Stochastic Systems, 2013 33

P-th order Autoregression

A p-th order autoregressive process (denotes as AR(p)) has the form:

Yk+1 = c+ Φ1Yk + Φ2Yk−1 + ...+ ΦpYk−p+1 + εk

written in lag notation:

Yk+1(1 − Φ1L− Φ2L2 − ...− ΦpL

p) = c+ εk

The AR(2) is stable if the roots of (1 − Φ1z − Φ2z2 − ...− Φpz

p) = 0 lie outside

the unit circle. The process has the mean value:

E[Yk+1] = c+ Φ1E[Yk] + Φ2E[Yk−1] + ...+ ΦpE[Yk−p+1] + E[εk]

µ = c+ Φ1µ+ Φ2µ+ ...+ Φpµ

µ =c

1 − Φ1 − Φ2 − ....− Φp

Stochastic Systems, 2013 34

P-th order Autoregression

The autocovariance is given as

νj =

{Φ1νj−1 + Φ2νj−2 + ...+ Φpνj−p j = 1, 2, 3..

Φ1ν1 + Φ2ν2 + ...+ Φpνp j = 0

The equation for the autocorrelation is given as

ρj = Φ1ρj−1 + Φ2ρj−2 + ...+ Φ2ρj−p j = 1, 2, 3..

which is known as the ”Yule-Walker” equation. The equation is again a p-th order

difference equation.

—————————————————————–

Stochastic Systems, 2013 35

Mixed Autoregressive Moving Average Process (ARMA)

A ARMA(p,q) process includes autoregressive elements as well as moving average

elements

Yk+1 = c+ Φ1Yk + Φ2Yk−1 + ...+ ΦpYk−p+1

+εk + θ1εk−1 + θ2εk−2 + ......+ θqεk−q .

The process is written in Lag form as

Yk+1 (1 − Φ1L− Φ2L2 − ...− ΦpL

p) =

+c+ εk(1 + θ1L+ θ2L2+ ......+ θqL

q) .

Provided that the roots of (1−Φ1L−Φ2L2 − ...−ΦpL

p) = 0 lie outside the unit

circle we obtain the following general form for an ARMA(p,q) process:

Yk+1 = µ+(1 + θ1L+ θ2L

2 + ......+ θqLq)

(1 − Φ1L− Φ2L2 − ...− ΦpLp)εk .

Stochastic Systems, 2013 36

Mixed Autoregressive Moving Average Process (ARMA)

The mean is given as

µ =c

(1 − Φ1L− Φ2L2 − ...− ΦpLp).

The stationarity of an ARMA process depends only on the autoregressive parameters

and not on the moving average parameters. The more convenient form of an ARMA(p,q)

process is given as

Yk+1 − µ = Φ1(Yk − µ) + Φ2(Yk−1 − µ) + ...+ Φp(Yk−p+1 − µ)

+εk + θ1εk−1 + θ2εk−2 + ......+ θqεk−q .

The autocovariance equation for j > q is given as

νj = Φ1νj−1 + Φ2νj−2 + ...+ Φpνj−p j = q + 1, q + 2, ...

Stochastic Systems, 2013 37

Mixed Autoregressive Moving Average Process (ARMA)

Thus after q lags, the autocovariance function νj follows a p-th order difference equation

governed by the the autoregressive part.

Note that last equation does not hold for j ≤ q since the correlation between θjεk−jand Yt−j cannot be neglected.

In order to analyze the auto-covariance, we introduce the auto-covariance generating

function

gY (z) =∞∑

j=−∞νjz

j

which is constructed as taking the j-th auto-covariance and multiplying with z to the

power of j and then summing over all possible values of j

Stochastic Systems, 2013 38

Parameter Identification

We start to examine the appropriate log-likelihood function with our frame work of the

discretized SDEs. Let ψ ∈ Rd be the vector of unknown parameters, which belongs to

the admissible parameter space Ψ, e.g. a, b, etc. The various system matrices of the

discrete models, as well as the variances of stochastic processes, given in the previous

section, depend on ψ. The likelihood function of the state space model is given by the

joint density of the observational data x = (xN , xN−1, . . . , x1)

l(x, ψ) = p(xN , xN−1, . . . , x1;ψ) (15)

which reflects how likely it would have been to have observed the date if ψ were the

true values of the parameters.

Stochastic Systems, 2013 39

Parameter Identification

Using the definition of condition probability and employing Bayes’s theorem recursively,

we now write the joint density as product of conditional densities

l(x, ψ) = p(xN |xN−1, . . . , x1;ψ) · . . . · p(xk|xk−1, . . . , x1;ψ) · . . . · p(x1;ψ) (16)

where we approximate the initial density function p(x1;ψ) by p(x1|x0;ψ). Further-

more, since we deal with a Markovian system, future values of xl with l > k only

depend on (xk, xk−1, . . . , x1) through the current values of xk, the expression in (16)

can be rewritten to depend on only the last observed values, thus it is reduced to

l(x, ψ) = p(xN |xN−1;ψ) · . . . · p(xk|xk−1;ψ) · . . . · p(x1;ψ) (17)

Stochastic Systems, 2013 40

Parameter Identification

A Euler discretized SDE is given as:

xk+1 = xk + f(xk, k)∆t+ g(xk, k)√∆tεk (18)

where εk is an i.i.d random variable with normal distribution, zero expectation and unit

variance. Since f and g are deterministic functions, xk+1 is conditional (on xk normal

distributed with

E[xk+1] = xk + f(xk, k)∆t (19)

Var[xk+1] = g(xk, k)g(xk, k)T∆t (20)

p(xk+1) =1

(2π)n2√

|Var[xk+1]|e−12(xk+1−E[xk+1])

TVar[xk+1]−1(xk+1−E[xk+1])(21)

Stochastic Systems, 2013 41

Parameter Identification

The parameter of an SDEs are estimated by maximizing the log-likelihood. To maximize

the log-likelihood is equivalent to maximizing the likelihood, since the log function is

strictly increasing function. The log likelihood is given as:

ln(l(x, ψ)) = ln(p(xN |xN−1;ψ) · . . . · p(yk|xk−1;ψ) · . . . · p(x1;ψ))(22)

ln(l(x, ψ)) =

N∑j=1

ln(p(xj|xj−1;ψ) (23)

ln(l(x, ψ)) =

N∑j=1

{ln((2π)

−n2 |Var[xk+1]|−

12)

−1

2(xk+1 − E[xk+1])

TVar[xk+1]

−1(xk+1 − E[xk+1])

}(24)

Stochastic Systems, 2013 42

Example: Parameter Identification

We want to identify the parameters of a AR(1) process without any constant:

xk+1 = Φxk + σεk (25)

which has the log-likelihood function:

ln(l(x, ψ)) =N∑j=1

ln((2π)−12σ

−1) −

(xk+1 − Φxk)2

2σ2. (26)

The function is maximized by taking the derivatives for Φ and σ. The log-likelihood

function has the same from as an so-called ordinary least-square (OLS) estimation.

Stochastic Systems, 2013 43

Example: Parameter Identification

The solution of the maximum likelihood estimations is:

Φ̂ =

∑Nj=1 xk+1xk∑N

j=1 x2k

(27)

σ̂ =1

N

N∑j=1

(xk+1 − Φxk)2

(28)

The solution for Φ is the same solution as for an OLS regression. Furthermore, the

solution is exactly the definition of autocorrelation with lag one. The estimation for σ

is the empirical standard deviation of the estimation of the empirical white noise, i.e.

ε̂k = xk+1 − Φxk

Stochastic Systems, 2013 44