All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

15
All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer

Transcript of All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Page 1: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

All of StatisticsChapter 5: Convergence

of Random Variables

Nick Schafer

Page 2: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Overview

What are we studying? Probability What is probability? The mathematical

language for quantifying uncertainty Why are we studying probability? Dan has his

reasons…

Page 3: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Motivation

These are the kind of questions that we hope to be able to answer.

Page 4: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Review Definition: Random Variable –

a random variable X is a mapping from the sample space to the real line

Example: Flip a fair coin twice. The sample space is every combination of heads and tails. Choose the random variable X to be the number of heads. Let our outcome be one head and one tail. X maps this outcome to 1.

Note: The notation can be confusing. In the book, usually X denotes the map and X with a subscript denotes a real number. However, this is not always the case, so you must examine the context to be sure of the meaning.

:X

{ , , , }HH HT TH TT

1XHT X

Page 5: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Sequences of Random Variables Much of probability theory is

concerned with large sequences of random variables.

This study is sometimes known as large sample theory, limit theory, or asymptotic theory.

What is a sequence of random variables? Simply a set of indexed set of random variables.

We will be interested in sequences that have some interesting limiting behavior (i.e. we can say something about them as n gets large)

1, , nX X

Page 6: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

A Special Kind of Sequence of Random Variables As it turns out, a very common and

particularly useful class of sequences of random variables is IID

Definition: IID – Independent and identically distributed Independent – essentially, the value of one

random variable doesn’t effect the value of any other; for instance, a coin doesn’t remember which side last landed up, so consecutive flips are said to be independent

Identically distributed – each random variable X has associated with it a cumulative distribution function (CDF) which is derived from the probability measure. The CDF gives the probability of the value of the random variable being less than or equal to a certain value. When two or more random variables have the same CDF, we say they are identically distributed.

[0,1]PA P A

( ) ( )XP CDF F x P X x

1, , nX X F

Page 7: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Statements about sequences of random variables Given a sequence of random variables, most likely IID, it would

be useful to be able to make statements of the form: The average of all Xi will be between two values with certain

probability. Example: Flip a fair coin n times. The average number of heads per

toss will be between .4 and .6 with probability greater than or equal to 70% if we flip 84 times (n=84).

Or, what is the probability that the average of Xi is less/greater than a certain value?

These kind of statements can be made with the help of the Weak Law of Large Numbers (WLLN) and the Central Limit Theorem (CLT), respectively. So why not state them now? “Hold your horses” there, Makarand. The statement of the WLLN and the CLT make use of a few different types of convergence which must be discussed first.

Page 8: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Types of Convergence There are two main types of convergence

Convergence in Probability (CIP) A sequence of random variables is said to

converge in probability to X if the probability of it differing from X goes to zero as n gets large.

Convergence in Distribution (CID) A sequence of random variables is said to

converge in distribution to X if the limit of the corresponding CDFs is the CDF of X.

There is also another type, called convergence in quadratic mean, which is used primarily because it is stronger than CIP or CID (it implies CIP and CID) and it can be computed relatively easily.

( ) 0n

Pn

P X X

X X

lim

~

nn

n

F t F t

X X

qmnX X

qm CIP CID

Page 9: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Weak Law of Large Numbers

If a sequence of random variables are IID, then the sample average converges in probability to the expectation value.

On the left we have information about many trials and on the right we have information about the relative likelihood of the different values a random variable can take on. In words, the WLLN says that the distribution of the sample average becomes more concentrated around the expectation as n gets large.

1( ) ( )PnX E X xf x dx

1n iiX X

n

Page 10: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Example of using the WLLN Consider flipping a coin for which the

probability of heads is p. Let Xi denote the outcome of a single toss (either 0 or 1). Hence p = P(Xi=1)=E(Xi). The first equality is a definition. The second equality is obtained by averaging over the distribution. The fraction of heads after n tosses is equal to the sample average. Note that the Xi are IID. Therefore the WLLN can be applied. The WLLN says that the sample average converges to p = E(Xi) in probability.

You may find yourself wondering, how many times must I flip this coin such that the sample average is between .4 and .6 with probability greater than or equal to 70%? The WLLN tells you that it is possible to find such an n. The inequalities that Justin presented on from Chapter 4 can be used to show that n=84 does the trick in this case, but I’ll spare you the details.

( ) ( )

0(1/ 2) 1(1/ 2) 1/ 2

i iiE X x f x

1n iiX X

n

1( )PnX E X

Page 11: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

The Central Limit Theorem Given a sequence of random variables with a mean and a

variance, the CLT says that the sample average has a distribution which is approximately Normal, and gives the new mean and variance.

21

2

, , ; ;

~ ( , / )

n

n

X X

X N n

Notice that nothing at all need be assumed about the P, CDF, or PDF associated with X, which could have any distribution from which a mean and variance can be derived.

Page 12: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Example of using the CLT Suppose that the number of errors per computer

program has a Poisson distribution with mean 5. We get 125 programs. Approximately what is the probability that the average number of errors per computer program is less than 5.5?

In this case 125 is the sample size, which we hope is large enough to make a good approximation. The approximation we are making here is that the sample average will have a Normal distribution. Taking the sample size, mean and variance into account, it is possible to show that the question asked is equivalent to the probability of the standard Normal distribution being less than 2.5, which turns out to be approximately 0.9983.

Page 13: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Topics in Chapter 5 not covered in this presentation

All proofs Slutzky’s theorem and related theorems

The effect of adding sequences of random variables on their convergence behavior

Multivariate central limit theorem CLT with IID random vectors instead of variables

The delta method The effect of applying a smooth function to a

sequence of random variables on its limiting behavior

Page 14: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Interesting problems

6 & 8

Page 15: All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Bibliography

Chapters 1-5 of All of Statistics by Larry Wasserman