Regression Variance-Bias Trade-off

7
Regression Variance-Bias Trade- off

description

Regression Variance-Bias Trade-off. Regression. We need a regression function h(x ) We need a loss function L(h(x),y ) We have a true distribution p(x,y ) Assume a quadratic loss, then:. Note: y t ; h(x)y(x ). e stimation error. n oise error. Regression: Learning. - PowerPoint PPT Presentation

Transcript of Regression Variance-Bias Trade-off

Page 1: Regression Variance-Bias Trade-off

RegressionVariance-Bias Trade-off

Page 2: Regression Variance-Bias Trade-off

Regression• We need a regression function h(x)

• We need a loss function L(h(x),y)

• We have a true distribution p(x,y)

• Assume a quadratic loss, then:

Note: yt; h(x)y(x)

estimation error noise error

Page 3: Regression Variance-Bias Trade-off

Regression: Learning• Assume h(x) is a parametric curve, e.g. h(x)=af(x)+b.

• Minimize loss over the parameters (e.g. a,b), where p(x,y) is replaced with a sum over data-cases (called a “Monte Carlo sum”):

• That is: we solve:

• The same results follows from posing a Gaussian model q(y|x) for p(y|x) with mean h(x) and maximizing the probability of the data over the parameters. (This approach is taken in 274; probabilistic learning).

Page 4: Regression Variance-Bias Trade-off

Back to overfitting• More parameters lead to more flexible functions which may lead to over-fitting.

• Formalize this by imagining very many datasets D, all of size N. Call h(x,D) the regression function estimated from a dataset D of size N, i.e. a(D)f(x)+b(D), then:

• Next, average over p(D)=p(x1)p(x2)….p(xN). Only first term depends on D:

0

Variance+bias2

Page 5: Regression Variance-Bias Trade-off

Bias/Variance Tradeoff

A

B

C

A: The label y label fluctuates (label variance).

B: The estimate of h fluctuates across different datasets (estimation variance).

C: The average estimate of h does not fit well to the true curve (squared estimation bias).

Page 6: Regression Variance-Bias Trade-off

Bias/Variance Illustration

Bias

Variance

Page 7: Regression Variance-Bias Trade-off

Relation to Over-fitting

Increasing regularization(less flexible models)

Decreasing regularization(more flexible models)

Training error is measuringbias, but ignoring variance.

Testing error / X-validation erroris measuring both bias and variance.