Fuzzy verification of fake cases

20
NCAR, 15 April 2008 1 Fuzzy verification of fake cases Beth Ebert Center for Australian Weather and Climate Research Bureau of Meteorology

description

Fuzzy verification of fake cases. Beth Ebert Center for Australian Weather and Climate Research Bureau of Meteorology. observation. forecast. Frequency. Forecast value. t - 1. t. Frequency. t + 1. Forecast value. Fuzzy (neighborhood) verification. - PowerPoint PPT Presentation

Transcript of Fuzzy verification of fake cases

Page 1: Fuzzy verification of fake cases

NCAR, 15 April 20081

Fuzzy verification of

fake cases

Beth Ebert

Center for Australian Weather and Climate Research

Bureau of Meteorology

Page 2: Fuzzy verification of fake cases

NCAR, 15 April 2008 2

• Look in a space / time neighborhood around the point of interest

– Evaluate using categorical, continuous, probabilistic scores / methods

– Will only consider spatial neighborhood for fake cases

Fuzzy (neighborhood) verification

t

t + 1

t - 1

Forecast value

Fre

qu

en

cy

Forecast value

Fre

qu

en

cy

forecast

observation

Page 3: Fuzzy verification of fake cases

NCAR, 15 April 2008 3

Fuzzy verification framework

Fuzzy methods use one of two approaches to compare forecasts and observations:

single observation – neighborhood forecast

(user-oriented)

neighborhood observation – neighborhood forecast

(model-oriented)

observation forecast

observation forecast

Page 4: Fuzzy verification of fake cases

NCAR, 15 April 2008 4

Fuzzy verification framework

good performance

poor performance

Page 5: Fuzzy verification of fake cases

NCAR, 15 April 2008 5

UpscalingNeighborhood observation - neighborhood forecast

Average the forecast and observations to successively larger grid resolutions, then verify as usual

% change in ETS

Weygandt et al. (2004)

Page 6: Fuzzy verification of fake cases

NCAR, 15 April 2008 6

Fractions skill scoreNeighborhood observation - neighborhood forecast

N

i

N

iobsfcst

N

iobsfcst

PP

PP

1 1

22

1

2

N1

N1

)(N1

1FSS

observed forecast

Compare forecast fractions with observed fractions (radar) in a probabilistic way over different sized neighbourhoods

Roberts and Lean (2008)

Page 7: Fuzzy verification of fake cases

NCAR, 15 April 2008 7

single threshold

ROC

Spatial multi-event contingency tableSingle observation - neighborhood forecast

Vary decision thresholds:

• magnitude (ex: 1 mm h-1 to 20 mm h-1)

• distance from point of interest (ex: within 10 km, .... , within 100 km)

• timing (ex: within 1 h, ... , within 12 h)

• anything else that may be important in interpreting the forecast

Fuzzy methodology – compute Hanssen and Kuipers score HK = POD – POFD

Measure how close the forecast is to the place / time / magnitude of interest.

Atger (2001)

Page 8: Fuzzy verification of fake cases

NCAR, 15 April 2008 8

Practically perfect hindcasts Single observation - neighborhood forecast

Q: If the forecaster had all of the observations in advance, what would the "practically perfect" forecast look like?

– Apply a smoothing function to the observations to get probability contours, choose yes/no threshold that maximizes CSI when verified against obs

– Did the actual forecast look like the practically perfect forecast?

– How did the performance of the actual forecast compare to the performance of the practically perfect forecast?

Fuzzy methodology – compute

forecast PracPerf

CSIforecast = 0.34 CSIPracPerf = 0.48

PracPerf

forecast

ETS

ETS

Kay and Brooks (2000)

Page 9: Fuzzy verification of fake cases

NCAR, 15 April 2008 9

1st geometric case50 pts to the right

bad

good

12.7 mm

25.4 mm

Page 10: Fuzzy verification of fake cases

NCAR, 15 April 2008 10

2nd geometric case200 pts to the right

bad

good

Page 11: Fuzzy verification of fake cases

NCAR, 15 April 2008 11

5th geometric case125 pts to the right and huge

bad

good

Page 12: Fuzzy verification of fake cases

NCAR, 15 April 2008 12

1st case vs. 5th case

~same

Case 1better

Case 5better

Page 13: Fuzzy verification of fake cases

NCAR, 15 April 2008 13

Perturbed cases

1000

km

"Observed"

(6) Shift 12 pts right, 20 pts down, intensity*1.5

(4) Shift 24 pts right, 40 pts down

Which forecast is better?

Page 14: Fuzzy verification of fake cases

NCAR, 15 April 2008 14

4th perturbed case24 pts right, 40 pts down

bad

good

Page 15: Fuzzy verification of fake cases

NCAR, 15 April 2008 15

6th perturbed case12 pts right, 20 pts down, intensity*1.5

bad

good

Page 16: Fuzzy verification of fake cases

NCAR, 15 April 2008 16

Difference between cases 6 and 4Case 4 - Shift 24 pts right, 40 pts down

Case 6 - Shift 12 pts right, 20 pts down, intensity*1.5

Case 6 – Case 46

4

Page 17: Fuzzy verification of fake cases

NCAR, 15 April 2008 17

How do fuzzy results for shift + amplification compare to results for the case of shifting only?

Case 6 - Shift 12 pts right, 20 pts down, intensity*1.5

Case 3 - Shift 12 pts right, 20 pts down, no intensity change

Case 6 – Case 3

3

Why does the case with incorrect amplitude sometimes perform better??Baldwin and Kain (2005): When the forecast is offset from the observations most scores can be improved by overestimating rain area, provided rain is less common than "no rain".

6

Page 18: Fuzzy verification of fake cases

NCAR, 15 April 2008 18

Some observations about methods

Traditional• Measures direct

correspondence of forecast and observed values at grid scale

• Hard to score well unless forecast is ~perfect

• Requires overlap of forecasts and obs

Entity-based (CRA)• Measures location

error and properties of blobs (size, mean/max intensity, etc.)

• Scores well if forecast looks similar to observations

• Does not require much overlap to score well

Fuzzy• Measures scale- and

intensity-dependent similarity of forecast to observations

• Forecast can score well at some scales and not at others

• Does not require overlap to score well

Page 19: Fuzzy verification of fake cases

NCAR, 15 April 2008 19

Some final thoughts…

Object-based and fuzzy verification seem to have different aims

Object-based methods• Focus on describing the error

• What is the error in this forecast?

• What is the cause of this error (wrong location, wrong size, wrong intensity, etc.)?

Fuzzy neighborhood methods• Focus on skill quantification

• What is the forecast skill at small scales? Large scales? Low/high intensities?

• What scales and intensities have reasonable skill?

• Different fuzzy methods emphasize different aspects of skill

Page 20: Fuzzy verification of fake cases

NCAR, 15 April 2008 20

Some final thoughts…

When can each type of method be used?

Object-based methods• When rain blobs are well defined (organized systems, longer

rain accumulations)

• When it is important to measure how well the forecast predicts the properties of systems

• When size of domain >> size of rain systems

Fuzzy neighborhood methods• Whenever high density observations are available over a

reasonable domain

• When knowing scale- and intensity-dependent skill is important

• When comparing forecasts at different resolutions