Interval Analysis Scott Ferson, [email protected] 11 September 2007, Stony Brook University, MAR 550,...

106
Interval Analysis Scott Ferson, [email protected] ember 2007, Stony Brook University, MAR 550, Challe

Transcript of Interval Analysis Scott Ferson, [email protected] 11 September 2007, Stony Brook University, MAR 550,...

Page 1: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Interval Analysis

Scott Ferson, [email protected] September 2007, Stony Brook University, MAR 550, Challenger 165

Page 2: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Outline of Interval Analysis• What they are and why they’re used• Forward calculations

– Elementary formulas– Outward-directed rounding– Example: contaminant plume travel time– Repeated parameters– Tricks and subinterval reconstitution– Propagating uncertainty through black boxes – Dependencies among intervals

• Specifying intervals• Updating

– Aggregation– Constraint propagation

• Backcalculation• Exercises

Page 3: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

An interval x = [x1, x2]

• Used when upper and lower bounds are known reliably• Sometimes absolutely sure, sometimes just a judgment

range

x1 x2

Page 4: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Why intervals are a good idea

• Intervals are easy to understand

• Calculations with them are easy

• They work with even the crappiest data

• Often sufficient for a decision

Page 5: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Rigorousness

• The computations are guaranteed to enclose the true results (so long as the inputs do)

• “Automatically verified calculations”

• You can still be wrong, but the method won’t be the reason if you are

Page 6: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

General interval arithmetic

x + y = [x1 + y1, x2 + y2]x y = [x1 y2, x2 y1]x y = [min(x1y1, x1y2, x2y1, x2y2), max(x1y1, x1y2, x2y1, x2y2)]x y = x [1/y2, 1/y1], if y doesn’t include zero

Examples: x = [3, 5]y = [1, 2]x + y = [2, 7]x y = [1, 6]x y = [5, 10]

Page 7: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Other simple arithmetic operations

min(x, y) = [min(x1, y1), min(x2, y2)]

max(x, y) = [max(x1, y1), max(x2, y2)]

env(x, y) = [min(x1, y1), max(x2, y2)]

intersect(x, y) = [max(x1, y1), min(x2, y2)], if x and y overlap

Examples: x = [0.3, 0.5], y = [0.1, 0.2]

min(x,y) = [0.1, 0.2]

env(x, y) = [0.1, 0.5]

intersect(x, y) doesn’t exist

Page 8: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Other functions

• If the function is monotone (e.g., exp, log, sqrt) just apply it to the two endpoints

• Trickier if it isn’t monotone (e.g., abs, sin, cos)

• Note that generally non-monotone functions may be monotone over a particular interval

Examples: x = [0.3, 0.5]exp(x) = [1.34, 1.65]sin(x) = [0.295, 0.480]abs(x) = [0.3, 0.5]

Page 9: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Outward-directed rounding

• To maintain the rigor of interval calculations, any necessary truncation of digits should be done with outward-directed rounding

Example:

[2,4] / 3 = [ 0.66666666666, 1.33333333333]

would be written as [0.66, 1.34]

even though 2/3 ordinarily rounds to 0.67,

and 4/3 ordinarily rounds to 1.33

_ _

Page 10: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Example: contaminant plume

• Hydrocarbon in groundwater near some wells• Constant, one-dimensional, uniform Darcian flow• Homogeneous properties (e.g., no pipes, conduits,

barriers or differential permeability among layers)• Linear retardation• No dispersion (just Darcian advection)

• How long before the contaminant reaches the wells?

Page 11: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Darcy’s Law

• Statistical description of fluid flow in porous media • Expresses conservation of momentum• Flow rate is proportional to

– cross-sectional area

– hydraulic conductivity and

– gradient in the hydraulic head

• Analogous to– Fick’s Law in diffusion

– Ohm’s Law (I=V/R)

• Derivable from Navier-Stokes by homogenization

Page 12: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Travel time (Lobascio)

iK

LKocfocBDnT

ParameterL source-receptor distancei hydraulic gradientK hydraulic conductivityn effective soil porosityBD soil bulk densityfoc fraction organic carbonKoc organic partition coefficient

Min800.00033000.215000.00015

Max1200.000830000.3517500.00520

Mean1000.0005510000.2516500.0025510

Stdv11.550.00014437500.051000.0014153

Unitsmm/mm/yr-kg/m3

-m3/kg

Page 13: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Results

• The most optimistic outcome would be that the plume takes ___ years to reach the well

• But, in the worst case, it could take __ years to reach the well

Answers: 233800 years, 31.6 years

Page 14: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

[31.6, 233800] years

• Is ‘6’ the last decimal digit of the lower bound?

• Did you check that the units balance?

• Did you include units in the answer?

Page 15: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

How to understand this result

• Highly reliable result, given the assumptions– Can’t get any worse

• Represents parametric uncertainty– Neglects (possibly big) model uncertainty

• Expresses only best and worst cases– How likely is 32 years? 50 years? 100 years?

Page 16: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Lobascio’s original formulation

Kd= foc Koc = [ 0.0005, 0.1] m3 kg-1

R = 1 + BD Kd / n = [ 3.143, 876] V = K i / (n R) = [ 0.000293, 3.82] m yr-1

T = L/V = [ 20.95, 408800] yr

Quickest plume reaches the well = 20.95 yrLongest plume reaches the well = 408,800 yr

What explains the difference? (hint: n is repeated above)

Page 17: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Repeated parameters

a = [1,2]b = [2,3]c = [2, 5]

z = a × (b + c) zz = a × b + a × c

b + c = [0, 8] a × b = [2, 6]z = [0, 16] a × c = [4, 10] zz = [2, 16]

inflated uncertainty

Page 18: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

What to do about repeated parameters

• Always rigorous, but maybe not best possible when uncertain parameters are repeated

• Inconsequential if all are non-negative and all operations are increasing (+, ×, but not – or ÷)

• Use cancellation to reduce repetitions, e.g., caia/m + cwiw/m + cdid/m = (caia + cwiw + cdid)/m

• Cancellation not always possible, e.g., (a + b) / (a + c) = ??

Page 19: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

If you can’t cancel

• Use tricks with algebrae.g., a² + a = (a +½)² – ¼

• Employ subinterval reconstitutionA brute-force and computationally intensive strategy

Workable if there aren’t too many repeated parameters

• Live with the suboptimalityDecisions may not require perfect precision

Page 20: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

TricksBasic identities

u + 0 = u

u – 0 = u

0 – u = –u

u 0 = 0

u 1 = u

u / 1 = u

u0 = 1

u1 = u

u & 1 = u

u | 1 = 1

u & 0 = 0

u | 0 = u

u & u = u

u | u = u

u & not(u) = 0

u | not(u) = 1

(u&a) | (u&b) = u&(a | b)

(u | a)&(u | b) = u | (a&b)

etc.

One repetitionu + u = 2uu – u = 0u u = u2

u / u = 1(1+u) / u = 1/u + 1(1+u)/(1–u) = (1/tan(acos(u)/2))2

au + bu = u(a + b)au – bu = u(a – b)a/u + b/u = (a + b) / ua/u – b/u = (a – b) / uu/a + u/b = u(b + a)/(ab)u/a – u/b = u(b – a)/(ab)aub + cub = (a + c) ub

aub cud = a c u(b + d)

au bu = exp(u (ln(a) + ln(b)))u2 + u = (u + ½)2 – ¼u2 – u = –¼ sin(2 asin(sqrt(u)))2

u2 + au = (u + a/(2))2 – a2/4etc.

u, v, etc. represent the uncertain numbers a, b, etc. represent arbitrary expressions

Two repetitionsu + v – uv = 1 – (1 – u) (1 – v)(u + v) / (1 – uv) = tan(atan(u) + atan(v))(u – v) / (1 + uv) = tan(atan(u) – atan(v))(1 + uv) / (u – v) = 1 / tan(atan(u) – atan(v))(1 – uv) / (u + v) = 1 / tan(atan(u) + atan(v))(uv – 1) / (u + v) = –1 / tan(atan(u) + atan(v))u sqrt(1 – v2) + v sqrt(1 – u2) = sin(asin(u) + asin(v))u sqrt(1 – v2) – v sqrt(1 – u2) = sin(asin(u) – asin(v))u v + sqrt(1 – u2) sqrt(1 – v2) = cos(acos(u) – acos(v))u v – sqrt((1 – u2) (1 – v2)) = cos(acos(u) + acos(v))u v – sqrt(1 – u2 – v2 + u2 v2) = cos(acos(u) + acos(v))sin(u) sqrt(1 – sin(v)2) + sin(v) sqrt(1 – sin(u)2) = sin(u + v)cos(u) cos(v) – sin(u) sin(v) = cos(u + v)sin(u) cos(v) – cos(u) sin(v) = sin(u – v)sqrt((1 + u) / (1 – u)) = 1 / tan(acos(u)/2)etc.

Page 21: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Subinterval reconstitution

• Partition each repeated interval into subintervals• Compute the function for every subinterval• The union of all the results contains the true range

i j k

kji zyxwvufzyxwvuf ),...,,,,...,,(...),...,,,...,,(

i j k

kji wwvvuu ...;;;

where u, v, …, w are repeated intervals and x, y,…, z are other interval and scalar inputs, and

Page 22: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Example: (a + b)a, a = [0.1, 1], b = [0,1]

Partition the repeated uncertain a:ai = [(i 1)w/m + a, iw/m + a] where i = 1,2,…, m, and m is the number of subintervals, w is thewidth of a, and a is its lower bound

m U(ai+b)ai

1 [ 0.1, 2]2 [ 0.282, 2]3 [ 0.398, 2]4 [ 0.473, 2]5 [ 0.525, 2]10 [ 0.624, 2]100 [ 0.686, 2]1,000 [ 0.692, 2]10,000 [ 0.692, 2]a

0 0.5 1

m U(ai+b)ai

1 [ 0.1, 2]2 [ 0.282, 2]3 [ 0.398, 2]4 [ 0.473, 2]5 [ 0.525, 2]10 [ 0.624, 2]100 [ 0.686, 2]1,000 [ 0.692, 2]10,000 [ 0.692, 2]

Page 23: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Cauchy-deviate method

• Propagates intervals through black-box model– Don’t need to know, but have to be able to query it

• “Sample” from around interval – Points not necessarily inside the interval!

• Scale results to get an asymptotically correct estimate of the interval uncertainty of the output

(Trejo and Kreinovich 2001)

Page 24: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Cauchy-deviate method

• Depends on the number of samples, not inputs– Works just as well for 2000 variables as 20– Similar in performance to Monte Carlo

• Need about 200 samples to obtain 20% relative accuracy of half-width of output range– With fewer samples, we’d get lower accuracy, but

we can compensate for this by scaling by N, which works under the linearity assumption

Page 25: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Limitations of the method

• Asymptotically correct, but not rigorous

• Intervals narrow relative to the nonlinearity– Function almost linear OR uncertainties small– Could combine with subinterval reconstitution

• Most efficient when dimensionality is high

• Only handles interval uncertainty

Page 26: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Dependence among intervals

Page 27: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Dependence in interval analysis

• Interval analysis automatically accounts for all possible dependencies

• Information about dependencies isn’t usually used to tighten results, but it could be

Page 28: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Dependencies

• Any restriction on the possible pairings of inputs (any subset of the unit square) May also require each value of u to match with at least v, and vice versa

• The null restriction is the full unit square“Nondependence” or “noninteractivity” rather than “independence”

• D denotes the set of all possible dependencies Set of all subsets (power set) of the unit square D = 2[0,1][0,1]

Page 29: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Three special cases

0 u 1 0 u 1 0 u 1

1

v

0

1

v

0

1

v

0

Perfect(comonotonic)

Nondependent(the Fréchet case)

Opposite(countermonotonic)

Page 30: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Correlation

• A model of dependence that’s parameterized by a (scalar) value called the “correlation coefficient”

: [1, +1] D = 2[0,1][0,1]

• A correlation model is called “complete” if

(1) = , (0) = , (+1) =

Page 31: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

r = 1 r = 0 r = +1

Corner-shaving dependence

D(r) = { (u,v) : max(0, ur, u1+r) v min(1, u+1r, u+2+r)} u [0,1], v [0,1] 

f (A, B) = { c : c = f (u (a2 – a1) + a1, v (b2 – b1) + b1), (u,v) D }A+B = [env(w(A, r)+b1, a1+w(B,r)), env(a2+w(B,1+r),w(A,1+r)+b2)]

a1 if p < 0w([a1,a2], p) = a2 if 1 < p p(a2a1)+a1 otherwise

Page 32: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Other complete correlation familiesr = 1 r = 0 r = +1

Page 33: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Elliptic dependence

• Not complete (because r = 0 isn’t nondependence)

A+B = [d, d] + (a1 + a2 + b1 + b2) / 2

d = (d12 + d2

2 + r d1 d2), d1 = (a2 a1)/2, d2 = (b2 b1)/2

r = 1 r = 0 r = +1

A+B = [d, d] + (a1 + a2 + b1 + b2) / 2

d = (d12 + d2

2 + r d1 d2), d1 = (a2 a1)/2, d2 = (b2 b1)/2

Page 34: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

[ 5, 14] Perfect

[ 8, 11] Opposite

[ 7.1, 11.9] Corner-shaving (r = 0.7)

[ 7.27, 11.73] Elliptic (r = 0.7)

[ 5, 14] Upper, left

[ 5, 11] Lower, left

[ 8, 14] Upper, right

[ 5, 14] Lower, right

[ 6.5, 12.5] Diamond

[ 5, 14] Nondependent

A + B

A = [2,5]B = [3,9]

Page 35: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Implications

• Intervals aren’t more widely used is because they’re thought to be (and often are) too wide to be helpful

• They can be much tighter if we used all available information, such as about intervariable dependence

• (Other information, like about moments, could also be used to tighten interval calculations)

Page 36: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Computing

• Sequence of binary operations– Need to deduce dependencies of intermediate results

with each other and the original inputs– Different calculation order can give different results

(which should be intersected)

• Do all at once in one multivariate calculation – Can be much more difficult computationally– Can produce much better tightening

Page 37: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Specifying input intervals

Page 38: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Interval uncertainty

• Statisticians often ignore this uncertainty

• “Interval uncertainty doesn’t exist in real life”(Tony O’Hagan et

al.)

Hammer salesmen saying screws don’t exist?

Page 39: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

When do intervals arise?

• Periodic observationsWhen did the fish in my aquarium die during the night?

• Plus-or-minus measurement uncertaintiesCoarse measurements, measurements from digital readouts

• Non-detects and data censoringChemical detection limits, studies prematurely terminated

• Privacy requirementsEpidemiological or medical information, census data

• Theoretical constraintsConcentrations, solubilities, probabilities, survival rates

• Bounding studies Presumed or hypothetical limits in what-if calculations

Page 40: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Ways to characterize intervals

• Theoretical constraints

• Modeled from other intervals

• Expert assertions

• Discounting (widening) intervals (Shlyakhter)

• Confidence procedures (Grosof)– But 95% confidence isn’t the same as surety– Use in interval calculations requires an assumption

Page 41: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Problems with confidence intervals

• Cannot be combined in arithmetic or logical operations without an assumption

• Don’t measure epistemic belief anyway

Example (Walley): For instance, a 90% confidence interval could have zero chance of containing the value. For example, suppose X ~ normal(, 1), where 0 < . If the sample mean happens by chance to be 1.3, the 90% confidence interval on the mean is the empty set (Walley, P. 1991, Statistical Reasoning with Imprecise Probabilities, Chapman & Hall, page 379).

Page 42: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Why we have to be careful

• Interval analysis yields contingent results• Results are contingent on assumptions that model inputs are within

their respective intervals• But all analysis results are contingent on similar assumptions that the

models they came from are true

• Naïve elicitation has big problems• Intervals are usually unrealistically narrow• People make incoherent statements

• Can’t mix together different kinds• Not clear how to translate data into intervals

Page 43: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Determining endpoints

• The largest observed may not be the largest possible (and it usually isn’t)

• Sampling theory >> theory of extremes

• Rigor of analysis is contingent on inputs

• If you’re nervous, just widen the bounds

Page 44: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Point sample data

Range (envelope)

Extreme value model

Tolerance interval

Confidence interval

Central value and width

P-box

Level cutcut

Model

Output rangesimulation, etc.Plus-minus interval

Shlyakhter widening

Envelope

Interval function

Distribution

Support

Percentile range

Trejo-KreinovichCauchy deviates

Tolerance solutionbackcalculation

Certain and tenable ranges

Intersection

Credibility interval

Prediction interval

Percentile range

Support

Page 45: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Eliciting dependence

• As hard as getting intervals (maybe a bit worse)

• Theoretical or “physics-based” arguments

• Inference from empirical data – Risk of loss of rigor at this step (just as there is

when we try to infer intervals from data)

Page 46: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Updating

Page 47: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Aggregation (updating)

• How do you combine different sources?

• If you trust them all, take the intersection– [max(x1, y1, z1, …), min(x2, y2, z2, …)]– What if there is no intersection (right<left)?

• If you’re not sure which is right, use the envelope– [min(x1, y1, z1, …), max(x2, y2, z2, …)]– But are you sure this is wide enough?

Page 48: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Example

• Suppose we have two rigorous interval estimates of the same quantity: [1,7] & [4,10]

• Their intersection [4,7] is also a rigorous interval for the quantity

AB

0 5 10

Page 49: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Constraint analysis (updating)

• Using knowledge of how variables are related to tighten their estimates

• Removes internal inconsistency and explicates unrecognized knowledge

• Also called ‘constraint updating’ or ‘editing’

• Also called ‘natural extension’

Page 50: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Example

• Suppose we know

W = [23, 33] m

H = [112, 150] m

A = [2000, 3200] m2

• Does knowing W H = A let us to say any more?

Page 51: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Answer

• Yes! We can infer that

W = [23, 28.57]

H = [112, 139.13]

A = [2576, 3200]

• The formulas are just W = intersect(W, A/H), etc.

To get the largest possible W, for instance, let A be as large as possible and H as small as possible, and solve for W =A/H.

Page 52: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Updating with p-boxes

2000 3000 40000

1

20 30 400

1

120 140 1600

1 A W H

Page 53: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

2000 3000 40000

1

20 30 400

1

120 140 1600

1

Answers

intersect(W, A/H) intersect(H, A/W) intersect(A, WH)

A W H

Page 54: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Bayesian strategy

20003200

])3200,2000[(

112150

])150,112[(

2333

])33,23[(),,Pr(

AIHIWI

AHW

)(),,|( HWAAHWHWAL

),,Pr()()|,,( AHWHWAHWAAHWf

Prior

Likelihood

Posterior

Page 55: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Bayes’ rule

• Concentrates mass onto the manifold of feasible combinations of W, H, and A

• Answers have the same supports as intervals

• Computationally complex• Needs specification of priors• Yields distributions that are not justified

(coming from the choice of priors)• Expresses less uncertainty than is present

Page 56: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Backcalculation

Page 57: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Backcalculation

• Needed for cleanup and remediation planning

• Untangles an equation in uncertain numbers when we know all but one of the variables

• For instance, backcalculation finds B such that A+B = C, from estimates for A and C

Page 58: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Can’t just invert the equation

Dose = Concentration × Intake

Concentration = Dose / Intake

When concentration is put back into the forward equation, the resulting dose is wider than planned

prescribed knownunknown

Page 59: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Exampledose = [0, 2] milligram per kilogramintake = [1, 2.5] litermass = [60, 96] kilogram

conc = dose * mass / intake [ 0, 192] milligram liter-1

dose = conc * intake / mass [ 0, 8] milligram kilogram-1

Doses four times larger than tolerable levels we planned

Page 60: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Untangling backcalculations

• Solving for B given A + B = C

B = backcalc(A, C) = [C1 A1, C2 A2]

• Solving for B given A B = C

B = factor(A, C) = [C1 / A1, C2 / A2]

• Sometimes called “tolerance solutions”

Page 61: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

1

2

3

4

5

6

1 1.5 2A

Bsh

ell

Kernal versus shell

Given A [1,2] C [2,6] C A BThere are two different ways to solve for B

kern

el

Shell (united solution)B C A [0,5]

Kernel (tolerance solution)B backcalc(A,C) [1,4]

Page 62: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

When you need for

A + B C

A – B C

A B C

A / B C

A ^ B C

2A C

A² C

And you have estimates for

A, BA, CB ,CA, BA, CB ,CA, BA, CB ,CA, BA, CB ,CA, BA, CB ,C

ACAC

Use this formulato find the unknownC = A + BB = backcalc(A,C)A = backcalc (B,C)C = A – BB = –backcalc(A,C)A = backcalc(–B,C)C = A * BB = factor(A,C)A = factor(B,C)C = A / BB = 1/factor(A,C)A = factor(1/B,C)C = A ^ BB = factor(log A, log C)A = exp(factor(B, log C))C = 2 * AA = C / 2C = A ^ 2A = sqrt(C)

Page 63: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Interval algebra

• Commutativity a+b=b+a, a×b=b×a• Associativity (a+b)+c=a+(b+c), (a×b)×c=a×(b×c)• Neutral elements a+0=0+a=a, a×1=1×a=a

• Subdistributivity a×(b+c) a×b+a×c• Subcancellation a a+bb, a a×b/b• No inverse elements a+( a) 0, a×(1/a) 1

Page 64: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Conclusions

• Interval analysis is a worst case analysis (that also includes the best case)

• Repeated uncertain parameters can cause unnecessary inflation of uncertainy

• Results will always be rigorous, but might not be best possible

• Moving a uncertain parameter to the other side of an equal sign often requires backcalculation

Page 65: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Exercises

1. Do the inputs in the travel time example seem dependent?2. What does subinterval reconstitution with m=100 on the

original Lobascio formulation give for the travel time?3. What contaminant concentrations C in water will lead to

doses D no larger than 6 mg per kg per day if it comes from both drinking and eating fish as D = (Iwater C) / W + (Ifish B C) / W, where

Iwater = [1.5, 2.5] liters per day // water intake

Ifish = [0, 8] g per day // dietary ingestion of fishB = [0.9, 2.1] liters per mg // bioaccumulation factorW = [60, 90] kg // receptor biomass

How do you check the solution?4. Is there a Bayesian analog of backcalculation?

////////////////////////////////////////////////////////////////////////////////////////////////////////////////////Answer to #1

L source-receptor distancei hydraulic gradientK hydraulic conductivityn effective soil porosityBD soil bulk densityfoc fraction organic carbonKoc organic partition coefficient

Porosity and conductivity would likely be positively relatedPorosity and density would likely be inversely relatedThus, conductivity and density should be inversely related tooFraction organic carbon and the partition coefficient would likely be positively related

////////////////////////////////////////////////////////////////////////////////////////////////////////////////////Answer to #2

L = [80,120] m // source-receptor distancei = [0.0003,0.0008] m per m // hydraulic gradientK = [300,3000] m per yr // hydraulic conductivityn = [0.2,0.35] // effective soil porosityBD = [1500,1750] kg per m{3} // soil bulk densityfoc = [0.0001,0.005] // fraction organic carbonKoc = [5 ,20] m{3} per kg // organic partition coefficient Kd = foc * KocR = 1 + BD * Kd / n V = K * i / (n * R)T = L/V T [ 20.952380952, 408800] yr

func TT() return L/(K * i / ($1 * (1 + BD * foc * Koc / $1)))

b = left(n)e = TT(b)w = width(n)/100for k=1 to 100 do begin s = [ (k-1), k] * w + b e = env(e, TT(s)) ende [ 31.480562448, 235352] yr

////////////////////////////////////////////////////////////////////////////////////////////////////////////////////Answer to #3

Iwater = [1.5, 2.5] liters per day // water intakeIfish = [0, 8] g per day // dietary ingestion of fish tissueB = [0.9, 2.1] liters per g // bioaccumulation factorW = [60, 90] kg // receptor biomass

D = [0, 6] mg per kg per day // tolerable dose

// forward equation// D = (Iwater * C) / W + (Ifish * B * C) / W

// condense repeated parameters// D = C * (Iwater + Ifish * B) / W

// solve for C by factoring the quotient (Iwater+Ifish*B)/W out of DC = factor((Iwater + Ifish * B) / W, D) C [ 0, 18.652849741] mg liters-1

// put C back into original forward equation (and ensure correct units)d = (Iwater * C) / W + (Ifish * C * B) / W + 0 mg per kg per day

d [ 0, 6] mg kg-1 day-1

Page 66: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

(break)

Page 67: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Interval probability

Page 68: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Outline for Interval Probability

• Logical operations under independence• Dependence among events

– Perfect and opposite dependence– Frank and Pearson correlation– Fréchet inequalities– Signed dependence

• Example: rupture of a pump system• Exercises• Conclusions• References

Page 69: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Interval arithmetic for probabilities

x + y = [x1, x2] + [y1, y2] = [x1 + y1, x2 + y2]

x y = [x1, x2] [y1, y2] = [x1 y2, x2 y1]

x y = [x1, x2] [y1, y2] = [x1 y1, x2 y2]

x y = [x1, x2] [y1, y2] = [x1 y2, x2 y1]

Examples: x = [0.3, 0.5]y = [0.1, 0.2]x + y = [0.4, 0.7]x y = [0.1, 0.4]

Rules are simpler because intervals confined to [0,1]

Page 70: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Probabilistic logic

• Conjunctions (and)

• Disjunctions (or)

• Negations (not)

• Exclusive disjunctions (xor)

• Modus ponens (if-then)

• etc.

Page 71: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Conjunction (and)

P(A |&| B) = P(A) P(B)

Example: P(A) = [0.3, 0.5]

P(B) = [0.1, 0.2]

A and B are independent

P(A |&| B) = [0.03, 0.1]

Page 72: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Disjunction (or)

P(A || B) = = 1 (1 P(A))(1 P(B))

Example: P(A) = [0.3, 0.5]

P(B) = [0.1, 0.2]

A and B are independent

P(A || B) = [0.37, 0.6]

Page 73: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Disjunction (or)

P(A || B) = P(A) + P(B) P(A) P(B)

Example: P(A) = [0.3, 0.5]

P(B) = [0.1, 0.2]

A and B are independent

P(A || B) = [0.3, 0.67]

repeatedparameters!

Page 74: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Negation (not)

P(not A) = 1 P(A)

Example: P(A) = [0.3, 0.5]

P(not A) = [0.5, 0.7]

Page 75: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

1 2 3 4 5

Dependence

Independent

Probabilities are depicted hereas the areas in Venn diagrams

Perfect Opposite

Page 76: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Perfect dependence

P(A /&/ B) = min(P(A), P(B)) P(A // B) = max(P(A), P(B))

Examples: P(A) = [0.3, 0.5]P(B) = [0.1, 0.2]A and B are perfectly dependentP(A /&/ B) = [0.1, 0.2]P(A // B) = [0.3, 0.5]

Page 77: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Opposite dependence

P(A \&\ B) = max(P(A) + P(B) 1, 0)

P(A \\ B) = min(1, P(A) + P(B))

Examples: P(A) = [0.3, 0.5]

P(B) = [0.1, 0.2]

A and B are oppositely dependent

P(A \&\ B) = 0

P(A \\ B) = [0.4, 0.7]certain

Page 78: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Correlated events (Frank)

where r is the correlation, s = tan((1r)/4), a = P(A), b = P(B)

otherwise

1 if

0 if

1 if

)1/()1)(1(1log

)0,1max(

),min(

),,(and)&P(

s

Frank

r

r

r

sss

ba

ab

ba

rbaBA

ba

)&P( BAr

otherwise.

1 if

0 if

1 if

)1/()1)(1(1log-1

)1,min(

)1)(1(1

),max(

),,(or) P(

11s

Frank

r

r

r

sss

ba

ba

ba

rbaBA

ba

) ∨P( BA r

Page 79: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

0

0.1

0.2

0.3

0.4

0.5

0.6

1 0.75 0.5 0.25 0 0.25 0.5 0.75 1

Frank correlation coefficient, r

Pro

babi

lity

orFrank(0.29, 0.22, r)

andFrank(0.29, 0.22, r)

Examples of Frank correlations

Page 80: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Other correlation models

• Pearson correlation (on indicator functions)

– doesn’t allow all possible correlation coefficients– permits negative and larger-than-one probabilities

• But any complete copula would work fine– Frank, Clayton, Mardia, etc.– Nelsen (1999) is the essential reference on copulas

Page 81: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

P(A & B) = [max(0,P(A)+P(B)–1), min(P(A), P(B))]

P(A or B) = [max(P(A), P(B)), min(1, P(A)+ P(B))]

• Makes no assumption about the dependence

• Rigorous (guaranteed to enclose true value)

• Best possible (cannot be any tighter)

Fréchet case

Page 82: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

The proofs are elementaryP(A  B) = P(A) + P(B) P(A & B) implies P(A) + P(B) P(A  B) = P(A & B).

P(A  B) 1, since probabilities are no bigger than 1, so P(A) + P(B) 1 P(A & B).

0 P(A & B), since probabilities are positive, so max(0, P(A) + P(B) 1) P(A & B).

This gives the lower bound on the conjunction.

To get the upper bound, recall that P(A & B) = P(A|B) P(B) = P(B|A) P(A).

P(A|B) 1 and P(B|A) 1, so P(A & B) P(A) and P(A & B) P(B).

Therefore, P(A & B) min(P(A), P(B)), which is the upper bound.

The best-possible nature of these bounds follows from observing that

they are realized by some dependency between the events A and B.

Comparable bounds on the disjunction are similarly derived.

Page 83: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Fréchet examples

Examples: P(A) = [0.3, 0.5]P(B) = [0.1, 0.2]P(A & B) = [0, 0.2]P(A B) = [0.3, 0.7]

P(C) = 0.29P(D) = 0.22P(C & D) = [0, 0.22]P(C D) = [0.29, 0.51]

}certain

}uncertain

Page 84: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Sign of the dependence

P(A &+ B) = [ab, min(a,b)]

P(A + B) = [1(1a)(1b), max(a, b)]

P(A & B) = [max(a+b1, 0), ab]

P(A B) = [1(1a)(1b), min(1, a+b)]

where a = P(A), b = P(B)

Page 85: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

RelayK2

Pressureswitch S

Timerrelay R

Pump

Motor

RelayK1

Fromreservoir

Pressure tank T Outletvalve

SwitchS1

S tartupSw itc h S1 c lo sedR e lay K1 c lo sedR e lay K2 c lo sedT im e r re la y c lo sedP re ssu re sw it ch c lo sed

Start pum pR e lay K2 c lo sedT im e r re la y s ta r tsP re ssu re sw it ch c lo sed

ReadyS w itch S1 o p enR e lay K1 c lo sedR ela y K2 o p enT im e r re la y c lo sedPre ssure sw itch o p en

System offS w itch S1 o p enR ela y K1 o p enR ela y K2 o p enT im e r re la y c lo sedP re ssu re sw it ch c lo sed

Shut downR ela y K1 o p enR ela y K2 o p enT im er re la y o p enP re ssu re sw it ch c lo sed

P um pingS w itch S1 o p enR e lay K1 c lo sedR e lay K2 c lo sedT im e r re la y c lo sedP re ssu re sw it ch c lo sed

S top pum pR ela y K2 o p enT im er re lay re se tsPre ssure sw itch o p en

Example: pump system

What’s the chance thetank ruptures under pumping?

Page 86: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

E1 = “tank rupturing under pressurization”

Fault tree

Vesely et al. (1981)

E1

T

K2

E2

E3

S

E4

S1

E5

K1

or and

or

or

or

R

Page 87: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Boolean expression

E1 = T (K2 (S & (S1 (K1 R))))

ComponentPressure tank T Relay K2 Pressure switch S Relay K1 Timer relay R On-switch S1 

Probability5 106

3 105

1 104

3 105

1 104

3 105

Page 88: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Different dependency models

Vesely et al. (all variables precise, independent)E1 = T || (K2 || (S |&| (S1 || (K1 || R))))

Mixed dependencies

E1 = T || (K2 (S &r (S1 || (K1 // R))))Frank to Fréchet

E1 = T || (K2 (S & (S1 || (K1 // R))))All Fréchet

E1 = T (K2 (S & (S1 (K1 R))))

Page 89: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Interval probabilities

ComponentPressure tank T Relay K2 Pressure switch S Relay K1 Timer relay R On-switch S1 

Probability interval[4.5 106, 5.5 106][2.5 105, 3.5 105][0.5 104, 1.5 104][2.5 105, 3.5 105][0.5 104, 1.5 104][2.5 105, 3.5 105]

Page 90: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

105 104 103

Probability of tank rupturing under pumping

Mixed dependencies

Frank to Fréchet

All Fréchet

Mixed dependence with intervals

Vesely et al. (all independent)

Results

3.5105

[3.499105, 3.504105]

[3.50105, 1.35104]

[3105, 1.4104]

[2.9105, 4.1105]

Page 91: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Strategies to handle repetitions

• Interval analysis is always rigorous, but maybe not best possible when parameters are repeated

• Use cancellation to reduce repetitions, e.g.,(A & B) (A & C) (A & D) = A & (B C D)

• When cancellation is not possible, mathematical programming may be necessary to get best possible result

Page 92: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Subtle dependencies

• May also require mathematical programming to obtain the best possible result

• But rigorous bounds are always easy to get with the artful use of the Fréchet case

Page 93: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Advances in practices

“It practice, independence among events and event probabilities is often assumed for the sake of convenience because it makes otherwise intractable problems easier to solve.” (Vick 2002, page 412)

“…Independent and conditional relationships [among events] could not be established without the subjective elements of one’s state of knowledge, making it otherwise all but impossible to apply the [probability] calculus to most situations of practical importance.” (ibid.)

Page 94: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Conclusions

• Easy to compute rigorous bounds

• Mathematical programming may be needed to get answers that are also best possible

• Rigor of analysis is contingent on inputs

• If you’re nervous, just widen the bounds

Page 95: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Exercises

5. Calculate the probability of tank rupture under pumping that assumes the interval inputs and makes no assumption about the dependencies among the events.

6. Develop an fault tree for establishment of snake populations on a Hawaiian island (or a star exploding).

7. Compute the probability of the conjunction of two events having probabilities 0.29 and 0.22, assuming a Pearson correlation of 1.0. Compare the result to the Fréchet range for such probabilities. What’s going on?

8. Derive an algorithm to compute the probability that n of k events occur, given intervals for the probability of each event, assuming they’re independent. Derive an analogous algorithm for the Fréchet case.

Vesely et al. (all independent)3.5105

[3.499105, 3.504105]

[3.50105, 1.35104]

[3105, 1.4104]

[2.9105, 4.1105]

[ 2.5105, 1.905 104]

Page 96: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Snakes on a plane (or boat)Introduction

Stowaway on plane OR Stowaway on ship OR Arrives on natural raft OR Brought as pet OR Brought for zoo

AND Escape and ReproductionPregnant snake escapes OR Male and female escape AND Successful reproduction OR Female escapes AND Parthenogenic reproduction

AND Survives predationDomestic and feral dogs ANDFeral cats ANDWild pigs

AND Prey is accessibleGround nesting bird eggsOR Rodents

AND Survives eradication programEludes detection by biologistsOR Reproduction too quickOR Dispersal too quick

Page 97: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Overall conclusions

Page 98: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Why bounding?

• Often sufficient to specify a decision • Possible even when estimates are impossible• Usually easy to compute and simple to combine• Rigorous, rather than an approximation

(after N.C. Rowe 1988)

Page 99: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Reasons to use interval analysis

• Requires very little data• Applicable to all kinds of uncertainty• Can be comprehensive• Fast and easy to compute answers• Conservative when correlations unknown • Can be made “best possible”• Backcalculations easy• Updating relatively easy

Page 100: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Reasons not to use it

• Same thing as worst case analysis

• Doesn't say how likely extreme event is

• Repeated parameters are cumbersome

• Not optimal when there’s a lot of data

• Can't use distribution information

• Doesn't use correlation information

Page 101: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Interval (worst case ) analysisHow?

– bound inputs, a = [a1, a2], where a1 a2

– addition: [a1, a2] + [b1, b2] = [a1+b1, a2+b2]– subtraction: [a1, a2] – [b1, b2] = [a1–b2, a2–b1]– multiplication, division, etc. are a little more complex

Why?– natural for scientists and easy to explain to others– works no matter where uncertainty comes from

Why not?– paradoxical: can’t give exact value but can give exact bounds– ranges could grow quickly, yielding very wide results– doesn’t give probabilities of extreme outcomes (tail risks)

Page 102: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

Interval probabilityHow?

– bound event probabilities, p = [p1, p2], where 0 p1 p2 1

– evaluate event trees as composition of ANDs, ORs, etc.

– standard probabilistic rules if events are independent

– Fréchet rules if their dependence is unknown

– other dependency relations can also be represented

Why?

– can capture incertitude about event probabilities

Why not?

– paradoxical: can’t give exact value but can give exact bounds

– ranges can grow quickly, especially without independence

Page 103: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

ReferencesDwyer, P. 1951. Linear Computations. John Wiley & Sons, New York.Ferson, S. 2002. RAMAS Risk Calc 4.0: Risk Assessment with Uncertain Numbers. Lewis Publishers, Boca Raton.Grosof, B.N. 1986. An inequality paradigm for probabilistic knowledge: the logic of conditional probability intervals.

Uncertainty in Artificial Intelligence. L.N. Kanal and J.F. Lemmer (eds.), Elsevier Science Publishers, Amsterdam.Hailperin, T. 1986. Boole’s Logic and Probability. North-Holland, Amsterdam.Kyburg, H.E., Jr. 1998. “Interval Valued Probabilities,” http://ippserv.rug.ac.be/documentation/interval_prob/interval_prob.html,

The Imprecise Probabilities Project, edited by G. de Cooman and P. Walley, http://ippserv.rug.ac.be/home/ipp.html.Lobascio, M.C. 1993. Uncertainty analysis tools for environmental modeling: application of Crystal Ball® to predict

groundwater plume traveling times. ENVIRONews 1: 6-10.Loui, R.P. 1986. Interval based decisions for reasoning systems. Uncertainty in Artificial Intelligence. L.N. Kanal and

J.F. Lemmer (eds.), Elsevier Science Publishers, Amsterdam.Moore, R.E. 1966. Interval Analysis. Prentice-Hall, Englewood Cliffs, New Jersey.Moore, R. 1979. Methods and Applications of Interval Analysis. SIAM, Philadelphia.Rowe, N.C. 1988. Absolute bounds on the mean and standard deviation of transformed data for constant-sign-derivative

transformations. SIAM Journal of Scientific Statistical Computing 9: 1098–1113. Shlyakhter A. 1994. Improved framework for uncertainty analysis: accounting for unsuspected errors. Risk Analysis

14(4):441-447.Tessem, B. 1992. Interval probability propagation. International Journal of Approximate Reasoning 7: 95-120.Trejo, R. and V. Kreinovich. 2001. Error estimations for indirect measurements: randomized vs. deterministic

algorithms for ‘black-box’ programs. Handbook on Randomized Computing, S. Rajasekaran, P. Pardalos, J. Reif, and J. Rolim (eds.), Kluwer, 673–729. http://www.cs.utep.edu/vladik/2000/tr00-17.pdf

Vesely, W.E., F.F. Goldberg, N.H. Roberts, D.F. Haasl. 1981. Fault Tree Handbook. Nuclear Regulatory Commission, Washington, DC.

Vick, S.G. 2002. Degrees of Belief: Subjective Probability and Engineering Judgment. ASCE Press, Reston, Virginia.

Page 104: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

End

Page 105: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.
Page 106: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165.

For next time

• Do exercise #1 and either #2 or #3

• Try to do one of the exercises #5, #6 or #7