2. Overview
• Outputs
• Tools and concepts
• Data sets used
• Methods
• Results
• Case Study
• Conclusions
• Future Work
3. Uncertainty in Weather Forecasts
It is being increasingly recognized that the
uncertainty in weather forecasts should be
quantified and furnished to users along with the
single value forecasts usually provided.
MDL’s goal is to provide probabilistic
guidance for all surface weather
variables in gridded form in the
National Digital Guidance Database
(NDGD).
4. Outputs
How do we provide probabilistic forecasts to our
customers and partners?
• Fit a parametric distribution (e. g., Normal).
– Economical, but restrictive
• Enumerate Probability Density Function (PDF) or
Cumulative Distribution Function (CDF) by computing
probabilities for chosen values of the weather element.
– Values must “work” everywhere
• Enumerate Quantile Function (QF) by
giving values of the weather element for
chosen exceedence probabilities.
5. Sample Forecast as Quantile Function
25
30
35
40
45
50
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Probability
Temperature
72-h T Fcst KBWI 12/14/2004
6. Sample Forecast as Quantile Function
25
30
35
40
45
50
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Probability
Temperature
One percent chance of temperature
below 29.8 degrees F.
20% chance of temperature
below 35.2 degrees F.
Median of the distribution
38.3 degrees F.
50% Confidence Interval
(35.8, 40.7) degrees F.
90% Confidence Interval
(32.2,44.3) degrees F.
72-h T Fcst KBWI 12/14/2004
Chance of temperature below
40.0 degrees F is 67.9%.
7. Sample Forecast as Probability Density Function
0
0.02
0.04
0.06
0.08
0.1
0.12
25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50
Temperature
Probability
Density
8. Tools and Concepts
We have combined the following tools in a
variety of ways to take advantage of linear
regression and ensemble modeling of the
atmosphere.
– Error estimation in linear regression
– Kernel Density Fitting (Estimation; KDE)
A brief overview of these tools follows.
9. Error Estimation in Linear Regression
• The linear regression theory used to
produce MOS guidance forecasts includes
error estimation.
• The Confidence Interval quantifies
uncertainty in the position of the regression
line.
• The Prediction Interval quantifies
uncertainty in predictions made using the
regression line.
The prediction interval can be used to
estimate uncertainty each time a MOS
equation is used to make a forecast.
10. Estimated Variance of a Single New
Independent Value
• Estimated variance
• Where
2
2
)
(
2 1
1
ˆ
X
X
X
X
n
MSE
Y
s
i
h
new
h
2
ˆ 2
n
Y
Y
MSE i
i
11. Computing the Prediction Interval
The prediction bounds for a new prediction is
where
t(1-α/2;n-2) is the t distribution n-2 degrees of freedom at the 1-α
(two-tailed) level of significance, and
s(Ŷh(new)) can be approximated by
where
s2 is variance of the predictand
r2 is the reduction of variance
)
(
)
(
ˆ
2
;
2
/
1
ˆ
new
h
new
h Y
s
n
t
Y
2
2
1 r
s
12. Multiple Regression (3-predictor case)
n
n
y
y
y
y
y
4
3
2
1
1
Y
3
2
1
43
42
41
33
32
31
23
22
21
13
12
11
4
1
1
1
1
1
n
n
n
n
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
X
3
2
1
0
1
4
a
a
a
a
A
Predictand
Vector
3-predictor
Matrix
Coefficient
Vector
13. Multiple Regression, Continued
Error bounds can be put around the new value of Y with
where
– s2 is the variance of the predictands,
– R2 is the reduction of variance,
– X’ is the matrix transpose of X, and
– ()-1 indicates the matrix inverse.
2
/
1
1
4
1
4
4
4
1
2
2
)
( 1
1
ˆ x
X
X
x
n
new
h R
s
Y
16. Advantages of MOS Techniques
for Assessing Uncertainty
• Single valued forecasts and probability
distributions come from a single consistent
source.
• Longer development sample can better
model climatological variability.
• Least squares technique is effective at
producing reliable distributions.
17. Kernel Density Fitting
• Used to estimate the Probability Density
Function (PDF) of a random variable, given a
sample of its population.
• A kernel function is centered at each data point.
• The kernels are then summed to generate a
PDF.
• Various kernel functions
can be used. Smooth,
unimodal functions with
a peak at zero are most
common.
18. Kernel Density Fitting
A common problem is choosing the shape and width of the
kernel functions. We’ve used the Normal Distribution
and Prediction Interval, respectively.
19. Spread Adjustment
Combination of prediction interval and spread in the
ensembles can yield too much spread.
Spread Adjustment attempts to correct over dispersion.
20. Weather Elements
• Temperature and dew point,
developed simultaneously
– 3-h time projections for 7 days
– Model data at 6-h time projections
– 1650 stations, generally the same
as GFS MOS
• Maximum and minimum
temperature
– 15 days
– Same stations
21. Cool
Season
2004/05
00Z
06Z
12Z
18Z
11-member era 15-member era
Warm
Season
2005
Cool
Season
2005/06
May 30,
2006
Warm Season
04/01 – 09/30
Cool Season
10/01 – 03/31
Warm
Season
2006
March 27,
2007
Cool
Season
2006/07
21-mem.
Warm
Season
2007
Global Ensemble Forecasting System Data
Available for Ensemble MOS Development
Development Data
Independent Data
22. Methods
We explored a number of methods. Three are
presented here.
Label Equation
Development
Equation
Evaluation
Post
Processing
Ctl-Ctl-N Control member
only
Control member
only
Use a Normal
Distribution
Mn-Mn-N Mean of all
ensemble
members
Mean of all
ensemble members
Use a Normal
Distribution
Mn-Ens-KDE Mean of all
ensemble
members
Each member
individually
Apply KDE,
and adjust
spread
26. Results
• Will present results for cool season temperature
forecasts developed with two seasons of development
data and verified against one season of independent
data.
• Results center on reliability and accuracy.
• The 0000 UTC cycle of the Global Ensemble Forecast
System is the base model.
• Results for dew point are available
and very similar to temperature.
• Results for maximum/minimum
temperature are in process, and they are
similar so far.
27. Probability Integral Transform (PIT)
Histogram
• Graphically assesses reliability for a set of probabilistic
forecasts. Visually similar to Ranked Histogram.
• Method
– For each forecast-
observation pair,
probability associated
with observed event
is computed.
– Frequency of
occurrence for each
probability is recorded
in histogram as a ratio.
– Histogram boundaries
set to QF probability
values.
T=34F;
p=.663
Ratio of 1.795 indicates ~9% of the
observations fell into this category,
rather than the desired 5%.
Ratio of .809 indicates ~8% of the
observations fell into this category,
rather than the desired 10%.
28. Probability Integral Transform (PIT)
Histogram, Continued
• Assessment
– Flat histogram at unity
indicates reliable,
unbiased forecasts.
– U-shaped histogram
indicates under-
dispersion in the
forecasts.
– O-shaped histogram
indicates over-
dispersion.
– Higher values in higher
percentages indicate
a bias toward lower
forecast values.
29. Squared Bias in Relative Frequency
• Weighted average of
squared differences
between actual
height and unity for
all histogram bars.
• Zero is ideal.
• Summarizes
histogram with one
value.
Sq Bias in RF = 0.057
30. Squared Bias in Relative Frequency
• Diurnal cycle evident in early projections.
• Use of ensemble mean as a predictor improves reliability
at most time projections.
• KDE technique seems to degrade reliability.
• Model resolution change evident in latest projections.
32. Cumulative Reliability Diagram (CRD)
• Graphically assesses reliability for a set of probabilistic
forecasts. Visually similar to reliability diagrams for event-
based probability forecasts.
• Method
– For each forecast-
observation pair,
probability associated
with observed event
is computed.
– Cumulative distribution
of verifying probabilities
is plotted against the
cumulative distribution
of forecasts.
63.5% of the observations occurred
when forecast probability was 70%
for that temperature or colder.
36. Continuous Ranked Probability Score
The formula for CRPS is
where P(x) and Pa(x) are both CDFs
and
dx
x
P
x
P
x
P
CRPS
CRPS a
a
2
)
(
)
(
,
x
dy
y
x
P )
(
)
(
)
(
)
( a
a x
x
H
x
P
0
for
1
0
for
0
)
(
x
x
x
H
37. Continuous Ranked Probability Score
• Proper score that
measures the
accuracy of a set
of probabilistic
forecasts.
• Squared differ-
ence between
the forecast CDF
and a perfect
single value
forecast, inte-
grated over all
possible values
of the variable.
Units are those of the variable.
• Zero indicates perfect accuracy. No upper bound.
dx
x
P
x
P
x
P
CRPS
a
a
2
)
(
)
(
,
38. Continuous Rank Probability Score
• All techniques show considerable accuracy.
• After Day 5 the 2 techniques that use ensembles show
~0.5 deg F improvement (~12 h).
40. Dependent data; No
spread adjustment
Dependent data; With
spread adjustment
Independent data; With
spread adjustment
Independent data; No
spread adjustment
Effects of Spread Adjustment
41. Grids
• Temperature forecasts for 1650 stations
can be used to generate grids.
– Technique is identical to that used currently
for gridded MOS.
• Each grid is associated with an
exceedence probability.
43. Case Study
• 120-h Temperature
forecast based on 0000
UTC 11/26/2006, valid
0000 UTC 12/1/2006.
• Daily Weather Map at
right is valid 12 h before
verification time.
• Cold front, inverted trough
suggests a tricky forecast,
especially for Day 5.
• Ensembles showed
considerable divergence.
44. Skew in Forecast Distributions
(T50-T10); Cold Tail (T90-T50); Warm Tail
Mn-
Ens-
KDE
Mn-
Mn-
N
0 5 10° F
46. Conclusions
• These techniques can capture the uncertainty in
temperature forecasts and routinely forecast probability
distributions.
• Linear regression alone can be used to generate
probability distributions from a single model run.
• Means of ensemble output variables are useful
predictors.
• The Mn-Ens-KDE technique shows considerable
promise, and it would be relatively easy to implement
within the current MOS framework.
• Enumerating the points of the quantile function is an
effective way to disseminate probability distributions.
47. Future Work
• Improve spread adjustment technique.
• Examine characteristics of forecast distributions
and their variation.
• Verify individual stations.
• Extend temperature, dew point, maximum/
minimum temperature development to four
forecast cycles and two seasons.
• Consider forecast sharpness and convergence
as well as reliability and accuracy.
• Create forecast distributions of QPF and wind
speed.
• Explore dissemination avenues.