Title: IN SEARCH FOR THE BEST FORECAST 2
1IN SEARCH FOR THE BEST FORECAST - 2
1) ESTIMATING THE FIRST MOMENT OF THE FORECAST
PDF BASED ON INFORMATION FROM HIGH RESOLUTION
CONTROL LOWER RESOLUTION ENSEMBLE
FORECASTS2) ENSEMBLE BASED ESTIMATES OF
FORECAST UNCERTAINTY AND THEIR DOWNSCALING TO
HIGHER RESOLUTION GRIDS
- Zoltan Toth, Jun Du(1), Yuejian Zhu
- Environmental Modeling Center
- NOAA/NWS/NCEP
- Ackn. Bo Cui, David Unger, Malaquias Pena
- (1) SAIC at EMC/NCEP/NWS/NOAA
- http//wwwt.emc.ncep.noaa.gov/gmb/ens/index.html
2OUTLINE / SUMMARY
- ENSEMBLE BASED ESTIMATES OF FORECAST UNCERTAINTY
AND THEIR DOWNSCALING TO HIGHER RESOLUTION GRIDS - ESTIMATING THE FORECAST PDF BASED ON
- SINGLE FORECASTS
- ENSEMBLES
- STATISTICAL POSTPROCESSING
- BIAS REDUCTION WRT OPERATIONAL ANALYSIS
- DOWNSCALING
- FORECAST REPRESENTATIVE OF SMALLER SCALES
- FORECAST FORMAT
- ALL ENSEMBLE MEMBERS
- DERIVED PRODUCTS
- NDGD PROPOSAL
- REPRESENTING FORECAST UNCERTAINTY
- 10, 50, 90 PERCENTILE OF FORECAST PDF
3INTRODUCTION
- GOAL
- Best estimate of future state
- Verify by standard probabilistic statistics
- APPROACH
- Use all available information, including
- High resolution control forecast
- Lower resolution ensemble forecasts
- Climatology of observations/analysis and
forecasts if available - FORMAT
- PDF for single variables
- Ensemble traces to carry temporal/spatial/cross-va
riable covariance info
4GENERAL FORECAST APPROACH
- BASED ON A SINGLE FORECAST
- In nonlinear regime, sub-optimal performance
- Does not provide best estimate for expected value
- Does not provide case dependent estimates of
forecast uncertainty - In near-linear regime, can offer case dependent
corrections for - Lower resolution ensemble
- BASED ON AN ENSEMBLE
- Better estimate of expected value
- Case dependent estimates of forecast uncertainty
- DATABASE
- Collection of all relevant information
- High resolution control forecast
- Lower resolution ensemble forecasts
- Climatology of observations/analysis and
forecasts if available - Short range
- SREF ensemble control forecasts
- Medium- extended range
5FORECASTING IN A CHAOTIC ENVIRONMENT
PROBABILISTIC FORECASTING BASED A ON SINGLE
FORECAST One integration with an NWP model,
combined with past verification statistics
DETERMINISTIC APPROACH - PROBABILISTIC FORMAT
- Does not contain all forecast information
- Not best estimate for future evolution of system
- UNCERTAINTY CAPTURED IN TIME AVERAGE SENSE -
- NO ESTIMATE OF CASE DEPENDENT VARIATIONS IN FCST
UNCERTAINTY
6FORECASTING IN A CHAOTIC ENVIRONMENT -
3DETERMINISTIC APPROACH - PROBABILISTIC FORMAT
- MONTE CARLO APPROACH ENSEMBLE FORECASTING
- IDEA Sample sources of forecast error
- Generate initial ensemble perturbations
- Represent model related uncertainty
- PRACTICE Run multiple NWP model integrations
- Advantage of perfect parallelization
- Use lower spatial resolution if short on
resources - USAGE Construct forecast pdf based on finite
sample - Ready to be used in real world applications
- Verification of forecasts
- Statistical post-processing (remove bias in 1st,
2nd, higher moments) - CAPTURES FLOW DEPENDENT VARIATIONS
- IN FORECAST UNCERTAINTY
7CONFIGURATION, OUTPUT CHARACTERISTICS
2005, 2006, 2007, 2008
8PROCESS
- Collection of relevant information (model grid)
- SREF
- NAEFS
- Statistical post-processing (model grid)
- Bias correction
- Statistical
- Case dependent (using hires control)
- Weights
- Verify added value
- Manual modification of forecast guidance (model
grid) - Verify added value
- Downscaling (NDGD grid)
- Observation locations
- NDGD grid
- Manual modification (NDGD grid)
- Verify added value
9RAW BASIC PRODUCT AVAILABILITY
2005, 2006, 2007, 2008
10STATISTICAL POST-PROCESSING
- PURPOSE
- Make all forecasts look like model analysis
- Independent of lead time
- Selected set of often used variables
- All ensemble members and hi-lower resolution
controls - Assign weights to each member, corresponding to
its performance - DATA
- Medium- extended range application
- 1x1 global grid, 50 variables
- NAEFS ensemble control forecasts
- Operational analyses
- TWO METHODS
- Frequentists approach
- Compare recent statistics of forecasts and
analyses - In collaboration with David Unger of CPC
- Bayesian approach
- In collaboration with Roman Krzysztofowicz
11POST-PROCESSING METHODS - 1
- FREQUENTISTS APPROACH
- First moment (expected value) June 2006
- Estimate bias in first moment
- 35 selected variables
- Compare weighted mean of recent forecasts with
that of verifying analysis - Kalman-filter type adaptive technique works well
for short lead time - Remove estimated bias from each forecast member
- Improved 1st moment Oct 07
- More efficient for longer lead times
- Second moment bias correction Oct 07
- Weights Oct 07
- Consider case dependent reduction of bias in
lowres ensemble - Jun Dus hybrid approach
- Compare high lower resolution (ensemble)
control forecasts - Difference interpreted as resolution vector
- Adjust all members by resolution vector
- Potential for short lead times, linear regime
12(No Transcript)
13POST-PROCESSING METHODS - 2
- BAYESIAN APPROACH
- Based on single forecast PQPF processor of
Krzysztofowicz et al - Modified for ensemble forecast application
- Interpret/correct latest ensemble forecast in
context of prior information - Apply procedure on each variable to be corrected
separately - Use NCEP/NCAR reanalysis climatology as prior
- Good estimate, based on 50 years, 2.5x2.5 grid
- Adjust reanalysis climate (2.5x2.5) to estimate
operational analysis climate (1x1) - See downscaling vector below
- Represent climate pdf with parametric
distribution (2-3 pars) - Compare forecast distributions with observations,
conditioned on fcst pdf? - Represent forecast ensemble distribution with
parametric distribution (2-3 vars) - Use shorter time period (90 days) with current
model/ensemble configuration - Use Kalman-filter type approach, update prior
with most recent comparison - Transform statistics to multinormal space
- Form posterior pdf, based on prior and fcst-obs
comparison - Estimate approximate weights for each member
- Based on Bayesian statistics combined for a set
of selected variables - Retain temporal/spatial/cross-variable rank
correlations present in ensemble
14DOWNSCALING METHODS
- PURPOSE
- Provide bias-corrected forecasts at small scales
of interest - Observation points
- High resolution grid
- National Digital Forecast (or Guidance) Database
(NDFD, NDGD) - DATA
- Input
- Bias corrected ensemble forecasts
- Output
- Observation locations
- NDGD grid
- TWO METHODS
- Climate anomalies June 2006
- Downscaling vector - Oct 2007
15DOWNSCALING CLIMATE ANOMALIES
- Apply procedure on each ensemble member
- 19 selected bias-corrected variables
- Adjust bias-corrected forecast to look like
reanalysis - Use standard Kalman-filter type bias-correction
algorithm - Evaluate systematic difference between
- CDAS analysis (2.5x2.5 grid) and
- Operational analyses (1x1 grid)
- Remove systematic difference from bias-corrected
forecast (2.5x2.5) - Compare adjusted bias-corrected forecast to
reanalysis climate pdf - Represent climate distribution by parametric pdf
(2-3 vars) - Determine climate percentile corresponding to
forecast value
16DOWNSCALING WITH DOWNSCALING VECTOR
- Determine downscaling vector
- Systematic difference between
- Operational analysis (1x1 grid)
- Interpolate to NDGD grid using standard routine
- Real Time Meso-scale Analysis (RTMA)
- Based on RUC, not cycled analysis
- Quality measured against independent data
- 4-5 variables (2m temp, dewpoint, 10m winds,
precip) - 5x5km NDGD grid
- Use standard Kalman-filter type bias-estimation
correction algorithm - Independent of lead time
- Ship downscaling vector to users
- 5x5 km grid send only for area of interest, 1
field per variable /day - Ship bias-corrected forecast information
- 1x1 lat/lon grid, all lead times of interest
- Interpolate on-site to 5x5km grid, using same
standard routine as above - Add downscaling vector to interpolated
bias-corrected forecast - Performance measure
17FORECAST FORMAT
- Different users have different requirements
- Multiple formats must be offered
- Full format
- All bias-corrected ensemble members
- In-house, NCEP service centers, interested ftp
users - All other formats
- Can be derived from full format (all members)
- Ensemble functionalities
- Different products can be generated and
distributed as needed - NDGD proposal
- Important special application
- In addition to most likely scenario
- Provide information on forecast uncertainty
18ENSEMBLE PRODUCTS - FUNCTIONALITIES
List of centrally/locally/interactively generated
products required by NCEP Service Centers for
each functionality are provided in attached
tables (eg., MSLP, Z,T,U,V,RH, etc, at
925,850,700,500, 400, 300, 250, 100, etc hPa)
Potentially useful functionalities that need
further development - Mean/Spread/Median/Ranges
for amplitude of specific features -
Mean/Spread/Median/Ranges for phase of specific
features
Additional basic GUI functionalities - Ability
to manually select/identify members - Ability to
weight selected members Sept. 2005
19ENSEMBLE PRODUCT REQUEST LIST NCEP SERVICE
CENTERS, OTHER PROJECTS
20NDGD PROPOSAL
- NDFD current status
- 5x5km (or 2.5x2.5km) grid, 15-20 variables, out
to 7 days - Most likely value (expected value of forecast
pdf) provided only - Format of official NWS forecast
- NDGD new vehicle to provide guidance to be used
in NDFD process - Same format as NDFD
- Proposal to add forecast uncertainty information
in NDGD - Forecast values corresponding to 10, 50, 90
percentile of forecast pdf - Bias corrected forecast data, 1x1 grid June 06
- Downscaling vector for 1-2 variables, 5x5km
grid Oct 07 - Quality to be assessed
- Can be expanded to include more NDFD variables
21MANUAL MANIPULATION OF FORECASTS
- Manipulation needed for
- Bias-corrected forecast on model grid (national
guidance) - Downscaled forecast on NDGD grid (local guidance)
- Suggestion
- Initially modify only forecast value
corresponding to 50 percentile of fcst pdf - Use standard IFPS, NAWIPS tools
- Verify added value
- Move whole forecast distribution (10 90
percentiles) with same adjustment - Later attempt to modify forecast values
corresponding to 10 90 perc. - Verify added value
22BACKGROUND
23PRACTICAL APPROACH
- BIAS CORRECTION
- Adjust both hires control lores ensemble
- Ensures maximum utility of all forecasts
- Limits hires advantage to
- Flow dependent bias correction that statistical
method cannot achieve - IDENTIFY CROSS-OVER POINT
- Compare measures normalized by ens spread
- Measure of nonlinearity
- Control minus ens mean over ens spread
- Measure of hires advantage
- Error reduction in hires vs. lores controls over
ens spread - Alternatively, assess ens mean error dropping
below hires control error - USE OF HIRES CONTROL
- Before turnover point
- Quantitatively
- Develop methods to objectively combine info from
hires control lores ens - Bayesian approach?
24INTRODUCTION
- GOAL
- Best estimate of future state
- Verify by standard probabilistic statistics
- FORMAT
- PDF for single variables
- Ensemble traces to carry temporal/spatial/cross-va
riable covariance info - DATABASE
- Collection of all relevant information
- High resolution control forecast
- Lower resolution ensemble forecasts
- Climatology of observations/analysis and
forecasts if available - Short range
- SREF ensemble control forecasts
- Medium- extended range
- North American Ensemble Forecast System (NAEFS)
- Current Canadian NCEP global ensembles (80
members/day) - Planned FNMOC, JMA added by FY08 (160
members/day)
25OUTLINE / SUMMARY
- ENSEMBLE BASED ESTIMATES OF FORECAST UNCERTAINTY
AND THEIR DOWNSCALING TO HIGHER RESOLUTION GRIDS - UNCERTAINTY RELATED TO THE MOST LIKELY SOLUTION
- LINK WITH TRADITIONAL FORECAST APPROACH
- STATISTICAL METHODS OF ASSESSING FORECAST
UNCERTAINTY - CHOICE OF THE PAST
- NOT CASE DEPENDENT NEED ANOTHER APPROACH
- MUST USE (POSSIBLY LOWER RESOLUTION) ENSEMBLE
- CAPTURE UNCERTAINTY ON RESOLVED SCALES
- COMMUNICATING FORECAST UNCERTAINTY
- MULTI-TIER APPROACH DEPENDING ON LEVEL OF USER
SOPHISTICATION - 10, 50, 90 PERCENTILE FORECAST VALUES FOR GENERAL
USERS - MORE DETAILED INFO FOR SOPHISTICATED USERS
- MULTIPLE SCENARIOS (MODES)
- FULL PDF
26STATISTICAL ESTIMATION OF FORECAST UNCERTAINTY
- BIAS CORRECTION
- Adjust both hires control lores ensemble
- Ensures maximum utility of all forecasts
- Limits hires advantage to
- Flow dependent bias correction that statistical
method cannot achieve - IDENTIFY CROSS-OVER POINT
- Compare measures normalized by ens spread
- Measure of nonlinearity
- Control minus ens mean over ens spread
- Measure of hires advantage
- Error reduction in hires vs. lores controls over
ens spread - Alternatively, assess ens mean error dropping
below hires control error - USE OF HIRES CONTROL
- Before turnover point
- Quantitatively
- Develop methods to objectively combine info from
hires control lores ens - Bayesian approach?
27IN SEARCH FOR THE BEST FORECAST - 1
1) ESTIMATING THE FIRST MOMENT OF THE FORECAST
PDF BASED ON INFORMATION FROM HIGH RESOLUTION
CONTROL LOWER RESOLUTION ENSEMBLE
FORECASTS2) ENSEMBLE BASED ESTIMATES OF
FORECAST UNCERTAINTY AND THEIR DOWNSCALING TO
HIGHER RESOLUTION GRIDS
- Zoltan Toth, Jun Du(1), Yuejian Zhu
- Environmental Modeling Center
- NOAA/NWS/NCEP
- Ackn. Bo Cui, David Unger, Malaquias Pena
- (1) SAIC at EMC/NCEP/NWS/NOAA
- http//wwwt.emc.ncep.noaa.gov/gmb/ens/index.html
28OUTLINE / SUMMARY
- ESTIMATING THE FIRST MOMENT OF THE FORECAST PDF
BASED ON INFORMATION FROM HIGH RESOLUTION CONTROL
LOWER RESOLUTION ENSEMBLE FORECASTS - RECONCILING INFO FROM HIRES CONTROL VS. LORES ENS
- ADVISE USER COMMUNITY
- REDESIGN NWP SUITE IF NEEDED
- ATTRIBUTES OF FORECAST SYSTEMS
- POSITIVE ATTRIBUTES OF HIRES CONTROL VS. LORES
ENSEMBLE - FIRST MOMENT ESTIMATION (MOST LIKELY STATE)
- WHAT LIMITES USE OF HIRES CONTROL
- LEVEL OF DIFFERENCE IN QUALITY OF HIRES VS. LORES
CONTROLS - EMERGENCE OF NONLINEARITIES
- PRACTICAL APPROACH
- BIAS CORRECT ALL FORECASTS
- USE OF HIRES CONTROL
- LINEAR PHASE
- Objectively combine hires control lores ens
29INTRODUCTION
- PROBLEM
- Given high resolution single control and lower
resolution ensemble forecasts - What is best way to construct forecast
probability density function (pdf)? - Estimate first moment of distribution
- Estimate second and higher moments
- How to modify operational suite of NWP forecasts
for best performance? - Future choices to be made based on evaluation
results - ATTRIBUTES OF FORECAST SYSTEMS
- RELIABILITY
- Statistically, forecasts look like reality,
irrespective of how skilful they are - With large enough sample of forecast-observation
pairs, can be improved - RESOLUTION
- Sequence of observed events captured,
irrespective of realism of forecasts - Cannot be improved by statistical methods
- USER REQUIREMENTS
- Both forecast attributes important
30HIGHRES CONTROL VS. LOWRES ENSEMBLE
- RELIABILITY
- Highres expected to be better, especially at low
levels affected by terrain etc - Reduced bias
- Some phenomena simulated with more fidelity (eg,
diurnal cycle, frontal structure, etc) - Quantify advantage of highres model before and
after bias correction - Bias correction benefits more the forecasts with
larger bias gt - Effect of bias correction expected to reduce
advantage of highres forecast - RESOLUTION
- At short lead time
- Highres may offer advantage?
- Never tested, needs to be evaluated
- At later lead time
- Ensemble may become advantageous
- Filtering of nonlinearly growing errors
- HOW TO RECONCILE FORECAST INFORMATION
- FROM HIRES CONTROL LOWRES ENS?
31RECONCILING HIGHRES CONTROL VS. LOWRES ENS
- WMO/CBS Ensemble Expert Team meeting (Feb. 2006)
- Guidance sought by WMO members as to relevance of
hires vs. lowres ensemble forecasts for - First moment estimation
- Forecast uncertainty estimation
- Some related studies from various viewpoints in
literature - Hires control vs. ens mean (Zhu, Atger, Du, etc)
- Probabilistic forecasts based on hires control
vs. lores ens - Zhu et al, Atger, Du
- Combined use of hires control lores ensemble
- Smith et al., J. Du
- Following careful evaluation of strength
weaknesses of both systems - Advise user community on best practices
- Develop new tools if necessary to facilitate best
use - Redesign NWP forecast suite if necessary
32ESTIMATING FIRST MOMENT
- EMERGENCE OF NONLINEARITIES
- Initially symmetric cloud of ensemble becomes
distorted with lead time - Measure level of nonlinearity by deviation of
lores control and ens mean - Normalized measure Control minus ens mean over
ens spread - What is critical level? 20 or what? (0linear,
1nonlinear) - Linearity is lost sooner for smaller scale / more
unstable phenomena - Once linearity is critically violated, relying on
single control (even hires) not appropriate - All ensemble solutions must be considered
- BEFORE NONLINEARITY BECOMES DOMINANT
- Hires control can have value beyond lores
ensemble - Value beyond ensemble is limited
- Resolution is typically only twice that of
ensemble - Bias correction of all forecasts further reduces
value of hires control - Quantify advantage of hires control statistically
- Error reduction (hires vs. lores controls) as
function of lead time - Normalize by ensemble spread?
- After cross-over time, only very special,
qualitative use of hires model
33PRACTICAL APPROACH
- BIAS CORRECTION
- Adjust both hires control lores ensemble
- Ensures maximum utility of all forecasts
- Limits hires advantage to
- Flow dependent bias correction that statistical
method cannot achieve - IDENTIFY CROSS-OVER POINT
- Compare measures normalized by ens spread
- Measure of nonlinearity
- Control minus ens mean over ens spread
- Measure of hires advantage
- Error reduction in hires vs. lores controls over
ens spread - Alternatively, assess ens mean error dropping
below hires control error - USE OF HIRES CONTROL
- Before turnover point
- Quantitatively
- Develop methods to objectively combine info from
hires control lores ens - Bayesian approach?
34SCIENTIFIC BACKGROUND WEATHER FORECASTS ARE
UNCERTAIN
Buizza 2002
35(No Transcript)
36BACKGROUND
37COMMENTS ON JUNS APPROACH
- Bias correction / post-processing algorithm based
on - Difference between hires and lores controls
- Useful as long as
- Controls evolve similarly gt
- Difference dominated by systematic differences
between 2 model resolutions - Once 2 controls diverge chaotically, random
errors introduced - How to quantify?
- Measure difference between 2 controls, normalized
by ens spread? - Related to limit of linearity discussed earlier?
- Stop using method beyond a point (use correction
with tapering weights?) - Characteristics
- Positive
- Completely flow dependent corrections
- Negative
- Correction is limited by resolution/performance
of hires model gt - Statistical bias correction still needed
38EVALUATION OF FORECAST SYSTEMS
- Some statistics based on forecast system only
- Other statistics based on comparison of forecast
and observed systems gt - FORECAST SYSTEM ATTRIBUTES
- Abstract concepts (like length)
- Reliability and Resolution
- Both can be measured through different statistics
- Statistical properties
- Interpreted for large set of forecasts (ie,
describe behavior of forecast system), - not for a single forecast
- For their definition
- Assume that forecasts
- Can be of any format
- Take a finite number of different classes
- Consider empirical frequency distribution of
- Verifying observations corresponding to large
number of forecasts of same class gt - Observed Frequency Distribution (ofd)
39STATISTICAL RELIABILITY STATISTICAL CONSISTENCY
OF FORECASTS WITH OBSERVATIONS
- BACKGROUND
- Consider particular forecast class Fa
- Consider distribution of observations Oa that
follow forecasts Fa - DEFINITION
- If forecast Fa has the exact same form as Oa, for
all forecast classes, - the forecast system is statistically consistent
with observations gt - The forecast system is perfectly reliable
- MEASURES OF RELIABILITY
- Based on different ways of comparing Fa and Oa
40STATISTICAL RESOLUTION ABILITY TO DISTINGUISH,
AHEAD OF TIME, AMONG DIFFERENT OUTCOMES
- BACKGROUND
- Assume observed events are classified into finite
number of classes - DEFINITION
- If all observed classes are preceded by
distinctly different forecasts, the forecasts
resolve the problem gt - The forecast system has perfect resolution
- MEASURES OF RELIABILITY
- Based on degree of separation of distributions of
observations that follow various forecast classes
- Measured by difference between ofds climate
distribution - Measures differ by how differences between
distributions are quantified
FORECASTS
OBSERVATIONS
EXAMPLES
41CHARACTERISTICS OF FORECAST SYSTEM ATTRIBUTES
- Reliability resolution are general forecast
attributes - Valid for any forecast format (single,
categorical, probabilistic, etc) - Reliability
- Can be statistically imposed at one time level
- If both natural forecast systems are stationary
in time, and - If there is a large enough set of
observed-forecast pairs - Replace forecast by corresponding observed
frequency distribution - Not related to time sequence of forecast/observed
systems - Resolution reflects inherent value of forecast
system - Can be improved only through more knowledge about
time sequence of events - Statistical consistency at one time level
(reliability) is irrelevant - Reliability resolution are independent
attributes - Climate pdf forecast is perfectly reliable, yet
has no resolution - Reversed rain / no-rain forecast can have perfect
resolution and no reliability - Perfect reliability and perfect resolution
perfect forecast system
42FORECAST METHODS
- Empirically based
- Based on record of observations gt
- Possibly very good reliability
- Will fail in new (not yet observed) situations
(eg., climate trend, etc) - Resolution (forecast skill) depends on length of
observations - Useful for now-casting, climate applications
- Not practical for typical weather forecasting
- Theoretically based
- Based on general scientific principles
- Incomplete/approximate knowledge in NWP models gt
- Prone to statistical inconsistency
- Run-of-the-mill cases can be statistically
calibrated to insure reliability - For forecasting rare/extreme events, statistical
consistency of model must be improved - Predictability limited by
- Gaps in knowledge about system
- Errors in initial state of system
43USER REQUIREMENTSPROBABILISTIC FORECAST
INFORMATION IS CRITICAL
44FORECASTING IN A CHAOTIC ENVIRONMENT
PROBABILISTIC FORECASTING BASED A ON SINGLE
FORECAST One integration with an NWP model,
combined with past verification statistics
DETERMINISTIC APPROACH - PROBABILISTIC FORMAT
- Does not contain all forecast information
- Not best estimate for future evolution of system
- UNCERTAINTY CAPTURED IN TIME AVERAGE SENSE -
- NO ESTIMATE OF CASE DEPENDENT VARIATIONS IN FCST
UNCERTAINTY
45- FORECASTING IN A CHAOTIC ENVIRONMENT - 2
- DETERMINISTIC APPROACH - PROBABILISTIC FORMAT
-
- PROBABILISTIC FORECASTING -
- Based on Liuville Equations
- Continuity equation for probabilities, given
dynamical eqs. of motion - Initialize with probability distribution
function (pdf) at analysis time - Dynamical forecast of pdf based on conservation
of probability values - Prohibitively expensive -
- Very high dimensional problem (state space x
probability space) - Separate integration for each lead time
- Closure problems when simplified solution sought
46FORECASTING IN A CHAOTIC ENVIRONMENT -
3DETERMINISTIC APPROACH - PROBABILISTIC FORMAT
- MONTE CARLO APPROACH ENSEMBLE FORECASTING
- IDEA Sample sources of forecast error
- Generate initial ensemble perturbations
- Represent model related uncertainty
- PRACTICE Run multiple NWP model integrations
- Advantage of perfect parallelization
- Use lower spatial resolution if short on
resources - USAGE Construct forecast pdf based on finite
sample - Ready to be used in real world applications
- Verification of forecasts
- Statistical post-processing (remove bias in 1st,
2nd, higher moments) - CAPTURES FLOW DEPENDENT VARIATIONS
- IN FORECAST UNCERTAINTY
47NCEP GLOBAL ENSEMBLE FORECAST SYSTEM
MARCH 2004 CONFIGURATION
48MOTIVATION FOR ENSEMBLE FORECASTING
- FORECASTS ARE NOT PERFECT - IMPLICATIONS FOR
- USERS
- Need to know how often / by how much forecasts
fail - Economically optimal behavior depends on
- Forecast error characteristics
- User specific application
- Cost of weather related adaptive action
- Expected loss if no action taken
- EXAMPLE Protect or not your crop against
possible frost - Cost 10k, Potential Loss 100k gt Will protect
if P(frost) gt Cost/Loss0.1 - NEED FOR PROBABILISTIC FORECAST INFORMATION
- DEVELOPERS
- Need to improve performance - Reduce error in
estimate of first moment - Traditional NWP activities (I.e., model, data
assimilation development) - Need to account for uncertainty - Estimate higher
moments - New aspect How to do this?
- Forecast is incomplete without information on
forecast uncertainty - NEED TO USE PROBABILISTIC FORECAST FORMAT
49HOW TO DEAL WITH FORECAST UNCERTAINTY?
- No matter what / how sophisticated forecast
methods we use - Forecast skill limited
- Skill varies from case to case
- Forecast uncertainty must be assessed by
meteorologists
THE PROBABILISTIC APPROACH
50SOCIO-ECONOMIC BENEFITS OFSEAMLESS
WEATHER/CLIMATE FORECAST SUITE
Commerce Energy
Ecosystem Health
Hydropower Agriculture
Boundary Condition Sensitivity
Reservoir control Recreation
Transportation Fire weather
Initial Condition Sensitivity
Flood mitigation Navigation
Protection of Life/Property
Weeks
Minutes
Days
Hours
Years
Seasons
Months
51(No Transcript)
52(No Transcript)
53144 hr forecast
Poorly predictable large scale wave Eastern
Pacific Western US
Highly predictable small scale wave Eastern US
Verification
54(No Transcript)
55(No Transcript)
56(No Transcript)
57FORECAST PERFORMANCE MEASURES
COMMON CHARACTERISTIC Function of both forecast
and observed values
MEASURES OF RELIABILITY DESCRIPTION Statisticall
y compares any sample of forecasts with sample of
corresponding observations GOAL To assess
similarity of samples (e.g., whether 1st and 2nd
moments match) EXAMPLES Reliability component
of Brier Score Ranked Probability
Score Analysis Rank Histogram Spread vs. Ens.
Mean error Etc.
MEASURES OF RESOLUTION DESCRIPTION Compares the
distribution of observations that follows
different classes of forecasts with the climate
distribution GOAL To assess how well the
observations are separated when grouped by
different classes of preceding fcsts EXAMPLES Res
olution component of Brier Score Ranked
Probability Score Information content Relative
Operational Characteristics Relative Economic
Value Etc.
COMBINED (RELRES) MEASURES Brier, Ranked
Probab. Scores, rmse, PAC, etc
58EXAMPLE PROBABILISTIC FORECASTS
RELIABILITY Forecast probabilities for given
event match observed frequencies of that event
(with given prob. fcst) RESOLUTION Many
forecasts fall into classes corresponding to high
or low observed frequency of given
event (Occurrence and non-occurrence of event is
well resolved by fcst system)
59(No Transcript)
60PROBABILISTIC FORECAST PERFORMANCE MEASURES
TO ASSESS TWO MAIN ATTRIBUTES OF PROBABILISTIC
FORECASTS RELIABILITY AND RESOLUTION Univariate
measures Statistics accumulated point by
point in space Multivariate measures Spatial
covariance is considered
BRIER SKILL SCORE (BSS)
EXAMPLE
COMBINED MEASURE OF RELIABILITY AND RESOLUTION
61BRIER SKILL SCORE (BSS)
COMBINED MEASURE OF RELIABILITY AND RESOLUTION
- METHOD
- Compares pdf against analysis
- Resolution (random error)
- Reliability (systematic error)
- EVALUATION
- BSS Higher better
- Resolution Higher better
- Reliability Lower better
- RESULTS
- Resolution dominates initially
- Reliability becomes important later
- ECMWF best throughout
- Good analysis/model?
- NCEP good days 1-2
- Good initial perturbations?
- No model perturb. hurts later?
- CANADIAN good days 8-10
May-June-July 2002 average Brier skill score for
the EC-EPS (grey lines with full circles), the
MSC-EPS (black lines with open circles) and the
NCEP-EPS (black lines with crosses). Bottom
resolution (dotted) and reliability(solid)
contributions to the Brier skill score. Values
refer to the 500 hPa geopotential height over the
northern hemisphere latitudinal band 20º-80ºN,
and have been computed considering 10
equally-climatologically-likely intervals (from
Buizza, Houtekamer, Toth et al, 2004)
62BRIER SKILL SCORE
COMBINED MEASURE OF RELIABILITY AND RESOLUTION
63RANKED PROBABILITY SCORE
COMBINED MEASURE OF RELIABILITY AND RESOLUTION
64ANALYSIS RANK HISTOGRAM (TALAGRAND DIAGRAM)
MEASURE OF RELIABILITY
65ENSEMBLE MEAN ERROR VS. ENSEMBLE SPREAD
MEASURE OF RELIABILITY
Statistical consistency between the ensemble and
the verifying analysis means that the verifying
analysis should be statistically
indistinguishable from the ensemble members
gt Ensemble mean error (distance between ens.
mean and analysis) should be equal to ensemble
spread (distance between ensemble mean and
ensemble members)
In case of a statistically consistent ensemble,
ens. spread ens. mean error, and they are both
a MEASURE OF RESOLUTION. In the presence of bias,
both rms error and PAC will be a combined measure
of reliability and resolution
66INFORMATION CONTENT
MEASURE OF RESOLUTION
67RELATIVE OPERATING CHARACTERISTICS
MEASURE OF RESOLUTION
68ECONOMIC VALUE OF FORECASTS
MEASURE OF RESOLUTION
69PERTURBATION VS. ERROR CORRELATION ANALYSIS (PECA)
MULTIVATIATE COMBINED MEASURE OF RELIABILITY
RESOLUTION
- METHOD Compute correlation between ens
perturbtns and error in control fcst for - Individual members
- Optimal combination of members
- Each ensemble
- Various areas, all lead time
- EVALUATION Large correlation indicates ens
captures error in control forecast - Caveat errors defined by analysis
- RESULTS
- Canadian best on large scales
- Benefit of model diversity?
- ECMWF gains most from combinations
- Benefit of orthogonalization?
- NCEP best on small scale, short term
- Benefit of breeding (best estimate initial
error)? - PECA increases with lead time
- Lyapunov convergence
- Nonlilnear saturation
- Higher values on small scales
70WHAT WE NEED FOR POSTPROCESSING TO WORK?
- LARGE SET OF FCST OBS PAIRS
- Consistency defined over large sample need same
for post-processing - Larger the sample, more detailed corrections can
be made - BOTH FCST AND REAL SYSTEMS MUST BE STATIONARY IN
TIME - Otherwise can make things worse
- Subjective forecasts difficult to calibrate
HOW WE MEASURE STATISTICAL INCONSISTENCY?
- MEASURES OF STATIST. RELIABILITY
- Time mean error
- Analysis rank histogram (Talagrand diagram)
- Reliability component of Brier etc scores
- Reliability diagram
71SOURCES OF STATISTICAL INCONSISTENCY
- TOO FEW FORECAST MEMBERS
- Single forecast inconsistent by definition,
unless perfect - MOS fcst hedged toward climatology as fcst skill
is lost - Small ensemble sampling error due to limited
ensemble size - (Houtekamer 1994?)
- MODEL ERROR (BIAS)
- Deficiencies due to various problems in NWP
models - Effect is exacerbated with increasing lead time
- SYSTEMATIC ERRORS (BIAS) IN ANALYSIS
- Induced by observations
- Effect dies out with increasing lead time
- Model related
- Bias manifests itself even in initial conditions
- ENSEMBLE FORMATION (INPROPER SPREAD)
- Not appropriate initial spread
- Lack of representation of model related
uncertainty in ensemble - I. E., use of simplified model that is not able
to account for model related uncertainty
72HOW TO IMPROVE STATISTICAL CONSISTENCY?
- MITIGATE SOURCES OF INCONSISTENCY
- TOO FEW MEMBERS
- Run large ensemble
- MODEL ERRORS
- Make models more realistic
- INSUFFICIENT ENSEMBLE SPREAD
- Enhance models so they can represent model
related forecast uncertainty - OTHERWISE gt
- STATISTICALLY ADJUST FCST TO REDUCE INCONSISTENCY
- Unpreferred way of doing it
- What we learn can feed back into development to
mitigate problem at sources - Can have LARGE impact on (inexperienced) users
73(No Transcript)
74SUMMARY
- WHY DO WE NEED PROBABILISTIC FORECASTS?
- Isnt the atmosphere deterministic? YES, but
its also CHAOTIC - FORECASTERS PERSPECTIVE USERS PERSPECTIVE
- Ensemble techniques Probabilistic description
- WHAT ARE THE MAIN ATTRIBUTES OF FORECAST SYSTEMS?
- RELIABILITY Stat. consistency with distribution
of corresponding observations - RESOLUTION Different events are preceded by
different forecasts - WHAT ARE THE MAIN TYPES OF FORECAST METHODS?
- EMPIRICAL Good reliability, limited resolution
(problems in new situations) - THEORETICAL Potentially high resolution, prone to
inconsistency - ENSEMBLE METHODS
- Only practical way of capturing fluctuations in
forecast uncertainty due to - Case dependent dynamics acting on errors in
- Initial conditions
- Forecast methods
- HOW CAN PROBABILSTIC FORECAST PERFORMANCE BE
MEASURED?
75Toth, Z., O. Talagrand, and Y. Zhu, 2005 The
Attributes of Forecast Systems A Framework for
the Evaluation and Calibration of Weather
Forecasts. In Predictability Seminars, 9-13
September 2002, Ed. T. Palmer, ECMWF, in press.
Toth, Z., O. Talagrand, G. Candille, and Y. Zhu,
2003 Probability and ensemble forecasts. In
Environmental Forecast Verification A
practitioner's guide in atmospheric science. Ed.
I. T. Jolliffe and D. B. Stephenson. Wiley, p.
137-164.