Bias Correction and Downscaling
CEVE 543 - Fall 2025
- Motivation and GCM biases (minutes 0-4)
- Weather vs climate distinction + supervised learning (minutes 4-9)
- Delta method (minutes 9-21)
- QQ-mapping (minutes 21-33)
- Quantile delta mapping (QDM) (minutes 33-41)
- Uncertainty and critical considerations (minutes 41-44)
- Wrap-up and looking ahead (minutes 44-45)
1 Motivation: Why Bias Correction?
Minutes 0-4
- Climate models have systematic biases that directly affect impact assessments.
- Impacts depend on distributions (frequency, intensity, duration) — not just means or totals.
1.1 Two Types of GCM Biases
- Drizzle bias (pervasive):
- Too many light rainfall events with intensities too low relative to stations
- Root causes: spatial averaging within large grid cells (≈100–300 km) and sub-grid parameterizations
- Impacts: distorts frequency/intensity; affects crop growth, soil moisture dynamics, infiltration/runoff, and floods even when monthly totals match
- Dynamical biases (circulation errors):
- Examples: ENSO amplitude/frequency, monsoon onset/timing, midlatitude storm tracks
- Impacts: shifts climatology and interannual variability; affects predictability and teleconnections
Pick one impact domain (agriculture, floods, urban drainage, or energy). Which bias is more damaging in your domain—drizzle bias or dynamical bias—and why? What observable metrics would you compute to diagnose and quantify the bias?
Suggested format: 2 min pair + 2 min share.
1.2 Weather vs Climate: A Critical Distinction
Minutes 4-9
- Weather forecasting: initial value problem with paired forecast–observation data → supervised learning works well.
- Climate projection: boundary value problem with no paired future observations → compare distributions under like-forcing; different statistical tools.
Aspect | Weather forecasting | Climate projection |
---|---|---|
Mathematical framing | Initial value problem | Boundary value problem |
Training data | Paired forecasts and observations | No pairs for the future; rely on historical distributions under same forcings |
Target | Next state/field at high resolution | Changes in distributions, moments, quantiles, extremes |
Horizon | Hours–weeks | Decades–centuries |
Typical methods | Supervised ML, data assimilation, super-resolution | Bias correction, quantile mapping, QDM, emulators respecting physics |
Validation | Skill vs paired observations | Hindcast distributional match under historical forcings |
Failure modes | Error growth, chaos | Nonstationarity, regime shifts, process changes |
Outputs | Deterministic/probabilistic fields | Corrected time series, deltas, distributional adjustments |
- Why progress is fast: large paired forecast–observation datasets enable powerful supervised models (e.g., diffusion models like CorrDiff).
- Why climate downscalers are harder: no paired future data, large distribution shift/nonstationarity, long-term physical constraints.
- Implication: higher risk; requires rigorous validation and uncertainty quantification.
When we do have paired data (predictor–predictand), supervised learning is natural:
- Super-resolution: coarse → fine fields (deterministic or probabilistic ensembles)
- Learned precipitation modules: use T, q, winds → estimate precip at fine scales; often patch-based, newer methods improve cross-patch consistency
- Simple bias correction for forecast systems: regress out mean/variance errors (e.g., +2°C warm bias)
- Key advantage: train directly on paired observations with standard ML losses
If a supervised diffusion model like CorrDiff excels at weather-scale downscaling, what specific obstacles arise when trying to train a similar model to downscale future climate projections? Which of those can you mitigate, and how?
Suggested format: 2 min pair + 2 min share.
1.3 Climate: The Hard Case
- No future observations → train/fit on historical distributions only (under same forcings)
- Core assumption: model and observed describe the same climate distribution historically
- Workflow: match GCM-hist to Obs-hist → derive transfer function → apply to GCM-fut
- Caveat: may fail under regime shifts/nonstationarity or novel process changes
2 Delta Method
Minutes 9-21
The delta method is simple but widely used (Hay, Wilby, and Leavesley 2000).
2.0.1 Delta method (at a glance)
- Inputs: Observed historical series; GCM historical and future series (or statistics) by month/season
- Use when: You trust the modeled change signal more than model absolute levels
- Outputs: Corrected future series/statistics anchored to observations, preserving modeled change
- Pitfalls: Stationary-bias assumption; no frequency/intensity correction; deterministic
- Quick steps: compute Δ (additive for T, multiplicative for P) by month → apply Δ to observed baseline
2.1 Assumption and When to Use
- GCM can have systematic absolute biases (e.g., radiation/clouds, ocean heat transport, topography)
- Assume the modeled change signal (trend/delta) is credible
- Example: model is +2°C warm bias but simulates +3°C future warming → apply +3°C to observed baseline
2.2 Mathematical Framework
Notation:
- X^\text{obs}_\text{hist} = observed variable in historical period
- X^\text{GCM}_\text{hist} = GCM-simulated variable in historical period
- X^\text{GCM}_\text{fut} = GCM-simulated variable in future period
- \bar{X} = temporal average (e.g., monthly mean)
Compute the change (delta): \Delta = \bar{X}^\text{GCM}_\text{fut} - \bar{X}^\text{GCM}_\text{hist}
Apply to observations: X^\text{corr}_\text{fut} = X^\text{obs}_\text{hist} + \Delta
Note on notation: in the delta method, we apply the modeled change signal (the delta) to an observed baseline series. That’s why the baseline term is X^\text{obs}_\text{hist} (not GCM). By contrast, QQ-mapping and QDM directly transform the GCM daily values to corrected values.
2.3 Variants
For temperature, we typically use an additive delta: T^\text{corr}(d, m, y) = T^\text{obs}(d, m, y_\text{hist}) + [\bar{T}^\text{GCM}_\text{fut}(m) - \bar{T}^\text{GCM}_\text{hist}(m)]
Here d is the day of month, m is the calendar month, y is the year, and y_\text{hist} is a year from the historical period. We take a day from the historical record and add the monthly-mean temperature change.
For precipitation, we use a multiplicative delta: P^\text{corr}(d, m, y) = P^\text{obs}(d, m, y_\text{hist}) \times \frac{\bar{P}^\text{GCM}_\text{fut}(m)}{\bar{P}^\text{GCM}_\text{hist}(m)}
2.3.1 Applying delta to derived statistics (quantiles, IDF curves)
The same idea applies when the target is a derived statistic rather than the raw daily series. Two common cases:
Quantiles (e.g., monthly temperature quantiles), additive form: Q^\text{corr}_{X}(p, m) = Q^\text{obs}_{X,\,\text{hist}}(p, m) + \Big[Q^\text{GCM}_{X,\,\text{fut}}(p, m) - Q^\text{GCM}_{X,\,\text{hist}}(p, m)\Big]
IDF curves (precipitation intensity–duration–frequency), multiplicative form: \text{IDF}^\text{corr}(D, r) = \text{IDF}^\text{obs}_\text{hist}(D, r) \times \frac{\text{IDF}^\text{GCM}_\text{fut}(D, r)}{\text{IDF}^\text{GCM}_\text{hist}(D, r)}
Here p is a percentile, m a month, D a duration, and r a return period (or equivalently a high quantile). This “delta-on-statistics” approach preserves the modeled change in the statistic of interest while anchoring to observed historical levels.
Practical note: to reduce noise and month-boundary discontinuities, practitioners sometimes compute deltas on moving windows (e.g., 15–31 days) or seasonal aggregates; we use monthly deltas here for clarity and speed in class.
Why use additive correction for temperature but multiplicative for precipitation? What would go wrong if you did it the other way?
Hint: Think about what happens if temperature drops by 5°C vs if precipitation drops by 50%.
2.4 Example Calculation
Suppose for July we have:
- Observed historical mean: \bar{T}^\text{obs}_\text{hist} = 25°\text{C}
- GCM historical mean: \bar{T}^\text{GCM}_\text{hist} = 27°\text{C} (2°C warm bias)
- GCM future mean: \bar{T}^\text{GCM}_\text{fut} = 30°\text{C}
The delta is \Delta = 30 - 27 = 3°\text{C}.
The corrected future temperature is T^\text{corr}_\text{fut} = 25 + 3 = 28°\text{C}.
We preserve the GCM’s 3°C warming signal while removing the 2°C warm bias.
2.5 Advantages and Limitations
- Advantages:
- Simple, transparent; preserves modeled change signal
- Works for many statistics (means, quantiles, variances)
- Seasonal specificity: apply deltas by month/season
- Limitations:
- Deterministic; no explicit uncertainty
- Assumes stationary bias structure across climate states
- Doesn’t correct frequency/intensity structure directly
The delta method assumes the bias is additive (for temperature) or multiplicative (for precipitation) and constant across climate states.
What physical processes might violate these assumptions? For example:
- What if precipitation shifts from frontal to convective?
- What if cloud feedbacks change with warming?
- What if the seasonal cycle shifts?
Think about: process-level changes, nonlinear responses, threshold effects.
3 Quantile-Quantile (QQ) Mapping
Minutes 21-33
QQ-mapping is more sophisticated and directly addresses the drizzle bias (Ines and Hansen 2006).
3.0.1 QQ-mapping (at a glance)
- Inputs: Daily/ sub-daily GCM and observed series (historical), by month/season
- Use when: Large-scale dynamics credible; need to fix wet-day frequency and intensity distribution
- Outputs: Bias-corrected series matching observed distribution (hist), applied to future with assumed stationarity
- Pitfalls: Can distort modeled change signal when applied to future; deterministic; tail extrapolation
- Quick steps: adjust wet-day frequency via threshold → map intensities via CDF-to-CDF transform (by month)
3.1 Assumption and When to Use
- Assumes large-scale dynamics are credible but frequency/intensity at sub-grid scales are biased
- Historical GCM–Obs distributional relationship is assumed to hold in the future
- Reasonable when GCM gets large-scale patterns right; risky if key processes are missing
3.2 Motivation
Precipitation is the product of frequency and intensity: \bar{X}_m = \pi_m \times \mu_{I,m}
where \pi_m is the wet-day frequency in calendar month m (dimensionless) and \mu_{I,m} is the mean wet-day intensity in month m (mm/day).
Correcting only the monthly mean \bar{X}_m doesn’t fix the frequency/intensity distortion.
A GCM simulates 20 days with 5 mm/day rain (monthly total = 100 mm). Observations show 10 days with 10 mm/day rain (also 100 mm total).
If you only correct the monthly mean, what problems might arise for:
- A soil infiltration model?
- A crop water stress model?
Think about: infiltration capacity, runoff generation, dry spell length, water stress timing.
3.3 Visual Illustration
The two figures below illustrate how QQ-mapping works:
3.4 Two-Step Process
Notation:
- P^\text{GCM}(d, m, y) = daily GCM precipitation on day d of month m in year y
- P^\text{obs}(d, m, y) = daily observed precipitation
- F(\cdot) = cumulative distribution function (CDF)
- F^{-1}(\cdot) = inverse CDF (quantile function)
- \tilde{x} = threshold for defining “wet day” (typically 0.1 mm)
Step 1: Frequency correction
Find threshold \tilde{x}^\text{GCM} such that GCM wet-day frequency matches observed: F^\text{GCM}(\tilde{x}^\text{GCM}) = F^\text{obs}(\tilde{x}^\text{obs})
where \tilde{x}^\text{obs} = 0.1 mm is the observed wet-day threshold.
Invert to get: \tilde{x}^\text{GCM} = (F^\text{GCM})^{-1}\left(F^\text{obs}(\tilde{x}^\text{obs})\right)
Discard all GCM values below \tilde{x}^\text{GCM}. This removes excess drizzle events.
Step 2: Intensity correction
For each wet day i with x_i \geq \tilde{x}^\text{GCM}, map from GCM intensity distribution to observed: P^\text{corr}_i = (F^\text{obs}_I)^{-1}\left(F^\text{GCM}_I(P^\text{GCM}_i)\right)
where F^\text{GCM}_I is the CDF of GCM intensities above threshold \tilde{x}^\text{GCM} and F^\text{obs}_I is the CDF of observed intensities above threshold \tilde{x}^\text{obs}.
The intuition is to find where P^\text{GCM}_i falls in the GCM distribution, then sample the same percentile from the observed distribution.
3.5 Distribution Choices
For observed intensities (F^\text{obs}_I), we typically use a gamma distribution: f(x; \alpha, \beta) = \frac{1}{\beta^\alpha \Gamma(\alpha)} x^{\alpha-1} \exp\left(-\frac{x}{\beta}\right), \quad x > 0
The shape parameter \alpha is dimensionless and the scale parameter \beta has the same units as x. Both are estimated by maximum likelihood estimation (MLE).
For GCM intensities (F^\text{GCM}_I), two approaches are common. Gamma-Gamma (GG) fits a gamma distribution to both GCM and observed data. Empirical-Gamma (EG) uses the empirical CDF for GCM data but gamma for observed. Ines and Hansen (2006) found EG slightly better for their crop modeling application.
3.6 Why This Matters
- Impact models are sensitive to:
- Wet/dry timing → dry spell lengths
- Event intensities → infiltration/runoff generation
- Temporal clustering → water stress and extreme flows
- Monthly-total-only correction can mislead impacts; frequency/intensity corrections matter (Ines and Hansen (2006))
3.7 Limitations
- Deterministic; no explicit uncertainty
- Assumes quantile relationship is time-invariant; process changes can break it
- Plain QM can distort modeled climate-change signals when applied to future data → motivation for QDM (Cannon, Sobie, and Murdock 2015)
- Terminology: “quantile mapping (QM)” ≈ “QQ-mapping” in common use
- QDM preserves quantile-specific change by construction (Cannon, Sobie, and Murdock 2015)
4 Quantile Delta Mapping (QDM)
Minutes 33-41
QDM blends the strengths of delta change and quantile mapping (Cannon, Sobie, and Murdock 2015). It corrects historical bias while preserving the GCM’s projected change at each quantile. This avoids “erasing” the modeled climate-change signal during bias correction.
4.1 Assumptions and Idea
- Historical quantile relation (model ↔︎ obs) is informative
- Modeled change signal by quantile is credible
- Procedure: quantile-map using historical relation → adjust by modeled change at the same percentile
4.2 Notation
- F^\text{mod}_\text{hist}, F^\text{obs}_\text{hist}: historical CDFs (model, observed)
- Q^\text{mod}_\text{hist}(p), Q^\text{mod}_\text{fut}(p), Q^\text{obs}_\text{hist}(p): corresponding quantile functions
- For a modeled future value y_i (from the GCM future series), define its modeled future percentile p_i = F^\text{mod}_\text{fut}(y_i)
4.3 QDM formula
Additive variables (e.g., temperature): x^{\text{QDM}}_{\text{fut}, i} = Q^\text{obs}_\text{hist}(p_i) + \Big[Q^\text{mod}_\text{fut}(p_i) - Q^\text{mod}_\text{hist}(p_i)\Big]
Multiplicative variables (e.g., precipitation): x^{\text{QDM}}_{\text{fut}, i} = Q^\text{obs}_\text{hist}(p_i) \times \frac{Q^\text{mod}_\text{fut}(p_i)}{Q^\text{mod}_\text{hist}(p_i)}
Intuition: map the modeled future value to a percentile p_i, take the observed historical value at that percentile, and then apply the model’s projected change at that percentile (absolute for temperature, relative for precipitation) (Cannon, Sobie, and Murdock 2015).
4.4 Pros and caveats
- QDM (at a glance)
- Inputs: Historical GCM/Obs distributions and modeled future distribution
- Use when: You need bias correction that preserves quantile-specific modeled change
- Outputs: Corrected future series with preserved change at each percentile
- Pitfalls: Assumes historical quantile relation persists; zero-inflation care; tail handling
- Quick steps: find percentile of modeled future value → map to observed-hist quantile → apply modeled change at that percentile (additive/multiplicative)
- Preserves modeled change signal across the distribution; mitigates “delta erosion.”
- Handles non-uniform changes (e.g., heavier tails) better than plain delta.
- Still assumes historical quantile relationships carry forward; tail extrapolation remains tricky.
- Requires care for zero-inflation and frequency changes for precipitation.
4.5 Method comparison (cheat sheet)
Feature | Delta | QQ-mapping (QM) | QDM |
---|---|---|---|
Preserves modeled change signal | Yes (mean/target stat) | No (can distort) | Yes (by quantile) |
Corrects wet-day frequency | No | Yes (with threshold) | Yes (if combined with frequency step) |
Corrects intensity distribution | No (only target stat) | Yes | Yes |
Assumes stationarity of bias | Yes | Yes (quantile relation) | Yes (quantile relation + change factors) |
Handles non-uniform change across quantiles | No | Not guaranteed | Yes |
Uncertainty representation | Deterministic | Deterministic | Deterministic |
Data requirements | Means/quantiles by month | Full distribution by month | Full distribution by month (hist + fut percentiles) |
When would you prefer QDM over plain QQ-mapping or the basic delta method? Give one example where preserving quantile-specific change matters for impacts, and one case where QDM’s assumptions might be risky.
Suggested format: 1 min think + 2 min pair + 2 min share.
5 Uncertainty in Bias Correction
Minutes 41-43
Bias correction is not a panacea and introduces additional uncertainty.
5.1 Sources of Uncertainty
- Sampling: limited records → few extremes → uncertain tail parameters
- Structural: method choice (delta/QQ/QDM), distributional form (gamma/empirical), temporal aggregation (daily/monthly), wet-day thresholds
- Extrapolation: does the transfer function hold in future climates with no historical analog? New circulation regimes, unprecedented T–q combinations, process changes
Suppose you calibrate QQ-mapping using 1980-2010 data, when the wettest month had 250 mm. In 2080, the GCM simulates a month with 400 mm.
- What will QQ-mapping do?
- What assumptions are you making?
- When might this fail?
Think about: extrapolation methods, physical realism, process changes.
5.2 Methods Can Disagree Strongly
- Hay, Wilby, and Leavesley (2000): delta vs statistical downscaling for three US basins → markedly different runoff projections
- Differences largest in complex terrain; “better” method depends on validation context
Applying bias correction doesn’t necessarily improve impact predictions. It transforms one type of uncertainty (GCM bias) into another (correction uncertainty). Multiple methods should always be compared.
5.3 Quantifying Uncertainty
- Use multiple methods and multiple GCMs; generate ensembles; validate on independent data when available
- Validation is hardest where it matters most (future conditions)
6 Critical Considerations
Minutes 43-44
6.1 Extrapolation
- All methods must extrapolate beyond calibration range
- QM: often uses constant (cap at max obs) or linear tail extrapolation → risk of unrealistic extremes or violations (e.g., negative P)
- Delta: assumes constant absolute (T) or relative (P) bias → may break if processes change (e.g., frontal → convective)
- No-analog futures: new circulation regimes, unprecedented T–q combos, shifts in precipitation mechanisms → transfer functions may fail
6.2 Missing Processes
You cannot correct for processes the GCM doesn’t simulate. Garbage in, garbage out still applies.
Consider tropical cyclones as an example. Many global GCMs don’t resolve tropical cyclones, which require roughly 50 km horizontal resolution. Observations clearly show TCs in coastal regions like Houston and Miami, but the GCM shows smooth, large-scale precipitation instead.
What happens with QQ-mapping? - It will create intense events to match the observed distribution, but: - Timing wrong (seasonally smeared vs clustered TC events) - Location wrong (tracks absent/shifted) - Dynamics wrong (no rotation/RI/structure)
Your GCM simulates no tropical cyclones but observations have them. You apply QQ-mapping for a coastal location.
- What happens to the bias-corrected rainfall?
- Is the result useful for impact assessment?
- What alternatives might you consider?
Think about: statistical properties versus physical realism, temporal clustering, spatial structure.
6.3 Credibility Check and Scale Considerations
The GCM must credibly simulate the target variable at some level. - Good uses: correct drizzle bias when large-scale patterns are realistic; adjust temperature when circulation is credible; tweak timing/magnitude when seasonal cycle is captured - Poor uses: “fixing” missing processes; correcting where circulation is unrealistic; extrapolating far beyond historical range
You’re studying flood risk in a monsoon region. The GCM simulates monsoon onset 2 weeks late and underestimates peak rainfall by 30%.
- Can bias correction fix these problems?
- Which aspects can be corrected and which cannot?
- What diagnostic checks would you perform before trusting the bias-corrected output?
Think about: timing versus magnitude errors, process representation, validation strategies.
Sometimes a hierarchical approach is better. We can bias-correct the GCM at coarse scale for large-scale precipitation, then run a specialized process model like a TC model or convective model, and finally combine outputs for local impacts. This preserves GCM information where it is credible while using process models for sub-grid phenomena.
7 Looking Ahead
Minutes 44-45
7.1 Beyond Basic Methods
The delta method and QQ-mapping are starting points. More sophisticated methods exist and will be covered later in the course.
- Stochastic weather generators: multiple realizations conditioned on GCM; represent sub-monthly variability and uncertainty (Module 2)
- Weather typing, HMMs: classify states and link to local distributions; naturally probabilistic (next two weeks)
- Multivariate bias correction: adjust multiple variables jointly while preserving correlations (active research)
- Spatial consistency: univariate QM can break spatial coherence → rank-based approaches, optimal transport, spatial QDM
7.2 Preparing for Wednesday
On Wednesday we will discuss Ines and Hansen (2006) in detail.
Key questions to consider while reading:
- How well does bias correction improve crop yield simulations?
- What limitations remain even after bias correction?
- Why are yields under-predicted even with “perfect” rainfall distributions?
The paper provides a concrete case study of the methods we discussed today and highlights remaining challenges.