Nonhomogeneous Boosting for Predictor Selection in Ensemble Postprocessing
read more
Citations
Various Versatile Variances: An Object-Oriented Implementation of Clustered Covariances in R
Neural Networks for Postprocessing Ensemble Weather Forecasts
Neural networks for post-processing ensemble weather forecasts.
BAMLSS: Bayesian Additive Models for Location, Scale and Shape (and Beyond)
Statistical Postprocessing for Weather Forecasts -- Review, Challenges and Avenues in a Big Data World
References
R: A language and environment for statistical computing.
Regression Shrinkage and Selection via the Lasso
The Elements of Statistical Learning: Data Mining, Inference, and Prediction
A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting
Statistical Methods in the Atmospheric Sciences
Related Papers (5)
Frequently Asked Questions (11)
Q2. What is the effect of boosting on the coe cients?
In addition to automatically selecting the most important input variables, boosting also regularizes the non-zero coe cients, i.e., the coe cients are shrunk compared to their maximumlikelihood values.
Q3. What is the CRPSS for dierent training data lengths?
For shorter training data lengths the number of selected input variables decreases but is still proportionally high compared to the training data length.
Q4. What is the coe cient of the daily mean maximum temperature ensemble mean?
After all coe cients being zero in the beginning, the daily mean maximum temperature ensemble mean (tmax2m dmean mean) is the first variable that gets a non-zero coe cients which indicates that it explains the observations best.
Q5. What is the coe cient for the log-scale?
After approximately 20 iterations the ensemble standard deviation of daily maximum evaporation (ske dmax sd) enters with a negative coe cient for the log-scale.
Q6. What is the RMSE of the forecasts?
While the RMSE shows the deterministic performance, the authors employ the continuous ranked probability score (CRPS; Hersbach 2000) to measure the probabilistic quality of the forecasts.
Q7. What is the main purpose of ensemble forecasts?
The set of potentially useful input variables is huge and includes, among others, ensemble forecasts for other variables or locations, deterministic forecasts, current observations, transformations and interactions of all of these.
Q8. What is the optimum forecast for the log-scale?
For the log-scale (Figure 3 bottom), ensemble standard deviations of various variables are selected but also ensemble mean forecasts (e.g., of 1000 hPa divergence d1000 dmax mean) seem to contain forecast uncertainty information.
Q9. What is the direct predictor of the temperature ensemble?
The direct predictor, the daily minimum minimum temperature ensemble mean, is clearly the most relevant variable over all lead times unlike for maximum temperatures.
Q10. What is the RMSE of the location forecasts?
Figure 5 shows the root mean squared error (RMSE) of the location forecasts (µ in Equation 2) of NGB, NGR, and the subset NGR, which is an NGR with the non-zero coe cients from boosting as input.
Q11. What are the main characteristics of the weather forecasting model?
many variables such as temperatures, have strongly pronounced seasonal patterns that probably a↵ect the statistical properties of forecasts and observations.