the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Remember the past: a comparison of timeadaptive training schemes for nonhomogeneous regression
Sebastian Lerch
Georg J. Mayr
Thorsten Simon
Reto Stauffer
Achim Zeileis
Nonhomogeneous regression is a frequently used postprocessing method for increasing the predictive skill of probabilistic ensemble weather forecasts. To adjust for seasonally varying error characteristics between ensemble forecasts and corresponding observations, different timeadaptive training schemes, including the classical sliding training window, have been developed for nonhomogeneous regression. This study compares three such training approaches with the slidingwindow approach for the application of postprocessing nearsurface air temperature forecasts across central Europe. The predictive performance is evaluated conditional on three different groups of stations located in plains, in mountain foreland, and within mountainous terrain, as well as on a specific change in the ensemble forecast system of the European Centre for MediumRange Weather Forecasts (ECMWF) used as input for the postprocessing.
The results show that timeadaptive training schemes using data over multiple years stabilize the temporal evolution of the coefficient estimates, yielding an increased predictive performance for all station types tested compared to the classical slidingwindow approach based on the most recent days only. While this may not be surprising under fully stable model conditions, it is shown that “remembering the past” from multiple years of training data is typically also superior to the classical slidingwindow approach when the ensemble prediction system is affected by certain model changes. Thus, reducing the variance of the nonhomogeneous regression estimates due to increased training data appears to be more important than reducing its bias by adapting rapidly to the most current training data only.
 Article
(2501 KB)  Fulltext XML

Supplement
(453 KB)  BibTeX
 EndNote
The need for accurate probabilistic weather forecasts is steadily increasing, because reliable information about the expected uncertainty is crucial for optimal risk assessment in agriculture and industry or for personal planning of outdoor activities. Therefore, most forecast centers nowadays issue probabilistic forecasts based on ensemble prediction systems (EPSs). To quantify the uncertainty of a specific forecast, an EPS provides a set of numerical weather predictions using slightly perturbed initial conditions and different model parameterizations (Palmer, 2002). However, due to various constraints and required simplifications in the EPS, these forecasts often show systematic biases and capture only parts of the expected uncertainty, especially when EPS forecasts are directly compared to point measurements (Gneiting and Katzfuss, 2014). In order to increase the predictive skill of the forecasts for specific locations, statistical postprocessing is often applied to correct for these systematic errors in the forecasts' expectation and uncertainty.
One of the most frequently used parametric postprocessing methods is “ensemble model output statistics” (EMOS) introduced by Gneiting et al. (2005). To emphasize that not only the errors in the mean but also the errors in the uncertainty are corrected, the method is often referred to as “nonhomogeneous regression” (NR). In the statistical literature, this type of model is also known as distributional regression (Klein et al., 2014) since all parameters of a specific response distribution are optimized simultaneously conditional on respective sets of covariates.
As the error characteristics between the covariates, typically provided by the EPS, and the observations often show seasonal dependencies and might change interannually over time, different timeadaptive training schemes have been developed for NR models. Gneiting et al. (2005) proposed the socalled “sliding training window” approach where the training data set consists of EPS forecasts and observations of the most recent 30–60 d only. As soon as new data become available, the training data set and the statistical model are updated so that the estimated coefficients automatically evolve over time and adjust to changing error characteristics. This makes it very handy for operational use; however, little training data can sometimes yield unrealistic jumps in the estimated coefficients over time, especially if events which show a significantly different error characteristic enter the training data set. Therefore, to stabilize the temporal variability of the coefficient estimates, several approaches have been proposed in the literature. Scheuerer (2014) regularizes the estimation by only allowing the optimizer to slightly adjust the coefficient from day to day. In an alternative approach, Möller et al. (2018) extend the training data by using not only the days prior to estimation, but also the days centered around the same calendar day over all previous years available. This idea of using a rolling centered training data set over multiple years is similar to the concept of using annual cyclic smooth functions to capture seasonality as employed by Lang et al. (2019). These smooth functions are also known as regression splines (Wood, 2017), where the estimate of each point in the function only depends on data in its closer neighborhood; this allows for a smooth and stable evolution of the coefficients over the year.
Alternative timeadaptive models are based on historical analogs or nonparametric approaches. For approaches employing analogs (Junk et al., 2015; Barnes et al., 2019), training sets are selected to consist of past forecast cases with atmospheric conditions similar to those on the day of interest. Such methods may lead to models that are able to account for the flow dependency of EPS errors (Pantillon et al., 2018; Rodwell et al., 2018). However, the definition and computation of similarity measures are far from straightforward, and substantial methodological developments may be required to obtain suitably extensive training data sets for stable model estimation (Hamill et al., 2008; Lerch and Baran, 2017). For nonparametric approaches (Taillardat et al., 2016; Henzi et al., 2019) or semiparametric approaches (Rasp and Lerch, 2018; Schlosser et al., 2019), timeadaptive choices of the training data are typically abandoned as well, as interactions between the day of the year and other covariates can capture the potential time adaptiveness. Therefore, analogbased and nonparametric approaches will not be pursued further in the context of this work.
In addition to the training scheme employed, an important dataspecific aspect which has to be considered in postprocessing is that the EPS may change over time (Hamill, 2018). This also motivates the recent study of Demaeyer and Vannitsem (2019), which introduces the promising concept of a postprocessing method specifically dealing with model changes in a simplified physical setup. However, as stated by the authors, more research would be required to transfer their findings to real case scenarios. When using data of an operational EPS, changes in the underlying numerical model, e.g., an increased horizontal resolution, can typically lead to sudden transitions in the predictive performance of the EPS and hence affect the error characteristics of the data. If the training data set used to estimate the statistical postprocessing model contains data of a previous EPS version which significantly differs from the current one, it can result in a loss of the predictive performance.
This paper presents a comparison of four widely used different timeadaptive training schemes proposed in the literature that employ alternative strategies to account for varying error characteristics in the data. To show a wide spectrum of possible approaches in a unified setup – rather than finding the universally best method – we consider typical basic applications of these training schemes and refrain from more elaborate tuning or combinations. A case study is shown for postprocessed 2 m temperature forecasts for three different groups of stations across central Europe at the midlatitudes, namely, stations in the plain, in the foreland, and within mountainous terrain (Fig. 1). The study highlights the advantages and drawbacks of the different approaches in different topographical environments and investigates the impact of a change in the horizontal resolution of the EPS, which is expected to have a particularly pronounced effect on the predictive performance.
The structure of the paper is as follows: Sect. 2 explains the different methods and the comparison setup including the underlying data. In Sect. 3, the different timeadaptive training schemes are compared in terms of their coefficient paths and their predictive performance. Finally, a summary and conclusion are given in Sect. 4.
The different training schemes for NR models proposed in the literature try to adapt to various kinds of error sources that can occur in postprocessing, both in space and time. In order to provide a unifying view and to fix jargon, we first discuss these different error sources and then introduce the training schemes considered along with the comparison setup employed.
2.1 Sources of errors in postprocessing
NR models aim to adjust for errors and biases in EPS forecasts but, of course, the NR models can be affected by errors and misspecifications themselves. Therefore, we try to carefully distinguish between the two different models involved with their associated errors, i.e., the numerical weather prediction model underlying the EPS vs. the statistical NR model employed for postprocessing.
The skill of the EPS can be quantified in EPS forecast biases and variances, which (i) typically vary for different locations conditional on the surrounding terrain, (ii) often show cyclic seasonal patterns, and (iii) can experience nonseasonal temporal changes, e.g., due to changes in the EPS itself.
In addition to the error sources in the employed EPS, the performance of the statistical postprocessing itself will typically also (iv) differ at different measurement sites, (v) strongly depend on the amount of training data used, and (vi) whether it is affected by effects that are not accounted for in the NR specification.
Clearly, larger training samples (v) will lead to more reliable predictions when the NR specification (vi) – in terms of response distribution, covariates and corresponding effects, link functions, estimation method, etc. – appropriately captures the error characteristics in the relationship between EPS forecasts and actual observations. However, when these error characteristics differ in space (i and iv) and/or in time (ii and iii), it is not obvious what the best strategy for training the NR is. Extending the training data (v) in space or time will reduce the variance of the NR estimation but might also introduce bias if the NR specification (vi) is not adapted. Thus, this is a classical biasvariance tradeoff problem, and we investigate which strategies for dealing with this are most useful in a typical temperature forecasting situation.
To fix jargon, we employ the terms “model” and “bias” without further qualifiers when referring to the NR model in postprocessing, whereas when referring to the numerical weather prediction model we employ “EPS model“ and “EPS bias”. Moreover, we refer to a statistical model whose estimates have small bias and variance as stable.
2.2 Nonhomogeneous regression with timeadaptive training schemes
Nonhomogeneous regression as originally introduced by Gneiting et al. (2005) is a special case of distributional regression, where a response variable y is assumed to follow a specific probability distribution 𝒟 with distribution parameters ${\mathit{\theta}}_{k},k=\mathrm{1},\mathrm{\dots},K$:
where each parameter of the distribution is linked to an additive predictor η_{k} via a link function h_{k} to ensure its appropriate codomain. In the case of postprocessing air temperatures, the normal distribution is typically employed (Gneiting and Katzfuss, 2014), and Eq. (1) can be rewritten as
In the classical NR (Gneiting et al., 2005), the two distribution parameters location μ and scale σ are expressed by the ensemble mean m and ensemble variance or standard deviation s, respectively:
with β_{•} and γ_{•} being the corresponding intercept and slope coefficients. Here, we use the logarithm link to ensure positivity of the scale parameter σ; however, a quadratic link with additional parameter constraints for the coefficients as used by Gneiting et al. (2005) would also be feasible. In this study, we regard the statistical model specifications according to Eqs. (2)–(4), but all concepts of timeadaptive training schemes could easily be transferred to other response distributions 𝒟, to alternative link functions h(⋅), or to more complex additive predictors η with additional covariates.
The regression coefficients β_{•} and γ_{•} are estimated by minimizing a loss function over a training data set containing historical pairs of observations and EPS forecasts. In this study, we employ maximum likelihood estimation, which performs very similarly to minimizing the continuous ranked probability score (CRPS, Gneiting and Raftery, 2007) as used by Gneiting et al. (2005) when the response distribution is well specified (Gebetsberger et al., 2018). For a single observation y, the loglikelihood L of the normal distribution is given by
where ϕ(⋅) is the probability density function of the normal distribution. The coefficients β_{•} and γ_{•}, specified in Eqs. (3) and (4), are derived by minimizing the sum of negative loglikelihood contributions L over the training data. The larger the training data, the more stable the estimation in case the statistical model is well specified; however, if the covariate's skill varies either seasonally or nonseasonally over time, this leads to the biasvariance tradeoff between preferable large training data sets for stable estimation and the benefit of shorter training periods which allow one to adjust more rapidly to changes in the data or, to be precise, in the error characteristics of the data (see Sect. 2.1). In the following, four approaches are discussed on how to gain informative timeadaptive training data sets while ensuring a stable estimation.
2.2.1 Slidingwindow
The slidingwindow approach originally introduced by Gneiting et al. (2005) uses the most recent days prior to the day of interest as training data for estimation. For postprocessing 2 m temperature forecasts, Gneiting et al. (2005) found the best predictive performance for training periods between 30 and 45 d with substantial gains in increasing the training period beyond 30 d and slow but steady performance losses for training lengths beyond 45 d. According to Gneiting et al. (2005), the latter is presumably a result of seasonally varying EPS forecast biases.
In this study, we use a period of 40 d for the slidingwindow approach, which is a frequently used compromise (e.g., Baran and Möller, 2017; Gneiting et al., 2005; Wilson et al., 2007). However, as discussed in Gneiting et al. (2005), different training periods might perform better for distinct weather variables, locations, forecast steps, or model specifications. Common choices in the literature include training lengths between 15 and 100 d, for example, depending on whether the estimation of regression coefficients is performed stationspecifically or jointly for multiple locations at once.
2.2.2 Regularized slidingwindow
A regularized adaption of the classical slidingwindow approach was introduced by Scheuerer (2014) in order to stabilize the estimation based on early stopping in statistical learning. The motivation is that gradientbased optimizers adjust the starting values by iteratively taking steps in the direction of the steepest descent of a distinct loss function until some convergence condition is fulfilled. These steps are largest in the first iteration and get smaller towards the optimum. Thus, the most important adjustments are made during the first steps, while further adjustments often improve the fit to unimportant or even random features in the data, which can lead to wiggly coefficient paths over time and ultimately to an overfitting (Scheuerer, 2014).
Therefore, Scheuerer (2014) proposes to use the coefficients of the previous day as starting values and to stop the optimizer after a single iteration to stabilize the evolution of the coefficient estimates. A drawback of his approach is that it implies that the estimation never converges and, in the case of poor starting values or strong truly observed temporal changes in the data, the obtained coefficients might be incorrect (Scheuerer, 2014). For postprocessing precipitation amounts employing a leftcensored generalized extreme value distribution, Scheuerer (2014) obtained better results with regularized coefficients than without regularization.
For the regularized slidingwindow approach used in this study, we employ the quasiNewton Broyden–Fletcher–Goldfarb–Shanno (BFGS) algorithm as in Scheuerer (2014) and stop the optimizer after one single iteration. For the first time, we let the BFGS algorithm perform 10 iterations and use $({\mathit{\beta}}_{\mathrm{0}},{\mathit{\beta}}_{\mathrm{1}}{)}^{\top}\phantom{\rule{0.125em}{0ex}}=\phantom{\rule{0.125em}{0ex}}(\mathrm{0},\mathrm{1}{)}^{\top}$ as starting values in the location parameter μ and $({\mathit{\gamma}}_{\mathrm{0}},{\mathit{\gamma}}_{\mathrm{1}}{)}^{\top}\phantom{\rule{0.125em}{0ex}}=\phantom{\rule{0.125em}{0ex}}(\mathrm{0.1},\mathrm{1}{)}^{\top}$ as starting values in the scale parameter σ. According to Scheuerer (2014) a single iteration might not always provide the optimum degree of regularization; however, for the presented comparison study a single iteration yields a regularized setup which is on the opposite side of the possible model spectrum compared to the classical slidingwindow approach which runs until convergence. In comparison to Scheuerer (2014), we perform maximum likelihood estimation instead of CRPS minimization.
2.2.3 Slidingwindow plus
As already pointed out by Gneiting et al. (2005), training data from previous years could additionally be included in the slidingwindow approach to address seasonal effects. This should reduce the variance in the estimation of the regression coefficients, which stabilizes the evolution of the coefficients similarly to the regularized slidingwindow approach.
This idea has recently been pursued by Vogel et al. (2018) for the construction of climatological reference forecasts and by Möller et al. (2018) for a postprocessing approach based on Dvine copulas in which many more coefficients than in classical NR need to be estimated, making a more extensive training data set necessary. Their socalled “refined training data set” consists of the most 45 recent days prior to the day of interest plus 91 d centered around the same calendar day over all previous years available. Including multiple years yields more stable estimates, while, on the other hand, there is the tradeoff of losing the ability to quickly adjust to nonseasonal temporal changes in the EPS forecast biases. The approach of Möller et al. (2018) can be seen as timeadaptive version of the seasonal training proposed by Hemri et al. (2016), who consider training data sets comprised of days from all previous years within the same season (winter/summer).
In this study, to be comparable to the slidingwindow approach, we use the most recent 40 d prior to estimation and a respective 81 d interval centered around the day of interest over the previous years available in the training data.
2.2.4 Smooth model
If we reformulate the slidingwindow plus approach, it is very similar to fitting an annual cyclic smooth function where the points of the function only depend on data points in the closer neighborhood, specified by the slidingwindow length.
Cyclic smooth functions belong to the broader model class of generalized additive models (GAMs, Hastie and Tibshirani, 1986), which allow one to include potentially nonlinear effects in the linear predictors η. Smooth functions are also referred to as regression splines and are directly linked to the model parameters as additive terms in η. Introductory material for cyclic smooth functions conditional on the day of the year can be found in Lang et al. (2019), and a comprehensive summary of GAMs is given in Wood (2017).
To account for seasonal variations we only need to fit one single model, here called the smooth model, over a training data set with several years of data. The effects included allow the coefficients to smoothly evolve over the year, which leads to the following adaptations in Eq. (3) and (4) for the location μ and scale σ, respectively:
with m and s being the ensemble mean and ensemble standard deviation, respectively; β_{•} and γ_{•} are regression coefficients, and f_{•}(doy) and g_{•}(doy) employ cyclic regression splines conditional on the day of the year (Wood, 2017). The regression coefficients β_{0} and γ_{0}, as well as β_{1} and γ_{1}, are unconditional on the day of the year and can be interpreted as global intercept or slope coefficients, respectively.
2.3 Comparison setup
2.3.1 NR training schemes
The NR training schemes presented in Sect. 2.2 deal with the potential temporal error sources from Sect. 2.1 in different ways (see Table 1 for an overview). The classic slidingwindow employs the basic NR model equations from Eqs. (3) to (4) and avoids potential biases in the NR model estimation by using only very recent data from the same year and season. Compared to this, the regularized slidingwindow and slidingwindow plus approaches both try to stabilize the coefficient estimates by reducing the variance – either through regularized estimation (vi) or by considering multiple years (v). The smooth model differs from all of these by modifying both the model (vi) and data (v) specification, using the extended model specification from Eqs. (6) to (7) fitted by penalized estimation to a large data set comprising several years and all seasons.
Potential spatial differences (i) and (iv) are handled for all training schemes in the same way: the NR models are estimated separately for each station and subsequently evaluated in groups of terrain types (plain, foreland, alpine). The underlying EPS data – described subsequently – are the same for all NR training schemes and are thus affected by the same seasonal (ii) and nonseasonal changes (iii).
2.3.2 Data sets
For validation of the training schemes, we consider 2 m temperature ensemble forecasts and corresponding observations at 15 measurement sites located across Austria, Germany, and Switzerland. The sites are chosen to investigate the impact of potential error sources in space (i) and (iv), e.g., through varying discrepancies between the real and EPS topography. The data comprises three groups of five stations located either in plains, in mountain foreland, or within mountainous terrain (see Fig. 1). The estimated statistical models for stations Hamburg and Innsbruck, highlighted by symbols with white borders, are discussed in more detail in Sect. 3.1.
As covariates for Eqs. (3)–(7), we employ the ensemble mean m and the ensemble standard deviation s of bilinearly interpolated 2 m temperature forecasts issued by the global 50member EPS of the European Centre for MediumRange Weather Forecasts (ECMWF). We assess forecast steps from +12 to +72 h ahead at a 12hourly temporal resolution for the EPS run initialized at 00:00 UTC and use data from 8 March 2010 to 7 March 2019.
This period has been selected in order to investigate the impact of nonseasonal longterm changes in the EPS model (iii) that is not reflected in the NR model specifications; i.e., the horizontal resolution of the ECMWF EPS changed from the previous version (cycle 36r1; 26 January 2010) to the new version on 8 March 2016 (cycle 41r2). This specific model change was chosen among various others as it modifies the height of the terrain and, thus, likely introduces an EPS bias for temperature forecasts directly affecting the coefficient estimates; other changes such as modified model parameterizations or improvements in the analysis scheme are expected to have a minor impact on the postprocessing of 2 m temperatures. It is of specific interest how the slidingwindow plus and the smooth model are affected if the training period comprises data from both the “old EPS version” before the change in the horizontal resolution as well as the “new EPS version”. Thus, we construct three data sets with different validation periods that are either (a) not affected by this EPS model change at all, (b) start immediately after the model change, or (c) have some time lag after change.
To understand how this affects the different training schemes, we first illustrate in Fig. 2a how training and validation periods are selected for each scheme. For the three slidingwindow approaches, the NR models are reestimated every day as the validation date rolls through the validation period (hatched area). In contrast, the smooth model is estimated only once for the entire validation period based on a fixed training data period of 4 years prior to the validation period. For a fair comparison, the training data for the slidingwindow plus model are also restricted to 4 years prior to each validation date.
Now Fig. 2b illustrates how the three data sets A, B, and C are selected in relation to the EPS change on 8 March 2016.

Data set A. All models are trained and evaluated without being affected by the EPS change.

Data set B. All models start with a training period entirely before the EPS change but a validation period entirely after the change. However, for the slidingwindow and regularized slidingwindow approaches, the training period quickly rolls across the change point, and after 40 d they are not affected by it anymore. For slidingwindow plus the training data also roll into the new EPS version but still partially use data from the old EPS version. Finally, as the smooth model is only estimated once, it cannot adapt at all to the new EPS version.

Data set C. Effects from A and B are mixed so that the smooth model and the slidingwindow plus model use data from both the old and new EPS versions, while the classical slidingwindow and regularized slidingwindow models already use only data from the new EPS version.
The validation period is 2 years for A and B and 1 year for C. A total number of 731/730/365 NR models has to be estimated for the three slidingwindow approaches, while only 1/1/1 smooth model is required for data sets A/B/C per station and forecast step. The computation time for the various slidingwindow approaches is in the order of seconds, whereas the estimation of the smooth model, including full Markov chain Monte Carlo (MCMC) sampling, is in the order of minutes on a standard computer.
This section assesses the performance of the different timeadaptive training schemes. First, the temporal evolutions of the estimated coefficients are shown for two stations representative of one measurement site in the plains and one in mountainous terrain. Afterwards, the predictive performance of the training schemes is evaluated in terms of the CRPS conditional on the three data sets with and without the change in the horizontal resolution of the EPS (Fig. 2) and grouped for stations classified as topographically plain, mountain foreland, and alpine sites (Fig. 1).
3.1 Coefficient paths
Figure 3 shows the estimated coefficients for Innsbruck at forecast step +36 h conditional on the day of the year. The coefficient paths are plotted for the different timeadaptive training schemes for 2 years included in the validation period of data set A. The pronounced seasonal evolution of the coefficients for all training schemes shows that the EPS' forecast bias and skill varies seasonally, which makes a timeadaptive training scheme mandatory to capture these characteristics in the postprocessing. During summer, a slope coefficient β_{1} close to 1 in the location parameter μ and a high slope coefficient γ_{1} in the scale parameter σ indicate a better performance of the EPS compared to the cold season.
In comparison to the other timeadaptive training schemes, the classical slidingwindow approach (Fig. 3a, d, g, j) shows very strong outliers and an unstable temporal evolution for all coefficients with distinct differences during the 2 subsequent validation years; this is more pronounced for the scale parameter σ where the estimates seem to be more volatile than for the location parameter μ. All strategies extending the classical slidingwindow approach smooth the temporal evolution of the coefficients to a certain extent while maintaining the overall seasonal cyclic pattern. For the regularized slidingwindow approach (Fig. 3b, e, h, k), the stabilization strongly differs for the individual coefficients, and some of the estimated coefficients seem to need rather long to adapt during the transition periods; the latter could indicate that a single iteration step might not be sufficient in this study. The coefficient paths for the slidingwindow plus approach (Fig. 3c, f, i, l) and for the smooth model (Fig. 3a–l; solid line) look very similar with minor distortions during the cold season. Due to the definition of the smooth model, its coefficient paths show the most stable evolution but with the lowest ability to react to abrupt changes in the error characteristics.
For Hamburg (Fig. 4) by contrast to Innsbruck, the information content of the mean EPS temperature forecast is quite high throughout the year. This yields a lower bias correction and an almost onetoone mapping of the ensemble mean to the location parameter μ indicated by a coefficient β_{1} close to 1. Despite the different postprocessing characteristics, the temporal evolution of the coefficient paths is similar to the one for Innsbruck, which confirms our previous findings: for the extended slidingwindow approaches the coefficients have indeed very little seasonal variability, while for the classical slidingwindow approach the coefficients show unrealistically strong fluctuations over time without a clear seasonal pattern (Fig. 4a, d, g, j). As for Innsbruck, the regularized slidingwindow approach has a rather unrealistic stepwise evolution for some coefficients (Fig. 4b, e, h, k). The coefficient paths for the slidingwindow plus approach (Fig. 4c, f, i, l) and the smooth model (Fig. 4; solid line) look comparable. These results support the biasvariance tradeoff where regularizing or smoothing stabilizes the coefficient paths while losing the ability to rapidly react to temporal changes in the data.
3.2 Predictive performance
After the illustrative evaluation of the coefficients' temporal evolution for the different timeadaptive training schemes, Fig. 5 shows aggregated CRPS skill scores for groups of five respective stations classified as topographically plain, mountain foreland, and alpine sites (Fig. 1) regarding data sets A, B, and C (Fig. 2). In all panels the regularized slidingwindow approach, the slidingwindow plus approach, and the smooth model are compared to the classical slidingwindow approach as a reference.

For data set A, the regularized slidingwindow approach shows only little improvements for the plain and foreland and an overall performance loss for alpine stations. By contrast, the slidingwindow plus and smooth model approaches show distinct improvements over the classical slidingwindow approach, with the largest values for alpine sites.

For data set B at stations in the plains and foreland, the mean predictive skill behaves similarly to data set A, except that the smooth model shows a slightly larger variance. For alpine stations, the regularized slidingwindow approach performs slightly worse than in data set A, while the two approaches using training data over multiple years no longer outperform the reference.

For data set C at stations in the plains and foreland, the predictive skill is again similar to data set A with slight performance losses. For alpine stations, the regularized slidingwindow approach shows even less skill than in data set B, while the two other approaches again outperform the slidingwindow approach and are on a similar level to that in data set A.
The validation of the different timeadaptive training schemes shows that the slidingwindow plus approach and the smooth model perform overall similarly and are clearly superior for all station types compared to the classical slidingwindow approach. However, the smooth model shows the highest variance in the predictive performance in the case of a change in the EPS, especially in mountainous terrain (data sets B and C); this is likely due to its reduced ability to adapt to temporal changes in the data. Furthermore, the validation shows that the regularized slidingwindow approach seems to have difficulties in mountainous terrain and yields only minor improvements for plain and foreland sites.
Nonhomogeneous regression (NR) is a widely used method to statistically postprocess ensemble weather forecasts. In its original version it was used for temperature forecasts employing a Gaussian response distribution, but over the last decade various statistical model extensions have been proposed for other quantities employing different response distributions or to enhance its predictive performance. When estimating NR models there is always a tradeoff between large enough training data sets to get stable estimates and still allowing the statistical model to adjust to temporal changes in the statistical error characteristics of the data. Therefore, different training schemes with specific advantages and drawbacks have been developed as presented in this paper. To show a wide spectrum of possible approaches in a unified setup, we consider typical basic applications of the training schemes and refrain from more elaborate tuning or combinations.
The classical slidingwindow approach has the advantage that no extensive training data set is required, which allows the statistical model to adjust itself rapidly to changing forecast biases, for example, in the case of changes in the EPS. On the other hand, statistical models trained on a small training data set have typically large variance in the estimation of the regression coefficients, which can yield unstable wiggly coefficient paths. Additional regularization allows one to stabilize the evolution of the regression coefficients without losing the simplicity of the classical slidingwindow approach. However, inappropriate settings of the optimizer, e.g., unrealistic starting values or insufficient update steps, can quickly lead to incorrect coefficients. The alternative slidingwindow plus strategy foregoes regularization but stabilizes the coefficients by using an extended training data set which includes data from the same season over several years. Compared to the classical approach the method requires historical data and partially loses its ability to rapidly adjust to changes in the error characteristics. The last approach presented in this paper can be seen as a generalization of the slidingwindow plus approach. Rather than using a training data set centered around the date of interest, the smooth model makes use of all historical data in combination with cyclic regression splines, which allows the coefficients to smoothly evolve over the year.
The differences between the methods presented can be seen in the coefficient paths shown in Figs. 3 and 4. The coefficients of the classical slidingwindow approach show strong fluctuations and pronounced peaks throughout the year. Regularization allows one to stabilize the evolution; however, strong stepwise changes in the coefficient paths still occur. The two methods using data from multiple years perform comparably similarly and show stable coefficient paths over the year. Figure 5 confirms that more stable estimates have a positive impact on the predictive performance. The slidingwindow plus approach and the smooth model show an overall improvement of about 3 %–5 % (in median) over the classical slidingwindow approach, while the regularized slidingwindow only partially outperforms the slidingwindow training scheme. Even in the case of the model change chosen to demonstrate the effect of nonseasonal longterm changes on the coefficient estimates, the training schemes using multiple years of data are still superior to the ones using the most recent days only, even if they technically allow adjustment to the EPS change more rapidly.
To conclude, all four training schemes shown in this paper have their advantages in particular applications. If only short periods of training data are available (<1 year), the classical slidingwindow approach may already provide sufficiently good estimates. However, as soon as one has access to longer historical data sets, the approaches using data from multiple years become superior due to a more stable coefficient evolution over time, which yields an overall improved performance. This even holds in the case of the EPS change considered in this study, but may be different for other changes or EPSs. While the slidingwindow plus approach is a natural extension of the classical slidingwindow approach and, therefore, can be estimated by the same software, the smooth model approach can be seen as a generalization, and only a single model has to be estimated for all seasons using all available data. The smooth model yields, by definition, the smoothest and most stable coefficient paths but with the lowest ability to adjust itself to a new error characteristic.
All computations are performed in R 3.6.1 (R Core Team, 2019) https://www.Rproject.org/ (last access: 10 December 2019). The statistical models using a slidingwindow approach are based on R package crch (Messner et al., 2016) (https://doi.org/10.32614/RJ2016012) employing a frequentist maximum likelihood approach. The statistical models using a timeadaptive training scheme by fitting cyclic smooth functions are fitted with R package bamlss (Umlauf et al., 2018) (https://doi.org/10.1080/10618600.2017.1407325). The package provides a flexible toolbox for distribution regression models in a Bayesian framework; introductory material can be found at http://BayesR.RForge.Rproject.org/ (last access: 10 December 2019). The computation of the CRPS is based on R package scoringRules (Jordan et al., 2019) (https://doi.org/10.18637/jss.v090.i12).
The supplement related to this article is available online at: https://doi.org/10.5194/npg27232020supplement.
This study is based on the PhD work of MNL under supervision of GJM and AZ. The majority of the work for this study was performed by MNL with the support of RS. All the authors worked closely together in discussing the results and commenting on the manuscript.
Sebastian Lerch is one of the editors of the special issue on “Advances in postprocessing and blending of deterministic and ensemble forecasts”. The remaining authors declare that they have no conflict of interest.
This article is part of the special issue “Advances in postprocessing and blending of deterministic and ensemble forecasts”. It is not associated with a conference.
We thank the Zentralanstalt für Meteorologie und Geodynamik (ZAMG) for providing access to the data.
This project was partly funded by the Austrian Research Promotion Agency (FFG, grant no. 858537) and by the Austrian Science Fund (FWF, grant no. P31836). Sebastian Lerch gratefully acknowledges support by the Deutsche Forschungsgemeinschaft (DFG) through SFB/TRR 165 “Waves to Weather”.
This paper was edited by Maxime Taillardat and reviewed by two anonymous referees.
Baran, S. and Möller, A.: Bivariate Ensemble Model Output Statistics Approach for Joint Forecasting of Wind Speed and Temperature, Meteorol. Atmos. Phys., 129, 99–112, https://doi.org/10.1007/s0070301604678, 2017. a
Barnes, C., Brierley, C. M., and Chandler, R. E.: New approaches to postprocessing of multimodel ensemble forecasts, Q. J. Roy. Meteor. Soc., 145, 3479–3498, https://doi.org/10.1002/qj.3632, 2019. a
Demaeyer, J. and Vannitsem, S.: Correcting for Model Changes in Statistical PostProcessing – An approach based on Response Theory, Nonlin. Processes Geophys. Discuss., https://doi.org/10.5194/npg201957, in review, 2019. a
Gebetsberger, M., Messner, J. W., Mayr, G. J., and Zeileis, A.: Estimation Methods for Nonhomogeneous Regression Models: Minimum Continuous Ranked Probability Score versus Maximum Likelihood, Mon. Weather Rev., 146, 4323–4338, https://doi.org/10.1175/MWRD170364.1, 2018. a
Gneiting, T. and Katzfuss, M.: Probabilistic Forecasting, Annu. Rev. Stat. Appl., 1, 125–151, https://doi.org/10.1146/annurevstatistics062713085831, 2014. a, b
Gneiting, T. and Raftery, A. E.: Strictly Proper Scoring Rules, Prediction, and Estimation, J. Am. Stat. Assoc., 102, 359–378, https://doi.org/10.1198/016214506000001437, 2007. a
Gneiting, T., Raftery, A. E., Westveld III, A. H., and Goldman, T.: Calibrated Probabilistic Forecasting Using Ensemble Model Output Statistics and Minimum CRPS Estimation, Mon. Weather Rev., 133, 1098–1118, https://doi.org/10.1175/MWR2904.1, 2005. a, b, c, d, e, f, g, h, i, j, k, l
Hamill, T. M.: Practical Aspects of Statistical Postprocessing, in: Statistical Postprocessing of Ensemble Forecasts, edited by: Vannitsem, S., Wilks, D. S., and Messner, J. W., 187–217, Elsevier, https://doi.org/10.1016/C20160032448, 2018. a
Hamill, T. M., Hagedorn, R., and Whitaker, J. S.: Probabilistic forecast calibration using ECMWF and GFS ensemble reforecasts. Part II: Precipitation, Mon. Weather Rev., 136, 2620–2632, https://doi.org/10.1175/2007MWR2411.1, 2008. a
Hastie, T. and Tibshirani, R.: Generalized Additive Models, Stat. Sci., 1, 297–310, 1986. a
Hemri, S., Haiden, T., and Pappenberger, F.: Discrete Postprocessing of Total Cloud Cover Ensemble Forecasts, Mon. Weather Rev., 144, 2565–2577, https://doi.org/10.1175/mwrd150426.1, 2016. a
Henzi, A., Ziegel, J. F., and Gneiting, T.: Isotonic Distributional Regression, arXiv 1909.03725, arXiv.org EPrint Archive, available at: http://arxiv.org/abs/1909.03725, last access: 10 December 2019. a
Jordan, A., Krüger, F., and Lerch, S.: Evaluating Probabilistic Forecasts with scoringRules, J. Stat. Softw., 90, 1–37, https://doi.org/10.18637/jss.v090.i12, 2019. a
Junk, C., Monache, L. D., and Alessandrini, S.: AnalogBased Ensemble Model Output Statistics, Mon. Weather Rev., 143, 2909–2917, https://doi.org/10.1175/mwrd150095.1, 2015. a
Klein, N., Kneib, T., Klasen, S., and Lang, S.: Bayesian Structured Additive Distributional Regression for Multivariate Responses, J. R. Stat. Soc. CAppl., 64, 569–591, https://doi.org/10.1111/rssc.12090, 2014. a
Lang, M. N., Mayr, G. J., Stauffer, R., and Zeileis, A.: Bivariate Gaussian models for wind vectors in a distributional regression framework, Adv. Stat. Clim. Meteorol. Oceanogr., 5, 115–132, https://doi.org/10.5194/ascmo51152019, 2019. a, b
Lerch, S. and Baran, S.: Similaritybased semilocal estimation of postprocessing models, J. R. Stat. Soc. CAppl., 66, 29–51, https://doi.org/10.1111/rssc.12153, 2017. a
Messner, J. W., Mayr, G. J., and Zeileis, A.: Heteroscedastic Censored and Truncated Regression with crch, R J., 8, 173–181, https://doi.org/10.32614/RJ2016012, 2016. a
Möller, A., Spazzini, L., Kraus, D., Nagler, T., and Czado, C.: Vine Copula Based PostProcessing of Ensemble Forecasts for Temperature, arXiv 1811.02255, arXiv.org EPrint Archive, available at: http://arxiv.org/abs/1811.02255 (last access: 10 December 2019), 2018. a, b, c
NASA JPL: NASA Shuttle Radar Topography Mission Global 30 Arc Second [Data Set], NASA EOSDIS Land Processes DAAC, https://doi.org/10.5067/MEaSUREs/SRTM/SRTMGL30.002, 2013. a
Palmer, T. N.: The Economic Value of Ensemble Forecasts as a Tool for Risk Assessment: From Days to Decades, Q. J. Roy. Meteor. Soc., 128, 747–774, https://doi.org/10.1256/0035900021643593, 2002. a
Pantillon, F., Lerch, S., Knippertz, P., and Corsmeier, U.: Forecasting Wind Gusts in Winter Storms Using a Calibrated ConvectionPermitting Ensemble, Q. J. Roy. Meteor. Soc., 144, 1864–1881, https://doi.org/10.1002/qj.3380, 2018. a
Rasp, S. and Lerch, S.: Neural Networks for Postprocessing Ensemble Weather Forecasts, Mon. Weather Rev., 146, 3885–3900, https://doi.org/10.1175/MWRD180187.1, 2018. a
R Core Team: R: A Language and Environment for Statistical Computing, R Foundation for Statistical Computing, Vienna, Austria, available at: https://www.Rproject.org/, last access: 10 December 2019. a
Rodwell, M. J., Richardson, D. S., Parsons, D. B., and Wernli, H.: Flowdependent reliability: A path to more skillful ensemble forecasts, B. Am. Meteorol. Soc., 99, 1015–1026, https://doi.org/10.1175/BAMSD170027.1, 2018. a
Scheuerer, M.: Probabilistic Quantitative Precipitation Forecasting Using Ensemble Model Output Statistics, Q. J. Roy. Meteor. Soc., 140, 1086–1096, https://doi.org/10.1002/qj.2183, 2014. a, b, c, d, e, f, g, h, i
Schlosser, L., Hothorn, T., Stauffer, R., and Zeileis, A.: Distributional Regression Forests for Probabilistic Precipitation Forecasting in Complex Terrain, Ann. Appl. Stat., 13, 1564–1589, https://doi.org/10.1214/19AOAS1247, 2019. a
Taillardat, M., Mestre, O., Zamo, M., and Naveau, P.: Calibrated Ensemble Forecasts Using Quantile Regression Forests and Ensemble Model Output Statistics, Mon. Weather Rev., 144, 2375–2393, https://doi.org/10.1175/mwrd150260.1, 2016. a
Umlauf, N., Klein, N., and Zeileis, A.: BAMLSS: Bayesian Additive Models for Location, Scale, and Shape (and Beyond), J. Comput. Graph. Stat., 27, 612–627, https://doi.org/10.1080/10618600.2017.1407325, 2018. a
Vogel, P., Knippertz, P., Fink, A. H., Schlueter, A., and Gneiting, T.: Skill of Global Raw and Postprocessed Ensemble Predictions of Rainfall over Northern Tropical Africa, Weather Forecast., 33, 369–388, https://doi.org/10.1175/wafd170127.1, 2018. a
Wilson, L. J., Beauregard, S., Raftery, A. E., and Verret, R.: Calibrated Surface Temperature Forecasts from the Canadian Ensemble Prediction System Using Bayesian Model Averaging, Mon. Weather Rev., 135, 1364–1385, https://doi.org/10.1175/MWR3347.1, 2007. a
Wood, S. N.: Generalized Additive Models: An Introduction with R, Chapman and Hall/CRC, https://doi.org/10.1201/9781315370279, 2017. a, b, c