<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing with OASIS Tables v3.0 20080202//EN" "https://jats.nlm.nih.gov/nlm-dtd/publishing/3.0/journalpub-oasis3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:oasis="http://docs.oasis-open.org/ns/oasis-exchange/table" xml:lang="en" dtd-version="3.0" article-type="research-article"><?xmltex \bartext{Research article}?>
  <front>
    <journal-meta><journal-id journal-id-type="publisher">NPG</journal-id><journal-title-group>
    <journal-title>Nonlinear Processes in Geophysics</journal-title>
    <abbrev-journal-title abbrev-type="publisher">NPG</abbrev-journal-title><abbrev-journal-title abbrev-type="nlm-ta">Nonlin. Processes Geophys.</abbrev-journal-title>
  </journal-title-group><issn pub-type="epub">1607-7946</issn><publisher>
    <publisher-name>Copernicus Publications</publisher-name>
    <publisher-loc>Göttingen, Germany</publisher-loc>
  </publisher></journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.5194/npg-30-503-2023</article-id><title-group><article-title>Robust weather-adaptive post-processing using model output statistics random forests</article-title><alt-title>MOS random forests</alt-title>
      </title-group><?xmltex \runningauthor{T. Muschinski et al.}?><?xmltex \runningtitle{MOS random forests}?>
      <contrib-group>
        <contrib contrib-type="author" corresp="yes" rid="aff1 aff3">
          <name><surname>Muschinski</surname><given-names>Thomas</given-names></name>
          <email>thomas.muschinski@kit.edu</email>
        <ext-link>https://orcid.org/0000-0001-9084-0721</ext-link></contrib>
        <contrib contrib-type="author" corresp="no" rid="aff1">
          <name><surname>Mayr</surname><given-names>Georg J.</given-names></name>
          
        <ext-link>https://orcid.org/0000-0001-6661-9453</ext-link></contrib>
        <contrib contrib-type="author" corresp="no" rid="aff2">
          <name><surname>Zeileis</surname><given-names>Achim</given-names></name>
          
        <ext-link>https://orcid.org/0000-0003-0918-3766</ext-link></contrib>
        <contrib contrib-type="author" corresp="no" rid="aff2">
          <name><surname>Simon</surname><given-names>Thorsten</given-names></name>
          
        <ext-link>https://orcid.org/0000-0002-3778-7738</ext-link></contrib>
        <aff id="aff1"><label>1</label><institution>Department of Atmospheric and Cryospheric Sciences, Universität Innsbruck, Innsbruck, Austria</institution>
        </aff>
        <aff id="aff2"><label>2</label><institution>Department of Statistics, Universität Innsbruck, Innsbruck, Austria</institution>
        </aff>
        <aff id="aff3"><label>3</label><institution>Department of Economics, Statistical Methods and Econometrics, Karlsruhe Institute of Technology, Karlsruhe, Germany</institution>
        </aff>
      </contrib-group>
      <author-notes><corresp id="corr1">Thomas Muschinski (thomas.muschinski@kit.edu)</corresp></author-notes><pub-date><day>20</day><month>November</month><year>2023</year></pub-date>
      
      <volume>30</volume>
      <issue>4</issue>
      <fpage>503</fpage><lpage>514</lpage>
      <history>
        <date date-type="received"><day>16</day><month>May</month><year>2023</year></date>
           <date date-type="rev-request"><day>30</day><month>May</month><year>2023</year></date>
           <date date-type="rev-recd"><day>25</day><month>September</month><year>2023</year></date>
           <date date-type="accepted"><day>26</day><month>September</month><year>2023</year></date>
      </history>
      <permissions>
        <copyright-statement>Copyright: © 2023 Thomas Muschinski et al.</copyright-statement>
        <copyright-year>2023</copyright-year>
      <license license-type="open-access"><license-p>This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this licence, visit <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link></license-p></license></permissions><self-uri xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023.html">This article is available from https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023.html</self-uri><self-uri xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023.pdf">The full text article is available as a PDF file from https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023.pdf</self-uri>
      <abstract><title>Abstract</title>

      <p id="d1e122">Physical numerical weather prediction models have biases and miscalibrations that can depend on the weather situation, which makes it difficult to post-process them effectively using the traditional model output statistics (MOS) framework based on parametric regression models. Consequently, much recent work has focused on using flexible machine learning methods that are able to take additional weather-related predictors into account during post-processing beyond the forecast of the variable of interest only. Some of these methods have achieved impressive results, but they typically require significantly more training data than traditional MOS and are less straightforward to implement and interpret.</p>

      <p id="d1e125">We propose MOS random forests, a new post-processing method that avoids these problems by fusing traditional MOS with a powerful machine learning method called random forests to estimate weather-adapted MOS coefficients from a set of predictors.  Since the assumed parametric base model contains valuable prior knowledge, much smaller training data sizes are required to obtain skillful forecasts, and model results are easy to interpret.  MOS random forests are straightforward to implement and typically work well, even with no or very little hyperparameter tuning. For the difficult task of post-processing daily precipitation sums in complex terrain, they outperform reference machine learning methods at most of the stations considered. Additionally, the method is highly robust in relation to changes in data size and works well even when less than 100 observations are available for training.</p>
  </abstract>
    
<funding-group>
<award-group id="gs1">
<funding-source>Austrian Science Fund</funding-source>
<award-id>P 31836</award-id>
</award-group>
</funding-group>
</article-meta>
  </front>
<body>
      

<sec id="Ch1.S1" sec-type="intro">
  <label>1</label><title>Introduction</title>
      <p id="d1e137">Although physically based numerical weather predictions (NWPs) have made significant improvements in recent decades <xref ref-type="bibr" rid="bib1.bibx3" id="paren.1"/>, statistical post-processing is still necessary to correct systematic errors in the forecasts and accurately quantify their uncertainty <xref ref-type="bibr" rid="bib1.bibx39" id="paren.2"/>. The popular model output statistics (MOS) framework introduced by <xref ref-type="bibr" rid="bib1.bibx8" id="text.3"/> post-processes NWPs using linear regressions between historical observations and their corresponding predictions. Since then, the idea behind MOS has been extended to ensemble post-processing (EMOS) using more flexible regression models that allow for heteroscedastic forecast errors <xref ref-type="bibr" rid="bib1.bibx10" id="paren.4"><named-content content-type="pre">NGR,</named-content></xref> or non-Gaussian responses <xref ref-type="bibr" rid="bib1.bibx25 bib1.bibx33" id="paren.5"><named-content content-type="pre">e.g.,</named-content></xref>.</p>
      <?pagebreak page504?><p id="d1e159">Post-processing with MOS or EMOS is intuitive and can work well but requires a dataset that is both sufficiently large to allow for stable estimation of model coefficients and homogeneous enough for a single model with constant coefficients to work well. This means that the numerical weather model which is to be post-processed must have relatively constant systematic biases and miscalibrations.  In order to obtain such a homogeneous dataset, it is standard practice to estimate separate MOSs for different atmospheric quantities, locations, and lead times. Seasonal changes in predictability can be accounted for using time-adaptive MOSs that employ sliding-window training schemes <xref ref-type="bibr" rid="bib1.bibx10" id="paren.6"/> or by replacing constant model coefficients with cyclical functions of the day of the year <xref ref-type="bibr" rid="bib1.bibx16" id="paren.7"/>. This approach also works with other univariate predictors such as altitude <xref ref-type="bibr" rid="bib1.bibx29" id="paren.8"/>.</p>
      <p id="d1e171">Weather-adaptive post-processing – i.e., allowing biases and miscalibrations of the NWP model to depend on the weather situation – is necessary to obtain optimal forecast performance but is made complicated by the large number of potentially relevant atmospheric variables whose interactions are unknown or poorly understood.  It is possible to include such additional predictors in a MOS model by using selection procedures based on expert knowledge <xref ref-type="bibr" rid="bib1.bibx35" id="paren.9"/> or gradient boosting <xref ref-type="bibr" rid="bib1.bibx22" id="paren.10"/>, but this requires that the interactions are either ignored or parameterized a priori.</p>
      <p id="d1e180">Machine learning (ML) methods have become increasingly popular post-processing tools in recent years because they are well suited to dealing with this high-dimensional predictor space <xref ref-type="bibr" rid="bib1.bibx30" id="paren.11"/>.  Neural networks (NNs), for example, have been used in parametric distributional regressions similar to EMOS <xref ref-type="bibr" rid="bib1.bibx23" id="paren.12"/> and semi-parametric quantile function regressions based on Bernstein polynomials <xref ref-type="bibr" rid="bib1.bibx6" id="paren.13"/>. The predictive skill of NNs can be impressive, but they typically require combining data from many different stations to effectively train the model.  Purely local (station-wise) ML-based post-processing is often performed using random forests, which generally assume either a parametric distribution for the response <xref ref-type="bibr" rid="bib1.bibx26" id="paren.14"/> or predict a collection of specified quantiles <xref ref-type="bibr" rid="bib1.bibx36 bib1.bibx7" id="paren.15"/>, although combinations of the two have been employed as well <xref ref-type="bibr" rid="bib1.bibx37" id="paren.16"/>. Random forests have the advantage of being straightforward to implement, but they generally can only approximate linear (or other very smooth) functions by combining many (highly non-linear) step functions from individual trees. This may prove to be somewhat of a disadvantage in MOS applications, where the relationship between observations and model outputs is typically close to linear.</p>
      <p id="d1e203">MOS random forests (MOS forests for short) fuse traditional and ML-based post-processing by first assuming an appropriate parametric MOS model and then adapting its coefficients to the weather situation at hand using random forests.  The split variables and corresponding split points in the individual trees of a MOS forest are not selected based on properties of the response variable directly (e.g., their mean, quantiles, or other parameters), as done in quantile forests or distributional forests. Instead, the splits are chosen based on changes in the MOS coefficients of the assumed model, which may reflect either changes in the marginal distribution of the response (e.g., captured by intercepts) or changes in the dependence on the model outputs (e.g., captured by slopes).  The predictor space is thus partitioned to ensure homogeneity with respect to the MOS coefficients, meaning that a single model with constant coefficients can be assumed to work well in each corresponding subsample of the data.  In order to decrease variance and to allow for smooth dependencies, a MOS forest combines the partitions from many different MOS trees grown using bootstrapped or subsampled data <xref ref-type="bibr" rid="bib1.bibx4" id="paren.17"/> and only random subsets of predictor variables for splitting at each node <xref ref-type="bibr" rid="bib1.bibx5" id="paren.18"/>. Weather-adapted MOS coefficients predicted by the MOS forest can then be interpreted and used for post-processing in the usual way.</p>
      <p id="d1e212">A detailed description of MOS forests can be found in Sect. <xref ref-type="sec" rid="Ch1.S2"/>. In the following Sect. <xref ref-type="sec" rid="Ch1.S3"/>, MOS forests and reference methods are used to post-process ensemble predictions of daily precipitation sums in complex terrain.  The results of this real-world application are presented in Sect. <xref ref-type="sec" rid="Ch1.S4"/>. The strengths and limitations of the proposed method are discussed in Sect. <xref ref-type="sec" rid="Ch1.S5"/>, and summarizing remarks conclude the paper in Sect. <xref ref-type="sec" rid="Ch1.S6"/>.</p>
</sec>
<sec id="Ch1.S2">
  <label>2</label><title>MOS random forests</title>
      <p id="d1e233">MOS forests adapt the regression coefficients of an assumed (non-adaptive) base MOS to some set of additional atmospheric variables that characterize the current weather situation. Thus, it is first necessary to choose a suitable base MOS for the specific post-processing task at hand (Sect. <xref ref-type="sec" rid="Ch1.S2.SS1"/>). Subsequently, individual MOS trees are grown from this base MOS using model-based recursive partitioning algorithms which seek to identify homogeneous weather partitions of the predictor space within the tree's terminal nodes (Sect. <xref ref-type="sec" rid="Ch1.S2.SS2"/>).  Individual MOS trees already allow for weather-adaptive post-processing but can only approximate smooth effects through step functions with many splits. To better capture smooth effects and improve predictive performance, MOS forests therefore combine the partitions from not just one but many different MOS trees learned on random subsamples of the full data, yielding the final weather-adapted MOS (Sect. <xref ref-type="sec" rid="Ch1.S2.SS3"/>). This model can then be used for post-processing as usual.</p>
<sec id="Ch1.S2.SS1">
  <label>2.1</label><title>Choosing a base MOS</title>
      <p id="d1e249">The goal of MOS is to improve upon the quality of physical NWP models by identifying their weather-related statistics using regression models trained on historical observations and corresponding predictions <xref ref-type="bibr" rid="bib1.bibx8" id="paren.19"/>. Since MOS was first introduced 50 years ago, there have been substantial changes in both (i) what is meant by weather-related statistics in the context of MOS and (ii) the flexibility of the regression methods used to identify these.</p>
      <?pagebreak page505?><p id="d1e255">In the simplest case – with a single (deterministic) forecast for an atmospheric quantity and forecast errors that may be assumed to be Gaussian – systematic biases in the NWPs can be identified using a classical linear regression. A classical example is to regress observed temperatures <inline-formula><mml:math id="M1" display="inline"><mml:mi>y</mml:mi></mml:math></inline-formula> on the corresponding temperature predictions <inline-formula><mml:math id="M2" display="inline"><mml:mi>x</mml:mi></mml:math></inline-formula>:
            <disp-formula id="Ch1.E1" content-type="numbered"><label>1</label><mml:math id="M3" display="block"><mml:mrow><mml:mi>E</mml:mi><mml:mo>(</mml:mo><mml:mi>y</mml:mi><mml:mo>∣</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>⋅</mml:mo><mml:mi>x</mml:mi><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
          MOS coefficients <inline-formula><mml:math id="M4" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M5" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> then describe how the temperature forecast from the physical model should be corrected to better match real-world observations. For the ideal case of an NWP with no systematic biases, these values would be <inline-formula><mml:math id="M6" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mn mathvariant="normal">0</mml:mn></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M7" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:math></inline-formula>. In the classical linear model, coefficients are estimated by minimizing the sum of the squared errors (OLS) on some set of training data, which is equivalent to minimizing the root mean square error (RMSE) of the residuals.</p>
      <p id="d1e364">This simple post-processing model not only allows biases in the NWP to be corrected but also implicitly estimates the uncertainty of the post-processed forecast. Namely, if <inline-formula><mml:math id="M8" display="inline"><mml:mi>y</mml:mi></mml:math></inline-formula> can be assumed to follow a Gaussian distribution conditionally on <inline-formula><mml:math id="M9" display="inline"><mml:mi>x</mml:mi></mml:math></inline-formula>, the minimum RMSE obtained during model estimation is an estimate of the standard deviation <inline-formula><mml:math id="M10" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula> of the forecast distribution, and Eq. (<xref ref-type="disp-formula" rid="Ch1.E1"/>) may be rewritten as
            <disp-formula id="Ch1.E2" content-type="numbered"><label>2</label><mml:math id="M11" display="block"><mml:mtable class="split" rowspacing="0.2ex" displaystyle="true" columnalign="right left"><mml:mtr><mml:mtd><mml:mrow><mml:mi>y</mml:mi><mml:mo>∼</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mi mathvariant="script">N</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">μ</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mi mathvariant="italic">σ</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>)</mml:mo><mml:mo>,</mml:mo><mml:mspace width="1em" linebreak="nobreak"/><mml:mi mathvariant="normal">where</mml:mi><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">μ</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>⋅</mml:mo><mml:mi>x</mml:mi><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd/><mml:mtd><mml:mrow><mml:mi mathvariant="normal">and</mml:mi><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi>log⁡</mml:mi><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>.</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
      <p id="d1e472">Generally though, weather forecasts do not have constant uncertainty, and many atmospheric variables do not follow Gaussian distributions, even conditionally. To allow for more flexibility in post-processing, modern implementations of MOS therefore often employ distributional regressions <xref ref-type="bibr" rid="bib1.bibx15" id="paren.20"/>, also known as generalized additive models for location, scale, and shape <xref ref-type="bibr" rid="bib1.bibx24" id="paren.21"><named-content content-type="pre">GAMLSS,</named-content></xref>. In distributional regression, the observation <inline-formula><mml:math id="M12" display="inline"><mml:mi>y</mml:mi></mml:math></inline-formula> can follow some other parametric distribution, and all parameters (not just the mean) of this distribution are modeled on appropriate predictors derived from the NWP (ensemble).</p>
      <p id="d1e491">Typically, coefficients of distributional regression models are estimated by maximizing the log-likelihood <inline-formula><mml:math id="M13" display="inline"><mml:mi mathvariant="normal">ℓ</mml:mi></mml:math></inline-formula> of the distributional parameters given the observations or by minimizing the continuous ranked probability score (CRPS). One prominent example in the post-processing literature is the nonhomogeneous Gaussian regression (NGR) of <xref ref-type="bibr" rid="bib1.bibx10" id="text.22"/>, also known as EMOS, where the parameters <inline-formula><mml:math id="M14" display="inline"><mml:mi mathvariant="italic">μ</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M15" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula> in Eq. (<xref ref-type="disp-formula" rid="Ch1.E2"/>) are modeled on the mean and spread of an NWP ensemble, respectively. Other examples include truncated Gaussian and generalized extreme value response distributions for forecasting wind speed <xref ref-type="bibr" rid="bib1.bibx38 bib1.bibx17" id="paren.23"/> and censored and shifted gamma distributions for forecasting precipitation <xref ref-type="bibr" rid="bib1.bibx2" id="paren.24"/>.</p>
      <p id="d1e527">In the subsequent sections, we therefore assume that the base MOS for <inline-formula><mml:math id="M16" display="inline"><mml:mi>y</mml:mi></mml:math></inline-formula> explained by <inline-formula><mml:math id="M17" display="inline"><mml:mi>x</mml:mi></mml:math></inline-formula> uses some parametric model with likelihood <inline-formula><mml:math id="M18" display="inline"><mml:mrow><mml:mi mathvariant="normal">ℓ</mml:mi><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:mi>y</mml:mi><mml:mo>,</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M19" display="inline"><mml:mi>r</mml:mi></mml:math></inline-formula>-dimensional parameter vector <inline-formula><mml:math id="M20" display="inline"><mml:mi mathvariant="bold-italic">θ</mml:mi></mml:math></inline-formula> that is estimated through likelihood maximization:
            <disp-formula id="Ch1.E3" content-type="numbered"><label>3</label><mml:math id="M21" display="block"><mml:mrow><mml:mover accent="true"><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo mathvariant="normal" stretchy="false">^</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="normal">argmax</mml:mi><mml:mi mathvariant="bold-italic">θ</mml:mi></mml:msub><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:mi mathvariant="normal">ℓ</mml:mi><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo>)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
          In the example from Eq. (<xref ref-type="disp-formula" rid="Ch1.E2"/>), the likelihood is Gaussian with parameter vector <inline-formula><mml:math id="M22" display="inline"><mml:mrow><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo>=</mml:mo><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, but other distributions, like the ones from the previous paragraph, could be used in the same way.</p>
</sec>
<sec id="Ch1.S2.SS2">
  <label>2.2</label><title>Growing individual MOS trees</title>
      <p id="d1e684">In order to adapt the coefficients of the base MOS chosen in Sect. <xref ref-type="sec" rid="Ch1.S2.SS1"/> to some additional weather-related predictors <inline-formula><mml:math id="M23" display="inline"><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mi mathvariant="normal">…</mml:mi><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>, a single MOS tree partitions the predictor space  <inline-formula><mml:math id="M24" display="inline"><mml:mrow><mml:msub><mml:mi>Z</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>×</mml:mo><mml:msub><mml:mi>Z</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msub><mml:mo>×</mml:mo><mml:mi mathvariant="normal">…</mml:mi><mml:mo>×</mml:mo><mml:msub><mml:mi>Z</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> into disjointed subsets that can each be considered to be homogeneous weather situations for the purpose of NWP post-processing – i.e., where constant MOS coefficients work well. It is grown using model-based recursive partitioning algorithms <xref ref-type="bibr" rid="bib1.bibx40 bib1.bibx31" id="paren.25"/> according to the following steps.</p>
</sec>
<sec id="Ch1.S2.SSx1" specific-use="unnumbered">
  <title>Step 1: Estimate coefficients of the base MOS</title>
      <p id="d1e754">MOS coefficients <inline-formula><mml:math id="M25" display="inline"><mml:mi mathvariant="bold-italic">θ</mml:mi></mml:math></inline-formula> are estimated through likelihood maximization on the <inline-formula><mml:math id="M26" display="inline"><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>,</mml:mo><mml:mi mathvariant="normal">…</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:math></inline-formula>, observations <inline-formula><mml:math id="M27" display="inline"><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> and corresponding predictions <inline-formula><mml:math id="M28" display="inline"><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> in the dataset. This is done by solving the first-order condition
            <disp-formula id="Ch1.E4" content-type="numbered"><label>4</label><mml:math id="M29" display="block"><mml:mrow><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:mi>s</mml:mi><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mn mathvariant="bold">0</mml:mn><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
          where
            <disp-formula id="Ch1.E5" content-type="numbered"><label>5</label><mml:math id="M30" display="block"><mml:mrow><mml:mi>s</mml:mi><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:msup><mml:mfenced close=")" open="("><mml:mrow><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mo>∂</mml:mo><mml:mi mathvariant="normal">ℓ</mml:mi><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>∂</mml:mo><mml:msub><mml:mi mathvariant="italic">θ</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo><mml:mi mathvariant="normal">…</mml:mi><mml:mo>,</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mo>∂</mml:mo><mml:mi mathvariant="normal">ℓ</mml:mi><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>∂</mml:mo><mml:msub><mml:mi mathvariant="italic">θ</mml:mi><mml:mi>r</mml:mi></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow></mml:mfenced><mml:mo>⊤</mml:mo></mml:msup></mml:mrow></mml:math></disp-formula>
          contains the partial derivatives of the log-likelihood with respect to each coefficient – i.e., the model scores – evaluated at the <inline-formula><mml:math id="M31" display="inline"><mml:mi>i</mml:mi></mml:math></inline-formula>th observation pair <inline-formula><mml:math id="M32" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>.</p>
</sec>
<sec id="Ch1.S2.SSx2" specific-use="unnumbered">
  <title>Step 2: Select the splitting variable</title>
      <?pagebreak page506?><p id="d1e1016">Scores with respect to each coefficient are again computed at all observations (Eq. <xref ref-type="disp-formula" rid="Ch1.E5"/>) and evaluated at the estimated coefficients <inline-formula><mml:math id="M33" display="inline"><mml:mrow><mml:mover accent="true"><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo mathvariant="normal" stretchy="false">^</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mo>(</mml:mo><mml:mover accent="true"><mml:mrow><mml:msub><mml:mi mathvariant="italic">θ</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub></mml:mrow><mml:mo stretchy="false" mathvariant="normal">^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mi mathvariant="normal">…</mml:mi><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:msub><mml:mi mathvariant="italic">θ</mml:mi><mml:mi>r</mml:mi></mml:msub></mml:mrow><mml:mo mathvariant="normal" stretchy="false">^</mml:mo></mml:mover><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> from step 1. Since the estimated coefficients were obtained using Eq. (<xref ref-type="disp-formula" rid="Ch1.E4"/>), each score vector has a mean of zero.  If the single MOS with constant coefficients fits well, the scores for each observation should randomly fluctuate around zero.  On the other hand, systematic departures of the scores from zero along some of the variables in <inline-formula><mml:math id="M34" display="inline"><mml:mi mathvariant="bold-italic">z</mml:mi></mml:math></inline-formula> suggest that predictions can be improved by splitting the data and estimating separate post-processing models based on the two resulting subsamples. Whether or not the scores fluctuate randomly or depend on one of the weather-related predictors can be assessed using an independence test between the scores and each of the variables in <inline-formula><mml:math id="M35" display="inline"><mml:mi mathvariant="bold-italic">z</mml:mi></mml:math></inline-formula> <xref ref-type="bibr" rid="bib1.bibx12 bib1.bibx13" id="paren.26"><named-content content-type="pre">see permutation tests of</named-content></xref>.  If there is a significant dependence with respect to at least one of the variables, then the most significant variable (i.e., with the smallest <inline-formula><mml:math id="M36" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula> value) is selected for splitting. The underlying test statistic captures the overall dependence in all score components (i.e., all MOS coefficients) simultaneously using a quadratic form.  To account for assessing multiple variables from <inline-formula><mml:math id="M37" display="inline"><mml:mi mathvariant="bold-italic">z</mml:mi></mml:math></inline-formula>, a Bonferroni correction for multiple testing is employed.</p>
</sec>
<sec id="Ch1.S2.SSx3" specific-use="unnumbered">
  <title>Step 3: Identify the optimal split point</title>
      <p id="d1e1104">Once the splitting variable <inline-formula><mml:math id="M38" display="inline"><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> has been selected, an exhaustive search is performed over all possible split points to identify the partition that improves the log-likelihood the most. For numerical splitting variables, up to <inline-formula><mml:math id="M39" display="inline"><mml:mrow><mml:mn mathvariant="normal">2</mml:mn><mml:mo>⋅</mml:mo><mml:mo>(</mml:mo><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> different MOSs are estimated in this step – separate models in both subsamples for each of the <inline-formula><mml:math id="M40" display="inline"><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:math></inline-formula> possible split points. The number of possible split points (and thus estimated models) decreases for each tie among the realizations of <inline-formula><mml:math id="M41" display="inline"><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>. For unordered categorical splitting variables, the number of possible split points is equal to the number of ways in which the different categories can be divided into two subgroups and thus increases exponentially with the number of distinct categories.</p>
</sec>
<sec id="Ch1.S2.SSx4" specific-use="unnumbered">
  <title>Repeat previous steps</title>
      <p id="d1e1167">The three steps described above split a dataset of size <inline-formula><mml:math id="M42" display="inline"><mml:mi>N</mml:mi></mml:math></inline-formula> into two disjoint subsamples that are then each post-processed using a separate MOS. In order to grow a MOS tree, these steps are repeated for each subsample until a stopping criterion has been reached. The terminal nodes of a MOS tree (i.e., those nodes that are not split any further) contain disjointed subsamples of the full data that correspond to different homogeneous weather situations for post-processing with MOS (Figs. <xref ref-type="fig" rid="Ch1.F1"/> and <xref ref-type="fig" rid="Ch1.F2"/>).</p>
      <p id="d1e1181">Coefficients <inline-formula><mml:math id="M43" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">θ</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mi mathvariant="normal">…</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="italic">θ</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> in each terminal node are obtained through likelihood maximization on the corresponding subsample.  Note that this can also be understood as a weighted estimated using the full data, where weights are either 0 or 1, indicating whether or not the respective observation is in the subsample of interest. In the following Sect. <xref ref-type="sec" rid="Ch1.S2.SS3"/>, this idea is extended to use weights that may change smoothly (rather than abruptly) between 0 and 1. This can express the degree of similarity (with respect to MOS coefficients) between some new weather situation and those historical weather situations in the training data.</p>
</sec>
<sec id="Ch1.S2.SS3">
  <label>2.3</label><title>Obtaining weather-adapted coefficients from a random forest of MOS trees</title>
      <p id="d1e1216">Individual MOS trees grown according to Sect. <xref ref-type="sec" rid="Ch1.S2.SS2"/> are easy to understand and interpret (see Sect. <xref ref-type="sec" rid="Ch1.S4.SS1"/>) but can be sensitive to small changes in the data and may have a suboptimal fit if the model parameters change smoothly with the weather situation variables.  To solve this problem and improve out-of-sample predictive skill, a MOS forest combines partitions from many different trees grown on bootstrap-aggregated (bagged) data and using only a randomly chosen subset of the atmospheric variables in <inline-formula><mml:math id="M44" display="inline"><mml:mi mathvariant="bold-italic">z</mml:mi></mml:math></inline-formula> for splitting at each node.</p>
      <p id="d1e1230">Given a MOS forest with <inline-formula><mml:math id="M45" display="inline"><mml:mi>T</mml:mi></mml:math></inline-formula> trees and <inline-formula><mml:math id="M46" display="inline"><mml:mrow><mml:msup><mml:mi>P</mml:mi><mml:mi>t</mml:mi></mml:msup></mml:mrow></mml:math></inline-formula> partitions in each tree <inline-formula><mml:math id="M47" display="inline"><mml:mi>t</mml:mi></mml:math></inline-formula>, MOS coefficients are adapted to a new weather situation <inline-formula><mml:math id="M48" display="inline"><mml:mrow><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>∈</mml:mo><mml:msub><mml:mi>Z</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>×</mml:mo><mml:msub><mml:mi>Z</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msub><mml:mo>×</mml:mo><mml:mi mathvariant="normal">…</mml:mi><mml:mo>×</mml:mo><mml:msub><mml:mi>Z</mml:mi><mml:mi>r</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> by maximizing the likelihood of the base MOS in relation to the full training data, as in Eq. (<xref ref-type="disp-formula" rid="Ch1.E3"/>):
            <disp-formula id="Ch1.E6" content-type="numbered"><label>6</label><mml:math id="M49" display="block"><mml:mrow><mml:mover accent="true"><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo stretchy="false" mathvariant="normal">^</mml:mo></mml:mover><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="normal">argmax</mml:mi><mml:mi mathvariant="bold-italic">θ</mml:mi></mml:msub><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:mi>w</mml:mi><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mi>i</mml:mi></mml:msup><mml:mo>)</mml:mo><mml:mo>⋅</mml:mo><mml:mi mathvariant="normal">ℓ</mml:mi><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo>)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
          but with observations <inline-formula><mml:math id="M50" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> weighted according to
            <disp-formula id="Ch1.E7" content-type="numbered"><label>7</label><mml:math id="M51" display="block"><mml:mrow><mml:mi>w</mml:mi><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mi>i</mml:mi></mml:msup><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mn mathvariant="normal">1</mml:mn><mml:mi>T</mml:mi></mml:mfrac></mml:mstyle><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>T</mml:mi></mml:munderover><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>p</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mi>P</mml:mi><mml:mi>t</mml:mi></mml:msup></mml:mrow></mml:munderover><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mn mathvariant="double-struck">1</mml:mn><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>∈</mml:mo><mml:msubsup><mml:mi mathvariant="script">P</mml:mi><mml:mi>p</mml:mi><mml:mi>t</mml:mi></mml:msubsup><mml:mo>)</mml:mo><mml:mo>∧</mml:mo><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mi>i</mml:mi></mml:msup><mml:mo>∈</mml:mo><mml:msubsup><mml:mi mathvariant="script">P</mml:mi><mml:mi>p</mml:mi><mml:mi>t</mml:mi></mml:msubsup><mml:mo>)</mml:mo><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>∣</mml:mo><mml:msubsup><mml:mi mathvariant="script">P</mml:mi><mml:mi>p</mml:mi><mml:mi>t</mml:mi></mml:msubsup><mml:mo>∣</mml:mo></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
          These weights thus capture how similar the new weather situation <inline-formula><mml:math id="M52" display="inline"><mml:mrow><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mo>⋆</mml:mo></mml:msup></mml:mrow></mml:math></inline-formula> is to any of the historical weather situations <inline-formula><mml:math id="M53" display="inline"><mml:mrow><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mi>i</mml:mi></mml:msup></mml:mrow></mml:math></inline-formula> from the training data by computing how often they end up in the same homogenous weather partition from the different trees in the forest. Thus, they characterize their similarity with respect to the MOS coefficients.</p>
      <p id="d1e1553">By using partitions from many different trees to estimate the weather-adapted MOS, model coefficients are not restricted to a discrete number of unique values at most equal to the number of terminal nodes (as can be seen with estimates for <inline-formula><mml:math id="M54" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula> from the MOS tree of Fig. <xref ref-type="fig" rid="Ch1.F3"/>).  Instead, coefficients are allowed to have smooth dependencies on the additional predictors, and, as a result, predictions are more stable (see estimates for <inline-formula><mml:math id="M55" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula> from the MOS forest of Fig. <xref ref-type="fig" rid="Ch1.F3"/>).</p>
      <p id="d1e1574">The MOS coefficients <inline-formula><mml:math id="M56" display="inline"><mml:mrow><mml:mover accent="true"><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo mathvariant="normal" stretchy="false">^</mml:mo></mml:mover><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> that have been adapted to the new weather situation <inline-formula><mml:math id="M57" display="inline"><mml:mrow><mml:msup><mml:mi>z</mml:mi><mml:mo>⋆</mml:mo></mml:msup></mml:mrow></mml:math></inline-formula> can be used to post-process the corresponding forecast <inline-formula><mml:math id="M58" display="inline"><mml:mrow><mml:msup><mml:mi>x</mml:mi><mml:mo>⋆</mml:mo></mml:msup></mml:mrow></mml:math></inline-formula> in the same way as coefficients obtained from a MOS tree or from the base MOS itself. That is, the (log-transformed) probability density function for the unknown observation <inline-formula><mml:math id="M59" display="inline"><mml:mrow><mml:msup><mml:mi>y</mml:mi><mml:mo>⋆</mml:mo></mml:msup></mml:mrow></mml:math></inline-formula> is given by <inline-formula><mml:math id="M60" display="inline"><mml:mrow><mml:mi mathvariant="normal">ℓ</mml:mi><mml:mo>(</mml:mo><mml:msup><mml:mi>y</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>∣</mml:mo><mml:msup><mml:mi>x</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mi mathvariant="bold-italic">θ</mml:mi><mml:mo mathvariant="normal" stretchy="false">^</mml:mo></mml:mover><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">z</mml:mi><mml:mo>⋆</mml:mo></mml:msup><mml:mo>)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, and the parameters of the response distribution are those values predicted by the MOS.</p>
      <p id="d1e1671">Using neighborhood weights as described above is commonplace in forests that contain more complex models rather than just a single scalar value in the terminal nodes <xref ref-type="bibr" rid="bib1.bibx26 bib1.bibx1" id="paren.27"><named-content content-type="pre">e.g.,</named-content></xref>. An alternative approach would be to obtain the weather-adapted MOS model by averaging over MOS coefficients predicted by the individual trees. In the application considered here, the two performed equally well, except in the case of smaller sample sizes, where using the weights was slightly better.</p>
</sec>
</sec>
<?pagebreak page507?><sec id="Ch1.S3">
  <label>3</label><title>Post-processing precipitation forecasts in complex terrain</title>
      <p id="d1e1688">The MOS forests described in Sect. <xref ref-type="sec" rid="Ch1.S2"/> are applied to the difficult task of obtaining reliable probabilistic precipitation forecasts in complex terrain. Individual topographical features cannot be resolved by NWP models, which means that predictions for these locations rely heavily on subgrid-scale parameterizations whose accuracy can depend on the weather situation.  Postprocessing models are trained and evaluated on the <monospace>RainTyrol</monospace> dataset described in Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/>, which contains observations of daily precipitation sums and various ensemble-derived predictor variables that can be used for weather-adaptive post-processing. The exact configuration of the MOS forest for this application and a description of the three reference methods are given in Sect. <xref ref-type="sec" rid="Ch1.S3.SS2"/>.</p>
<sec id="Ch1.S3.SS1">
  <label>3.1</label><title>Data</title>
      <p id="d1e1707">The <monospace>RainTyrol</monospace> dataset <xref ref-type="bibr" rid="bib1.bibx26" id="paren.28"/> is composed of observed daily precipitation sums from the Austrian National Hydrographical Service and NWPs from the 11-member global ensemble forecast system <xref ref-type="bibr" rid="bib1.bibx11" id="paren.29"><named-content content-type="pre">GEFS,</named-content></xref> of the US National Oceanic and Atmospheric Administration (NOAA). Data are available at 95 different stations in Tyrol, Austria, (and surrounding border regions) for all July days between 1985 and 2012, except the missing day of 19 July 2011. July is a month with some of the largest precipitation amounts and variability within the year, and both large-scale precipitation events and local convective events occur. To reduce skewness, daily precipitation sums observed at 06:00 UTC (<monospace>robs</monospace>) are power transformed using a parameter value of <inline-formula><mml:math id="M61" display="inline"><mml:mrow><mml:mn mathvariant="normal">1</mml:mn><mml:mo>/</mml:mo><mml:mn mathvariant="normal">1.6</mml:mn></mml:mrow></mml:math></inline-formula>, which corresponds to the median of the power coefficients estimated at all stations <xref ref-type="bibr" rid="bib1.bibx34" id="paren.30"/>.</p>
      <p id="d1e1740">There are 80 different predictor variables derived from the GEFS that can be used for post-processing. These include the direct predictor of the observation: the mean of the ensemble forecast of total (24 h) precipitation between <inline-formula><mml:math id="M62" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">6</mml:mn></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M63" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">30</mml:mn></mml:mrow></mml:math></inline-formula> h but also the ensemble spread and its minimum and maximum. To account for the fact that summertime rainfall in Tirol is often caused by convection during the late afternoon and evening hours, ensemble statistics for the four sub-daily 6 h precipitation forecasts (<inline-formula><mml:math id="M64" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">6</mml:mn></mml:mrow></mml:math></inline-formula> to <inline-formula><mml:math id="M65" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">12</mml:mn></mml:mrow></mml:math></inline-formula>, <inline-formula><mml:math id="M66" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">12</mml:mn></mml:mrow></mml:math></inline-formula> to <inline-formula><mml:math id="M67" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">18</mml:mn></mml:mrow></mml:math></inline-formula>, <inline-formula><mml:math id="M68" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">18</mml:mn></mml:mrow></mml:math></inline-formula> to <inline-formula><mml:math id="M69" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">24</mml:mn></mml:mrow></mml:math></inline-formula>, and <inline-formula><mml:math id="M70" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">24</mml:mn></mml:mrow></mml:math></inline-formula> to <inline-formula><mml:math id="M71" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">30</mml:mn></mml:mrow></mml:math></inline-formula> h) are also used as predictors. The same variations are also included for forecasts of the convective available potential energy (CAPE), a key ingredient in thunderstorms. Forecasts of temperature and temperature differences at and between different heights, as well as incoming solar radiation (i.e., sunshine), pressure, precipitable water, and total column-integrated condensate, are also added. Predictors derived from these atmospheric variables are not included for every sub-forecast, but the ensemble means and spreads are temporally aggregated using the minimum, maximum, or mean. For example, <monospace>pwat_mean_max</monospace> refers to the maximum ensemble mean of precipitable water forecasted by the GEFS for a lead time between <inline-formula><mml:math id="M72" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">6</mml:mn></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M73" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">30</mml:mn></mml:mrow></mml:math></inline-formula> h. A thorough description of all available predictor variables and their naming conventions can be found in Table 1 of <xref ref-type="bibr" rid="bib1.bibx26" id="text.31"/>.</p>
</sec>
<sec id="Ch1.S3.SS2">
  <label>3.2</label><title>Methods</title>
      <p id="d1e1879">The ensemble forecasts described in Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/> are post-processed using MOS forests, two other forest-based weather-adaptive reference methods, and a non-adaptive EMOS. An overview of the methods is given in Table <xref ref-type="table" rid="Ch1.T1"/>, and more details are supplied below.</p>

<?xmltex \floatpos{t}?><table-wrap id="Ch1.T1" specific-use="star"><?xmltex \currentcnt{1}?><label>Table 1</label><caption><p id="d1e1889">Overview of methods used to post-process precipitation forecasts from the <monospace>RainTyrol</monospace> dataset (see Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/>). In the dataset, variable names <monospace>tppow_mean</monospace> and <monospace>tppow_sprd</monospace> refer to the mean and standard deviation of the power-transformed ensemble forecasts of total precipitation, respectively.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="5">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="left"/>
     <oasis:colspec colnum="3" colname="col3" align="left"/>
     <oasis:colspec colnum="4" colname="col4" align="left"/>
     <oasis:colspec colnum="5" colname="col5" align="left"/>
     <oasis:thead>
       <oasis:row>
         <oasis:entry colname="col1">Model name</oasis:entry>
         <oasis:entry colname="col2">Forecast type</oasis:entry>
         <oasis:entry rowsep="1" namest="col3" nameend="col4">Prespecified regression model for </oasis:entry>
         <oasis:entry colname="col5">Splitting variables</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2"/>
         <oasis:entry colname="col3">Location: <inline-formula><mml:math id="M74" display="inline"><mml:mi mathvariant="italic">μ</mml:mi></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col4">Scale: <inline-formula><mml:math id="M75" display="inline"><mml:mrow><mml:mi>log⁡</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col5"/>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">MOS forest</oasis:entry>
         <oasis:entry colname="col2">censored Gaussian</oasis:entry>
         <oasis:entry colname="col3"><inline-formula><mml:math id="M76" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>⋅</mml:mo></mml:mrow></mml:math></inline-formula><monospace>tppow_mean</monospace></oasis:entry>
         <oasis:entry colname="col4"><inline-formula><mml:math id="M77" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col5">all, except <monospace>tppow_mean</monospace></oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Distributional forest</oasis:entry>
         <oasis:entry colname="col2">censored Gaussian</oasis:entry>
         <oasis:entry colname="col3"/>
         <oasis:entry colname="col4"/>
         <oasis:entry colname="col5">all</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Quantile regression forest</oasis:entry>
         <oasis:entry colname="col2">set of quantiles</oasis:entry>
         <oasis:entry colname="col3"/>
         <oasis:entry colname="col4"/>
         <oasis:entry colname="col5">all</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">EMOS</oasis:entry>
         <oasis:entry colname="col2">censored Gaussian</oasis:entry>
         <oasis:entry colname="col3"><inline-formula><mml:math id="M78" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>⋅</mml:mo></mml:mrow></mml:math></inline-formula><monospace>tppow_mean</monospace></oasis:entry>
         <oasis:entry colname="col4"><inline-formula><mml:math id="M79" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>⋅</mml:mo><mml:mi>log⁡</mml:mi></mml:mrow></mml:math></inline-formula>(<monospace>tppow_sprd</monospace>)</oasis:entry>
         <oasis:entry colname="col5"/>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table><?xmltex \gdef\@currentlabel{1}?></table-wrap>

<sec id="Ch1.S3.SS2.SSS1">
  <label>3.2.1</label><title>MOS forests</title>
      <p id="d1e2123">To deal with the fact that precipitation sums are strictly non-negative, we follow <xref ref-type="bibr" rid="bib1.bibx26" id="text.32"/> and assume a left-censored Gaussian response distribution with log-likelihood given by
              <disp-formula id="Ch1.E8" content-type="numbered"><label>8</label><mml:math id="M80" display="block"><mml:mrow><mml:mi mathvariant="normal">ℓ</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">μ</mml:mi><mml:mo>,</mml:mo><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>;</mml:mo><mml:mi>y</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mfenced open="{" close=""><mml:mtable class="cases" rowspacing="0.2ex" columnspacing="1em" columnalign="left left" framespacing="0em"><mml:mtr><mml:mtd><mml:mrow><mml:mi>log⁡</mml:mi><mml:mfenced open="(" close=")"><mml:mrow><mml:mstyle displaystyle="false"><mml:mstyle displaystyle="false"><mml:mfrac style="text"><mml:mn mathvariant="normal">1</mml:mn><mml:mi mathvariant="italic">σ</mml:mi></mml:mfrac></mml:mstyle></mml:mstyle><mml:mo>⋅</mml:mo><mml:mi mathvariant="italic">ϕ</mml:mi><mml:mfenced open="(" close=")"><mml:mstyle displaystyle="false"><mml:mstyle displaystyle="false"><mml:mfrac style="text"><mml:mrow><mml:mi>y</mml:mi><mml:mo>-</mml:mo><mml:mi mathvariant="italic">μ</mml:mi></mml:mrow><mml:mi mathvariant="italic">σ</mml:mi></mml:mfrac></mml:mstyle></mml:mstyle></mml:mfenced></mml:mrow></mml:mfenced><mml:mo>,</mml:mo><mml:mspace linebreak="nobreak" width="1em"/></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mtext>if </mml:mtext><mml:mi>y</mml:mi><mml:mo>&gt;</mml:mo><mml:mn mathvariant="normal">0</mml:mn></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mrow><mml:mi>log⁡</mml:mi><mml:mfenced open="(" close=")"><mml:mrow><mml:mi mathvariant="normal">Φ</mml:mi><mml:mfenced close=")" open="("><mml:mstyle displaystyle="false"><mml:mstyle displaystyle="false"><mml:mfrac style="text"><mml:mrow><mml:mo>-</mml:mo><mml:mi mathvariant="italic">μ</mml:mi></mml:mrow><mml:mi mathvariant="italic">σ</mml:mi></mml:mfrac></mml:mstyle></mml:mstyle></mml:mfenced></mml:mrow></mml:mfenced><mml:mo>,</mml:mo><mml:mspace width="1em" linebreak="nobreak"/></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mtext>if </mml:mtext><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">0</mml:mn><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mfenced></mml:mrow></mml:math></disp-formula>
            where <inline-formula><mml:math id="M81" display="inline"><mml:mi mathvariant="italic">ϕ</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M82" display="inline"><mml:mi mathvariant="normal">Φ</mml:mi></mml:math></inline-formula> are the probability density function and cumulative density function of a standard Gaussian distribution <inline-formula><mml:math id="M83" display="inline"><mml:mrow><mml:mi mathvariant="script">N</mml:mi><mml:mo>(</mml:mo><mml:mn mathvariant="normal">0</mml:mn><mml:mo>,</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, respectively.</p>
      <p id="d1e2266">The prespecified base MOS
              <disp-formula id="Ch1.E9" content-type="numbered"><label>9</label><mml:math id="M84" display="block"><mml:mrow><mml:mi mathvariant="italic">μ</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>⋅</mml:mo><mml:mi mathvariant="monospace">tppow</mml:mi><mml:mi mathvariant="monospace">_</mml:mi><mml:mi mathvariant="monospace">mean</mml:mi><mml:mo>,</mml:mo><mml:mspace width="2em" linebreak="nobreak"/><mml:mi>log⁡</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
            linearly models the distributional mean <inline-formula><mml:math id="M85" display="inline"><mml:mi mathvariant="italic">μ</mml:mi></mml:math></inline-formula> on the mean of the (power-transformed) daily precipitation sums predicted by the individual ensemble members – i.e., the direct predictor from the NWP model. The standard deviation of the response distribution <inline-formula><mml:math id="M86" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula> is modeled by an intercept.</p>
      <p id="d1e2334">MOS forests are able to flexibly model MOS coefficients <inline-formula><mml:math id="M87" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> on all additional predictors from the dataset. The direct predictor <monospace>tppow_mean</monospace> from the base MOS could also be included among the splitting variables, but this did not improve the forecast skill for our application. All model estimation is performed in <sans-serif>R</sans-serif> with the <monospace>model4you</monospace> <xref ref-type="bibr" rid="bib1.bibx32" id="paren.33"/> and <monospace>crch</monospace>
<xref ref-type="bibr" rid="bib1.bibx21" id="paren.34"/> packages using the same hyperparameters as the distributional forests of <xref ref-type="bibr" rid="bib1.bibx26" id="text.35"/>. In particular, this means that a node must have at least 50 samples in order to be split again (<monospace>minsplit</monospace> <inline-formula><mml:math id="M88" display="inline"><mml:mrow><mml:mo>=</mml:mo><mml:mn mathvariant="normal">50</mml:mn></mml:mrow></mml:math></inline-formula>) and that terminal nodes must have at least 20 samples (<monospace>minbucket</monospace> <inline-formula><mml:math id="M89" display="inline"><mml:mrow><mml:mo>=</mml:mo><mml:mn mathvariant="normal">20</mml:mn></mml:mrow></mml:math></inline-formula>).</p>
</sec>
<sec id="Ch1.S3.SS2.SSS2">
  <label>3.2.2</label><title>Distributional forests</title>
      <?pagebreak page508?><p id="d1e2420">Distributional forests <xref ref-type="bibr" rid="bib1.bibx26" id="paren.36"/> work in a similar fashion to MOS forests but do not contain a prespecified MOS model.  Instead, <inline-formula><mml:math id="M90" display="inline"><mml:mi mathvariant="italic">θ</mml:mi></mml:math></inline-formula> only contains the parameters of the assumed response distribution – i.e., in this case, <inline-formula><mml:math id="M91" display="inline"><mml:mi mathvariant="italic">μ</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M92" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula> of a censored Gaussian – rather than the MOS coefficients. Trees are split with respect to distributional parameters rather than MOS coefficients, and the forest estimates the post-processed response distribution rather than a weather-adapted MOS. Distributional forests are estimated with the <monospace>disttree</monospace> package in <sans-serif>R</sans-serif> following the model configuration chosen by <xref ref-type="bibr" rid="bib1.bibx26" id="text.37"/> and thus have the same hyperparameters as the MOS forests.</p>
</sec>
<sec id="Ch1.S3.SS2.SSS3">
  <label>3.2.3</label><title>Quantile regression forests</title>
      <p id="d1e2465">Both MOS forests and distributional forests require specifying a parametric response distribution a priori. Since this assumption may not always hold (even conditionally), a fully non-parametric method called quantile regression forests <xref ref-type="bibr" rid="bib1.bibx20 bib1.bibx36" id="paren.38"/> is also considered. Splits are chosen with respect to the response value as in the standard random forest algorithm <xref ref-type="bibr" rid="bib1.bibx5" id="paren.39"/>, but the partitions are subsequently used to perform weighted quantile regressions and to generate probabilistic forecasts. In this application, 99 quantiles are considered, corresponding to probabilities of <inline-formula><mml:math id="M93" display="inline"><mml:mrow><mml:mi>p</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">0.01</mml:mn><mml:mo>,</mml:mo><mml:mn mathvariant="normal">0.02</mml:mn><mml:mo>,</mml:mo><mml:mi mathvariant="normal">…</mml:mi><mml:mn mathvariant="normal">0.99</mml:mn></mml:mrow></mml:math></inline-formula>. Model estimation is performed using the <monospace>quantregForest</monospace>
<xref ref-type="bibr" rid="bib1.bibx19" id="paren.40"/> package in <sans-serif>R</sans-serif>.</p>
</sec>
<sec id="Ch1.S3.SS2.SSS4">
  <label>3.2.4</label><title>EMOS</title>
      <p id="d1e2514">All three methods described above incorporate additional predictors using forest-based algorithms to allow for weather-adaptive post-processing. In order to quantify the benefit that comes with this added model flexibility,  a simple fully parametric non-adaptive EMOS is also considered:
              <disp-formula id="Ch1.E10" content-type="numbered"><label>10</label><mml:math id="M94" display="block"><mml:mtable class="split" rowspacing="0.2ex" displaystyle="true" columnalign="right left"><mml:mtr><mml:mtd><mml:mrow><mml:mi mathvariant="italic">μ</mml:mi></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>⋅</mml:mo><mml:mi mathvariant="monospace">tppow</mml:mi><mml:mi mathvariant="monospace">_</mml:mi><mml:mi mathvariant="monospace">mean</mml:mi><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd/><mml:mtd><mml:mrow><mml:mi>log⁡</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>⋅</mml:mo><mml:mi>log⁡</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="monospace">tppow</mml:mi><mml:mi mathvariant="monospace">_</mml:mi><mml:mi mathvariant="monospace">sprd</mml:mi><mml:mo>)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
            This EMOS has the same mean model as the pre-specified MOS in the MOS forest, but it also linearly models <inline-formula><mml:math id="M95" display="inline"><mml:mrow><mml:mi>log⁡</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> on the log-transformed standard deviation of the ensemble precipitation forecasts.</p>
</sec>
</sec>
</sec>
<sec id="Ch1.S4">
  <label>4</label><title>Results</title>
      <p id="d1e2621">To illustrate how post-processing with MOS forests works in practice, first a single MOS tree is grown at the station of Axams <xref ref-type="bibr" rid="bib1.bibx26" id="paren.41"><named-content content-type="pre">for location, see Fig. 8 of</named-content></xref>. This MOS tree is analyzed in Sect. <xref ref-type="sec" rid="Ch1.S4.SS1"/>. Subsequently, separate MOS forests are grown and used to post-process forecasts at all stations. The quality of these forecasts is evaluated in Sect. <xref ref-type="sec" rid="Ch1.S4.SS2"/>.</p>
<sec id="Ch1.S4.SS1">
  <label>4.1</label><title>Interpreting a MOS tree</title>
      <p id="d1e2640">A MOS tree for Axams is grown from the first 24 years of data and is visualized in Fig. <xref ref-type="fig" rid="Ch1.F1"/>. The first split of the tree separates rare (<inline-formula><mml:math id="M96" display="inline"><mml:mrow><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">23</mml:mn></mml:mrow></mml:math></inline-formula>) weather situations with very high ensemble-averaged total column liquid condensate (<monospace>tcolc_mean_mean</monospace>) from the remainder of the data. The rest of the data are then split based on the maximum temperature predicted by the ensemble (<monospace>tmax_mean_mean</monospace>).  The lower temperature branch has two subsequent splits: first based on precipitable water (<monospace>pwat_mean_max</monospace>) and then on the ensemble spread of precipitation (<monospace>tppow_sprd</monospace>). This results in three terminal nodes (nodes 5, 6, and 7).  The higher temperature branch has three splits, the first based again on (<inline-formula><mml:math id="M97" display="inline"><mml:mrow><mml:mi mathvariant="monospace">tcolc</mml:mi><mml:mi mathvariant="monospace">_</mml:mi><mml:mi mathvariant="monospace">mean</mml:mi><mml:mi mathvariant="monospace">_</mml:mi><mml:mi mathvariant="monospace">mean</mml:mi></mml:mrow></mml:math></inline-formula>) and the other two based on the ensemble spreads of 500 hPa temperature (<monospace>t500_sprd_min</monospace>) and precipitation (<monospace>tppow_sprd1824</monospace>). This results in four terminal nodes (nodes 11, 12, 13, and 14).</p>
      <p id="d1e2692">MOS models for each terminal node (i.e., distinct weather situation) are visualized in Fig. <xref ref-type="fig" rid="Ch1.F2"/>.  The majority of observations are found in either node 5, 13, or 11. For nodes 5 and 13, the MOS are quite similar, the largest difference being that forecasts in node 13 are less certain (i.e., <inline-formula><mml:math id="M98" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> is greater).  In contrast, the MOS used to post-process NWPs in node 11 is very different, with a strongly negative intercept for the mean model (<inline-formula><mml:math id="M99" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mn mathvariant="normal">8.16</mml:mn></mml:mrow></mml:math></inline-formula>) and a high forecast uncertainty (<inline-formula><mml:math id="M100" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1.09</mml:mn></mml:mrow></mml:math></inline-formula>).  This is because node 11 contains many days where the ensemble mean is greater than zero – i.e., some ensemble members predict precipitation for Axams – although no precipitation is actually observed. To understand when the tree makes such a prediction, it is only necessary to consider the splits in Fig. <xref ref-type="fig" rid="Ch1.F1"/> that lead to node 11: high maximum<?pagebreak page509?> temperature, low column liquid condensate, and narrow ensemble spreads for minimum temperature at 500 hPa and accumulated precipitation between 18:00 and 24:00 UTC.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F1" specific-use="star"><?xmltex \currentcnt{1}?><?xmltex \def\figurename{Figure}?><label>Figure 1</label><caption><p id="d1e2744">A single MOS tree estimated for Axams. Ellipses represent nodes used for splitting and contain the name of the splitting variable along with the <inline-formula><mml:math id="M101" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula> value of the independence test. The corresponding split point is included in the two branches (lines) emanating from the node. Terminal nodes (which are not split again) are visualized by rectangles and contain the number of observations <inline-formula><mml:math id="M102" display="inline"><mml:mi>n</mml:mi></mml:math></inline-formula> and estimated MOS coefficients <inline-formula><mml:math id="M103" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="italic">γ</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula>.  The models fit into each terminal node are visualized in Fig. <xref ref-type="fig" rid="Ch1.F2"/>.</p></caption>
          <?xmltex \igopts{width=455.244094pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023-f01.png"/>

        </fig>

</sec>
<sec id="Ch1.S4.SS2">
  <label>4.2</label><title>Evaluating predictive skill</title>
      <p id="d1e2802">MOS forests are compared to the reference methods described in Sect. <xref ref-type="sec" rid="Ch1.S3.SS2"/> by evaluating the skill of post-processed forecasts using the widely used continuous ranked probability score <xref ref-type="bibr" rid="bib1.bibx18 bib1.bibx9" id="paren.42"><named-content content-type="pre">CRPS,</named-content></xref>. To replicate a true operational scenario, all evaluations are performed out of sample on data that were not used to train the models.  First, forecasts at Axams are evaluated using multiple replications of a randomized 7-fold cross-validation (Sect. <xref ref-type="sec" rid="Ch1.S4.SS2.SSS1"/>).  At all other stations, forecasts are issued for a single hold-out fold (containing the last 4 years), and the remaining six folds (containing the first 24 years) are used for model training (Sect. <xref ref-type="sec" rid="Ch1.S4.SS2.SSS2"/>).  Finally, models are also trained using different amounts of data (12, 6, and 3 years) to investigate their robustness in this respect (Sect. <xref ref-type="sec" rid="Ch1.S4.SS2.SSS3"/>).</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F2" specific-use="star"><?xmltex \currentcnt{2}?><?xmltex \def\figurename{Figure}?><label>Figure 2</label><caption><p id="d1e2820">Scatterplots of observations versus ensemble mean forecasts in each terminal node of Fig. <xref ref-type="fig" rid="Ch1.F1"/>. Numbers identifying the nodes are included in the top left of each plot. Dashed and solid lines are quantiles corresponding to probabilities of 2.5 %, 25 %, 50 %, 75 %, and 97.5 %, obtained from the MOS model fit in each node.</p></caption>
          <?xmltex \igopts{width=455.244094pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023-f02.png"/>

        </fig>

<sec id="Ch1.S4.SS2.SSS1">
  <label>4.2.1</label><title>Full cross-validation at individual stations</title>
      <p id="d1e2838">The Axams data are randomly split into seven disjoint folds that each contain observations and NWPs from 4 different years. MOS forests and the reference post-processing methods outlined in Sect. <xref ref-type="sec" rid="Ch1.S3.SS2"/> are trained on six out of the seven folds and are then used to make predictions based on the remaining fold. After seven rounds of this,  out-of-sample predictions are available for each day in the 28 years of data and are used to compute an average CRPS for each method. The entire process is then repeated 10 times, each with a different random choice for the seven folds. CRPS skill scores are computed relative to the EMOS model and visualized by boxplots in Fig. <xref ref-type="fig" rid="Ch1.F4"/>. MOS forests improve CRPS by more than 7 % at Axams and thus perform slightly better than both the distributional forest and the quantile regression forest, which each lead to improvements of around 6 %.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F3" specific-use="star"><?xmltex \currentcnt{3}?><?xmltex \def\figurename{Figure}?><label>Figure 3</label><caption><p id="d1e2847">Solid lines are out-of-sample predictions for the location (<inline-formula><mml:math id="M104" display="inline"><mml:mi mathvariant="italic">μ</mml:mi></mml:math></inline-formula>) and scale (<inline-formula><mml:math id="M105" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula>) parameters of the response distribution at Axams in July 2009, obtained from the MOS tree visualized in Figs. <xref ref-type="fig" rid="Ch1.F1"/> and <xref ref-type="fig" rid="Ch1.F2"/>, as well as a MOS forest. Dashed lines are corresponding predictions from the base MOS (Eq. <xref ref-type="disp-formula" rid="Ch1.E9"/>).</p></caption>
            <?xmltex \igopts{width=341.433071pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023-f03.png"/>

          </fig>

</sec>
<sec id="Ch1.S4.SS2.SSS2">
  <label>4.2.2</label><title>Hold-out validation at all stations</title>
      <p id="d1e2885">To investigate predictive performance at all 95 stations, all models are trained on the first 24 years of data (1985–2008), and out-of-sample predictions are made for the last 4 years (2009–2012).</p>
      <p id="d1e2888">CRPS skill scores relative to EMOS are computed for each method at each station and are visualized by boxplots in Fig. <xref ref-type="fig" rid="Ch1.F4"/>. MOS forests generally outperform the other forest-based post-processing methods and are noticeably more robust.  Distributional forests and quantile regression forests occasionally perform up to 5 % worse than a basic EMOS, and the quantile regression forest is outperformed by EMOS nearly 25 % of the time.  This is not the case for the MOS forests, which always perform at least as well as EMOS and improve the forecasts by more than 5 % at 75 % of the stations.</p>
      <p id="d1e2893">Regional differences in model performance can be seen in the map of Fig. <xref ref-type="fig" rid="Ch1.F5"/>. While MOS forests significantly outperform distributional forests and quantile regression forests in the northeast and southeast of the forecast region, results are less clear in the more mountainous regions further west and near the main Alpine crest. At these locations, quantile regression forests often perform slightly better. Such clear regional differences in model performance are not visible in Fig. 8 of <xref ref-type="bibr" rid="bib1.bibx26" id="text.43"/>, perhaps because all their post-processing methods assumed the same type of response distribution.</p>
      <p id="d1e2901">Overall, probabilistic forecasts obtained from the MOS forests not only have a better CRPS than those obtained from the other two methods but are also more statistically consistent with observations (i.e., calibrated).  Calibration across all stations is visualized by probability integral transform (PIT) histograms for MOS forests and distributional forests and with a rank histogram for the quantile regression forests (Fig. <xref ref-type="fig" rid="Ch1.F6"/>). For perfectly calibrated forecasts, these histograms should be approximately uniform. Although all methods somewhat overestimate probabilities for high-precipitation events, this overestimation is much less pronounced in the MOS forests.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F4" specific-use="star"><?xmltex \currentcnt{4}?><?xmltex \def\figurename{Figure}?><label>Figure 4</label><caption><p id="d1e2909"><bold>(a)</bold> CRPSS relative to EMOS at Axams based on 10 randomly chosen 7-fold cross-validations. <bold>(b)</bold> CRPSS relative to EMOS at each station for the time period 2009–2012. Individual stations are connected by thin gray lines. Scores for the station of Axams are indicated by filled black circles connected by black lines.</p></caption>
            <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023-f04.png"/>

          </fig>

      <?xmltex \floatpos{t}?><fig id="Ch1.F5" specific-use="star"><?xmltex \currentcnt{5}?><?xmltex \def\figurename{Figure}?><label>Figure 5</label><caption><p id="d1e2925">Map showing the post-processing method that performs best at each station. Three different circle sizes (small, medium, large) are used to indicate where the CRPSS with respect to the second best method is less than 0.2, between 0.2 and 0.4, and more than 0.4, respectively. Terrain elevation is indicated by background color.</p></caption>
            <?xmltex \igopts{width=341.433071pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023-f05.png"/>

          </fig>

      <?xmltex \floatpos{t}?><fig id="Ch1.F6" specific-use="star"><?xmltex \currentcnt{6}?><?xmltex \def\figurename{Figure}?><label>Figure 6</label><caption><p id="d1e2936">Probability integral transform (PIT) histograms for MOS forests and distributional forests and rank histogram for quantile regression forests across all stations for the time period 2009–2012. Dashed red lines are the 95 % confidence intervals for a uniform distribution.</p></caption>
            <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023-f06.png"/>

          </fig>

</sec>
<sec id="Ch1.S4.SS2.SSS3">
  <label>4.2.3</label><title>Sensitivity to size of training data</title>
      <p id="d1e2953">The methods compared above use 24 years of data for model training, but since such large datasets are not always available in post-processing – e.g., for newly erected observational sites – the hold-out evaluations for all stations in Sect. <xref ref-type="sec" rid="Ch1.S4.SS2.SSS2"/> are repeated using only 12, 6, and 3 years of data for training. The boxplots in Fig. <xref ref-type="fig" rid="Ch1.F7"/> show that MOS forests are very robust in relation to these changes and still perform significantly better than a non-adaptive EMOS even when trained using only 3 years of data (i.e., 93 observations).  In contrast, distributional forests nearly always perform significantly worse than EMOS in such cases and have a median skill score of <inline-formula><mml:math id="M106" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">10</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula> across all stations. Similarly, quantile regression forests are also outperformed by the non-adaptive EMOS at around half of the stations.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F7" specific-use="star"><?xmltex \currentcnt{7}?><?xmltex \def\figurename{Figure}?><label>Figure 7</label><caption><p id="d1e2975">As for the hold-out evaluation of all stations in Fig. <xref ref-type="fig" rid="Ch1.F4"/> but with models trained on the past 24, 6, and 3 years. Blue lines highlight the influence changing data size has on the median CRPSS of each method.</p></caption>
            <?xmltex \igopts{width=341.433071pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/503/2023/npg-30-503-2023-f07.png"/>

          </fig>

</sec>
</sec>
</sec>
<sec id="Ch1.S5">
  <label>5</label><title>Discussion</title>
      <p id="d1e2997">When compared to state-of-the-art weather-adaptive post-processing methods, MOS forests have the main advantage of being highly robust: they reliably outperform simple non-adaptive reference methods even when trained on very small sample sizes.  This is possible because, unlike state-of-the-art weather-adaptive methods that treat all predictors equally and use a data-driven approach to learn their relationships to the response, MOS forests directly incorporate prior (physically based) knowledge about the most important relationships in the form of a parametric model. One might think that robustness is not important in our current big-data era, but consider the fact that NWP models are continuously updated (e.g., with improved resolutions or parameterizations), and new stations (or measurement instruments) can always be installed. In the words of <xref ref-type="bibr" rid="bib1.bibx8" id="text.44"/>, “data<?pagebreak page511?> samples containing numerical model output are a perishable commodity”, and this is still true today.</p>
      <p id="d1e3003">In the application considered here, MOS forests are used to post-process NWP ensembles, and separate models are estimated for each station. Without any modifications, MOS forests also offer a powerful way to obtain probabilistic forecasts from deterministic NWPs, where no predictors explicitly characterizing the forecast uncertainty are available. Similarly, MOS forests could also be employed as spatial (rather than station-wise) post-processing models by including predictors that contain information about the individual grid points or stations within the training data.  Potentially relevant variables would then include latitude, longitude, and altitude but also surface roughness, land cover type, or other characteristics.</p>
      <p id="d1e3006">Despite their many advantages, MOS forests require specifying the same two things as all other MOS models: (i) a parametric distribution for the response and (ii) models linking the parameters of that distribution with appropriate predictors derived from the NWP. Not much can be done about the first point besides trying different response distributions or transformations of the data. As for the second point, in cases where no suitable models for the distributional parameters can be specified a priori, MOS forests have no advantage over distributional forests. In fact, MOS forests collapse to distributional forests if the assumed base MOS has intercept-only models for the parameters of the response distribution.</p>
</sec>
<sec id="Ch1.S6" sec-type="conclusions">
  <label>6</label><title>Conclusions</title>
      <p id="d1e3017">Since NWPs have errors that can depend on the weather situation, weather-adaptive post-processing methods are necessary to obtain optimal probabilistic forecasts.  By fusing traditional (non-adaptive) and modern (weather-adaptive) post-processing approaches, MOS forests retain the best of both worlds: a method that is flexible enough to allow for weather-adaptive post-processing but that is also robust, intuitive, and straightforward to implement. This is achieved by using random forests to adapt the regression coefficients of a prespecified parametric base MOS to a set of additional predictor<?pagebreak page512?> variables that characterize the current weather situation. In contrast to state-of-the-art post-processing methods, which typically directly estimate properties of the response from these predictors, MOS forests only use them to estimate the regression coefficients of the assumed base model. As a result, they can generate skillful forecasts even when only a very limited amount of data are available for training and when purely data-driven weather-adaptive methods fail to outperform a simple non-adaptive model.</p>
</sec>

      
      </body>
    <back><notes notes-type="codeavailability"><title>Code availability</title>

      <p id="d1e3024">Code with wrapper functions for training and evaluating postprocessing models on the RainTyrol dataset can be found at <uri>https://github.com/thomas-muschinski/mos_forests</uri> (last access: 15 November 2023).  MOS forests are estimated using the R package model4you <xref ref-type="bibr" rid="bib1.bibx32" id="paren.45"/> in combination with crch <xref ref-type="bibr" rid="bib1.bibx21" id="paren.46"/>. Distributional forests are estimated using disttree <xref ref-type="bibr" rid="bib1.bibx28" id="paren.47"/>. Quantile regression forests are estimated using quantregForest <xref ref-type="bibr" rid="bib1.bibx19" id="paren.48"/>.  EMOS models are estimated using crch <xref ref-type="bibr" rid="bib1.bibx21" id="paren.49"/>.  Forecast evaluation is performed using scoringRules <xref ref-type="bibr" rid="bib1.bibx14" id="paren.50"/>.</p>
  </notes><notes notes-type="dataavailability"><title>Data availability</title>

      <p id="d1e3052">The RainTyrol dataset used for training and evaluating the postprocessing models is available at <xref ref-type="bibr" rid="bib1.bibx27" id="text.51"/>.</p>
  </notes><notes notes-type="authorcontribution"><title>Author contributions</title>

      <?pagebreak page513?><p id="d1e3061">TM, GJM, AZ, and TS planned the research. TM wrote the original paper draft, and all the authors subsequently reviewed and revised it.</p>
  </notes><notes notes-type="competinginterests"><title>Competing interests</title>

      <p id="d1e3067">The contact author has declared that none of the authors has any competing interests.</p>
  </notes><notes notes-type="disclaimer"><title>Disclaimer</title>

      <p id="d1e3073">Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.</p>
  </notes><ack><title>Acknowledgements</title><p id="d1e3079">The authors thank the two anonymous reviewers for their helpful comments.</p></ack><notes notes-type="financialsupport"><title>Financial support</title>

      <p id="d1e3084">This research has been supported by the Austrian Science Fund (grant no. P 31836). Thomas Muschinski   was also supported by the Doktoratsstipendium of Universität Innsbruck. <?xmltex \hack{\newline}?><?xmltex \hack{\newline}?>The article processing charges for this open-access <?xmltex \notforhtml{\newline}?>publication were covered by the Karlsruhe Institute<?xmltex \notforhtml{\newline}?> of Technology (KIT).</p>
  </notes><notes notes-type="reviewstatement"><title>Review statement</title>

      <p id="d1e3098">This paper was edited by Takemasa Miyoshi and reviewed by two anonymous referees.</p>
  </notes><ref-list>
    <title>References</title>

      <ref id="bib1.bibx1"><?xmltex \def\ref@label{{Athey et~al.(2019)Athey, Tibshirani, and Wager}}?><label>Athey et al.(2019)Athey, Tibshirani, and Wager</label><?label athey2019?><mixed-citation>Athey, S., Tibshirani, J., and Wager, S.: Generalized random forests, Ann. Stat., 47, 1148–1178, <ext-link xlink:href="https://doi.org/10.1214/18-AOS1709" ext-link-type="DOI">10.1214/18-AOS1709</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx2"><?xmltex \def\ref@label{{Baran and Nemoda(2016)}}?><label>Baran and Nemoda(2016)</label><?label baran2016censored?><mixed-citation>Baran, S. and Nemoda, D.: Censored and shifted gamma distribution based EMOS model for probabilistic quantitative precipitation forecasting, Environmetrics, 27, 280–292, <ext-link xlink:href="https://doi.org/10.1002/env.2391" ext-link-type="DOI">10.1002/env.2391</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx3"><?xmltex \def\ref@label{{Bauer et~al.(2015)Bauer, Thorpe, and Brunet}}?><label>Bauer et al.(2015)Bauer, Thorpe, and Brunet</label><?label bauer2015quiet?><mixed-citation>Bauer, P., Thorpe, A., and Brunet, G.: The Quiet Revolution of Numerical Weather Prediction, Nature, 525, 47–55, <ext-link xlink:href="https://doi.org/10.1038/nature14956" ext-link-type="DOI">10.1038/nature14956</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bibx4"><?xmltex \def\ref@label{{Breiman(1996)}}?><label>Breiman(1996)</label><?label breiman1996bagging?><mixed-citation>Breiman, L.: Bagging Predictors, Mach. Learn., 24, 123–140, <ext-link xlink:href="https://doi.org/10.1007/bf00058655" ext-link-type="DOI">10.1007/bf00058655</ext-link>, 1996.</mixed-citation></ref>
      <ref id="bib1.bibx5"><?xmltex \def\ref@label{{Breiman(2001)}}?><label>Breiman(2001)</label><?label breiman2001random?><mixed-citation>Breiman, L.: Random Forests, Mach. Learn., 45, 5–32, <ext-link xlink:href="https://doi.org/10.1023/a:1010933404324" ext-link-type="DOI">10.1023/a:1010933404324</ext-link>, 2001.</mixed-citation></ref>
      <ref id="bib1.bibx6"><?xmltex \def\ref@label{{Bremnes(2020)}}?><label>Bremnes(2020)</label><?label bremnes2020ensemble?><mixed-citation>Bremnes, J. B.: Ensemble postprocessing using quantile function regression based on neural networks and Bernstein polynomials, Mon. Weather Rev., 148, 403–414, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-19-0227.1" ext-link-type="DOI">10.1175/MWR-D-19-0227.1</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx7"><?xmltex \def\ref@label{{Evin et~al.(2021)Evin, Lafaysse, Taillardat, and Zamo}}?><label>Evin et al.(2021)Evin, Lafaysse, Taillardat, and Zamo</label><?label evin2021?><mixed-citation>Evin, G., Lafaysse, M., Taillardat, M., and Zamo, M.: Calibrated ensemble forecasts of the height of new snow using quantile regression forests and ensemble model output statistics, Nonlin. Processes Geophys., 28, 467–480, <ext-link xlink:href="https://doi.org/10.5194/npg-28-467-2021" ext-link-type="DOI">10.5194/npg-28-467-2021</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx8"><?xmltex \def\ref@label{{Glahn and Lowry(1972)}}?><label>Glahn and Lowry(1972)</label><?label glahn1972use?><mixed-citation>Glahn, H. R. and Lowry, D. A.: The Use of Model Output Statistics (MOS) in Objective Weather Forecasting, J. Appl. Meteorol., 11, 1203–1211, <ext-link xlink:href="https://doi.org/10.1175/1520-0450(1972)011&lt;1203:tuomos&gt;2.0.co;2" ext-link-type="DOI">10.1175/1520-0450(1972)011&lt;1203:tuomos&gt;2.0.co;2</ext-link>, 1972.</mixed-citation></ref>
      <ref id="bib1.bibx9"><?xmltex \def\ref@label{{Gneiting and Raftery(2007)}}?><label>Gneiting and Raftery(2007)</label><?label gneiting2007proper?><mixed-citation>Gneiting, T. and Raftery, A. E.: Strictly Proper Scoring Rules, Prediction, and Estimation, J. Am. Stat. Assoc., 102, 359–378, <ext-link xlink:href="https://doi.org/10.1198/016214506000001437" ext-link-type="DOI">10.1198/016214506000001437</ext-link>, 2007.</mixed-citation></ref>
      <ref id="bib1.bibx10"><?xmltex \def\ref@label{{Gneiting et~al.(2005)Gneiting, Raftery, {Westveld III}, and
Goldman}}?><label>Gneiting et al.(2005)Gneiting, Raftery, Westveld III, and Goldman</label><?label gneiting2005calibrated?><mixed-citation>Gneiting, T., Raftery, A. E., Westveld III, A. H., and Goldman, T.: Calibrated Probabilistic Forecasting Using Ensemble Model Output Statistics and Minimum CRPS Estimation, Mon. Weather Rev., 133, 1098–1118, <ext-link xlink:href="https://doi.org/10.1175/mwr2904.1" ext-link-type="DOI">10.1175/mwr2904.1</ext-link>, 2005.</mixed-citation></ref>
      <?pagebreak page514?><ref id="bib1.bibx11"><?xmltex \def\ref@label{{Hamill et~al.(2013)Hamill, Bates, Whitaker, Murray, Fiorino,
Galarneau, Zhu, and Lapenta}}?><label>Hamill et al.(2013)Hamill, Bates, Whitaker, Murray, Fiorino, Galarneau, Zhu, and Lapenta</label><?label hamill2013noaa?><mixed-citation>Hamill, T. M., Bates, G. T., Whitaker, J. S., Murray, D. R., Fiorino, M., Galarneau, T. J., Zhu, Y., and Lapenta, W.: NOAA's second-generation global medium-range ensemble reforecast dataset, B. Am. Meteorol. Soc., 94, 1553–1565, <ext-link xlink:href="https://doi.org/10.1175/BAMS-D-12-00014.1" ext-link-type="DOI">10.1175/BAMS-D-12-00014.1</ext-link>, 2013.</mixed-citation></ref>
      <ref id="bib1.bibx12"><?xmltex \def\ref@label{{Hothorn et~al.(2006)Hothorn, Hornik, and Zeileis}}?><label>Hothorn et al.(2006)Hothorn, Hornik, and Zeileis</label><?label hothorn2006?><mixed-citation>Hothorn, T., Hornik, K., and Zeileis, A.: Unbiased Recursive Partitioning: A Conditional Inference Framework, J. Comput. Graph. Stat., 15, 651–674, <ext-link xlink:href="https://doi.org/10.1198/106186006X133933" ext-link-type="DOI">10.1198/106186006X133933</ext-link>, 2006.</mixed-citation></ref>
      <ref id="bib1.bibx13"><?xmltex \def\ref@label{{Hothorn et~al.(2008)Hothorn, Hornik, Van De~Wiel, and
Zeileis}}?><label>Hothorn et al.(2008)Hothorn, Hornik, Van De Wiel, and Zeileis</label><?label hothorn2008implementing?><mixed-citation> Hothorn, T., Hornik, K., Van De Wiel, M. A., and Zeileis, A.: Implementing a class of permutation tests: the coin package, J. Stat. Softw., 28, 1–23, 2008.</mixed-citation></ref>
      <ref id="bib1.bibx14"><?xmltex \def\ref@label{Jordan et~al.(2023)}?><label>Jordan et al.(2023)</label><?label scoringRules?><mixed-citation>Jordan, A. I., Krueger, F., Lerch, S., Allen, S., and Graeter, M.: scoringRules: Scoring Rules for Parametric and Simulated Distribution Forecasts, R package version 1.1.1, <uri>https://cran.r-project.org/web/packages/scoringRules/</uri>, 2023.</mixed-citation></ref>
      <ref id="bib1.bibx15"><?xmltex \def\ref@label{{Kneib et~al.(2021)Kneib, Silbersdorff, and
S{\"{a}}fken}}?><label>Kneib et al.(2021)Kneib, Silbersdorff, and Säfken</label><?label kneib2021rage?><mixed-citation>Kneib, T., Silbersdorff, A., and Säfken, B.: Rage against the mean–a review of distributional regression approaches, Econometrics and Statistics, 26, 99–123, <ext-link xlink:href="https://doi.org/10.1016/j.ecosta.2021.07.006" ext-link-type="DOI">10.1016/j.ecosta.2021.07.006</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx16"><?xmltex \def\ref@label{{Lang et~al.(2020)Lang, Lerch, Mayr, Simon, Stauffer, and
Zeileis}}?><label>Lang et al.(2020)Lang, Lerch, Mayr, Simon, Stauffer, and Zeileis</label><?label lang2020?><mixed-citation>Lang, M. N., Lerch, S., Mayr, G. J., Simon, T., Stauffer, R., and Zeileis, A.: Remember the past: a comparison of time-adaptive training schemes for non-homogeneous regression, Nonlin. Processes Geophys., 27, 23–34, <ext-link xlink:href="https://doi.org/10.5194/npg-27-23-2020" ext-link-type="DOI">10.5194/npg-27-23-2020</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx17"><?xmltex \def\ref@label{{Lerch and Thorarinsdottir(2013)}}?><label>Lerch and Thorarinsdottir(2013)</label><?label lerch2013comparison?><mixed-citation>Lerch, S. and Thorarinsdottir, T. L.: Comparison of non-homogeneous regression models for probabilistic wind speed forecasting, Tellus A, 65, 21206, <ext-link xlink:href="https://doi.org/10.3402/tellusa.v65i0.21206" ext-link-type="DOI">10.3402/tellusa.v65i0.21206</ext-link>, 2013.</mixed-citation></ref>
      <ref id="bib1.bibx18"><?xmltex \def\ref@label{{Matheson and Winkler(1976)}}?><label>Matheson and Winkler(1976)</label><?label matheson1976scoring?><mixed-citation> Matheson, J. E. and Winkler, R. L.: Scoring rules for continuous probability distributions, Manage. Sci., 22, 1087–1096, 1976.</mixed-citation></ref>
      <ref id="bib1.bibx19"><?xmltex \def\ref@label{{Meinshausen(2017)}}?><label>Meinshausen(2017)</label><?label meinshausen2017quantregforest?><mixed-citation>Meinshausen, N.: Quantregforest: quantile regression forests, R package version,  1.3-7, <uri>https://cran.r-project.org/web/packages/quantregForest/</uri> (last access: 15 November 2023), 2017.</mixed-citation></ref>
      <ref id="bib1.bibx20"><?xmltex \def\ref@label{{Meinshausen and Ridgeway(2006)}}?><label>Meinshausen and Ridgeway(2006)</label><?label meinshausen2006quantile?><mixed-citation> Meinshausen, N. and Ridgeway, G.: Quantile regression forests, J. Mach. Learn. Res., 7, 983–999, 2006.</mixed-citation></ref>
      <ref id="bib1.bibx21"><?xmltex \def\ref@label{{Messner et~al.(2016)Messner, Mayr, and
Zeileis}}?><label>Messner et al.(2016)Messner, Mayr, and Zeileis</label><?label messner2016heteroscedastic?><mixed-citation>Messner, J. W., Mayr, G. J., and Zeileis, A.: Heteroscedastic Censored and Truncated Regression with crch, R J., 8, 173–181, <ext-link xlink:href="https://doi.org/10.32614/RJ-2016-012" ext-link-type="DOI">10.32614/RJ-2016-012</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx22"><?xmltex \def\ref@label{{Messner et~al.(2017)Messner, Mayr, and
Zeileis}}?><label>Messner et al.(2017)Messner, Mayr, and Zeileis</label><?label messner2017nonhomogeneous?><mixed-citation>Messner, J. W., Mayr, G. J., and Zeileis, A.: Nonhomogeneous boosting for predictor selection in ensemble postprocessing, Mon. Weather Rev., 145, 137–147, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-16-0088.1" ext-link-type="DOI">10.1175/MWR-D-16-0088.1</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx23"><?xmltex \def\ref@label{{Rasp and Lerch(2018)}}?><label>Rasp and Lerch(2018)</label><?label rasp2018neural?><mixed-citation>Rasp, S. and Lerch, S.: Neural networks for postprocessing ensemble weather forecasts, Mon. Weather Rev., 146, 3885–3900, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-18-0187.1" ext-link-type="DOI">10.1175/MWR-D-18-0187.1</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx24"><?xmltex \def\ref@label{{Rigby and Stasinopoulos(2005)}}?><label>Rigby and Stasinopoulos(2005)</label><?label rigby2005generalized?><mixed-citation>Rigby, R. A. and Stasinopoulos, D. M.: Generalized additive models for location, scale and shape, J. Roy. Stat. Soc. C-App., 54, 507–554, <ext-link xlink:href="https://doi.org/10.1111/j.1467-9876.2005.00510.x" ext-link-type="DOI">10.1111/j.1467-9876.2005.00510.x</ext-link>, 2005.</mixed-citation></ref>
      <ref id="bib1.bibx25"><?xmltex \def\ref@label{{Scheuerer(2014)}}?><label>Scheuerer(2014)</label><?label scheuerer2014probabilistic?><mixed-citation>Scheuerer, M.: Probabilistic Quantitative Precipitation Forecasting Using Ensemble Model Output Statistics, Q. J. Roy. Meteor. Soc., 140, 1086–1096, <ext-link xlink:href="https://doi.org/10.1002/qj.2183" ext-link-type="DOI">10.1002/qj.2183</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx26"><?xmltex \def\ref@label{{Schlosser et~al.(2019)Schlosser, Hothorn, Stauffer, and
Zeileis}}?><label>Schlosser et al.(2019)Schlosser, Hothorn, Stauffer, and Zeileis</label><?label schlosser2019distributional?><mixed-citation>Schlosser, L., Hothorn, T., Stauffer, R., and Zeileis, A.: Distributional regression forests for probabilistic precipitation forecasting in complex terrain, Ann. Appl. Stat., 13, 1564–1589, <ext-link xlink:href="https://doi.org/10.1214/19-AOAS1247" ext-link-type="DOI">10.1214/19-AOAS1247</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx27"><?xmltex \def\ref@label{{Schlosser et~al.(2020)}}?><label>Schlosser et al.(2020)</label><?label RainTyrol?><mixed-citation>Schlosser, L., Stauffer, R., and Zeileis, A.: RainTyrol: Precipitation Observations and NWP Forecasts from GEFS, R package version 0.2-0/r2952, <uri>https://R-Forge.R-project.org/projects/partykit/</uri>  (last access: 15 November 2023), 2020.</mixed-citation></ref>
      <ref id="bib1.bibx28"><?xmltex \def\ref@label{{Schlosser et~al.(2021)}}?><label>Schlosser et al.(2021)</label><?label disttree?><mixed-citation>Schlosser, L.,  Lang, M. N.,  Hothorn, T., and Zeileis, A.: disttree: Trees and Forests for Distributional Regression, R package version 0.2-0/r3189, <uri>https://R-Forge.R-project.org/projects/partykit/</uri> (last access: 15 November 2023), 2021.</mixed-citation></ref>
      <ref id="bib1.bibx29"><?xmltex \def\ref@label{{Schoenach et~al.(2020)Schoenach, Simon, and Mayr}}?><label>Schoenach et al.(2020)Schoenach, Simon, and Mayr</label><?label schoenach2020?><mixed-citation>Schoenach, D., Simon, T., and Mayr, G. J.: Postprocessing ensemble forecasts of vertical temperature profiles, Adv. Stat. Clim. Meteorol. Oceanogr., 6, 45–60, <ext-link xlink:href="https://doi.org/10.5194/ascmo-6-45-2020" ext-link-type="DOI">10.5194/ascmo-6-45-2020</ext-link>, 2020. </mixed-citation></ref><?xmltex \hack{\newpage}?>
      <ref id="bib1.bibx30"><?xmltex \def\ref@label{{Schulz and Lerch(2022)}}?><label>Schulz and Lerch(2022)</label><?label schulz2022machine?><mixed-citation>Schulz, B. and Lerch, S.: Machine learning methods for postprocessing ensemble forecasts of wind gusts: A systematic comparison, Mon. Weather Rev., 150, 235–257, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-21-0150.1" ext-link-type="DOI">10.1175/MWR-D-21-0150.1</ext-link>, 2022.</mixed-citation></ref>
      <ref id="bib1.bibx31"><?xmltex \def\ref@label{{Seibold et~al.(2018)Seibold, Zeileis, and
Hothorn}}?><label>Seibold et al.(2018)Seibold, Zeileis, and Hothorn</label><?label seibold2018individual?><mixed-citation>Seibold, H., Zeileis, A., and Hothorn, T.: Individual treatment effect prediction for amyotrophic lateral sclerosis patients, Stat. Methods  Med. Res., 27, 3104–3125, <ext-link xlink:href="https://doi.org/10.1177/0962280217693034" ext-link-type="DOI">10.1177/0962280217693034</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx32"><?xmltex \def\ref@label{{Seibold et~al.(2019)Seibold, Zeileis, and
Hothorn}}?><label>Seibold et al.(2019)Seibold, Zeileis, and Hothorn</label><?label seibold2019model4you?><mixed-citation>Seibold, H., Zeileis, A., and Hothorn, T.: model4you: an R package for personalised treatment effect estimation, J. Open Res. Softw., 7, 17, <ext-link xlink:href="https://doi.org/10.5334/jors.219" ext-link-type="DOI">10.5334/jors.219</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx33"><?xmltex \def\ref@label{{Simon et~al.(2019)Simon, Mayr, Umlauf, and Zeileis}}?><label>Simon et al.(2019)Simon, Mayr, Umlauf, and Zeileis</label><?label simon2019nwp?><mixed-citation>Simon, T., Mayr, G. J., Umlauf, N., and Zeileis, A.: NWP-based lightning prediction using flexible count data regression, Adv. Stat. Clim. Meteorol. Oceanogr., 5, 1–16, <ext-link xlink:href="https://doi.org/10.5194/ascmo-5-1-2019" ext-link-type="DOI">10.5194/ascmo-5-1-2019</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx34"><?xmltex \def\ref@label{{Stauffer et~al.(2017{\natexlab{a}})Stauffer, Mayr, Messner, Umlauf,
and Zeileis}}?><label>Stauffer et al.(2017a)Stauffer, Mayr, Messner, Umlauf, and Zeileis</label><?label stauffer2017spatio?><mixed-citation>Stauffer, R., Mayr, G. J., Messner, J. W., Umlauf, N., and Zeileis, A.: Spatio-temporal precipitation climatology over complex terrain using a censored additive regression model, Int. J. Climatol., 37, 3264–3275, <ext-link xlink:href="https://doi.org/10.1002/joc.4913" ext-link-type="DOI">10.1002/joc.4913</ext-link>, 2017a.</mixed-citation></ref>
      <ref id="bib1.bibx35"><?xmltex \def\ref@label{{Stauffer et~al.(2017{\natexlab{b}})Stauffer, Umlauf, Messner, Mayr,
and Zeileis}}?><label>Stauffer et al.(2017b)Stauffer, Umlauf, Messner, Mayr, and Zeileis</label><?label stauffer2017ensemble?><mixed-citation>Stauffer, R., Umlauf, N., Messner, J. W., Mayr, G. J., and Zeileis, A.: Ensemble postprocessing of daily precipitation sums over complex terrain using censored high-resolution standardized anomalies, Mon. Weather Rev., 145, 955–969, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-16-0260.1" ext-link-type="DOI">10.1175/MWR-D-16-0260.1</ext-link>, 2017b.</mixed-citation></ref>
      <ref id="bib1.bibx36"><?xmltex \def\ref@label{{Taillardat et~al.(2016)Taillardat, Mestre, Zamo, and
Naveau}}?><label>Taillardat et al.(2016)Taillardat, Mestre, Zamo, and Naveau</label><?label taillardat2016calibrated?><mixed-citation>Taillardat, M., Mestre, O., Zamo, M., and Naveau, P.: Calibrated ensemble forecasts using quantile regression forests and ensemble model output statistics, Mon. Weather Rev., 144, 2375–2393, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-15-0260.1" ext-link-type="DOI">10.1175/MWR-D-15-0260.1</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx37"><?xmltex \def\ref@label{{Taillardat et~al.(2019)Taillardat, Foug{\`{e}}res, Naveau, and
Mestre}}?><label>Taillardat et al.(2019)Taillardat, Fougères, Naveau, and Mestre</label><?label taillardat2019forest?><mixed-citation>Taillardat, M., Fougères, A.-L., Naveau, P., and Mestre, O.: Forest-based and semiparametric methods for the postprocessing of rainfall ensemble forecasting, Weather Forecast., 34, 617–634, <ext-link xlink:href="https://doi.org/10.1175/WAF-D-18-0149.1" ext-link-type="DOI">10.1175/WAF-D-18-0149.1</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx38"><?xmltex \def\ref@label{{Thorarinsdottir and
Gneiting(2010)}}?><label>Thorarinsdottir and Gneiting(2010)</label><?label thorarinsdottir2010probabilistic?><mixed-citation>Thorarinsdottir, T. L. and Gneiting, T.: Probabilistic forecasts of wind speed: Ensemble model output statistics by using heteroscedastic censored regression, J. Roy. Stat. Soc. A Sta., 173, 371–388, <ext-link xlink:href="https://doi.org/10.1111/j.1467-985X.2009.00616.x" ext-link-type="DOI">10.1111/j.1467-985X.2009.00616.x</ext-link>, 2010.</mixed-citation></ref>
      <ref id="bib1.bibx39"><?xmltex \def\ref@label{{Vannitsem et~al.(2021)Vannitsem, Bremnes, Demaeyer, Evans, Flowerdew,
Hemri, Lerch, Roberts, Theis, Atencia et~al.}}?><label>Vannitsem et al.(2021)Vannitsem, Bremnes, Demaeyer, Evans, Flowerdew, Hemri, Lerch, Roberts, Theis, Atencia et al.</label><?label vannitsem2021statistical?><mixed-citation>Vannitsem, S., Bremnes, J. B., Demaeyer, J., Evans, G. R., Flowerdew, J., Hemri, S., Lerch, S., Roberts, N., Theis, S., Atencia, A., Bouallègue, Z. B., Bhend, J., Dabernig, M., De Cruz, L., Hieta, L., Mestre, O., Moret, L.,  Plenković, I. O., Schmeits, M., Taillardat, M., Van den Bergh, J., Van Schaeybroeck,  B., Whan, K., and Ylhaisi, J.: Statistical postprocessing for weather forecasts: Review, challenges, and avenues in a big data world, B. Am. Meteorol. Soc., 102, E681–E699, <ext-link xlink:href="https://doi.org/10.1175/BAMS-D-19-0308.1" ext-link-type="DOI">10.1175/BAMS-D-19-0308.1</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx40"><?xmltex \def\ref@label{{Zeileis et~al.(2008)Zeileis, Hothorn, and Hornik}}?><label>Zeileis et al.(2008)Zeileis, Hothorn, and Hornik</label><?label zeileis2008model?><mixed-citation>Zeileis, A., Hothorn, T., and Hornik, K.: Model-based recursive partitioning, J. Computat. Graph. Stat., 17, 492–514, <ext-link xlink:href="https://doi.org/10.1198/106186008X319331" ext-link-type="DOI">10.1198/106186008X319331</ext-link>, 2008.</mixed-citation></ref>

  </ref-list></back>
    <!--<article-title-html>Robust weather-adaptive post-processing using model output statistics random forests</article-title-html>
<abstract-html/>
<ref-html id="bib1.bib1"><label>Athey et al.(2019)Athey, Tibshirani, and Wager</label><mixed-citation>
      
Athey, S., Tibshirani, J., and Wager, S.: Generalized random forests,
Ann. Stat., 47, 1148–1178, <a href="https://doi.org/10.1214/18-AOS1709" target="_blank">https://doi.org/10.1214/18-AOS1709</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib2"><label>Baran and Nemoda(2016)</label><mixed-citation>
      
Baran, S. and Nemoda, D.: Censored and shifted gamma distribution based EMOS
model for probabilistic quantitative precipitation forecasting,
Environmetrics, 27, 280–292, <a href="https://doi.org/10.1002/env.2391" target="_blank">https://doi.org/10.1002/env.2391</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib3"><label>Bauer et al.(2015)Bauer, Thorpe, and Brunet</label><mixed-citation>
      
Bauer, P., Thorpe, A., and Brunet, G.: The Quiet Revolution of Numerical
Weather Prediction, Nature, 525, 47–55, <a href="https://doi.org/10.1038/nature14956" target="_blank">https://doi.org/10.1038/nature14956</a>, 2015.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib4"><label>Breiman(1996)</label><mixed-citation>
      
Breiman, L.: Bagging Predictors, Mach. Learn., 24, 123–140,
<a href="https://doi.org/10.1007/bf00058655" target="_blank">https://doi.org/10.1007/bf00058655</a>, 1996.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib5"><label>Breiman(2001)</label><mixed-citation>
      
Breiman, L.: Random Forests, Mach. Learn., 45, 5–32,
<a href="https://doi.org/10.1023/a:1010933404324" target="_blank">https://doi.org/10.1023/a:1010933404324</a>, 2001.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib6"><label>Bremnes(2020)</label><mixed-citation>
      
Bremnes, J. B.: Ensemble postprocessing using quantile function regression
based on neural networks and Bernstein polynomials, Mon. Weather Rev.,
148, 403–414, <a href="https://doi.org/10.1175/MWR-D-19-0227.1" target="_blank">https://doi.org/10.1175/MWR-D-19-0227.1</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib7"><label>Evin et al.(2021)Evin, Lafaysse, Taillardat, and Zamo</label><mixed-citation>
      
Evin, G., Lafaysse, M., Taillardat, M., and Zamo, M.: Calibrated ensemble forecasts of the height of new snow using quantile regression forests and ensemble model output statistics, Nonlin. Processes Geophys., 28, 467–480, <a href="https://doi.org/10.5194/npg-28-467-2021" target="_blank">https://doi.org/10.5194/npg-28-467-2021</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib8"><label>Glahn and Lowry(1972)</label><mixed-citation>
      
Glahn, H. R. and Lowry, D. A.: The Use of Model Output Statistics (MOS) in
Objective Weather Forecasting, J. Appl. Meteorol., 11,
1203–1211, <a href="https://doi.org/10.1175/1520-0450(1972)011&lt;1203:tuomos&gt;2.0.co;2" target="_blank">https://doi.org/10.1175/1520-0450(1972)011&lt;1203:tuomos&gt;2.0.co;2</a>, 1972.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib9"><label>Gneiting and Raftery(2007)</label><mixed-citation>
      
Gneiting, T. and Raftery, A. E.: Strictly Proper Scoring Rules, Prediction, and
Estimation, J. Am. Stat. Assoc., 102, 359–378,
<a href="https://doi.org/10.1198/016214506000001437" target="_blank">https://doi.org/10.1198/016214506000001437</a>, 2007.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib10"><label>Gneiting et al.(2005)Gneiting, Raftery, Westveld III, and
Goldman</label><mixed-citation>
      
Gneiting, T., Raftery, A. E., Westveld III, A. H., and Goldman, T.:
Calibrated Probabilistic Forecasting Using Ensemble Model Output Statistics
and Minimum CRPS Estimation, Mon. Weather Rev., 133, 1098–1118,
<a href="https://doi.org/10.1175/mwr2904.1" target="_blank">https://doi.org/10.1175/mwr2904.1</a>, 2005.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib11"><label>Hamill et al.(2013)Hamill, Bates, Whitaker, Murray, Fiorino,
Galarneau, Zhu, and Lapenta</label><mixed-citation>
      
Hamill, T. M., Bates, G. T., Whitaker, J. S., Murray, D. R., Fiorino, M.,
Galarneau, T. J., Zhu, Y., and Lapenta, W.: NOAA's second-generation global
medium-range ensemble reforecast dataset, B. Am.
Meteorol. Soc., 94, 1553–1565, <a href="https://doi.org/10.1175/BAMS-D-12-00014.1" target="_blank">https://doi.org/10.1175/BAMS-D-12-00014.1</a>,
2013.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib12"><label>Hothorn et al.(2006)Hothorn, Hornik, and Zeileis</label><mixed-citation>
      
Hothorn, T., Hornik, K., and Zeileis, A.: Unbiased Recursive Partitioning: A
Conditional Inference Framework,
J. Comput. Graph. Stat., 15, 651–674, <a href="https://doi.org/10.1198/106186006X133933" target="_blank">https://doi.org/10.1198/106186006X133933</a>, 2006.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib13"><label>Hothorn et al.(2008)Hothorn, Hornik, Van De Wiel, and
Zeileis</label><mixed-citation>
      
Hothorn, T., Hornik, K., Van De Wiel, M. A., and Zeileis, A.: Implementing a
class of permutation tests: the coin package, J. Stat.
Softw., 28, 1–23, 2008.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib14"><label>Jordan et al.(2023)</label><mixed-citation>
      
Jordan, A. I., Krueger, F., Lerch, S.,
Allen, S., and Graeter, M.: scoringRules: Scoring Rules for Parametric and Simulated Distribution
Forecasts,
R package version 1.1.1,
<a href="https://cran.r-project.org/web/packages/scoringRules/" target="_blank"/>, 2023.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib15"><label>Kneib et al.(2021)Kneib, Silbersdorff, and
Säfken</label><mixed-citation>
      
Kneib, T., Silbersdorff, A., and Säfken, B.: Rage against the mean–a
review of distributional regression approaches,
Econometrics and Statistics, 26, 99–123,
<a href="https://doi.org/10.1016/j.ecosta.2021.07.006" target="_blank">https://doi.org/10.1016/j.ecosta.2021.07.006</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib16"><label>Lang et al.(2020)Lang, Lerch, Mayr, Simon, Stauffer, and
Zeileis</label><mixed-citation>
      
Lang, M. N., Lerch, S., Mayr, G. J., Simon, T., Stauffer, R., and Zeileis, A.: Remember the past: a comparison of time-adaptive training schemes for non-homogeneous regression, Nonlin. Processes Geophys., 27, 23–34, <a href="https://doi.org/10.5194/npg-27-23-2020" target="_blank">https://doi.org/10.5194/npg-27-23-2020</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib17"><label>Lerch and Thorarinsdottir(2013)</label><mixed-citation>
      
Lerch, S. and Thorarinsdottir, T. L.: Comparison of non-homogeneous regression
models for probabilistic wind speed forecasting, Tellus A, 65, 21206, <a href="https://doi.org/10.3402/tellusa.v65i0.21206" target="_blank">https://doi.org/10.3402/tellusa.v65i0.21206</a>,
2013.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib18"><label>Matheson and Winkler(1976)</label><mixed-citation>
      
Matheson, J. E. and Winkler, R. L.: Scoring rules for continuous probability
distributions, Manage. Sci., 22, 1087–1096, 1976.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib19"><label>Meinshausen(2017)</label><mixed-citation>
      
Meinshausen, N.: Quantregforest: quantile regression forests, R package
version,  1.3-7, <a href="https://cran.r-project.org/web/packages/quantregForest/" target="_blank"/> (last access: 15 November 2023), 2017.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib20"><label>Meinshausen and Ridgeway(2006)</label><mixed-citation>
      
Meinshausen, N. and Ridgeway, G.: Quantile regression forests, J.
Mach. Learn. Res., 7, 983–999, 2006.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib21"><label>Messner et al.(2016)Messner, Mayr, and
Zeileis</label><mixed-citation>
      
Messner, J. W., Mayr, G. J., and Zeileis, A.: Heteroscedastic Censored and
Truncated Regression with crch, R J., 8, 173–181, <a href="https://doi.org/10.32614/RJ-2016-012" target="_blank">https://doi.org/10.32614/RJ-2016-012</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib22"><label>Messner et al.(2017)Messner, Mayr, and
Zeileis</label><mixed-citation>
      
Messner, J. W., Mayr, G. J., and Zeileis, A.: Nonhomogeneous boosting for
predictor selection in ensemble postprocessing, Mon. Weather Rev., 145,
137–147, <a href="https://doi.org/10.1175/MWR-D-16-0088.1" target="_blank">https://doi.org/10.1175/MWR-D-16-0088.1</a>, 2017.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib23"><label>Rasp and Lerch(2018)</label><mixed-citation>
      
Rasp, S. and Lerch, S.: Neural networks for postprocessing ensemble weather
forecasts, Mon. Weather Rev., 146, 3885–3900,
<a href="https://doi.org/10.1175/MWR-D-18-0187.1" target="_blank">https://doi.org/10.1175/MWR-D-18-0187.1</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib24"><label>Rigby and Stasinopoulos(2005)</label><mixed-citation>
      
Rigby, R. A. and Stasinopoulos, D. M.: Generalized additive models for
location, scale and shape, J. Roy. Stat. Soc. C-App., 54, 507–554, <a href="https://doi.org/10.1111/j.1467-9876.2005.00510.x" target="_blank">https://doi.org/10.1111/j.1467-9876.2005.00510.x</a>,
2005.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib25"><label>Scheuerer(2014)</label><mixed-citation>
      
Scheuerer, M.: Probabilistic Quantitative Precipitation Forecasting Using
Ensemble Model Output Statistics, Q. J. Roy.
Meteor. Soc., 140, 1086–1096, <a href="https://doi.org/10.1002/qj.2183" target="_blank">https://doi.org/10.1002/qj.2183</a>, 2014.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib26"><label>Schlosser et al.(2019)Schlosser, Hothorn, Stauffer, and
Zeileis</label><mixed-citation>
      
Schlosser, L., Hothorn, T., Stauffer, R., and Zeileis, A.: Distributional
regression forests for probabilistic precipitation forecasting in complex
terrain, Ann. Appl. Stat., 13, 1564–1589,
<a href="https://doi.org/10.1214/19-AOAS1247" target="_blank">https://doi.org/10.1214/19-AOAS1247</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib27"><label>Schlosser et al.(2020)</label><mixed-citation>
      
Schlosser, L., Stauffer, R., and Zeileis, A.:
RainTyrol: Precipitation Observations and NWP Forecasts from GEFS,
R package version 0.2-0/r2952,
<a href="https://R-Forge.R-project.org/projects/partykit/" target="_blank"/>  (last access: 15 November 2023), 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib28"><label>Schlosser et al.(2021)</label><mixed-citation>
      
Schlosser, L.,  Lang, M. N.,  Hothorn, T., and Zeileis, A.:
disttree: Trees and Forests for Distributional Regression,
R package version 0.2-0/r3189,
<a href="https://R-Forge.R-project.org/projects/partykit/" target="_blank"/> (last access: 15 November 2023), 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib29"><label>Schoenach et al.(2020)Schoenach, Simon, and Mayr</label><mixed-citation>
      
Schoenach, D., Simon, T., and Mayr, G. J.: Postprocessing ensemble forecasts of vertical temperature profiles, Adv. Stat. Clim. Meteorol. Oceanogr., 6, 45–60, <a href="https://doi.org/10.5194/ascmo-6-45-2020" target="_blank">https://doi.org/10.5194/ascmo-6-45-2020</a>, 2020.


    </mixed-citation></ref-html>
<ref-html id="bib1.bib30"><label>Schulz and Lerch(2022)</label><mixed-citation>
      
Schulz, B. and Lerch, S.: Machine learning methods for postprocessing ensemble
forecasts of wind gusts: A systematic comparison, Mon. Weather Rev.,
150, 235–257, <a href="https://doi.org/10.1175/MWR-D-21-0150.1" target="_blank">https://doi.org/10.1175/MWR-D-21-0150.1</a>, 2022.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib31"><label>Seibold et al.(2018)Seibold, Zeileis, and
Hothorn</label><mixed-citation>
      
Seibold, H., Zeileis, A., and Hothorn, T.: Individual treatment effect
prediction for amyotrophic lateral sclerosis patients,
Stat. Methods  Med. Res., 27, 3104–3125, <a href="https://doi.org/10.1177/0962280217693034" target="_blank">https://doi.org/10.1177/0962280217693034</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib32"><label>Seibold et al.(2019)Seibold, Zeileis, and
Hothorn</label><mixed-citation>
      
Seibold, H., Zeileis, A., and Hothorn, T.: model4you: an R package for
personalised treatment effect estimation, J. Open Res. Softw.,
7, 17, <a href="https://doi.org/10.5334/jors.219" target="_blank">https://doi.org/10.5334/jors.219</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib33"><label>Simon et al.(2019)Simon, Mayr, Umlauf, and Zeileis</label><mixed-citation>
      
Simon, T., Mayr, G. J., Umlauf, N., and Zeileis, A.: NWP-based lightning prediction using flexible count data regression, Adv. Stat. Clim. Meteorol. Oceanogr., 5, 1–16, <a href="https://doi.org/10.5194/ascmo-5-1-2019" target="_blank">https://doi.org/10.5194/ascmo-5-1-2019</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib34"><label>Stauffer et al.(2017a)Stauffer, Mayr, Messner, Umlauf,
and Zeileis</label><mixed-citation>
      
Stauffer, R., Mayr, G. J., Messner, J. W., Umlauf, N., and Zeileis, A.:
Spatio-temporal precipitation climatology over complex terrain using a
censored additive regression model, Int. J. Climatol., 37,
3264–3275, <a href="https://doi.org/10.1002/joc.4913" target="_blank">https://doi.org/10.1002/joc.4913</a>, 2017a.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib35"><label>Stauffer et al.(2017b)Stauffer, Umlauf, Messner, Mayr,
and Zeileis</label><mixed-citation>
      
Stauffer, R., Umlauf, N., Messner, J. W., Mayr, G. J., and Zeileis, A.:
Ensemble postprocessing of daily precipitation sums over complex terrain
using censored high-resolution standardized anomalies, Mon. Weather
Rev., 145, 955–969, <a href="https://doi.org/10.1175/MWR-D-16-0260.1" target="_blank">https://doi.org/10.1175/MWR-D-16-0260.1</a>, 2017b.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib36"><label>Taillardat et al.(2016)Taillardat, Mestre, Zamo, and
Naveau</label><mixed-citation>
      
Taillardat, M., Mestre, O., Zamo, M., and Naveau, P.: Calibrated ensemble
forecasts using quantile regression forests and ensemble model output
statistics, Mon. Weather Rev., 144, 2375–2393,
<a href="https://doi.org/10.1175/MWR-D-15-0260.1" target="_blank">https://doi.org/10.1175/MWR-D-15-0260.1</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib37"><label>Taillardat et al.(2019)Taillardat, Fougères, Naveau, and
Mestre</label><mixed-citation>
      
Taillardat, M., Fougères, A.-L., Naveau, P., and Mestre, O.: Forest-based
and semiparametric methods for the postprocessing of rainfall ensemble
forecasting, Weather Forecast., 34, 617–634,
<a href="https://doi.org/10.1175/WAF-D-18-0149.1" target="_blank">https://doi.org/10.1175/WAF-D-18-0149.1</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib38"><label>Thorarinsdottir and
Gneiting(2010)</label><mixed-citation>
      
Thorarinsdottir, T. L. and Gneiting, T.: Probabilistic forecasts of wind speed:
Ensemble model output statistics by using heteroscedastic censored
regression, J. Roy. Stat. Soc. A Sta., 173, 371–388, <a href="https://doi.org/10.1111/j.1467-985X.2009.00616.x" target="_blank">https://doi.org/10.1111/j.1467-985X.2009.00616.x</a>, 2010.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib39"><label>Vannitsem et al.(2021)Vannitsem, Bremnes, Demaeyer, Evans, Flowerdew,
Hemri, Lerch, Roberts, Theis, Atencia et al.</label><mixed-citation>
      
Vannitsem, S., Bremnes, J. B., Demaeyer, J., Evans, G. R., Flowerdew, J.,
Hemri, S., Lerch, S., Roberts, N., Theis, S., Atencia, A., Bouallègue, Z. B., Bhend, J., Dabernig, M., De Cruz, L., Hieta, L., Mestre, O., Moret, L.,  Plenković, I. O., Schmeits, M., Taillardat, M., Van den Bergh, J., Van Schaeybroeck,  B., Whan, K., and Ylhaisi, J.:
Statistical postprocessing for weather forecasts: Review, challenges, and
avenues in a big data world, B. Am. Meteorol. Soc.,
102, E681–E699, <a href="https://doi.org/10.1175/BAMS-D-19-0308.1" target="_blank">https://doi.org/10.1175/BAMS-D-19-0308.1</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib40"><label>Zeileis et al.(2008)Zeileis, Hothorn, and Hornik</label><mixed-citation>
      
Zeileis, A., Hothorn, T., and Hornik, K.: Model-based recursive partitioning,
J. Computat. Graph. Stat., 17, 492–514,
<a href="https://doi.org/10.1198/106186008X319331" target="_blank">https://doi.org/10.1198/106186008X319331</a>, 2008.

    </mixed-citation></ref-html>--></article>
