<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing with OASIS Tables v3.0 20080202//EN" "journalpub-oasis3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:oasis="http://docs.oasis-open.org/ns/oasis-exchange/table" xml:lang="en" dtd-version="3.0" article-type="research-article"><?xmltex \bartext{Research article}?>
  <front>
    <journal-meta><journal-id journal-id-type="publisher">NPG</journal-id><journal-title-group>
    <journal-title>Nonlinear Processes in Geophysics</journal-title>
    <abbrev-journal-title abbrev-type="publisher">NPG</abbrev-journal-title><abbrev-journal-title abbrev-type="nlm-ta">Nonlin. Processes Geophys.</abbrev-journal-title>
  </journal-title-group><issn pub-type="epub">1607-7946</issn><publisher>
    <publisher-name>Copernicus Publications</publisher-name>
    <publisher-loc>Göttingen, Germany</publisher-loc>
  </publisher></journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.5194/npg-30-13-2023</article-id><title-group><article-title>Guidance on how to improve vertical covariance localization based on a 1000-member ensemble</article-title><alt-title>Guidance on how to improve vertical covariance localization</alt-title>
      </title-group><?xmltex \runningtitle{Guidance on how to improve vertical covariance localization}?><?xmltex \runningauthor{T. Necker et al.}?>
      <contrib-group>
        <contrib contrib-type="author" corresp="yes" rid="aff1">
          <name><surname>Necker</surname><given-names>Tobias</given-names></name>
          <email>tobias.necker@univie.ac.at</email>
        <ext-link>https://orcid.org/0000-0002-7484-3372</ext-link></contrib>
        <contrib contrib-type="author" corresp="no" rid="aff1">
          <name><surname>Hinger</surname><given-names>David</given-names></name>
          
        </contrib>
        <contrib contrib-type="author" corresp="no" rid="aff1">
          <name><surname>Griewank</surname><given-names>Philipp Johannes</given-names></name>
          
        <ext-link>https://orcid.org/0000-0003-0906-3553</ext-link></contrib>
        <contrib contrib-type="author" corresp="no" rid="aff2">
          <name><surname>Miyoshi</surname><given-names>Takemasa</given-names></name>
          
        <ext-link>https://orcid.org/0000-0003-3160-2525</ext-link></contrib>
        <contrib contrib-type="author" corresp="no" rid="aff1">
          <name><surname>Weissmann</surname><given-names>Martin</given-names></name>
          
        <ext-link>https://orcid.org/0000-0003-4073-1791</ext-link></contrib>
        <aff id="aff1"><label>1</label><institution>Institut für Meteorologie und Geophysik, Universität Wien, Vienna, Austria</institution>
        </aff>
        <aff id="aff2"><label>2</label><institution>RIKEN Center for Computational Science, Kobe, Japan</institution>
        </aff>
      </contrib-group>
      <author-notes><corresp id="corr1">Tobias Necker (tobias.necker@univie.ac.at)</corresp></author-notes><pub-date><day>9</day><month>January</month><year>2023</year></pub-date>
      
      <volume>30</volume>
      <issue>1</issue>
      <fpage>13</fpage><lpage>29</lpage>
      <history>
        <date date-type="received"><day>2</day><month>June</month><year>2022</year></date>
           <date date-type="rev-request"><day>9</day><month>June</month><year>2022</year></date>
           <date date-type="rev-recd"><day>14</day><month>December</month><year>2022</year></date>
           <date date-type="accepted"><day>14</day><month>December</month><year>2022</year></date>
      </history>
      <permissions>
        <copyright-statement>Copyright: © 2023 Tobias Necker et al.</copyright-statement>
        <copyright-year>2023</copyright-year>
      <license license-type="open-access"><license-p>This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this licence, visit <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link></license-p></license></permissions><self-uri xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023.html">This article is available from https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023.html</self-uri><self-uri xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023.pdf">The full text article is available as a PDF file from https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023.pdf</self-uri>
      <abstract><title>Abstract</title>

      <p id="d1e125">The success of ensemble data assimilation systems substantially depends on localization, which is required to mitigate sampling errors caused by modeling background error covariances with undersized ensembles. However, finding an optimal localization is highly challenging, as covariances, sampling errors, and appropriate localization depend on various factors. Our study investigates vertical localization based on a unique convection-permitting 1000-member ensemble simulation; 1000-member ensemble correlations serve as truth for examining vertical correlations and their sampling error. We discuss requirements for vertical localization by deriving an empirical optimal localization (EOL) that minimizes the sampling error in 40-member subsample correlations with respect to the 1000-member reference. Our analysis covers temperature, specific humidity, and wind correlations on various pressure levels. Results suggest that vertical localization should depend on several aspects, such as the respective variable, vertical level, or correlation type (self- or cross-correlations). Comparing the empirical optimal localization with common distance-dependent localization approaches highlights that finding suitable localization functions bears substantial room for improvement. Furthermore, we examine approaches for achieving positive semi-definiteness for covariance localization that hardly affect the sampling error reduction. Finally, we discuss the gain of combining different localization approaches with an adaptive statistical sampling error correction.</p>
  </abstract>
    </article-meta>
  </front>
<body>
      

<sec id="Ch1.S1" sec-type="intro">
  <label>1</label><title>Introduction</title>
      <p id="d1e137">The accuracy of the initial conditions provided by data assimilation systems strongly determines the skill of numerical weather prediction (NWP). Data assimilation (DA) relies on accurate estimates of forecast errors and error covariances that determine the weighting and spreading of observational information. However, modeling suitable error covariances is intrinsically difficult given various atmospheric processes acting on different scales, leading to situation- and flow-dependent error covariance structures. A breakthrough in estimating background errors has been the development of ensemble and hybrid data assimilation algorithms <xref ref-type="bibr" rid="bib1.bibx18 bib1.bibx13 bib1.bibx8" id="paren.1"><named-content content-type="pre">e.g.,</named-content></xref>.</p>
      <p id="d1e145">Considering the large state space of atmospheric models with a hundred million or more degrees of freedom, estimating error covariances with an ensemble forecast is demanding. Computational restrictions usually limit the number of affordable ensemble members to about 20 to 80 members <xref ref-type="bibr" rid="bib1.bibx8 bib1.bibx23" id="paren.2"/>. Ensemble systems, therefore, suffer from severe undersampling and sampling errors. For this reason, all ensemble and hybrid data assimilation systems require some form of sampling error correction for horizontal and vertical covariances, usually referred to as localization. Localization mitigates spurious correlations that arise from undersampling. During the assimilation procedure, spurious correlations lead to sub-optimal analysis increments, resulting in a sub-optimal analysis and forecast as well as an inaccurate representation of forecast error by the ensemble. Horizontal and vertical localizations are both challenging topics. Since fundamentally different processes are acting in the horizontal and vertical directions, the two structures require different solutions. Depending on the specific data assimilation algorithm, localization may also be important for other reasons, such as computational efficiency or rank deficiency. However, in this study, we focus on mitigating sampling errors independent of algorithm-specific constraints.</p>
      <p id="d1e151">In the past decade, advanced high-performance computing systems such as the Japanese K-computer <xref ref-type="bibr" rid="bib1.bibx48 bib1.bibx49 bib1.bibx50" id="paren.3"/> enabled the first atmospheric ensemble simulations with thousands of ensemble members that can provide reliable error covariances <xref ref-type="bibr" rid="bib1.bibx38 bib1.bibx47 bib1.bibx37 bib1.bibx55" id="paren.4"/>. The assumption that such large ensembles provide covariances close to true covariances allows one to investigate sampling errors in smaller subsets. <xref ref-type="bibr" rid="bib1.bibx56" id="text.5"/>, for example, evaluated a statistical sampling error correction method based on a 1000-member ensemble. Preceding studies used a similar approach but with a smaller ensemble size or lower resolution <xref ref-type="bibr" rid="bib1.bibx25 bib1.bibx60 bib1.bibx9" id="paren.6"><named-content content-type="pre">e.g.,</named-content></xref>. <xref ref-type="bibr" rid="bib1.bibx67" id="text.7"/>, for example, showed the potential of a 256-member ensemble for studying sampling errors in a 40-member ensemble focusing on covariances of radar observations on convective scales. Our present study aims to guide advances in vertical localization by analyzing vertical error correlations and the empirical optimal vertical localization derived from the convection-permitting 1000-member ensemble simulation of <xref ref-type="bibr" rid="bib1.bibx55 bib1.bibx56" id="text.8"/>.</p>
      <p id="d1e175">In recent years, several approaches for vertical localization have been developed. The most frequently applied localization approach is a distance-dependent localization that dampens long-range correlations <xref ref-type="bibr" rid="bib1.bibx30 bib1.bibx31 bib1.bibx25 bib1.bibx46" id="paren.9"><named-content content-type="pre">e.g.,</named-content></xref>. For example, many data assimilation algorithms use the Gaspari–Cohn tapering function <xref ref-type="bibr" rid="bib1.bibx20" id="paren.10"/>, which has a cut-off at a defined distance to dampen correlations depending on the spatial distance. However, long-distance vertical error correlations often have a physical meaning. Vertically, e.g., radiative effects of clouds, deep convection, or hydrostatic balance can cause relevant correlations. Inappropriate localization can therefore eliminate meaningful error correlations <xref ref-type="bibr" rid="bib1.bibx47 bib1.bibx37" id="paren.11"/> or cause imbalances in the initial conditions <xref ref-type="bibr" rid="bib1.bibx35 bib1.bibx22 bib1.bibx40" id="paren.12"/>.</p>
      <p id="d1e193">Several studies investigated different aspects of optimal localization but often focused on horizontal localization. These studies cover fundamental research on sampling errors and their correction <xref ref-type="bibr" rid="bib1.bibx1 bib1.bibx5 bib1.bibx19" id="paren.13"><named-content content-type="pre">e.g.,</named-content></xref>. Besides, some studies discuss suitable tapering functions for localization <xref ref-type="bibr" rid="bib1.bibx20 bib1.bibx21 bib1.bibx12 bib1.bibx63" id="paren.14"><named-content content-type="pre">e.g.,</named-content></xref>. Distance-dependent localization always requires tuning of localization scales. Consequently, multiple studies aim to derive optimal localization scales and functions by minimizing the error in correlations or the subsequent analysis <xref ref-type="bibr" rid="bib1.bibx58 bib1.bibx2 bib1.bibx39 bib1.bibx36 bib1.bibx19" id="paren.15"><named-content content-type="pre">e.g.,</named-content></xref>.</p>
      <p id="d1e211">Localization approaches can roughly be grouped into two categories: adaptive and non-adaptive approaches. Non-adaptive approaches apply fixed domain- or variable-uniform localization functions and scales that do not change with time. Adaptive localization approaches, such as statistical sampling error correction methods, enable a flow- or error-correlation-dependent localization <xref ref-type="bibr" rid="bib1.bibx1 bib1.bibx10 bib1.bibx11 bib1.bibx5 bib1.bibx52 bib1.bibx53" id="paren.16"><named-content content-type="pre">e.g.,</named-content></xref>. A promising adaptive localization approach is the global group ensemble filter <xref ref-type="bibr" rid="bib1.bibx39" id="paren.17"><named-content content-type="pre">GGF;</named-content></xref>. The GGF enables adaptive vertical localization of satellite radiances <xref ref-type="bibr" rid="bib1.bibx41 bib1.bibx43" id="paren.18"/>. However, adaptive methods usually require additional computational resources, which can be a limiting factor in operational applications.</p>
      <p id="d1e227">Current regional NWP models exhibit a grid spacing of a few kilometers, allowing an explicit representation of deep convection <xref ref-type="bibr" rid="bib1.bibx14 bib1.bibx24 bib1.bibx23" id="paren.19"/>. Finding optimal localization scales or functions is challenging, particularly for convection-permitting simulations <xref ref-type="bibr" rid="bib1.bibx45 bib1.bibx51 bib1.bibx17" id="paren.20"/>. In these simulations, correlations and sampling errors depend on strongly nonlinear dynamics, the chaotic nature of convection, and uncertainties in microphysical processes that all contribute to rapid error growth <xref ref-type="bibr" rid="bib1.bibx27 bib1.bibx51 bib1.bibx67" id="paren.21"/>. However, little knowledge exists on the structure of short-term forecast errors in regions with atmospheric convection <xref ref-type="bibr" rid="bib1.bibx33" id="paren.22"/>. Consequently, better understanding of optimal vertical localization for convection-permitting simulations has the potential to improve forecasts of convective precipitation and related hazards.</p>
      <p id="d1e242">This paper investigates how vertical error covariances should be localized based on an existing convection-permitting 1000-member ensemble simulation <xref ref-type="bibr" rid="bib1.bibx55" id="paren.23"/>. Our study focuses on correlations instead of covariances as correlation sampling errors are the main contributor to covariance sampling error <xref ref-type="bibr" rid="bib1.bibx5" id="paren.24"/>. We will investigate domain-uniform vertical localization but will also partly address the potential of adaptive localization approaches by applying a statistical sampling error correction <xref ref-type="bibr" rid="bib1.bibx5 bib1.bibx6" id="paren.25"><named-content content-type="pre">SEC</named-content></xref>. Furthermore, we will analyze vertical correlations and empirically derive an optimal vertical localization that minimizes the sampling error in subsamples of the 1000-member ensemble. Since the optimal localization matrix is not necessarily symmetric positive semi-definite (SPSD), we explore methods to ensure SPSD. Our setup allows for general conclusions independent of a specific DA algorithm. Among different aspects of localization, we will address the following research questions.
<list list-type="bullet"><list-item>
      <p id="d1e258">How do vertical error correlations for humidity, temperature, or wind behave on average?</p></list-item><list-item>
      <p id="d1e262">How should we localize vertical error correlations from small ensembles?</p></list-item><list-item>
      <p id="d1e266">How much error reduction can be achieved with a domain-uniform vertical localization or by combining different localization approaches?</p></list-item></list></p>
      <p id="d1e269">The remainder of the paper is outlined as follows:
Sect. <xref ref-type="sec" rid="Ch1.S2"/> introduces the 1000-member ensemble, the experimental setup, and the weather period. Furthermore, we explain how vertical correlations and the empirical optimal localization are derived from the 1000-member ensemble using subsampling. Section <xref ref-type="sec" rid="Ch1.S3.SS1"/> evaluates vertical correlations and the empirical optimal localization for single variable pairs to explore requirements for a variable-dependent localization. In Sect. <xref ref-type="sec" rid="Ch1.S3.SS2"/>, we group variables and correlations based on similar behavior to derive an empirical optimal localization for self- and cross-correlations. Section <xref ref-type="sec" rid="Ch1.S3.SS3"/> evaluates the error reduction achieved by different localization approaches and settings. Section <xref ref-type="sec" rid="Ch1.S3.SS4"/> discusses different methods to ensure that the localization matrix is SPSD. Finally, we summarize our results in Sect. <xref ref-type="sec" rid="Ch1.S4"/> and discuss implications for improving vertical localization.</p>
</sec>
<sec id="Ch1.S2">
  <label>2</label><title>Methods and experiments</title>
<sec id="Ch1.S2.SS1">
  <label>2.1</label><title>1000-member ensemble simulation</title>
      <p id="d1e300">Our study uses an existing convective-scale 1000-member ensemble simulation described in detail by <xref ref-type="bibr" rid="bib1.bibx55" id="text.26"/>. The 1000-member ensemble applies the full-physics non-hydrostatic Scalable Computing for Advanced Library and Environment regional model (SCALE-RM) and the SCALE Localized Ensemble Transform Kalman Filter (SCALE-LETKF) DA system <xref ref-type="bibr" rid="bib1.bibx44" id="paren.27"/>. Using an offline nesting approach, the 1000-member ensemble setup couples two domains with different horizontal resolutions. Ensemble forecasts in the outer domain covering central Europe (15 km grid spacing) delivered the boundary and initial conditions for the convective-scale ensemble forecasts in the inner domain covering Germany (3 km grid spacing). High-resolution short-term forecasts from the inner domain will be analyzed to evaluate correlations and localization.</p>
      <p id="d1e309">Initial and boundary conditions: the data assimilation cycling has been performed in the coarse European domain assimilating conventional observations with a LETKF <xref ref-type="bibr" rid="bib1.bibx34" id="paren.28"/>. A set of 1000 independent and specifically constructed ensemble boundary conditions (BCs) drives the European-scale forecasts. These BCs combine 1000 climatologically scaled random perturbations with a 20-member analysis ensemble of the NCEP Global Ensemble Forecast System (GEFS). The GEFS 20-member analysis ensemble is used 50 times to reach 1000 BCs and afterwards combined with 1000 random climatologically scaled perturbations. This approach yields 1000 independent BCs that ensure sufficient ensemble spread when combined with relaxation to prior spread <xref ref-type="bibr" rid="bib1.bibx66" id="paren.29"><named-content content-type="pre">RTPS</named-content></xref>. The boundary and initial conditions for the inner and convective-scale forecast domain are downscaled from 15 to 3 km resolution based on simulations in the European domain.</p>
      <p id="d1e320">Our study uses the model output from the inner model domain with a <inline-formula><mml:math id="M1" display="inline"><mml:mrow><mml:mn mathvariant="normal">250</mml:mn><mml:mo>×</mml:mo><mml:mn mathvariant="normal">230</mml:mn></mml:mrow></mml:math></inline-formula> grid area centered over Germany with a 3 km horizontal resolution. This sub-domain excludes the Alps and regions within 10 grid points of the domain boundary. The model output has 30 vertical levels ranging from the surface to the model top at 16.9 km. The original vertical grid is terrain-following and has fixed height levels above the surface (m). For practical reasons, we extracted temperature (<inline-formula><mml:math id="M2" display="inline"><mml:mi>T</mml:mi></mml:math></inline-formula>), specific humidity (<inline-formula><mml:math id="M3" display="inline"><mml:mi>Q</mml:mi></mml:math></inline-formula>), and horizontal zonal (<inline-formula><mml:math id="M4" display="inline"><mml:mi>U</mml:mi></mml:math></inline-formula>) and meridional (<inline-formula><mml:math id="M5" display="inline"><mml:mi>V</mml:mi></mml:math></inline-formula>) wind components on 20 vertical pressure levels (<inline-formula><mml:math id="M6" display="inline"><mml:mn mathvariant="normal">100</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M7" display="inline"><mml:mn mathvariant="normal">150</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M8" display="inline"><mml:mn mathvariant="normal">200</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M9" display="inline"><mml:mn mathvariant="normal">250</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M10" display="inline"><mml:mn mathvariant="normal">300</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M11" display="inline"><mml:mn mathvariant="normal">350</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M12" display="inline"><mml:mn mathvariant="normal">400</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M13" display="inline"><mml:mn mathvariant="normal">450</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M14" display="inline"><mml:mn mathvariant="normal">500</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M15" display="inline"><mml:mn mathvariant="normal">550</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M16" display="inline"><mml:mn mathvariant="normal">600</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M17" display="inline"><mml:mn mathvariant="normal">650</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M18" display="inline"><mml:mn mathvariant="normal">700</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M19" display="inline"><mml:mn mathvariant="normal">750</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M20" display="inline"><mml:mn mathvariant="normal">800</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M21" display="inline"><mml:mn mathvariant="normal">850</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M22" display="inline"><mml:mn mathvariant="normal">900</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M23" display="inline"><mml:mn mathvariant="normal">925</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M24" display="inline"><mml:mn mathvariant="normal">950</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M25" display="inline"><mml:mn mathvariant="normal">975</mml:mn></mml:math></inline-formula> hPa). Performing our analysis on this modified grid allows horizontal averaging of data on pressure levels where needed. Overall, our examination includes 10 short-term forecasts that were initialized twice per day from 29 May to 2 June 2016 at 00:00 and 12:00 UTC. The 3 h forecasts valid at 03:00 and 15:00 UTC serve as a basis for computing and investigating background error correlations.</p>
</sec>
<sec id="Ch1.S2.SS2">
  <label>2.2</label><title>Weather period</title>
      <p id="d1e515">Atmospheric blocking over the Atlantic influenced the large-scale flow over Europe in the 5 d experimental period. The blocking led to a quasi-stationary weather pattern over central Europe with an upper-level trough over western Europe and a shallow surface low over central Europe. The low-pressure system was associated with a cold front and a warm front that moved over Germany during the period. A convergence zone over southern Germany caused large-scale lifting. Furthermore, mid-level winds advected warm and moist air masses from southern Europe towards Germany at the beginning of the experimental period. Combined with the convergence zone, atmospheric conditions led to intense convection and heavy precipitation, including hail. Weak pressure gradients and slowly moving convective cells resulted in high local precipitation rates and flash flooding. Due to these severe weather events, several studies focused on this exceptional period <xref ref-type="bibr" rid="bib1.bibx59" id="paren.30"><named-content content-type="pre">e.g.,</named-content></xref>. <xref ref-type="bibr" rid="bib1.bibx55 bib1.bibx56" id="text.31"/>, <xref ref-type="bibr" rid="bib1.bibx57" id="text.32"/>, and <xref ref-type="bibr" rid="bib1.bibx15" id="text.33"/> provide further details on the weather situation in this period as these studies also explore the 1000-member ensemble simulation with a different purpose.</p>
</sec>
<sec id="Ch1.S2.SS3">
  <label>2.3</label><title>Vertical localization</title>
      <p id="d1e540">Error covariances are a key component in data assimilation and determine how assimilated information is weighted and distributed in state space. Given a sample of state vectors <inline-formula><mml:math id="M26" display="inline"><mml:mrow><mml:msup><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mi>n</mml:mi></mml:msup></mml:mrow></mml:math></inline-formula> provided by a background forecast ensemble, the flow-dependent sample error covariance matrix <inline-formula><mml:math id="M27" display="inline"><mml:mi mathvariant="bold">P</mml:mi></mml:math></inline-formula> can be computed as follows:
            <disp-formula id="Ch1.E1" content-type="numbered"><label>1</label><mml:math id="M28" display="block"><mml:mrow><mml:mi mathvariant="bold">P</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mn mathvariant="normal">1</mml:mn><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:mfrac></mml:mstyle><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mi>n</mml:mi></mml:msup><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mo mathvariant="normal">‾</mml:mo></mml:mover><mml:mo>)</mml:mo><mml:mo>(</mml:mo><mml:msup><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mi>n</mml:mi></mml:msup><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mo mathvariant="normal">‾</mml:mo></mml:mover><mml:msup><mml:mo>)</mml:mo><mml:mi>T</mml:mi></mml:msup><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
          where <inline-formula><mml:math id="M29" display="inline"><mml:mi>N</mml:mi></mml:math></inline-formula> is the ensemble size and <inline-formula><mml:math id="M30" display="inline"><mml:mover accent="true"><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mo mathvariant="normal">‾</mml:mo></mml:mover></mml:math></inline-formula> is the ensemble mean state. The covariance matrix <inline-formula><mml:math id="M31" display="inline"><mml:mi mathvariant="bold">P</mml:mi></mml:math></inline-formula> by definition is a symmetric, positive semi-definite matrix with variances on its diagonal entries and covariances on its off-diagonal entries. Each off-diagonal element contains a sample covariance (cov) of two state variables <inline-formula><mml:math id="M32" display="inline"><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>:
            <disp-formula id="Ch1.E2" content-type="numbered"><label>2</label><mml:math id="M33" display="block"><mml:mrow><mml:mi mathvariant="normal">cov</mml:mi><mml:mo>(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msub><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mi>r</mml:mi><mml:mo>(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msub><mml:mo>)</mml:mo><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:mo>)</mml:mo><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msub><mml:mo>)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
          where <inline-formula><mml:math id="M34" display="inline"><mml:mrow><mml:mi>r</mml:mi><mml:mo>∈</mml:mo><mml:mo>[</mml:mo><mml:mo>-</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>,</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> is the sample correlation and <inline-formula><mml:math id="M35" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula> is the sample standard deviation.</p>
      <p id="d1e766">Usually, the number of affordable ensemble members is limited in NWP due to a huge state space and computational restrictions. This deficit causes severe sampling errors. Consequently, all ensemble filters require a correction of sampling errors, often referred to as localization. For example, <xref ref-type="bibr" rid="bib1.bibx5" id="text.34"/> highlighted that the sampling error in covariances is dominated by sampling error in the sample correlation <inline-formula><mml:math id="M36" display="inline"><mml:mi>r</mml:mi></mml:math></inline-formula>, not by sampling error in the variances. Therefore, our analysis will solely tackle sampling errors in sample correlations. Sample correlations are normalized with standard deviations and possess no unit. The normalization allows comparison or combination of correlations of different variables, facilitating the interpretation.</p>
      <p id="d1e779">The implementation of localization depends on various factors determined by the type of ensemble filter. Usually, localization is applied directly to the background error covariance matrix using a Schur product:
            <disp-formula id="Ch1.E3" content-type="numbered"><label>3</label><mml:math id="M37" display="block"><mml:mrow><mml:msub><mml:mi mathvariant="bold">P</mml:mi><mml:mi mathvariant="normal">loc</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mi mathvariant="bold">C</mml:mi><mml:mo>∘</mml:mo><mml:mi mathvariant="bold">P</mml:mi><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
          where <inline-formula><mml:math id="M38" display="inline"><mml:mi mathvariant="bold">C</mml:mi></mml:math></inline-formula> is the localization matrix <xref ref-type="bibr" rid="bib1.bibx20 bib1.bibx7" id="paren.35"/>. The matrix <inline-formula><mml:math id="M39" display="inline"><mml:mi mathvariant="bold">C</mml:mi></mml:math></inline-formula> consists of tapering factors <inline-formula><mml:math id="M40" display="inline"><mml:mi mathvariant="italic">α</mml:mi></mml:math></inline-formula> that are determined using the localization approach of choice. Based on the Schur product theorem <xref ref-type="bibr" rid="bib1.bibx28" id="paren.36"><named-content content-type="post">theorem 7.5.3</named-content></xref>, positive semi-definite matrices <inline-formula><mml:math id="M41" display="inline"><mml:mi mathvariant="bold">C</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M42" display="inline"><mml:mi mathvariant="bold">P</mml:mi></mml:math></inline-formula> guarantee positive semi-definiteness of the localized covariance matrix <inline-formula><mml:math id="M43" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="bold">P</mml:mi><mml:mi mathvariant="normal">loc</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>. Furthermore, localization matrices should feature ones on the diagonal similar to a correlation matrix to avoid undesired inflation of variances <xref ref-type="bibr" rid="bib1.bibx19" id="paren.37"/>.</p>
<sec id="Ch1.S2.SS3.SSS1">
  <label>2.3.1</label><title>Distance-dependent localization</title>
      <p id="d1e868">The most common localization approach is a distance-dependent localization that determines tapering factors <inline-formula><mml:math id="M44" display="inline"><mml:mi mathvariant="italic">α</mml:mi></mml:math></inline-formula> based on distance <xref ref-type="bibr" rid="bib1.bibx30 bib1.bibx31" id="paren.38"/>. The vertical separation distance in our study is defined in <inline-formula><mml:math id="M45" display="inline"><mml:mrow><mml:mi>ln⁡</mml:mi><mml:mo>(</mml:mo><mml:mi>p</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. We consider the widely used Gaspari–Cohn function <xref ref-type="bibr" rid="bib1.bibx20" id="paren.39"><named-content content-type="pre">GC; Eq. 4.10,</named-content></xref> for comparison with other methods. Applying a GC function always requires the selection of the separation distance but guarantees positive definiteness. The separation distance is often referred to as the localization scale, while the cut-off radius is usually twice the localization scale. In our study, we apply vertical localization according to the definition of the Deutscher Wetterdienst (DWD) <xref ref-type="bibr" rid="bib1.bibx62" id="paren.40"/>. For the DWD, the localization scale is determined by a pre-selected localization length that is multiplied by a factor of <inline-formula><mml:math id="M46" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:msqrt><mml:mrow><mml:mn mathvariant="normal">10</mml:mn><mml:mo>/</mml:mo><mml:mn mathvariant="normal">3</mml:mn></mml:mrow></mml:msqrt><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. Operationally, the localization length of the DWD is height-dependent and increases linearly in <inline-formula><mml:math id="M47" display="inline"><mml:mrow><mml:mi>ln⁡</mml:mi><mml:mo>(</mml:mo><mml:mi>p</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> from the surface (<inline-formula><mml:math id="M48" display="inline"><mml:mn mathvariant="normal">0.075</mml:mn></mml:math></inline-formula>) to 300 hPa (<inline-formula><mml:math id="M49" display="inline"><mml:mn mathvariant="normal">0.5</mml:mn></mml:math></inline-formula>).</p>
      <p id="d1e950">In Sect. <xref ref-type="sec" rid="Ch1.S3.SS3"/>, we apply three different domain-uniform GC localization setups. (a) <italic>GC</italic>: an optimally tuned GC localization scale that applies a uniform localization scale for all variables and heights. (b) <italic>GCLEV</italic>: a height-dependent optimally tuned GC localization scale that is uniform for all variables. (c) <italic>DWD</italic>: a localization setting similar to the DWD as described above that is also domain- and variable-uniform.</p>
</sec>
<sec id="Ch1.S2.SS3.SSS2">
  <label>2.3.2</label><title>SEC</title>
      <p id="d1e973"><xref ref-type="bibr" rid="bib1.bibx56" id="text.41"/> showed that an adaptive statistical SEC <xref ref-type="bibr" rid="bib1.bibx5 bib1.bibx6" id="paren.42"/> substantially reduces the sampling error in sample correlations and ensemble sensitivities. The SEC is a look-up table-based approach and corrects the overestimation of correlations caused by sampling noise. The look-up table is computed offline and is based on Monte Carlo simulations that consider the likelihood of the correlation <inline-formula><mml:math id="M50" display="inline"><mml:mi>r</mml:mi></mml:math></inline-formula>. The SEC depends on the sample correlation, the ensemble size, and the assumed prior distribution of correlations. Here, we assume a uniform default prior distribution and apply the SEC table that is provided within the Data Assimilation Research Testbed <xref ref-type="bibr" rid="bib1.bibx3" id="paren.43"><named-content content-type="pre">DART;</named-content></xref>. In Sect. <xref ref-type="sec" rid="Ch1.S3.SS3"/>, we will compare the benefit of the SEC with different localization approaches. The comparison includes combinations of the SEC with these approaches.</p>

<?xmltex \floatpos{t}?><table-wrap id="Ch1.T1" specific-use="star"><?xmltex \currentcnt{1}?><label>Table 1</label><caption><p id="d1e998">Analyzed correlation pairs. Self-correlations on the diagonal and cross-correlations on the off-diagonal of the table. The first variable of each pair represents the ensemble at the reference level.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="5">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="center"/>
     <oasis:colspec colnum="3" colname="col3" align="center"/>
     <oasis:colspec colnum="4" colname="col4" align="center"/>
     <oasis:colspec colnum="5" colname="col5" align="center"/>
     <oasis:thead>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Variable</oasis:entry>
         <oasis:entry colname="col2">Temperature (<inline-formula><mml:math id="M51" display="inline"><mml:mi>T</mml:mi></mml:math></inline-formula>)</oasis:entry>
         <oasis:entry colname="col3">Humidity (<inline-formula><mml:math id="M52" display="inline"><mml:mi>Q</mml:mi></mml:math></inline-formula>)</oasis:entry>
         <oasis:entry colname="col4">Zon. wind (<inline-formula><mml:math id="M53" display="inline"><mml:mi>U</mml:mi></mml:math></inline-formula>)</oasis:entry>
         <oasis:entry colname="col5">Mer. wind (<inline-formula><mml:math id="M54" display="inline"><mml:mi>V</mml:mi></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">Temperature  (<inline-formula><mml:math id="M55" display="inline"><mml:mi>T</mml:mi></mml:math></inline-formula>)</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M56" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col3"><inline-formula><mml:math id="M57" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col4"><inline-formula><mml:math id="M58" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col5"><inline-formula><mml:math id="M59" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>V</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Humidity  (<inline-formula><mml:math id="M60" display="inline"><mml:mi>Q</mml:mi></mml:math></inline-formula>)</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M61" display="inline"><mml:mrow><mml:mi>Q</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col3"><inline-formula><mml:math id="M62" display="inline"><mml:mrow><mml:mi>Q</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col4"><inline-formula><mml:math id="M63" display="inline"><mml:mrow><mml:mi>Q</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col5"><inline-formula><mml:math id="M64" display="inline"><mml:mrow><mml:mi>Q</mml:mi><mml:mi>V</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Zon. wind  (<inline-formula><mml:math id="M65" display="inline"><mml:mi>U</mml:mi></mml:math></inline-formula>)</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M66" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col3"><inline-formula><mml:math id="M67" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col4"><inline-formula><mml:math id="M68" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col5"><inline-formula><mml:math id="M69" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>V</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Mer. wind  (<inline-formula><mml:math id="M70" display="inline"><mml:mi>V</mml:mi></mml:math></inline-formula>)</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M71" display="inline"><mml:mrow><mml:mi>V</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col3"><inline-formula><mml:math id="M72" display="inline"><mml:mrow><mml:mi>V</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col4"><inline-formula><mml:math id="M73" display="inline"><mml:mrow><mml:mi>V</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col5"><inline-formula><mml:math id="M74" display="inline"><mml:mrow><mml:mi>V</mml:mi><mml:mi>V</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

</sec>
</sec>
<sec id="Ch1.S2.SS4">
  <label>2.4</label><title>Subsampling and vertical correlations</title>
      <p id="d1e1306">The sampling noise expected for zero correlation estimates and sample size <inline-formula><mml:math id="M75" display="inline"><mml:mi>N</mml:mi></mml:math></inline-formula> is <inline-formula><mml:math id="M76" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:msqrt><mml:mi>N</mml:mi></mml:msqrt><mml:msup><mml:mo>)</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> <xref ref-type="bibr" rid="bib1.bibx30" id="paren.44"/>. For the 1000-member ensemble (<inline-formula><mml:math id="M77" display="inline"><mml:mrow><mml:mi>N</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1000</mml:mn></mml:mrow></mml:math></inline-formula>), this estimation yields a very small sampling noise of approximately <inline-formula><mml:math id="M78" display="inline"><mml:mrow><mml:mn mathvariant="normal">3</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>. In comparison, a 40-member ensemble reveals an expected sampling noise of approximately <inline-formula><mml:math id="M79" display="inline"><mml:mrow><mml:mn mathvariant="normal">16</mml:mn><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>. Throughout this study, correlations computed using the full 1000-member ensemble serve as truth (<inline-formula><mml:math id="M80" display="inline"><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn mathvariant="normal">1000</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula>) for the interpretation of vertical correlations and the evaluation of sampling errors and localization in smaller subsamples of the full ensemble. We focus on vertical correlations and sampling errors in 40-member subsamples as this is a typical ensemble size applied by operational weather services such as, e.g., the Deutscher Wetterdienst. Preceding studies applied a similar approach for studying sampling errors <xref ref-type="bibr" rid="bib1.bibx25 bib1.bibx60 bib1.bibx9 bib1.bibx55 bib1.bibx56" id="paren.45"/>.</p>
      <p id="d1e1387">The present study will adopt the subsampling approach from <xref ref-type="bibr" rid="bib1.bibx55" id="text.46"/> and <xref ref-type="bibr" rid="bib1.bibx15" id="text.47"/>. The 1000-member ensemble provides 25 randomly drawn 40-member subsamples without repetition of members (illustrated in Fig. <xref ref-type="fig" rid="Ch1.F1"/>a). We assume that the 40-member sub-ensembles of the 1000-member ensemble statically have sampling errors similar to what those independent 40-member ensemble EnKF systems would have. As mentioned above, we will analyze 10 3 h forecasts. This setup results in a sample of 250 ensemble forecasts with 40 members that we can compare to the 10 ensemble forecasts with 1000 members. The model domain has <inline-formula><mml:math id="M81" display="inline"><mml:mrow><mml:mn mathvariant="normal">250</mml:mn><mml:mo>×</mml:mo><mml:mn mathvariant="normal">230</mml:mn></mml:mrow></mml:math></inline-formula> grid points yielding 57 500 vertical columns in our domain. We will, therefore, analyze approximately <inline-formula><mml:math id="M82" display="inline"><mml:mrow><mml:mn mathvariant="normal">11.5</mml:mn><mml:mo>×</mml:mo><mml:msup><mml:mn mathvariant="normal">10</mml:mn><mml:mn mathvariant="normal">6</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> true and <inline-formula><mml:math id="M83" display="inline"><mml:mrow><mml:mn mathvariant="normal">287.5</mml:mn><mml:mo>×</mml:mo><mml:msup><mml:mn mathvariant="normal">10</mml:mn><mml:mn mathvariant="normal">6</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> 40-member vertical correlation profiles per variable pair, accounting for all 20 reference levels. This data set allows robust statistical analysis of error correlations, but it should be noted that error correlations may differ for other periods and regions.</p>
      <p id="d1e1441">In the present study, we will analyze four prognostic variables: temperature (<inline-formula><mml:math id="M84" display="inline"><mml:mi>T</mml:mi></mml:math></inline-formula>), specific humidity (<inline-formula><mml:math id="M85" display="inline"><mml:mi>Q</mml:mi></mml:math></inline-formula>), zonal wind (<inline-formula><mml:math id="M86" display="inline"><mml:mi>U</mml:mi></mml:math></inline-formula>), and meridional wind (<inline-formula><mml:math id="M87" display="inline"><mml:mi>V</mml:mi></mml:math></inline-formula>). This setup yields 16 correlation pairs (Table <xref ref-type="table" rid="Ch1.T1"/>) that we will inspect for different reference levels. Furthermore, we will group correlations as “self” (e.g., temperature–temperature as shown in Fig. <xref ref-type="fig" rid="Ch1.F1"/>) or “cross” (e.g., temperature–humidity) correlations to highlight common behavior. Subsequently, we will use the correlation coding shown in Table <xref ref-type="table" rid="Ch1.T1"/>. For example, “<inline-formula><mml:math id="M88" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:math></inline-formula>” combines all temperature correlations from the reference level to specific humidity at all other vertical levels in a column. Throughout the paper, we will mainly present results for the <inline-formula><mml:math id="M89" display="inline"><mml:mi>U</mml:mi></mml:math></inline-formula>-wind component as conclusions for the <inline-formula><mml:math id="M90" display="inline"><mml:mi>V</mml:mi></mml:math></inline-formula>-wind component are similar.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F1" specific-use="star"><?xmltex \currentcnt{1}?><?xmltex \def\figurename{Figure}?><label>Figure 1</label><caption><p id="d1e1506">Vertical temperature–temperature correlations and empirical optimal localization for reference level 500 hPa on 29 May 2016, 15:00 UTC. <bold>(a)</bold> Single random column. <bold>(b)</bold> Domain average. <bold>(c)</bold> Estimated domain-uniform EOL and absolute error reduction. The sample includes the correlations from all 25 40-member subsamples. Shading indicates spatial variability.</p></caption>
          <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f01.png"/>

        </fig>

<sec id="Ch1.S2.SS4.SSSx1" specific-use="unnumbered">
  <title>Example of vertical correlations</title>
      <p id="d1e1529">Figure <xref ref-type="fig" rid="Ch1.F1"/>a shows an example of vertical self-correlations of temperature (<inline-formula><mml:math id="M91" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula>) from reference level 500 hPa to all other levels in a single random vertical column. The 1000-member correlation (also referred to as the true correlation) is 1 at the reference level and drops to 0.5 after approximately 100 hPa vertical distance. Given the true correlation, the temperature at 500 hPa weakly correlates with the temperature in the boundary layer. This weak correlation is linked to cloud shadowing by mid-tropospheric clouds and resulting colder near-surface temperatures. Almost no correlation is visible to levels above the tropopause, which lies around 200 hPa. Most 40-member sample correlations strongly deviate from the true correlation, highlighting the severe undersampling issue. Sampling errors appear to be larger with increasing distance and smaller correlation values. This behavior motivates most distance-based localization approaches with predefined tapering functions that damp distant correlations. However, such an approach might cut off significant non-zero correlations, as seen for the boundary layer close to the surface in this example.</p>
      <p id="d1e1544">Throughout this paper we will analyze the 1000-member horizontally averaged absolute vertical correlation to support the discussion of the empirical optimal localization. Averaged absolute correlations are computed as follows:
              <disp-formula id="Ch1.E4" content-type="numbered"><label>4</label><mml:math id="M92" display="block"><mml:mrow><mml:mover accent="true"><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn mathvariant="normal">1000</mml:mn></mml:msup></mml:mrow><mml:mo mathvariant="normal">‾</mml:mo></mml:mover><mml:mo>(</mml:mo><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mi>z</mml:mi><mml:mo>,</mml:mo><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>A</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mn mathvariant="normal">1</mml:mn><mml:mi>K</mml:mi></mml:mfrac></mml:mstyle><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:mo>(</mml:mo><mml:mo>|</mml:mo><mml:msubsup><mml:mi>r</mml:mi><mml:mi>k</mml:mi><mml:mn mathvariant="normal">1000</mml:mn></mml:msubsup><mml:mo>|</mml:mo><mml:mo>)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
            where <inline-formula><mml:math id="M93" display="inline"><mml:mi>K</mml:mi></mml:math></inline-formula> is the number of vertical columns in the domain. This analysis will be done separately for different forecasts <inline-formula><mml:math id="M94" display="inline"><mml:mi>t</mml:mi></mml:math></inline-formula>, reference levels <inline-formula><mml:math id="M95" display="inline"><mml:mi>z</mml:mi></mml:math></inline-formula>, pressure levels <inline-formula><mml:math id="M96" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula>, and variable pairs <inline-formula><mml:math id="M97" display="inline"><mml:mi>A</mml:mi></mml:math></inline-formula>.</p>
      <p id="d1e1654">Figure <xref ref-type="fig" rid="Ch1.F1"/>b displays an example of a mean absolute temperature self-correlation (<inline-formula><mml:math id="M98" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula>) for reference level 500 hPa and a single date. On average, the mean absolute correlation of all 40-member subsamples well captures the shape of the true mean absolute correlation. However, 40-member ensembles overestimate the absolute correlation due to sampling error for weaker correlations and larger distances. Furthermore, the 40-member correlations reveal a larger variance. Plotted in <inline-formula><mml:math id="M99" display="inline"><mml:mrow><mml:mi>ln⁡</mml:mi><mml:mo>(</mml:mo><mml:mi>p</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, true and 40-member mean absolute correlations decay nearly symmetrically with increasing vertical distance from the reference level. This behavior explains why distance-dependent vertical localization scales are defined in logarithmic pressure coordinates.</p>
</sec>
</sec>
<sec id="Ch1.S2.SS5">
  <label>2.5</label><title>Empirical optimal localization (EOL)</title>
      <p id="d1e1693">Our goal is to empirically find the optimal localization factor <inline-formula><mml:math id="M100" display="inline"><mml:mi mathvariant="italic">α</mml:mi></mml:math></inline-formula> that minimizes the sampling error or cost function <inline-formula><mml:math id="M101" display="inline"><mml:mi>J</mml:mi></mml:math></inline-formula>:
            <disp-formula id="Ch1.E5" content-type="numbered"><label>5</label><mml:math id="M102" display="block"><mml:mrow><mml:mi>J</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">α</mml:mi><mml:mo>,</mml:mo><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mi>z</mml:mi><mml:mo>,</mml:mo><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>A</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:msqrt><mml:mrow><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>s</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>S</mml:mi></mml:munderover><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:msup><mml:mfenced open="(" close=")"><mml:mrow><mml:mi mathvariant="italic">α</mml:mi><mml:msubsup><mml:mi>r</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mn mathvariant="normal">40</mml:mn></mml:msubsup><mml:mo>-</mml:mo><mml:msubsup><mml:mi>r</mml:mi><mml:mi>k</mml:mi><mml:mn mathvariant="normal">1000</mml:mn></mml:msubsup></mml:mrow></mml:mfenced><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:msqrt><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
          where the minimization is done separately for each forecast time <inline-formula><mml:math id="M103" display="inline"><mml:mi>t</mml:mi></mml:math></inline-formula>, reference level <inline-formula><mml:math id="M104" display="inline"><mml:mi>z</mml:mi></mml:math></inline-formula>, pressure level <inline-formula><mml:math id="M105" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula>, and variable pair <inline-formula><mml:math id="M106" display="inline"><mml:mi>A</mml:mi></mml:math></inline-formula>. <inline-formula><mml:math id="M107" display="inline"><mml:mi>S</mml:mi></mml:math></inline-formula> is the number of 40-member ensembles (<inline-formula><mml:math id="M108" display="inline"><mml:mrow><mml:mi>S</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">25</mml:mn></mml:mrow></mml:math></inline-formula>). This is
equivalent to finding the <inline-formula><mml:math id="M109" display="inline"><mml:mi mathvariant="italic">α</mml:mi></mml:math></inline-formula> that minimizes
            <disp-formula id="Ch1.E6" content-type="numbered"><label>6</label><mml:math id="M110" display="block"><mml:mrow><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>s</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>S</mml:mi></mml:munderover><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:mfenced close="]" open="["><mml:mrow><mml:msup><mml:mi mathvariant="italic">α</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:msup><mml:mfenced open="(" close=")"><mml:mrow><mml:msubsup><mml:mi>r</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mn mathvariant="normal">40</mml:mn></mml:msubsup></mml:mrow></mml:mfenced><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>-</mml:mo><mml:mn mathvariant="normal">2</mml:mn><mml:mi mathvariant="italic">α</mml:mi><mml:msubsup><mml:mi>r</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mn mathvariant="normal">40</mml:mn></mml:msubsup><mml:msubsup><mml:mi>r</mml:mi><mml:mi>k</mml:mi><mml:mn mathvariant="normal">1000</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msup><mml:mfenced close=")" open="("><mml:mrow><mml:msubsup><mml:mi>r</mml:mi><mml:mi>k</mml:mi><mml:mn mathvariant="normal">1000</mml:mn></mml:msubsup></mml:mrow></mml:mfenced><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:mfenced><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
          Taking a derivative with respect to <inline-formula><mml:math id="M111" display="inline"><mml:mi mathvariant="italic">α</mml:mi></mml:math></inline-formula> and finding the minimum gives us
            <disp-formula id="Ch1.E7" content-type="numbered"><label>7</label><mml:math id="M112" display="block"><mml:mrow><mml:mi mathvariant="italic">α</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:msubsup><mml:mo>∑</mml:mo><mml:mrow><mml:mi>s</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>S</mml:mi></mml:msubsup><mml:msubsup><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:msubsup><mml:msubsup><mml:mi>r</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mn mathvariant="normal">40</mml:mn></mml:msubsup><mml:msubsup><mml:mi>r</mml:mi><mml:mi>k</mml:mi><mml:mn mathvariant="normal">1000</mml:mn></mml:msubsup></mml:mrow><mml:mrow><mml:msubsup><mml:mo>∑</mml:mo><mml:mrow><mml:mi>s</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>S</mml:mi></mml:msubsup><mml:msubsup><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:msubsup><mml:msup><mml:mfenced open="(" close=")"><mml:mrow><mml:msubsup><mml:mi>r</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mn mathvariant="normal">40</mml:mn></mml:msubsup></mml:mrow></mml:mfenced><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
          In other words, the EOL minimizes the root mean square difference (RMSD) between the 1000-member correlation and all 25 40-member subsample correlations for a chosen setting. For technical reasons, we minimized the cost function using the Brents method as implemented in scipy.optimize <xref ref-type="bibr" rid="bib1.bibx65" id="paren.48"/>. Note that the range of localization is not confined to <inline-formula><mml:math id="M113" display="inline"><mml:mrow><mml:mo>[</mml:mo><mml:mn mathvariant="normal">0</mml:mn><mml:mo>,</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula>, which means that the EOL could increase correlations if required. Values larger than 1 can occur when the true correlation is larger than the sample correlation. For example, this can happen when estimating the EOL after applying other localization approaches. Negative EOL values can be observed when the EOL is computed for a small correlation sample (e.g., a single vertical column), which is not the case in the present study. However, we suggest setting negative EOL values to 0 in case they might occur.</p>
      <p id="d1e2076">Applying the EOL by construction yields a symmetric but not necessarily positive semi-definite localization matrix. In our case, constructed localization matrices were not positive semi-definite. Depending on the data assimilation algorithm, additional steps could be required to apply the EOL results to guarantee the positive semi-definiteness of the localized covariance matrix. For this purpose, Sect. <xref ref-type="sec" rid="Ch1.S3.SS4"/> will assess approaches for achieving positive semi-definiteness of the EOL. The assessment includes a numerical approach to approximate the EOL matrix with the nearest correlation matrix, which is SPSD.</p>
      <p id="d1e2081">Our approach for empirically estimating localization is inspired by <xref ref-type="bibr" rid="bib1.bibx39" id="text.49"/>, who compare two methods: the GGF and empirical localization functions (ELFs). The GGF minimizes the RMSD between the estimated regression coefficients in subsets of the ensemble using a hierarchical ensemble filter <xref ref-type="bibr" rid="bib1.bibx1 bib1.bibx41" id="paren.50"/>. ELFs are derived from an observing system simulation experiment (OSSE) by minimizing the RMSD between the true values of the state variables and the posterior ensemble mean <xref ref-type="bibr" rid="bib1.bibx2" id="paren.51"/>. In contrast to ELFs, the GGF and EOL purely judge localization based on ensemble sampling error without an OSSE. Furthermore, in contrast to the GGF, the EOL assumes the large ensemble correlation as truth for minimizing the sampling error. The EOL presented in our study corresponds to a non-adaptive distant-dependent domain-uniform vertical localization that is common for operational convective-scale regional data assimilation systems.</p>
      <p id="d1e2093">Figure <xref ref-type="fig" rid="Ch1.F1"/>c displays the EOL <inline-formula><mml:math id="M114" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:mi mathvariant="italic">α</mml:mi><mml:mo>(</mml:mo><mml:mi>p</mml:mi><mml:mo>)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> as estimated for the example of <inline-formula><mml:math id="M115" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula> correlations introduced above and reference level 500 hPa. The domain-uniform EOL equals 1 at the reference level 500 hPa as no correction is needed. The EOL appears broader and follows the shape of the mean absolute correlation. For example, this localization behavior was also described by <xref ref-type="bibr" rid="bib1.bibx19" id="text.52"/>. The error reduction is largest for weak and distant correlations.</p>
</sec>
</sec>
<sec id="Ch1.S3">
  <label>3</label><title>Results</title>
      <p id="d1e2138">This section presents mean absolute 1000-member vertical correlations and EOLs for various settings. First, we will evaluate how vertical localization for various single variable pairs should be constructed. Afterward, we will group variable pairs based on similar behavior. Finally, at the end of the results section, we will evaluate the error reduction of all discussed localization approaches, including combinations with the SEC.</p>
<sec id="Ch1.S3.SS1">
  <label>3.1</label><title>Vertical localization for single variable pairs</title>
      <p id="d1e2148">As discussed in Sect. <xref ref-type="sec" rid="Ch1.S2.SS4"/>, the domain-averaged absolute vertical correlation can aid the interpretation of the EOL. For this reason, we will first evaluate the mean absolute vertical correlation and then the EOL. Figure <xref ref-type="fig" rid="Ch1.F2"/> shows the mean absolute vertical correlation for all possible variable combinations and reference level 500 hPa. Self-correlations of the same variable all peak at the reference level. In contrast, cross-correlations are weaker and do not always exhibit a maximum correlation at 500 hPa. The <inline-formula><mml:math id="M116" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula> correlation, for example, peaks around the tropopause, while the <inline-formula><mml:math id="M117" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula> correlation reveals a minimum at that height. The mean vertical correlation length is variable-dependent, shortest for specific humidity and longest for wind. The domain-averaged absolute vertical correlation only exhibits a fairly small variability within the 5 d experimental period. The variability between day and night also appears to be small (not shown). Results could, however, differ for other conditions or seasons, e.g., situations with strong atmospheric stability.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F2" specific-use="star"><?xmltex \currentcnt{2}?><?xmltex \def\figurename{Figure}?><label>Figure 2</label><caption><p id="d1e2177">Domain-averaged absolute 1000-member (true) vertical correlations for reference level 500 hPa and different variable pairs. <bold>(a)</bold> Humidity. <bold>(b)</bold> Temperature. <bold>(c)</bold> <inline-formula><mml:math id="M118" display="inline"><mml:mi>u</mml:mi></mml:math></inline-formula>-wind. Mean and standard deviation over 10 forecasts from 29 May to 2 June 2016.</p></caption>
          <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f02.png"/>

        </fig>

      <p id="d1e2202">Next, we focus on the EOL derived for 40-member subsamples from all forecasts. Figure <xref ref-type="fig" rid="Ch1.F3"/> displays the EOL for all variable combinations and reference level 500 hPa. The EOL depends on the prevailing correlation but has a different shape and vertical extent. As seen for the single forecast example in Sect. <xref ref-type="sec" rid="Ch1.S2.SS5"/>, weaker correlations are more affected by sampling errors and require stronger correction. Consequently, all cross-correlations require a stronger localization. The localization for cross-correlations reveals an amplitude smaller than 1 at the reference level. Given this behavior, tapering functions for cross-correlations should not be 1 at zero distance when applying a distance-dependent localization. Self-correlations are less affected by sampling error and require only a weaker correction, especially close to the reference level.</p>
      <p id="d1e2210">EOLs for humidity correlations all peak at the reference level 500 hPa (Fig. <xref ref-type="fig" rid="Ch1.F3"/>a). However, temperature and wind EOLs behave differently (Fig. <xref ref-type="fig" rid="Ch1.F3"/>b, c) and do not peak at the reference level following the correlation pattern (Fig. <xref ref-type="fig" rid="Ch1.F2"/>b, c). For example, the <inline-formula><mml:math id="M119" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula> EOL peaks around the tropopause, where winds are typically strongest. Wind correlations (e.g., <inline-formula><mml:math id="M120" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula>; Fig. <xref ref-type="fig" rid="Ch1.F3"/>c) require only a small correction. The EOL for <inline-formula><mml:math id="M121" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>V</mml:mi></mml:mrow></mml:math></inline-formula> correlations is almost constant with height and does not show a distinct maximum. All self- and cross-correlations involving humidity peak at the reference level 500 hPa (for example, see <inline-formula><mml:math id="M122" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:math></inline-formula> localization).</p>
      <p id="d1e2262">Overall, the variability of domain-averaged correlations from forecast to forecast is small (Fig. <xref ref-type="fig" rid="Ch1.F2"/>). EOLs exhibit a larger variability than domain-averaged correlations. For most variables, the variability is larger close to the surface, especially for temperature correlations (Fig. <xref ref-type="fig" rid="Ch1.F3"/>b). Results should be treated with caution where changes in the EOL with height are smaller than the variability from forecast to forecast.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F3" specific-use="star"><?xmltex \currentcnt{3}?><?xmltex \def\figurename{Figure}?><label>Figure 3</label><caption><p id="d1e2271">EOL for vertical sample correlations of 40-member ensembles: <bold>(a)</bold> humidity; <bold>(b)</bold> temperature; <bold>(c)</bold> <inline-formula><mml:math id="M123" display="inline"><mml:mi>u</mml:mi></mml:math></inline-formula>-wind. Mean and standard deviation over 10 forecasts from 29 May to 2 June 2016.</p></caption>
          <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f03.png"/>

        </fig>

      <?xmltex \floatpos{t}?><fig id="Ch1.F4" specific-use="star"><?xmltex \currentcnt{4}?><?xmltex \def\figurename{Figure}?><label>Figure 4</label><caption><p id="d1e2298">Same as Fig. <xref ref-type="fig" rid="Ch1.F3"/> but for reference level 300 hPa.</p></caption>
          <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f04.png"/>

        </fig>

      <?xmltex \floatpos{t}?><fig id="Ch1.F5" specific-use="star"><?xmltex \currentcnt{5}?><?xmltex \def\figurename{Figure}?><label>Figure 5</label><caption><p id="d1e2312">Same as Fig. <xref ref-type="fig" rid="Ch1.F3"/> but for reference level 900 hPa.</p></caption>
          <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f05.png"/>

        </fig>

      <p id="d1e2323">Subsequently, we will discuss the EOL for two additional reference levels to highlight changes in height within the troposphere. Figure <xref ref-type="fig" rid="Ch1.F4"/> shows the EOL for a reference level 300 hPa. For reference level 300 hPa, EOLs appear to be broader compared to 500 hPa. This height dependence is in line with larger vertical correlation length scales found for the upper troposphere, in contrast to the lower troposphere, boundary layer, or close to the surface. Similar to other reference levels in the middle and upper troposphere, EOLs for correlations between wind and temperature reveal a maximum (<inline-formula><mml:math id="M124" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula>; Fig. <xref ref-type="fig" rid="Ch1.F4"/>b) and minimum (<inline-formula><mml:math id="M125" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula>; Fig. <xref ref-type="fig" rid="Ch1.F4"/>c) above the tropopause level.</p>
      <p id="d1e2352">All reference levels within the boundary layer show similar behavior of the EOL (see, for example, Fig. <xref ref-type="fig" rid="Ch1.F5"/> using reference level 900 hPa). The EOL shows a narrow optimal vertical localization for reference levels close to the surface. In contrast to higher reference levels, the EOL of cross-correlations also peaks at the reference level (Fig. <xref ref-type="fig" rid="Ch1.F5"/>). The EOL drops to different constant values with increasing distance. For wind and humidity, the EOL reveals an almost constant value above 550 hPa. In contrast, the EOL for temperature steadily declines with increasing distance until the domain top. Temperature self-correlations (<inline-formula><mml:math id="M126" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula>) exhibit a second peak in the upper troposphere.
EOLs do not converge to 0 for large vertical distances. Separation distances where the EOL converges to a small constant value could indicate suitable cut-off distances. A common aspect of the choice of cut-off distance is the signal-to-noise ratio that depends on the ensemble size and correlation strength.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F6" specific-use="star"><?xmltex \currentcnt{6}?><?xmltex \def\figurename{Figure}?><label>Figure 6</label><caption><p id="d1e2371">Root mean square difference before and after the EOL was applied to each vertical correlation. Shading and numbers (%) indicate the change in RMSD analyzed for each variable pair averaged over all reference levels, columns, subsamples, and 10 forecasts from 29 May to 2 June 2016. Self correlations are highlighted via hatching.</p></caption>
          <?xmltex \igopts{width=426.791339pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f06.png"/>

        </fig>

<sec id="Ch1.S3.SS1.SSSx1" specific-use="unnumbered">
  <title>Error reduction for different variables</title>
      <p id="d1e2385">Assessing the EOL for single variable pairs revealed several requirements for vertical localization. Now, we evaluate the error reduction by the EOL, considering each possible correlation pair separately. The 1000-member ensemble correlation serves as truth to compute the RMSD of each 40-member subsample correlation. Figure <xref ref-type="fig" rid="Ch1.F6"/> displays the RMSD before and after applying the EOL. The applied EOL varies for each forecast and height level for the error evaluation. The final result shows the average RMSD of all 40-member subsamples, forecasts, and height levels. The results can be interpreted as a benchmark of the maximum possible correlation error reduction achieved by a domain-uniform height and variable-dependent localization. Results for optimizing the analysis may lead to different optimal localization values under some circumstances, but this analysis is beyond the scope of this paper.</p>
      <p id="d1e2390">The sampling error of the 40-member correlation of most correlations lies within the expected range and close to <inline-formula><mml:math id="M127" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:msqrt><mml:mn mathvariant="normal">40</mml:mn></mml:msqrt><mml:msup><mml:mo>)</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> (Fig. <xref ref-type="fig" rid="Ch1.F6"/>). Self-correlations exhibit a smaller sampling error as, on average, they are stronger and less affected by spurious correlations. The error reduction achieved by the EOL ranges approximately from 10 % to 40 %, depending on the variable pair. The <inline-formula><mml:math id="M128" display="inline"><mml:mrow><mml:mi>Q</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:math></inline-formula> self-correlation benefits most from localization, whereas the <inline-formula><mml:math id="M129" display="inline"><mml:mrow><mml:mi>U</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula> self-correlation benefits least. Correlations involving humidity are weaker and, therefore, benefit most from localization. On the other hand, correcting temperature correlations seems most challenging. Temperature correlations exhibit the largest RMSD, even after applying the EOL. The error is larger than for wind correlations, which is surprising considering a larger correlation strength and length for wind. This result could originate from a larger variability of vertical temperature correlations within the domain, given strong convective processes and the associated latent heat release. Temperature correlations, consequently, could benefit from an adaptive localization that applies different localization scales within the domain depending on, e.g., vertical velocity. The first tests showed promising results for such a situation-dependent approach, but a thorough evaluation will be left for subsequent study.</p>
</sec>
</sec>
<sec id="Ch1.S3.SS2">
  <label>3.2</label><title>Vertical localization for grouped variable pairs</title>
      <p id="d1e2443">Some operational DA systems apply a uniform distance-based vertical localization that does not change with time, height, variable, or observation type. In this case, appropriate localization needs to meet several requirements using a suitable uniform localization approach. Results in Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/> showed that cross-correlations systematically behave differently than self-correlations. For this reason, we will now evaluate the mean absolute correlation and EOL of three groups of variables: self-correlations, cross-correlations, or all correlations combined. Derived EOLs now minimize the sampling error for all gathered correlations of each group.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F7" specific-use="star"><?xmltex \currentcnt{7}?><?xmltex \def\figurename{Figure}?><label>Figure 7</label><caption><p id="d1e2450">Domain mean absolute 1000-member (true) vertical correlations for different variable combinations (self, cross, and all): reference levels <bold>(a)</bold> 300 hPa, <bold>(b)</bold> 500 hPa, and <bold>(c)</bold> 900 hPa. Mean and standard deviation over 10 forecasts from 29 May  to 2 June 2016. Note that the standard deviations are small and hardly distinguishable from the curves themselves.</p></caption>
          <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f07.png"/>

        </fig>

      <p id="d1e2468">Figure <xref ref-type="fig" rid="Ch1.F7"/> displays the mean absolute correlation for the three groups of correlations. The results show the average correlation and its variability over the 10 forecasts. Self-correlations again highlight the height dependence of the vertical correlation length and always exhibit a peak that is 1. Cross-correlations are weaker and only exhibit a narrow peak at the reference level. For all correlations combined, the peak amplitude is closer to the peak of cross-correlations as there are more cross-correlations than self-correlations. Combining all correlations or only cross-correlations results in a peak amplitude smaller than 1 at the reference level.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F8" specific-use="star"><?xmltex \currentcnt{8}?><?xmltex \def\figurename{Figure}?><label>Figure 8</label><caption><p id="d1e2476">EOL for vertical sample correlations of 40-member ensembles and different variable combinations (self, cross, and all): reference levels <bold>(a)</bold> 300 hPa, <bold>(b)</bold> 500 hPa, and <bold>(c)</bold> 900 hPa. Mean and standard deviation over 10 forecasts from 29 May to 2 June 2016.</p></caption>
          <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f08.png"/>

        </fig>

      <p id="d1e2494">In contrast, the peak amplitude of the EOL for all correlations is closer to the peak of self-correlation (Fig. <xref ref-type="fig" rid="Ch1.F8"/>). The shape of EOLs substantially differs from the single variable pair cases. The EOL is weaker due to wind correlations that account for half of all correlations. The change in the shape of the EOL indicates that different tapering functions could be needed for different variables. Minimizing the error for grouped correlations, the strength of the EOL is always weaker than 0.4. Finally, domain-averaged absolute correlations reveal a small variability from forecast to forecast (Fig. <xref ref-type="fig" rid="Ch1.F7"/>). The same applies to EOLs. Only the EOL of self-correlations exhibits a slightly larger variability, especially far from the reference level (Fig. <xref ref-type="fig" rid="Ch1.F8"/>).</p>
</sec>
<sec id="Ch1.S3.SS3">
  <label>3.3</label><title>Evaluation of error reduction</title>
<sec id="Ch1.S3.SS3.SSS1">
  <label>3.3.1</label><title>Setting</title>
      <p id="d1e2518">As discussed in Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/>, the maximum reduction of sampling errors achieved by an EOL ranges from 11 % to 44 % depending on the variable pair. Now, we will compare the performance of the EOL with different localization setups that use two common localization approaches, a distance-dependent localization using a Gaspari–Cohn tapering function (GC; <xref ref-type="bibr" rid="bib1.bibx30 bib1.bibx20" id="altparen.53"/>) and a statistical sampling error correction <xref ref-type="bibr" rid="bib1.bibx5" id="paren.54"/>. Furthermore, we investigate the benefit of combining non-adaptive localization approaches with the adaptive SEC. Compared to Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/>, the improvement will be evaluated using 1000-member correlations from independent background forecasts. Again we will analyze the improvement relative to uncorrected 40-member ensemble subsample correlations (REF40, Fig. <xref ref-type="fig" rid="Ch1.F9"/>). The first eight forecasts (29 May to 1 June 2016) serve as training data for estimating EOLs. Similarly, localization scales for distance-dependent localization are tuned using the same training period. We then verified the performance using the last two independent forecasts on 2 June 2016.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F9" specific-use="star"><?xmltex \currentcnt{9}?><?xmltex \def\figurename{Figure}?><label>Figure 9</label><caption><p id="d1e2535">Root mean square difference before and after localization of 40-member vertical subsample correlations. EOL and Gaspari–Cohn scales are obtained and tuned using the first eight forecasts. Errors are evaluated using two independent forecasts on 2 June 2016: 03:00 UTC (opaque) and 15:00 UTC (hatched). Numbers (%) indicate the average change in RMSD analyzed for different settings (<inline-formula><mml:math id="M130" display="inline"><mml:mi>x</mml:mi></mml:math></inline-formula>-axis labels).</p></caption>
            <?xmltex \igopts{width=426.791339pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f09.png"/>

          </fig>

</sec>
<sec id="Ch1.S3.SS3.SSS2">
  <label>3.3.2</label><title>EOL</title>
      <p id="d1e2559">Figure <xref ref-type="fig" rid="Ch1.F9"/> displays the error reduction achieved by all considered vertical localization setups. <italic>REF40</italic> shows the RMSD found when modeling error correlations using small 40-member ensembles without localization. First, we will evaluate the performance of different EOL settings. Applying a different EOL for each variable pair and height (as presented in Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/>) gives the largest error reduction of all the setups (<italic>SINGLE</italic>, 26.7 %). Only small differences are visible between day and night (not shown). Using different EOLs only for self- and cross-correlations leads to a slightly reduced performance but still gives about 23 % error reduction (<italic>SELF</italic>). Applying an EOL that was estimated for all correlations at once reduces the error by 17 % (<italic>ALL</italic>). Given these results, treating variable pairs and self- or cross-correlations differently enables substantial improvements. Finally, we tested the error reduction for applying the EOL estimated for self-correlations to both self- and cross-correlations of each variable (e.g., EOL derived from <inline-formula><mml:math id="M131" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula> applied to <inline-formula><mml:math id="M132" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>T</mml:mi></mml:mrow></mml:math></inline-formula>, <inline-formula><mml:math id="M133" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:math></inline-formula>, <inline-formula><mml:math id="M134" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>U</mml:mi></mml:mrow></mml:math></inline-formula>, and <inline-formula><mml:math id="M135" display="inline"><mml:mrow><mml:mi>T</mml:mi><mml:mi>V</mml:mi></mml:mrow></mml:math></inline-formula>). For this setting, the error reduction was similar to <italic>ALL</italic> or <italic>SEC</italic> (not shown), which underlines the benefit of treating self- and cross-correlations differently.</p>
</sec>
<sec id="Ch1.S3.SS3.SSS3">
  <label>3.3.3</label><title>Distance-dependent localization</title>
      <p id="d1e2644">Now, we will compare the performance of EOLs to three different domain-uniform distance-dependent localization approaches using Gaspari–Cohn functions. Section <xref ref-type="sec" rid="Ch1.S2.SS3.SSS1"/> provides more details on distance-dependent Gaspari–Cohn localization and details on all three considered localization setups (<italic>GC</italic>, <italic>GCLEV</italic>, and <italic>DWD</italic>). We will first evaluate two optimized setups with tuned localization scales (<italic>GC</italic> and <italic>GCLEV</italic>) and then compare them to a non-tuned setup (<italic>DWD</italic>).</p>
      <p id="d1e2668"><italic>GC</italic> uses a uniform localization scale for all levels and variable pairs, and <italic>GCLEV</italic> uses a height-dependent optimal localization scale that changes with the reference level. <italic>GC</italic> reduces the sampling error by about 10 %. Using height-dependent localization scales (<italic>GCLEV</italic>) slightly improves the performance further by about 1 %. However, the small gain of the height-dependent localization is partly associated with a sub-optimal shape of the tapering function, given the error reduction achieved by the uniform EOL (<italic>ALL</italic>). This comparison highlights that finding suitable tapering scales and functions bears great potential for improving vertical localization.</p>
      <p id="d1e2685">In contrast, a vertical localization constructed similarly to the regional DA system of the DWD increases the difference of the 40-member ensemble correlation with respect to the 1000-member ensemble. The increased difference originates from the damping of meaningful error correlations. The DWD system employs a LETKF that uses observation-space localization, tuned to function in all seasons and weather situations that may differ from our investigation period. Furthermore, it needs to be considered that localization in the LETKF also affects the degrees of freedom of the analysis <xref ref-type="bibr" rid="bib1.bibx29" id="paren.55"/>. The DWD setup illustrates that an appropriate localization depends on various aspects. Consequently, our findings will likely have different implications for different DA algorithms. However, the LETKF, for example, could benefit from applying different localization scales for different observed variables. Based on the results in Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/>, humidity, temperature, or wind require different vertical localization scales.</p>
</sec>
<sec id="Ch1.S3.SS3.SSS4">
  <label>3.3.4</label><title>SEC</title>
      <p id="d1e2701">Now, we will evaluate the benefit of using a look-up table-based SEC that adjusts correlations based on predefined statistical assumptions. The SEC is an adaptive localization approach that corrects sampling errors as a function of the correlation value. Therefore, the SEC applies an individual correction for each correlation within the domain. An adaptive localization (<italic>SEC</italic>) achieves 17.5 % error reduction and outperforms a optimal domain-uniform GC localization. The SEC exhibits a similar error reduction to that seen for <italic>ALL</italic> but cannot outperform the <italic>SELF</italic> or <italic>SINGLE</italic> setup. An optimal domain-uniform localization can compete with an adaptive statistical sampling error correction for the evaluated period.</p>
</sec>
<sec id="Ch1.S3.SS3.SSS5">
  <label>3.3.5</label><title>Combined approaches</title>
      <p id="d1e2725">Finally, we investigate the benefit of combining the statistical SEC with an EOL or a distance-dependent localization. For this analysis, EOLs have been estimated after applying the SEC to highlight the maximum error reduction achieved by combining SEC with an optimal localization. The localization scale of the distance-dependent localization is kept the same as for the <italic>GC</italic> setup to emphasize required changes for the localization scale. <italic>SEC</italic> <inline-formula><mml:math id="M136" display="inline"><mml:mo>+</mml:mo></mml:math></inline-formula> <italic>GC</italic> reveals a similar performance to the SEC alone but outperforms the <italic>GC</italic> setup. Combining <italic>SEC</italic> with <italic>GC</italic> requires a re-tuning of localization scales to larger values (not shown). Combining the SEC with a uniform EOL (<italic>SEC</italic> <inline-formula><mml:math id="M137" display="inline"><mml:mo>+</mml:mo></mml:math></inline-formula> <italic>ALL</italic>) reduces the sampling error by about 20 %. However, combining the SEC with the <italic>SELF</italic> or <italic>SINGLE</italic> EOL leads to less error reduction. The poor performance could originate from sub-optimal assumptions made in the derivation of the SEC <xref ref-type="bibr" rid="bib1.bibx6 bib1.bibx56" id="paren.56"/>. For example, the EOL exhibited values larger than 1 when estimated after applying the SEC, compensating for an overcorrection of sampling errors, especially close to the reference level (not shown). In this study, we apply the most general SEC look-up table as provided in DART <xref ref-type="bibr" rid="bib1.bibx3" id="paren.57"/>, which assumes that each correlation value is equally likely. Studying more informed prior assumptions in the SEC may lead to better results but is beyond the scope of the present study.</p>
</sec>
</sec>
<sec id="Ch1.S3.SS4">
  <label>3.4</label><title>Covariance localization and positive semi-definiteness</title>
      <p id="d1e2790">The EOL approach empirically yields an optimal localization by minimizing differences between sample correlations and a defined true correlation. By design, the results discussed above exclude algorithm-specific requirements to better understand how vertical localization should behave in different situations. However, further steps might be required to apply EOL results depending on the data assimilation algorithm. As mentioned in Sect. <xref ref-type="sec" rid="Ch1.S2.SS5"/>, using EOL estimates does not necessarily yield a positive semi-definite localization matrix. Consequently, EOL localization could cause non-proper mathematical covariance matrices and undermine the data assimilation process. For example, positive semi-definite covariance matrices are elemental when solving a quadratic cost function as they ensure a global minimum. For this reason, we will now discuss different approaches that allow one to achieve positive semi-definiteness of the EOL.</p>
      <p id="d1e2795"><xref ref-type="bibr" rid="bib1.bibx52" id="text.58"/> applied a Gaussian fitting based on the Gaspari–Cohn function <xref ref-type="bibr" rid="bib1.bibx20" id="paren.59"/> to ensure SPSD. However, fitting a correlation function restricts the localization to a specific function shape. This restriction can diminish the error reduction. In our case, the error reduction achieved by an optimally tuned Gaspari–Cohn function is substantially smaller than by the EOL, as discussed in Sect. <xref ref-type="sec" rid="Ch1.S3.SS3"/> and Fig. <xref ref-type="fig" rid="Ch1.F9"/>. This finding motivates exploration of other localization functions <xref ref-type="bibr" rid="bib1.bibx16 bib1.bibx12" id="paren.60"/>, including multivariate localization functions <xref ref-type="bibr" rid="bib1.bibx63" id="paren.61"/>, to improve fitting approaches.</p>
      <p id="d1e2813">Besides function fitting, re-conditioning of matrices can help achieve positive definiteness <xref ref-type="bibr" rid="bib1.bibx64" id="paren.62"/>. Given its aim, a localization matrix should have similar properties to a correlation matrix, which exhibits ones on the diagonal. <xref ref-type="bibr" rid="bib1.bibx26" id="text.63"><named-content content-type="post">theorem 2.5</named-content></xref> states that for a symmetric matrix with <inline-formula><mml:math id="M138" display="inline"><mml:mi>t</mml:mi></mml:math></inline-formula> non-positive eigenvalues and diagonal elements <inline-formula><mml:math id="M139" display="inline"><mml:mrow><mml:mo>≥</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:math></inline-formula>, the nearest correlation matrix has at least <inline-formula><mml:math id="M140" display="inline"><mml:mi>t</mml:mi></mml:math></inline-formula> zero eigenvalues. Following this theorem, we attempted to set all negative eigenvalues to zero using an EOL-based localization matrix and eigendecomposition. Forcing negative eigenvalues to zero ensured positive semi-definiteness but led to values larger than one on the diagonal of the matrix. Further matrix transformations were required to achieve a proper correlation matrix.</p>
      <p id="d1e2848">Finally, our most successful attempt at achieving positive definiteness aiming for the least changes in the EOL was by searching for the nearest correlation matrix using specifically designed mathematical algorithms. For example, <xref ref-type="bibr" rid="bib1.bibx26" id="text.64"/> highlighted that the convexity properties of the problem allow one to find a unique nearest correlation matrix (NCM) for a given symmetric matrix, in our case the EOL-based localization matrix. The distance between matrices can be measured using weighted Frobenius norms. We apply the NCM algorithm <xref ref-type="bibr" rid="bib1.bibx26" id="paren.65"/> for the <italic>SINGLE</italic> EOL case as a proof of concept. Employing four state variables on 20 vertical levels yields an <inline-formula><mml:math id="M141" display="inline"><mml:mrow><mml:mn mathvariant="normal">80</mml:mn><mml:mo>×</mml:mo><mml:mn mathvariant="normal">80</mml:mn></mml:mrow></mml:math></inline-formula> EOL-based vertical localization matrix for one vertical column (Fig. <xref ref-type="fig" rid="Ch1.F10"/>a). The eigenvalues of this EOL matrix range from about <inline-formula><mml:math id="M142" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.5</mml:mn></mml:mrow></mml:math></inline-formula> to 50, with approximately half the eigenvalues being negative. The EOL localization matrix features ones on the diagonal and reconfirms that cross-correlations require stronger tapering. The nearest correlation matrix computed by the NCM algorithm exhibits small changes in off-diagonal elements compared to the EOL-based matrix (Fig. <xref ref-type="fig" rid="Ch1.F10"/>b, c). Zooming in, the SPSD matrix appears to be smoother. Comparing the error reduction (not shown), the nearest SPSD matrix performs only marginally worse than the EOL matrix. The relative difference in sampling error reduction for a test case was lower than 1 %. However, the EOL achieves the larger error reduction as it is designed to minimize the sampling error without any constraints.</p>
      <p id="d1e2888">This example suggests that ensuring SPSD can be achieved with minor changes to the EOL estimate. However, providing a general answer on how the EOL needs to be adapted is difficult as changes will depend on the construction of the localization matrix and its unique nearest correlation matrix. NCM algorithms can iteratively determine the nearest correlation matrix for a symmetric matrix and could be a useful tool for data assimilation. Choosing the best approach to guaranteeing SPSD is likely case-dependent given changing properties of the problem and potentially very large matrices in NWP.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F10" specific-use="star"><?xmltex \currentcnt{10}?><?xmltex \def\figurename{Figure}?><label>Figure 10</label><caption><p id="d1e2893">Examples of EOL-based localization matrices for a single vertical column: <bold>(a)</bold> matrix C constructed based on the <italic>SINGLE</italic> case, <bold>(b)</bold> the resulting nearest correlation matrix C following the NCM algorithm, and <bold>(c)</bold> changes due to enforcing positive definiteness.</p></caption>
          <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://npg.copernicus.org/articles/30/13/2023/npg-30-13-2023-f10.png"/>

        </fig>

</sec>
</sec>
<sec id="Ch1.S4" sec-type="conclusions">
  <label>4</label><title>Conclusions and discussion</title>
      <p id="d1e2923">Current ensemble data assimilation systems suffer from severe undersampling requiring vertical localization of error covariances. Our study analyzes vertical correlations from an existing convection-permitting 1000-member ensemble simulation <xref ref-type="bibr" rid="bib1.bibx55 bib1.bibx56" id="paren.66"/>. The 1000-member ensemble correlation is assumed as truth for studying reliable vertical correlations and optimal vertical localization in 40-member subsamples. The unique convective-scale simulation covers 10 forecasts in a 5 d mid-latitude summer period. Our analysis includes four prognostic variables (humidity, temperature, and two horizontal wind components) on 20 pressure levels. We apply the 1000-member ensemble and various 40-member subsamples to derive an empirical optimal localization (EOL) for different settings. Those settings include localization for single variable pairs and variables grouped by common behavior. Presented EOLs minimize the sampling error in sample correlations assuming the 1000-member correlation as truth and provide insights into how to construct an optimal vertical localization independent of algorithm-specific constraints.</p>
      <p id="d1e2929">Furthermore, we use the 1000-member ensemble to evaluate the error reduction achieved by different localization approaches. These approaches include EOLs, distance-dependent localization approaches using a Gaspari–Cohn tapering function <xref ref-type="bibr" rid="bib1.bibx30 bib1.bibx20" id="paren.67"/>, and an adaptive statistical sampling error correction <xref ref-type="bibr" rid="bib1.bibx5" id="paren.68"/>. Overall, our results lead to the following conclusions for vertical localization.
<?xmltex \hack{\newpage}?>
<list list-type="bullet"><list-item>
      <p id="d1e2942"><italic>Localization scales.</italic> All investigated variables reveal different average correlation scales, which result in different EOL scales. Within the troposphere, EOL scales increase with height. Humidity requires the strongest localization with short scales. EOL scales for temperature appear to be larger than for humidity and exhibit the largest variability from forecast to forecast. Given a high variability, temperature correlations could benefit most from using adaptive localization. Our results indicate that winds can be vertically correlated throughout the troposphere, resulting in the largest localization scales. Given this outcome, it could be beneficial not to cut off wind correlations within the troposphere.</p></list-item><list-item>
      <p id="d1e2948"><italic>Localization shape.</italic> The EOL provides insights into the required shape of localization functions. Correlations of different variable pairs require differently shaped localization functions. Localization functions should not necessarily be symmetric in <inline-formula><mml:math id="M143" display="inline"><mml:mrow><mml:mi>ln⁡</mml:mi><mml:mo>(</mml:mo><mml:mi>p</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, as seen for wind. Furthermore, the optimal center of a distance-dependent localization can deviate from the reference level. For example, correlations of temperature and wind peak below the tropopause if the reference level is above the boundary layer. The maximum vertical correlation could indicate a suitable positioning of distance-dependent tapering functions. Finally, EOLs do not reveal a clear localization cut-off distance for tropospheric correlations. However, other considerations, e.g., continuity, computational efficiency, or matrix rank, may also need to be considered when deciding on a cut-off.</p></list-item><list-item>
      <p id="d1e2968"><italic>Self- and cross-correlations.</italic> Self- (e.g., temperature–temperature) and cross-correlations (e.g., temperature–humidity) should be localized differently. This fact could allow the development of correlation-dependent localization approaches. For example, self-correlations require no localization at zero distance, while the amplitude of cross-correlations should be tapered by at least 25 %. Differently treating self- and cross-correlations resulted in performance close to a variable-dependent localization.</p></list-item><list-item>
      <p id="d1e2974"><italic>Domain-uniform localization.</italic> A tuned uniform distance-dependent localization using Gaspari–Cohn functions reduces the sampling error by about 10 %. Using tapering functions with an optimal shape could improve the localization substantially. The maximum error reduction was found for domain-uniform, variable, and height-dependent EOLs with about 27 % improvement. Distinguishing between self- and cross-correlations leads to a similar but slightly smaller error reduction.</p></list-item><list-item>
      <p id="d1e2980"><italic>Adaptive localization.</italic> A statistical sampling error correction (SEC) achieves similar error reduction to a variable- and domain-uniform localization. Combining the SEC with a Gaspari–Cohn localization improves the error reduction. However, combining distance-dependent and statistical approaches requires re-tuning of localization scales. Combining SEC and EOLs led to an overcorrection of correlations, which slightly degraded the error reduction. This change could be related to sub-optimal prior assumptions when deriving SEC, as discussed by <xref ref-type="bibr" rid="bib1.bibx6" id="text.69"/> and <xref ref-type="bibr" rid="bib1.bibx56" id="text.70"/>.</p></list-item></list></p>
      <p id="d1e2991">Our results allow a better understanding of the requirements for vertical localization. When employing these conclusions, it is important to consider the specific demands of different ensemble filter algorithms. In ensemble transform Kalman filters, localization increases the degrees of freedom of the analysis and thereby enables the assimilation of more observations <xref ref-type="bibr" rid="bib1.bibx29" id="paren.71"/>. Furthermore, our evaluation excluded considerations about the rank of the error covariance matrix and computational efficiency. Hence, our findings might need to be adapted to improve the analysis performance depending on the data assimilation system. Localization in operational NWP has many system-dependent requirements and is tuned to avoid bad signal-to-noise ratios during assimilation. For example, while we find no strong support for a vertical cut-off within the troposphere for some variables, this could be beneficial for the reasons discussed above.</p>
      <p id="d1e2997">How to apply EOL estimates will vary with the data assimilation algorithm as the application of localization is highly algorithm-specific. In case of covariance localization, constructing a generally non-SPSD localization matrix based on the EOL does not guarantee a symmetric positive semi-definite localized covariance matrix. However, different approaches allow one to achieve positive semi-definiteness of localization matrices. Applying an NCM algorithm <xref ref-type="bibr" rid="bib1.bibx26" id="paren.72"/> to achieve positive semi-definiteness resulted in only very minor changes in the EOL that hardly affect the error reduction.</p>
      <p id="d1e3004">For a serial filter (e.g., the ensemble adjustment Kalman filter (EAKF) by <xref ref-type="bibr" rid="bib1.bibx4" id="altparen.73"/>), an EOL-based localization can be applied directly, and it is planned to test this in follow-on studies. The EAKF does not involve a Schur product localization of a covariance matrix as each single observation is assimilated at a time, serially. Instead, the EAKF localizes the increment or gain. The gain between the observation and each state variable is multiplied by a scalar between 0 and 1. This localization factor can be provided by the EOL.</p>
      <p id="d1e3010">Our study solely judges localization based on ensemble sampling error, assuming the 1000-member ensemble correlation as truth. It is difficult to predict the number of ensembles needed to apply our method, as it will vary for differing scenarios. However, we do not expect our results to change drastically if we had a larger ensemble. Besides, it would be interesting to compare the EOL with the ELF or GGF approach. For example, comparing ELF and EOL could allow us to investigate other error sources in the assimilation that can influence localization <xref ref-type="bibr" rid="bib1.bibx2" id="paren.74"/>. However, a proper comparison would require an OSSE with a sufficiently large ensemble.</p>
      <p id="d1e3016">We have found robust results for a mid-latitude convective summer period. The ever-increasing computational capabilities will enable extended data sets and a higher vertical resolution that is comparatively coarse in the current setup. Furthermore, our approach can be easily applied to other large ensemble simulations to study additional aspects, including horizontal localization. Extending this analysis is desirable given that localization can depend on the underlying weather condition <xref ref-type="bibr" rid="bib1.bibx40 bib1.bibx17" id="paren.75"/>. For example, using a global simulation with a higher model top would allow us to study different geographical regions, seasons, and stratospheric correlations that are particularly important for satellite data assimilation <xref ref-type="bibr" rid="bib1.bibx42 bib1.bibx61" id="paren.76"/>.</p>
</sec>

      
      </body>
    <back><notes notes-type="codedataavailability"><title>Code and data availability</title>

      <p id="d1e3030">Code and processed data such as derived empirical optimal localizations are shared on Zenodo: <ext-link xlink:href="https://doi.org/10.5281/zenodo.7254119" ext-link-type="DOI">10.5281/zenodo.7254119</ext-link> <xref ref-type="bibr" rid="bib1.bibx54" id="paren.77"/>. The 1000-member ensemble data set and derived covariances and correlations (approximately 60 TB of data) are too large for an upload but are available upon request to the corresponding author.</p>
  </notes><notes notes-type="authorcontribution"><title>Author contributions</title>

      <p id="d1e3042">TN and MW were responsible for the conceptualization and formal analysis. All the authors from the University of Vienna contributed to the development of the methodology. TN developed the software code and was responsible for data curation and visualization of results. TM supported the research with important computational resources. TN and MW wrote and prepared the original paper draft. DH, PJG, and TM helped during the review and editing.</p>
  </notes><notes notes-type="competinginterests"><title>Competing interests</title>

      <p id="d1e3048">At least one of the (co-)authors is a member of the editorial board of <italic>Nonlinear Processes in Geophysics</italic>. The peer-review process was guided by an independent editor, and the authors also have no other competing interests to declare.</p>
  </notes><notes notes-type="disclaimer"><title>Disclaimer</title>

      <p id="d1e3057">Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.</p>
  </notes><ack><title>Acknowledgements</title><p id="d1e3063">Many thanks to Juan Ruiz, Jago Silberbauer, Jeffrey Anderson, and other colleagues at the University of Vienna, RIKEN RCC-S in Kobe, and LMU in Munich, who contributed to this research. Furthermore, we want to thank the two reviewers and the editor for their helpful comments that allowed us to improve the manuscript. The open-source project and Python package “xarray” (<xref ref-type="bibr" rid="bib1.bibx32" id="altparen.78"/>) has been used to process ensemble data. Open-access funding was provided by the University of Vienna.</p></ack><notes notes-type="financialsupport"><title>Financial support</title>

      <p id="d1e3071">Open-access funding was provided by the University of Vienna.</p>
  </notes><notes notes-type="reviewstatement"><title>Review statement</title>

      <p id="d1e3077">This paper was edited by Olivier Talagrand and reviewed by Pavel Sakov and one anonymous referee.</p>
  </notes><ref-list>
    <title>References</title>

      <ref id="bib1.bibx1"><?xmltex \def\ref@label{{Anderson(2007)}}?><label>Anderson(2007)</label><?label Anderson2007?><mixed-citation>Anderson, J.: Exploring the need for localization in ensemble data
assimilation using a hierarchical ensemble filter, Physica D, 230, 99–111, <ext-link xlink:href="https://doi.org/10.1016/j.physd.2006.02.011" ext-link-type="DOI">10.1016/j.physd.2006.02.011</ext-link>, 2007.</mixed-citation></ref>
      <ref id="bib1.bibx2"><?xmltex \def\ref@label{{Anderson and Lei(2013)}}?><label>Anderson and Lei(2013)</label><?label Anderson2013?><mixed-citation>Anderson, J. and Lei, L.: Empirical Localization of Observation Impact in
Ensemble Kalman Filters, Mon. Weather Rev., 141, 4140–4153,
<ext-link xlink:href="https://doi.org/10.1175/MWR-D-12-00330.1" ext-link-type="DOI">10.1175/MWR-D-12-00330.1</ext-link>, 2013.</mixed-citation></ref>
      <ref id="bib1.bibx3"><?xmltex \def\ref@label{{Anderson et~al.(2009)Anderson, Hoar, Raeder, Liu, Collins, Torn, and
Avellano}}?><label>Anderson et al.(2009)Anderson, Hoar, Raeder, Liu, Collins, Torn, and
Avellano</label><?label Anderson2009?><mixed-citation>Anderson, J., Hoar, T., Raeder, K., Liu, H., Collins, N., Torn, R., and
Avellano, A.: The Data Assimilation Research Testbed: A Community Facility,
B. Am. Meteor. Soc., 90, 1283–1296,
<ext-link xlink:href="https://doi.org/10.1175/2009BAMS2618.1" ext-link-type="DOI">10.1175/2009BAMS2618.1</ext-link>, 2009.</mixed-citation></ref>
      <ref id="bib1.bibx4"><?xmltex \def\ref@label{{Anderson(2001)}}?><label>Anderson(2001)</label><?label Anderson2001?><mixed-citation>Anderson, J. L.: An Ensemble Adjustment Kalman Filter for Data Assimilation,
Mon. Weather Rev., 129, 2884–2903,
<ext-link xlink:href="https://doi.org/10.1175/1520-0493(2001)129&lt;2884:AEAKFF&gt;2.0.CO;2" ext-link-type="DOI">10.1175/1520-0493(2001)129&lt;2884:AEAKFF&gt;2.0.CO;2</ext-link>, 2001.</mixed-citation></ref>
      <ref id="bib1.bibx5"><?xmltex \def\ref@label{{Anderson(2012)}}?><label>Anderson(2012)</label><?label Anderson2012?><mixed-citation>Anderson, J. L.: Localization and Sampling Error Correction in Ensemble Kalman
Filter Data Assimilation, Mon. Weather Rev., 140, 2359–2371,
<ext-link xlink:href="https://doi.org/10.1175/MWR-D-11-00013.1" ext-link-type="DOI">10.1175/MWR-D-11-00013.1</ext-link>, 2012.</mixed-citation></ref>
      <ref id="bib1.bibx6"><?xmltex \def\ref@label{{Anderson(2016)}}?><label>Anderson(2016)</label><?label Anderson2016?><mixed-citation>Anderson, J. L.: Reducing Correlation Sampling Error in Ensemble Kalman Filter
Data Assimilation, Mon. Weather Rev., 144, 913–925,
<ext-link xlink:href="https://doi.org/10.1175/MWR-D-15-0052.1" ext-link-type="DOI">10.1175/MWR-D-15-0052.1</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx7"><?xmltex \def\ref@label{{Bannister(2008)}}?><label>Bannister(2008)</label><?label Bannister2008?><mixed-citation>Bannister, R. N.: A review of forecast error covariance statistics in
atmospheric variational data assimilation. II: Modelling the forecast error
covariance statistics,
Q. J. Roy. Meteor. Soc.,
134, 1971–1996, <ext-link xlink:href="https://doi.org/10.1002/qj.340" ext-link-type="DOI">10.1002/qj.340</ext-link>, 2008.</mixed-citation></ref>
      <ref id="bib1.bibx8"><?xmltex \def\ref@label{{Bannister(2017)}}?><label>Bannister(2017)</label><?label Bannister2017?><mixed-citation>Bannister, R. N.: A review of operational methods of variational and
ensemble-variational data assimilation, Q. J. R. Meteorol. Soc., 143,
607–633, <ext-link xlink:href="https://doi.org/10.1002/qj.2982" ext-link-type="DOI">10.1002/qj.2982</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx9"><?xmltex \def\ref@label{{Bannister et~al.(2017)Bannister, Migliorini, Rudd, and
Baker}}?><label>Bannister et al.(2017)Bannister, Migliorini, Rudd, and
Baker</label><?label Bannister2017b?><mixed-citation>Bannister, R. N., Migliorini, S., Rudd, A. C., and Baker, L. H.: Methods of investigating forecast error sensitivity to ensemble size in a limited-area convection-permitting ensemble, Geosci. Model Dev. Discuss. [preprint], <ext-link xlink:href="https://doi.org/10.5194/gmd-2017-260" ext-link-type="DOI">10.5194/gmd-2017-260</ext-link>, in review, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx10"><?xmltex \def\ref@label{{Bishop and Hodyss(2009{\natexlab{a}})}}?><label>Bishop and Hodyss(2009a)</label><?label Bishop2009a?><mixed-citation>Bishop, C. H. and Hodyss, D.: Ensemble covariances adaptively localized with
ECO-RAP. Part 1: tests on simple error models, Tellus A, 61, 84–96,
<ext-link xlink:href="https://doi.org/10.1111/j.1600-0870.2008.00371.x" ext-link-type="DOI">10.1111/j.1600-0870.2008.00371.x</ext-link>, 2009a.</mixed-citation></ref>
      <ref id="bib1.bibx11"><?xmltex \def\ref@label{{Bishop and Hodyss(2009{\natexlab{b}})}}?><label>Bishop and Hodyss(2009b)</label><?label Bishop2009b?><mixed-citation>Bishop, C. H. and Hodyss, D.: Ensemble covariances adaptively localized with
ECO-RAP. Part 2: a strategy for the atmosphere, Tellus A, 61, 97–111,
<ext-link xlink:href="https://doi.org/10.1111/j.1600-0870.2008.00372.x" ext-link-type="DOI">10.1111/j.1600-0870.2008.00372.x</ext-link>, 2009b.</mixed-citation></ref>
      <ref id="bib1.bibx12"><?xmltex \def\ref@label{{Bolin and Wallin(2016)}}?><label>Bolin and Wallin(2016)</label><?label Bolin2016?><mixed-citation>Bolin, D. and Wallin, J.: Spatially adaptive covariance tapering, Spatial
Statistics, 18, 163–178, <ext-link xlink:href="https://doi.org/10.1016/j.spasta.2016.03.003" ext-link-type="DOI">10.1016/j.spasta.2016.03.003</ext-link>,
2016.</mixed-citation></ref>
      <ref id="bib1.bibx13"><?xmltex \def\ref@label{{Bonavita et~al.(2016)Bonavita, Hólm, Isaksen, and
Fisher}}?><label>Bonavita et al.(2016)Bonavita, Hólm, Isaksen, and
Fisher</label><?label Bonavita2016?><mixed-citation>Bonavita, M., Hólm, E., Isaksen, L., and Fisher, M.: The evolution of the
ECMWF hybrid data assimilation system, Q. J. Roy. Meteor. Soc., 142, 287–303, <ext-link xlink:href="https://doi.org/10.1002/qj.2652" ext-link-type="DOI">10.1002/qj.2652</ext-link>,
2016.</mixed-citation></ref>
      <ref id="bib1.bibx14"><?xmltex \def\ref@label{{Bouttier et~al.(2016)Bouttier, Raynaud, Nuissier, and
Menetrier}}?><label>Bouttier et al.(2016)Bouttier, Raynaud, Nuissier, and
Menetrier</label><?label Bouttier2016?><mixed-citation>Bouttier, F., Raynaud, L., Nuissier, O., and Menetrier, B.: Sensitivity of the
AROME ensemble to initial and surface perturbations during HyMeX, Q. J. Roy. Meteor. Soc., 142, 390–403,
<ext-link xlink:href="https://doi.org/10.1002/qj.2622" ext-link-type="DOI">10.1002/qj.2622</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx15"><?xmltex \def\ref@label{{Craig et~al.(2022)Craig, Puh, Keil, Tempest, Necker, Ruiz, Weissmann,
and Miyoshi}}?><label>Craig et al.(2022)Craig, Puh, Keil, Tempest, Necker, Ruiz, Weissmann,
and Miyoshi</label><?label Craig2022?><mixed-citation>Craig, G. C., Puh, M., Keil, C., Tempest, K., Necker, T., Ruiz, J., Weissmann,
M., and Miyoshi, T.: Distributions and convergence of forecast variables in
a 1000 member convection-permitting ensemble, Q. J. Roy.
Meteor. Soc., 148, 2325–2343, <ext-link xlink:href="https://doi.org/10.1002/qj.4305" ext-link-type="DOI">10.1002/qj.4305</ext-link>, 2022.</mixed-citation></ref>
      <ref id="bib1.bibx16"><?xmltex \def\ref@label{{Daley et~al.(2015)Daley, Porcu, and Bevilacqua}}?><label>Daley et al.(2015)Daley, Porcu, and Bevilacqua</label><?label Daley2015?><mixed-citation>Daley, D., Porcu, E., and Bevilacqua, M.: Classes of compactly supported
covariance functions for multivariate random fields,
Stoch. Env. Res. Risk A., 29, 1249–1263,
<ext-link xlink:href="https://doi.org/10.1007/s00477-014-0996-y" ext-link-type="DOI">10.1007/s00477-014-0996-y</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bibx17"><?xmltex \def\ref@label{{Destouches et~al.(2021)Destouches, Montmerle, Michel, and
Ménétrier}}?><label>Destouches et al.(2021)Destouches, Montmerle, Michel, and
Ménétrier</label><?label Destouches2021?><mixed-citation>Destouches, M., Montmerle, T., Michel, Y., and Ménétrier, B.: Estimating
optimal localization for sampled background-error covariances of hydrometeor
variables, Q. J. Roy. Meteor. Soc., 147,
74–93, <ext-link xlink:href="https://doi.org/10.1002/qj.3906" ext-link-type="DOI">10.1002/qj.3906</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx18"><?xmltex \def\ref@label{{Evensen(1994)}}?><label>Evensen(1994)</label><?label Evensen1994?><mixed-citation>Evensen, G.: Sequential data assimilation with a nonlinear quasi-geostrophic
model using Monte Carlo methods to forecast error statistics, J.
Geophys. Res.-Oceans, 99, 10143–10162, <ext-link xlink:href="https://doi.org/10.1029/94JC00572" ext-link-type="DOI">10.1029/94JC00572</ext-link>,
1994.</mixed-citation></ref>
      <ref id="bib1.bibx19"><?xmltex \def\ref@label{{Flowerdew(2015)}}?><label>Flowerdew(2015)</label><?label Flowerdew2015?><mixed-citation>Flowerdew, J.: Towards a theory of optimal localisation, Tellus A, 67, 25257, <ext-link xlink:href="https://doi.org/10.3402/tellusa.v67.25257" ext-link-type="DOI">10.3402/tellusa.v67.25257</ext-link>,
2015.</mixed-citation></ref>
      <ref id="bib1.bibx20"><?xmltex \def\ref@label{{Gaspari and Cohn(1999)}}?><label>Gaspari and Cohn(1999)</label><?label Gaspari1999?><mixed-citation>Gaspari, G. and Cohn, S. E.: Construction of correlation functions in two and
three dimensions, Q. J. Roy. Meteor. Soc., 125,
723–757, <ext-link xlink:href="https://doi.org/10.1002/qj.49712555417" ext-link-type="DOI">10.1002/qj.49712555417</ext-link>, 1999.</mixed-citation></ref>
      <ref id="bib1.bibx21"><?xmltex \def\ref@label{{Gaspari et~al.(2006)Gaspari, Cohn, Guo, and Pawson}}?><label>Gaspari et al.(2006)Gaspari, Cohn, Guo, and Pawson</label><?label Gaspari2006?><mixed-citation>Gaspari, G., Cohn, S. E., Guo, J., and Pawson, S.: Construction and application
of covariance functions with variable length-fields, Q. J.
Roy. Meteor. Soc., 132, 1815–1838,
<ext-link xlink:href="https://doi.org/10.1256/qj.05.08" ext-link-type="DOI">10.1256/qj.05.08</ext-link>, 2006.</mixed-citation></ref>
      <ref id="bib1.bibx22"><?xmltex \def\ref@label{{Greybush et~al.(2011)Greybush, Kalnay, Miyoshi, Ide, and
Hunt}}?><label>Greybush et al.(2011)Greybush, Kalnay, Miyoshi, Ide, and
Hunt</label><?label Greybush2011?><mixed-citation>Greybush, S. J., Kalnay, E., Miyoshi, T., Ide, K., and Hunt, B. R.: Balance and
Ensemble Kalman Filter Localization Techniques, Mon. Weather Rev., 139,
511–522, <ext-link xlink:href="https://doi.org/10.1175/2010MWR3328.1" ext-link-type="DOI">10.1175/2010MWR3328.1</ext-link>, 2011.</mixed-citation></ref>
      <ref id="bib1.bibx23"><?xmltex \def\ref@label{{Gustafsson et~al.(2018)Gustafsson, Janji{\'{c}}, Schraff,
Leuenberger, Weissman, Reich, Brousseau, Montmerle, Wattrelot,
Bu{\v{c}}{\'{a}}nek, Mile, Hamdi, Lindskog, Barkmeijer, Dahlbom, Macpherson,
Ballard, Inverarity, Carley, Alexander, Dowell, Liu, Ikuta, and
Fujita}}?><label>Gustafsson et al.(2018)Gustafsson, Janjić, Schraff,
Leuenberger, Weissman, Reich, Brousseau, Montmerle, Wattrelot,
Bučánek, Mile, Hamdi, Lindskog, Barkmeijer, Dahlbom, Macpherson,
Ballard, Inverarity, Carley, Alexander, Dowell, Liu, Ikuta, and
Fujita</label><?label Gustafsson2018?><mixed-citation>Gustafsson, N., Janjić, T., Schraff, C., Leuenberger, D., Weissman, M.,
Reich, H., Brousseau, P., Montmerle, T., Wattrelot, E., Bučánek,
A., Mile, M., Hamdi, R., Lindskog, M., Barkmeijer, J., Dahlbom, M.,
Macpherson, B., Ballard, S., Inverarity, G., Carley, J., Alexander, C.,
Dowell, D., Liu, S., Ikuta, Y., and Fujita, T.: Survey of data assimilation
methods for convective-scale numerical weather prediction at operational
centres, Q. J. Roy. Meteor. Soc., 144,
1218–1256, <ext-link xlink:href="https://doi.org/10.1002/qj.3179" ext-link-type="DOI">10.1002/qj.3179</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx24"><?xmltex \def\ref@label{{Hagelin et~al.(2017)Hagelin, Son, Swinbank, McCabe, Roberts, and
Tennant}}?><label>Hagelin et al.(2017)Hagelin, Son, Swinbank, McCabe, Roberts, and
Tennant</label><?label Hagelin2017?><mixed-citation>Hagelin, S., Son, J., Swinbank, R., McCabe, A., Roberts, N., and Tennant, W.:
The Met Office convective-scale ensemble, MOGREPS-UK, Q. J. Roy. Meteor. Soc., 143, 2846–2861, <ext-link xlink:href="https://doi.org/10.1002/qj.3135" ext-link-type="DOI">10.1002/qj.3135</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx25"><?xmltex \def\ref@label{{Hamill et~al.(2001)Hamill, Whitaker, and Snyder}}?><label>Hamill et al.(2001)Hamill, Whitaker, and Snyder</label><?label Hamill2001?><mixed-citation>Hamill, T., Whitaker, J., and Snyder, C.: Distance-Dependent Filtering of
Background Error Covariance Estimates in an Ensemble Kalman Filter, Mon. Weather Rev., 129, 2776–2790,
<ext-link xlink:href="https://doi.org/10.1175/1520-0493(2001)129&lt;2776:DDFOBE&gt;2.0.CO;2" ext-link-type="DOI">10.1175/1520-0493(2001)129&lt;2776:DDFOBE&gt;2.0.CO;2</ext-link>, 2001.</mixed-citation></ref>
      <ref id="bib1.bibx26"><?xmltex \def\ref@label{{Higham(2002)}}?><label>Higham(2002)</label><?label Higham2002?><mixed-citation>Higham, N. J.: Computing the nearest correlation matrix–a problem from
finance, IMA J. Numer. Anal., 22, 329–343,
<ext-link xlink:href="https://doi.org/10.1093/imanum/22.3.329" ext-link-type="DOI">10.1093/imanum/22.3.329</ext-link>, 2002.</mixed-citation></ref>
      <ref id="bib1.bibx27"><?xmltex \def\ref@label{{Hohenegger and Schaer(2007)}}?><label>Hohenegger and Schaer(2007)</label><?label Hohenegger2007?><mixed-citation>Hohenegger, C. and Schaer, C.: Predictability and Error Growth Dynamics in
Cloud-Resolving Models, J. Atmos. Sci., 64, 4467–4478,
<ext-link xlink:href="https://doi.org/10.1175/2007JAS2143.1" ext-link-type="DOI">10.1175/2007JAS2143.1</ext-link>, 2007.</mixed-citation></ref>
      <ref id="bib1.bibx28"><?xmltex \def\ref@label{{Horn and Johnson(2012)}}?><label>Horn and Johnson(2012)</label><?label Horn2012?><mixed-citation>Horn, R. A. and Johnson, C. R.: Matrix Analysis, Cambridge University Press, 2
edn., <ext-link xlink:href="https://doi.org/10.1017/CBO9781139020411" ext-link-type="DOI">10.1017/CBO9781139020411</ext-link>, 2012.</mixed-citation></ref>
      <ref id="bib1.bibx29"><?xmltex \def\ref@label{{Hotta and Ota(2021)}}?><label>Hotta and Ota(2021)</label><?label Hotta2021?><mixed-citation>Hotta, D. and Ota, Y.: Why does EnKF suffer from analysis overconfidence? An
insight into exploiting the ever-increasing volume of observations, Q. J. Roy. Meteor. Soc., 147, 1258–1277,
<ext-link xlink:href="https://doi.org/10.1002/qj.3970" ext-link-type="DOI">10.1002/qj.3970</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx30"><?xmltex \def\ref@label{{Houtekamer and Mitchell(1998)}}?><label>Houtekamer and Mitchell(1998)</label><?label Houtekamer1998?><mixed-citation>Houtekamer, P. L. and Mitchell, H. L.: Data Assimilation Using an Ensemble
Kalman Filter Technique, Mon. Weather Rev., 126, 796–811,
<ext-link xlink:href="https://doi.org/10.1175/1520-0493(1998)126&lt;0796:DAUAEK&gt;2.0.CO;2" ext-link-type="DOI">10.1175/1520-0493(1998)126&lt;0796:DAUAEK&gt;2.0.CO;2</ext-link>, 1998.</mixed-citation></ref>
      <ref id="bib1.bibx31"><?xmltex \def\ref@label{{Houtekamer and Mitchell(2001)}}?><label>Houtekamer and Mitchell(2001)</label><?label Houtekamer2001?><mixed-citation>Houtekamer, P. L. and Mitchell, H. L.: A Sequential Ensemble Kalman Filter for
Atmospheric Data Assimilation, Mon. Weather Rev., 129, 123–137,
<ext-link xlink:href="https://doi.org/10.1175/1520-0493(2001)129&lt;0123:ASEKFF&gt;2.0.CO;2" ext-link-type="DOI">10.1175/1520-0493(2001)129&lt;0123:ASEKFF&gt;2.0.CO;2</ext-link>, 2001.</mixed-citation></ref>
      <ref id="bib1.bibx32"><?xmltex \def\ref@label{{Hoyer and Hamman(2017)}}?><label>Hoyer and Hamman(2017)</label><?label Hoyer2017?><mixed-citation>Hoyer, S. and Hamman, J. J.: xarray: N-D labeled Arrays and Datasets in
Python, Journal of Open Research Software, 5, 1–6, <ext-link xlink:href="https://doi.org/10.5334/jors.148" ext-link-type="DOI">10.5334/jors.148</ext-link>,
2017.</mixed-citation></ref>
      <ref id="bib1.bibx33"><?xmltex \def\ref@label{{Hu et~al.(2023)Hu, Dance, Bannister, Chipilski, Guillet, Macpherson,
Weissmann, and Yussouf}}?><label>Hu et al.(2023)Hu, Dance, Bannister, Chipilski, Guillet, Macpherson,
Weissmann, and Yussouf</label><?label Hu2022?><mixed-citation>Hu, G., Dance, S. L., Bannister, R. N., Chipilski, H., Guillet, O., Macpherson,
B., Weissmann, M., and Yussouf, N.: Progress, challenges, and future steps in data assimilation for convection-permitting numerical weather prediction: Report on the virtual meeting held on 10 and 12 November 2021, Atmos. Sci. Lett., 24, e1130, <ext-link xlink:href="https://doi.org/10.1002/asl.1130" ext-link-type="DOI">10.1002/asl.1130</ext-link>, 2023.</mixed-citation></ref>
      <ref id="bib1.bibx34"><?xmltex \def\ref@label{{Hunt et~al.(2007)Hunt, Kostelich, and Szunyogh}}?><label>Hunt et al.(2007)Hunt, Kostelich, and Szunyogh</label><?label Hunt2007?><mixed-citation>Hunt, B. R., Kostelich, E. J., and Szunyogh, I.: Efficient data assimilation
for spatiotemporal chaos: A local ensemble transform Kalman filter, Physica
D, 230, 112–126, <ext-link xlink:href="https://doi.org/10.1016/j.physd.2006.11.008" ext-link-type="DOI">10.1016/j.physd.2006.11.008</ext-link>,
2007.</mixed-citation></ref>
      <ref id="bib1.bibx35"><?xmltex \def\ref@label{{Kepert(2009)}}?><label>Kepert(2009)</label><?label Kepert2009?><mixed-citation>Kepert, J. D.: Covariance localisation and balance in an Ensemble Kalman
Filter, Q. J. Roy. Meteor. Soc., 135,
1157–1176, <ext-link xlink:href="https://doi.org/10.1002/qj.443" ext-link-type="DOI">10.1002/qj.443</ext-link>, 2009.</mixed-citation></ref>
      <ref id="bib1.bibx36"><?xmltex \def\ref@label{{Kirchgessner et~al.(2014)Kirchgessner, Nerger, and
Bunse-Gerstner}}?><label>Kirchgessner et al.(2014)Kirchgessner, Nerger, and
Bunse-Gerstner</label><?label Kirchgessner2014?><mixed-citation>Kirchgessner, P., Nerger, L., and Bunse-Gerstner, A.: On the Choice of an
Optimal Localization Radius in Ensemble Kalman Filter Methods, Mon. Weather Rev., 142, 2165–2175, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-13-00246.1" ext-link-type="DOI">10.1175/MWR-D-13-00246.1</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx37"><?xmltex \def\ref@label{{Kondo and Miyoshi(2016)}}?><label>Kondo and Miyoshi(2016)</label><?label Kondo2016?><mixed-citation>Kondo, K. and Miyoshi, T.: Impact of Removing Covariance Localization in an
Ensemble Kalman Filter: Experiments with 10 240 Members Using an Intermediate
AGCM, Mon. Weather Rev., 144, 4849–4865, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-15-0388.1" ext-link-type="DOI">10.1175/MWR-D-15-0388.1</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx38"><?xmltex \def\ref@label{{Kunii(2014)}}?><label>Kunii(2014)</label><?label Kunii2014?><mixed-citation>Kunii, M.: The 1000-Member Ensemble Kalman Filtering with the JMA
Nonhydrostatic Mesoscale Model on the K Computer,
J. Meteorol. Soc. Jpn., 92, 623–633,
<ext-link xlink:href="https://doi.org/10.2151/jmsj.2014-607" ext-link-type="DOI">10.2151/jmsj.2014-607</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx39"><?xmltex \def\ref@label{{Lei and Anderson(2014)}}?><label>Lei and Anderson(2014)</label><?label Lei2014a?><mixed-citation>Lei, L. and Anderson, J. L.: Comparisons of Empirical Localization Techniques
for Serial Ensemble Kalman Filters in a Simple Atmospheric General
Circulation Model, Mon. Weather Rev., 142, 739–754,
<ext-link xlink:href="https://doi.org/10.1175/MWR-D-13-00152.1" ext-link-type="DOI">10.1175/MWR-D-13-00152.1</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx40"><?xmltex \def\ref@label{{Lei et~al.(2015)Lei, Anderson, and Romine}}?><label>Lei et al.(2015)Lei, Anderson, and Romine</label><?label Lei2015?><mixed-citation>Lei, L., Anderson, J. L., and Romine, G. S.: Empirical Localization Functions
for Ensemble Kalman Filter Data Assimilation in Regions with and without
Precipitation, Mon. Weather Rev., 143, 3664–3679,
<ext-link xlink:href="https://doi.org/10.1175/MWR-D-14-00415.1" ext-link-type="DOI">10.1175/MWR-D-14-00415.1</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bibx41"><?xmltex \def\ref@label{{Lei et~al.(2016)Lei, Anderson, and Whitaker}}?><label>Lei et al.(2016)Lei, Anderson, and Whitaker</label><?label Lei2016?><mixed-citation>Lei, L., Anderson, J. L., and Whitaker, J. S.: Localizing the impact of
satellite radiance observations using a global group ensemble filter, J. Adv. Model. Earth Sy., 8, 719–734,
<ext-link xlink:href="https://doi.org/10.1002/2016MS000627" ext-link-type="DOI">10.1002/2016MS000627</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx42"><?xmltex \def\ref@label{{Lei et~al.(2018)Lei, Whitaker, and Bishop}}?><label>Lei et al.(2018)Lei, Whitaker, and Bishop</label><?label Lei2018?><mixed-citation>Lei, L., Whitaker, J. S., and Bishop, C.: Improving Assimilation of Radiance
Observations by Implementing Model Space Localization in an Ensemble Kalman
Filter, J. Adv. Model. Earth Sy., 10, 3221–3232,
<ext-link xlink:href="https://doi.org/10.1029/2018MS001468" ext-link-type="DOI">10.1029/2018MS001468</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx43"><?xmltex \def\ref@label{{Lei et~al.(2020)Lei, Whitaker, Anderson, and Tan}}?><label>Lei et al.(2020)Lei, Whitaker, Anderson, and Tan</label><?label Lei2020?><mixed-citation>Lei, L., Whitaker, J. S., Anderson, J. L., and Tan, Z.: Adaptive Localization
for Satellite Radiance Observations in an Ensemble Kalman Filter, J.
Adv. Model. Earth Sy., 12, e2019MS001693,
<ext-link xlink:href="https://doi.org/10.1029/2019MS001693" ext-link-type="DOI">10.1029/2019MS001693</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx44"><?xmltex \def\ref@label{{Lien et~al.(2017)Lien, Miyoshi, Nishizawa, Yoshida, Yashiro, Adachi,
Yamaura, and Tomita}}?><label>Lien et al.(2017)Lien, Miyoshi, Nishizawa, Yoshida, Yashiro, Adachi,
Yamaura, and Tomita</label><?label Lien2017?><mixed-citation>Lien, G.-Y., Miyoshi, T., Nishizawa, S., Yoshida, R., Yashiro, H., Adachi,
S. A., Yamaura, T., and Tomita, H.: The Near-Real-Time SCALE-LETKF System: A
Case of the September 2015 Kanto-Tohoku Heavy Rainfall, SOLA, 13, 1–6,
<ext-link xlink:href="https://doi.org/10.2151/sola.2017-001" ext-link-type="DOI">10.2151/sola.2017-001</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx45"><?xmltex \def\ref@label{{Michel et~al.(2011)Michel, Auligné, and Montmerle}}?><label>Michel et al.(2011)Michel, Auligné, and Montmerle</label><?label Michel2011?><mixed-citation>Michel, Y., Auligné, T., and Montmerle, T.: Heterogeneous Convective-Scale
Background Error Covariances with the Inclusion of Hydrometeor Variables,
Mon. Weather Rev., 139, 2994–3015, <ext-link xlink:href="https://doi.org/10.1175/2011MWR3632.1" ext-link-type="DOI">10.1175/2011MWR3632.1</ext-link>, 2011.</mixed-citation></ref>
      <ref id="bib1.bibx46"><?xmltex \def\ref@label{{Miyoshi and Yamane(2007)}}?><label>Miyoshi and Yamane(2007)</label><?label Miyoshi2007?><mixed-citation>Miyoshi, T. and Yamane, S.: Local Ensemble Transform Kalman Filtering with an
AGCM at a T159/L48 Resolution, Mon. Weather Rev., 135, 3841–3861,
<ext-link xlink:href="https://doi.org/10.1175/2007MWR1873.1" ext-link-type="DOI">10.1175/2007MWR1873.1</ext-link>, 2007.</mixed-citation></ref>
      <ref id="bib1.bibx47"><?xmltex \def\ref@label{{Miyoshi et~al.(2014)Miyoshi, Kondo, and Imamura}}?><label>Miyoshi et al.(2014)Miyoshi, Kondo, and Imamura</label><?label Miyoshi2014?><mixed-citation>Miyoshi, T., Kondo, K., and Imamura, T.: The 10,240-member ensemble Kalman
filtering with an intermediate AGCM, Geophys. Res. Lett., 41,
5264–5271, <ext-link xlink:href="https://doi.org/10.1002/2014GL060863" ext-link-type="DOI">10.1002/2014GL060863</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx48"><?xmltex \def\ref@label{{Miyoshi et~al.(2015)Miyoshi, Kondo, and Terasaki}}?><label>Miyoshi et al.(2015)Miyoshi, Kondo, and Terasaki</label><?label Miyoshi2015?><mixed-citation>Miyoshi, T., Kondo, K., and Terasaki, K.: Big Ensemble Data Assimilation in
Numerical Weather Prediction, Computer, 48, 15–21,
<ext-link xlink:href="https://doi.org/10.1109/MC.2015.332" ext-link-type="DOI">10.1109/MC.2015.332</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bibx49"><?xmltex \def\ref@label{{Miyoshi et~al.(2016{\natexlab{a}})Miyoshi, Kunii, Ruiz, Lien, Satoh,
Ushio, Bessho, Seko, Tomita, and Ishikawa}}?><label>Miyoshi et al.(2016a)Miyoshi, Kunii, Ruiz, Lien, Satoh,
Ushio, Bessho, Seko, Tomita, and Ishikawa</label><?label Miyoshi2016a?><mixed-citation>Miyoshi, T., Kunii, M., Ruiz, J., Lien, G.-Y., Satoh, S., Ushio, T., Bessho,
K., Seko, H., Tomita, H., and Ishikawa, Y.: Big Data Assimilation
Revolutionizing Severe Weather Prediction, B. Am. Meteor. Soc., 97, 1347–1354, <ext-link xlink:href="https://doi.org/10.1175/BAMS-D-15-00144.1" ext-link-type="DOI">10.1175/BAMS-D-15-00144.1</ext-link>,
2016a.</mixed-citation></ref>
      <ref id="bib1.bibx50"><?xmltex \def\ref@label{{Miyoshi et~al.(2016{\natexlab{b}})Miyoshi, Lien, Satoh, Ushio,
Bessho, Tomita, Nishizawa, Yoshida, Adachi, Liao, Gerofi, Ishikawa, Kunii,
Ruiz, Maejima, Otsuka, Otsuka, Okamoto, and Seko}}?><label>Miyoshi et al.(2016b)Miyoshi, Lien, Satoh, Ushio,
Bessho, Tomita, Nishizawa, Yoshida, Adachi, Liao, Gerofi, Ishikawa, Kunii,
Ruiz, Maejima, Otsuka, Otsuka, Okamoto, and Seko</label><?label Miyoshi2016b?><mixed-citation>Miyoshi, T., Lien, G., Satoh, S., Ushio, T., Bessho, K., Tomita, H., Nishizawa,
S., Yoshida, R., Adachi, S. A., Liao, J., Gerofi, B., Ishikawa, Y., Kunii,
M., Ruiz, J., Maejima, Y., Otsuka, S., Otsuka, M., Okamoto, K., and Seko, H.:
“Big Data Assimilation” Toward Post-Petascale Severe Weather Prediction:
An Overview and Progress, P. IEEE, 104, 2155–2179,
<ext-link xlink:href="https://doi.org/10.1109/JPROC.2016.2602560" ext-link-type="DOI">10.1109/JPROC.2016.2602560</ext-link>, 2016b.</mixed-citation></ref>
      <ref id="bib1.bibx51"><?xmltex \def\ref@label{{Ménétrier et~al.(2014)Ménétrier, Montmerle, Berre, and
Michel}}?><label>Ménétrier et al.(2014)Ménétrier, Montmerle, Berre, and
Michel</label><?label Menetrier2014?><mixed-citation>Ménétrier, B., Montmerle, T., Berre, L., and Michel, Y.: Estimation and
diagnosis of heterogeneous flow-dependent background-error covariances at the
convective scale using either large or small ensembles, Q. J.
Roy. Meteor. Soc., 140, 2050–2061, <ext-link xlink:href="https://doi.org/10.1002/qj.2267" ext-link-type="DOI">10.1002/qj.2267</ext-link>,
2014.</mixed-citation></ref>
      <ref id="bib1.bibx52"><?xmltex \def\ref@label{{Ménétrier et~al.(2015{\natexlab{a}})Ménétrier, Montmerle, Michel,
and Berre}}?><label>Ménétrier et al.(2015a)Ménétrier, Montmerle, Michel,
and Berre</label><?label Menetrier2015a?><mixed-citation>Ménétrier, B., Montmerle, T., Michel, Y., and Berre, L.: Linear Filtering of
Sample Covariances for Ensemble-Based Data Assimilation. Part I: Optimality
Criteria and Application to Variance Filtering and Covariance Localization,
Mon. Weather Rev., 143, 1622–1643, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-14-00157.1" ext-link-type="DOI">10.1175/MWR-D-14-00157.1</ext-link>,
2015a.</mixed-citation></ref>
      <ref id="bib1.bibx53"><?xmltex \def\ref@label{{Ménétrier et~al.(2015{\natexlab{b}})Ménétrier, Montmerle, Michel,
and Berre}}?><label>Ménétrier et al.(2015b)Ménétrier, Montmerle, Michel,
and Berre</label><?label Menetrier2015b?><mixed-citation>Ménétrier, B., Montmerle, T., Michel, Y., and Berre, L.: Linear Filtering of
Sample Covariances for Ensemble-Based Data Assimilation. Part II: Application
to a Convective-Scale NWP Model, Mon. Weather Rev., 143, 1644–1664,
<ext-link xlink:href="https://doi.org/10.1175/MWR-D-14-00156.1" ext-link-type="DOI">10.1175/MWR-D-14-00156.1</ext-link>, 2015b.</mixed-citation></ref>
      <ref id="bib1.bibx54"><?xmltex \def\ref@label{{Necker(2022)}}?><label>Necker(2022)</label><?label necker_tobias_2022_7254119?><mixed-citation>Necker, T.: Empirical optimal localization/Vertical covariance
localization, Zenodo [data set, code], <ext-link xlink:href="https://doi.org/10.5281/zenodo.7254119" ext-link-type="DOI">10.5281/zenodo.7254119</ext-link>, 2022.</mixed-citation></ref>
      <ref id="bib1.bibx55"><?xmltex \def\ref@label{{Necker et~al.(2020{\natexlab{a}})Necker, Geiss, Weissmann, Ruiz,
Miyoshi, and Lien}}?><label>Necker et al.(2020a)Necker, Geiss, Weissmann, Ruiz,
Miyoshi, and Lien</label><?label Necker2020a?><mixed-citation>Necker, T., Geiss, S., Weissmann, M., Ruiz, J., Miyoshi, T., and Lien, G.-Y.: A
Convective-scale 1,000-member Ensemble Simulation and Potential Applications,
Q. J. Roy. Meteor. Soc., 146, 1423–1442, <ext-link xlink:href="https://doi.org/10.1002/qj.3744" ext-link-type="DOI">10.1002/qj.3744</ext-link>,
2020a.</mixed-citation></ref>
      <ref id="bib1.bibx56"><?xmltex \def\ref@label{{Necker et~al.(2020{\natexlab{b}})Necker, Weissmann, Ruckstuhl,
Anderson, and Miyoshi}}?><label>Necker et al.(2020b)Necker, Weissmann, Ruckstuhl,
Anderson, and Miyoshi</label><?label Necker2020b?><mixed-citation>Necker, T., Weissmann, M., Ruckstuhl, Y., Anderson, J., and Miyoshi, T.:
Sampling Error Correction Evaluated Using a Convective-Scale
1000-Member Ensemble, Mon. Weather Rev., 148, 1229–1249,
<ext-link xlink:href="https://doi.org/10.1175/MWR-D-19-0154.1" ext-link-type="DOI">10.1175/MWR-D-19-0154.1</ext-link>, 2020b.</mixed-citation></ref>
      <ref id="bib1.bibx57"><?xmltex \def\ref@label{{Nomokonova et~al.(2022)Nomokonova, Griewank, Löhnert, Miyoshi,
Necker, and Weissmann}}?><label>Nomokonova et al.(2022)Nomokonova, Griewank, Löhnert, Miyoshi,
Necker, and Weissmann</label><?label Nomokonova2022?><mixed-citation>Nomokonova, T., Griewank, P. J., Löhnert, U., Miyoshi, T., Necker, T., and
Weissmann, M.: Estimating the benefit of Doppler wind lidars for short-term low-level wind ensemble forecasts, Q. J. Roy. Meteor. Soc., 1–19, <ext-link xlink:href="https://doi.org/10.1002/qj.4402" ext-link-type="DOI">10.1002/qj.4402</ext-link>, 2022.</mixed-citation></ref>
      <ref id="bib1.bibx58"><?xmltex \def\ref@label{{Perianez et~al.(2014)Perianez, Reich, and Potthast}}?><label>Perianez et al.(2014)Perianez, Reich, and Potthast</label><?label Perianez2014?><mixed-citation>Perianez, A., Reich, H., and Potthast, R.: Optimal Localization for Ensemble
Kalman Filter Systems, J. Meteor. Soc. Jpn., 92, 585–597, <ext-link xlink:href="https://doi.org/10.2151/jmsj.2014-605" ext-link-type="DOI">10.2151/jmsj.2014-605</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx59"><?xmltex \def\ref@label{{Piper et~al.(2016)Piper, Kunz, Ehmele, Mohr, M{\"{u}}hr, Kron, and
Daniell}}?><label>Piper et al.(2016)Piper, Kunz, Ehmele, Mohr, Mühr, Kron, and
Daniell</label><?label Piper2016?><mixed-citation>Piper, D., Kunz, M., Ehmele, F., Mohr, S., Mühr, B., Kron, A., and Daniell, J.: Exceptional sequence of severe thunderstorms and related flash floods in May and June 2016 in Germany – Part 1: Meteorological background, Nat. Hazards Earth Syst. Sci., 16, 2835–2850, <ext-link xlink:href="https://doi.org/10.5194/nhess-16-2835-2016" ext-link-type="DOI">10.5194/nhess-16-2835-2016</ext-link>, 2016.
</mixed-citation></ref><?xmltex \hack{\newpage}?>
      <ref id="bib1.bibx60"><?xmltex \def\ref@label{{Poterjoy et~al.(2014)}}?><label>Poterjoy et al.(2014)</label><?label Poterjoy2014?><mixed-citation>Poterjoy, J., Zhang, F., and Weng, Y.: The Effects of Sampling Errors on the
EnKF Assimilation of Inner-Core Hurricane Observations, Mon. Weather Rev., 142, 1609–1630, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-13-00305.1" ext-link-type="DOI">10.1175/MWR-D-13-00305.1</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx61"><?xmltex \def\ref@label{{Scheck et~al.(2020)Scheck, Weissmann, and Bach}}?><label>Scheck et al.(2020)Scheck, Weissmann, and Bach</label><?label Scheck2020?><mixed-citation>Scheck, L., Weissmann, M., and Bach, L.: Assimilating visible satellite images
for convective-scale numerical weather prediction: A case-study, Q. J. Roy. Meteor. Soc., 146, 3165–3186,
<ext-link xlink:href="https://doi.org/10.1002/qj.3840" ext-link-type="DOI">10.1002/qj.3840</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx62"><?xmltex \def\ref@label{{Schraff et~al.(2016)Schraff, Reich, Rhodin, Schomburg, Stephan,
Periáñez, and Potthast}}?><label>Schraff et al.(2016)Schraff, Reich, Rhodin, Schomburg, Stephan,
Periáñez, and Potthast</label><?label Schraff2015?><mixed-citation>Schraff, C., Reich, H., Rhodin, A., Schomburg, A., Stephan, K., Periáñez, A.,
and Potthast, R.: Kilometre-scale ensemble data assimilation for the COSMO
model (KENDA), Q. J. Roy. Meteor. Soc., 142,
1453–1472, <ext-link xlink:href="https://doi.org/10.1002/qj.2748" ext-link-type="DOI">10.1002/qj.2748</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx63"><?xmltex \def\ref@label{{Stanley et~al.(2021)Stanley, Grooms, and Kleiber}}?><label>Stanley et al.(2021)Stanley, Grooms, and Kleiber</label><?label Stanley2021?><mixed-citation>Stanley, Z., Grooms, I., and Kleiber, W.: Multivariate localization functions for strongly coupled data assimilation in the bivariate Lorenz 96 system, Nonlin. Processes Geophys., 28, 565–583, <ext-link xlink:href="https://doi.org/10.5194/npg-28-565-2021" ext-link-type="DOI">10.5194/npg-28-565-2021</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx64"><?xmltex \def\ref@label{{Tabeart et~al.(2019)Tabeart, Dance, Lawless, Nichols, and
Waller}}?><label>Tabeart et al.(2019)Tabeart, Dance, Lawless, Nichols, and
Waller</label><?label Tabeart2019?><mixed-citation>Tabeart, J. M., Dance, S. L., Lawless, A. S., Nichols, N. K., and Waller,
J. A.: Improving the condition number of estimated covariance matrices,
Tellus, 72, 1–19, <ext-link xlink:href="https://doi.org/10.1080/16000870.2019.1696646" ext-link-type="DOI">10.1080/16000870.2019.1696646</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx65"><?xmltex \def\ref@label{{Virtanen et~al.(2020)Virtanen, Gommers, Oliphant, Haberland, Reddy,
Cournapeau, Burovski, Peterson, Weckesser, Bright, {van der Walt}, Brett,
Wilson, Millman, Mayorov, Nelson, Jones, Kern, Larson, Carey, Polat, Feng,
Moore, {VanderPlas}, Laxalde, Perktold, Cimrman, Henriksen, Quintero, Harris,
Archibald, Ribeiro, Pedregosa, {van Mulbregt}, and {SciPy 1.0
Contributors}}}?><label>Virtanen et al.(2020)Virtanen, Gommers, Oliphant, Haberland, Reddy,
Cournapeau, Burovski, Peterson, Weckesser, Bright, van der Walt, Brett,
Wilson, Millman, Mayorov, Nelson, Jones, Kern, Larson, Carey, Polat, Feng,
Moore, VanderPlas, Laxalde, Perktold, Cimrman, Henriksen, Quintero, Harris,
Archibald, Ribeiro, Pedregosa, van Mulbregt, and SciPy 1.0
Contributors</label><?label Scipy2020?><mixed-citation>Virtanen, P., Gommers, R., Oliphant, T. E., Haberland, M., Reddy, T.,
Cournapeau, D., Burovski, E., Peterson, P., Weckesser, W., Bright, J., van
der Walt, S. J., Brett, M., Wilson, J., Millman, K. J., Mayorov, N., Nelson,
A. R. J., Jones, E., Kern, R., Larson, E., Carey, C. J., Polat, I., Feng, Y.,
Moore, E. W., VanderPlas, J., Laxalde, D., Perktold, J., Cimrman, R.,
Henriksen, I., Quintero, E. A., Harris, C. R., Archibald, A. M., Ribeiro,
A. H., Pedregosa, F., van Mulbregt, P., and SciPy 1.0 Contributors:
SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python,
Nat. Methods, 17, 261–272, <ext-link xlink:href="https://doi.org/10.1038/s41592-019-0686-2" ext-link-type="DOI">10.1038/s41592-019-0686-2</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx66"><?xmltex \def\ref@label{{Whitaker and Hamill(2012)}}?><label>Whitaker and Hamill(2012)</label><?label Whitaker2012?><mixed-citation>Whitaker, J. S. and Hamill, T. M.: Evaluating Methods to Account for System
Errors in Ensemble Data Assimilation, Mon. Weather Rev., 140, 3078–3089, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-11-00276.1" ext-link-type="DOI">10.1175/MWR-D-11-00276.1</ext-link>, 2012.</mixed-citation></ref>
      <ref id="bib1.bibx67"><?xmltex \def\ref@label{{Wu et~al.(2020)Wu, Yang, Tsai, and Cheng}}?><label>Wu et al.(2020)Wu, Yang, Tsai, and Cheng</label><?label Wu2020?><mixed-citation>Wu, P.-Y., Yang, S.-C., Tsai, C.-C., and Cheng, H.-W.: Convective-Scale
Sampling Error and Its Impact on the Ensemble Radar Data Assimilation System:
A Case Study of a Heavy Rainfall Event on 16 June 2008 in Taiwan, Mon. Weather Rev., 148, 3631–3652, <ext-link xlink:href="https://doi.org/10.1175/MWR-D-19-0319.1" ext-link-type="DOI">10.1175/MWR-D-19-0319.1</ext-link>, 2020.</mixed-citation></ref>

  </ref-list></back>
    <!--<article-title-html>Guidance on how to improve vertical covariance localization based on a 1000-member ensemble</article-title-html>
<abstract-html/>
<ref-html id="bib1.bib1"><label>Anderson(2007)</label><mixed-citation>
Anderson, J.: Exploring the need for localization in ensemble data
assimilation using a hierarchical ensemble filter, Physica D, 230, 99–111, <a href="https://doi.org/10.1016/j.physd.2006.02.011" target="_blank">https://doi.org/10.1016/j.physd.2006.02.011</a>, 2007.
</mixed-citation></ref-html>
<ref-html id="bib1.bib2"><label>Anderson and Lei(2013)</label><mixed-citation>
Anderson, J. and Lei, L.: Empirical Localization of Observation Impact in
Ensemble Kalman Filters, Mon. Weather Rev., 141, 4140–4153,
<a href="https://doi.org/10.1175/MWR-D-12-00330.1" target="_blank">https://doi.org/10.1175/MWR-D-12-00330.1</a>, 2013.
</mixed-citation></ref-html>
<ref-html id="bib1.bib3"><label>Anderson et al.(2009)Anderson, Hoar, Raeder, Liu, Collins, Torn, and
Avellano</label><mixed-citation>
Anderson, J., Hoar, T., Raeder, K., Liu, H., Collins, N., Torn, R., and
Avellano, A.: The Data Assimilation Research Testbed: A Community Facility,
B. Am. Meteor. Soc., 90, 1283–1296,
<a href="https://doi.org/10.1175/2009BAMS2618.1" target="_blank">https://doi.org/10.1175/2009BAMS2618.1</a>, 2009.
</mixed-citation></ref-html>
<ref-html id="bib1.bib4"><label>Anderson(2001)</label><mixed-citation>
Anderson, J. L.: An Ensemble Adjustment Kalman Filter for Data Assimilation,
Mon. Weather Rev., 129, 2884–2903,
<a href="https://doi.org/10.1175/1520-0493(2001)129&lt;2884:AEAKFF&gt;2.0.CO;2" target="_blank">https://doi.org/10.1175/1520-0493(2001)129&lt;2884:AEAKFF&gt;2.0.CO;2</a>, 2001.
</mixed-citation></ref-html>
<ref-html id="bib1.bib5"><label>Anderson(2012)</label><mixed-citation>
Anderson, J. L.: Localization and Sampling Error Correction in Ensemble Kalman
Filter Data Assimilation, Mon. Weather Rev., 140, 2359–2371,
<a href="https://doi.org/10.1175/MWR-D-11-00013.1" target="_blank">https://doi.org/10.1175/MWR-D-11-00013.1</a>, 2012.
</mixed-citation></ref-html>
<ref-html id="bib1.bib6"><label>Anderson(2016)</label><mixed-citation>
Anderson, J. L.: Reducing Correlation Sampling Error in Ensemble Kalman Filter
Data Assimilation, Mon. Weather Rev., 144, 913–925,
<a href="https://doi.org/10.1175/MWR-D-15-0052.1" target="_blank">https://doi.org/10.1175/MWR-D-15-0052.1</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib7"><label>Bannister(2008)</label><mixed-citation>
Bannister, R. N.: A review of forecast error covariance statistics in
atmospheric variational data assimilation. II: Modelling the forecast error
covariance statistics,
Q. J. Roy. Meteor. Soc.,
134, 1971–1996, <a href="https://doi.org/10.1002/qj.340" target="_blank">https://doi.org/10.1002/qj.340</a>, 2008.
</mixed-citation></ref-html>
<ref-html id="bib1.bib8"><label>Bannister(2017)</label><mixed-citation>
Bannister, R. N.: A review of operational methods of variational and
ensemble-variational data assimilation, Q. J. R. Meteorol. Soc., 143,
607–633, <a href="https://doi.org/10.1002/qj.2982" target="_blank">https://doi.org/10.1002/qj.2982</a>, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib9"><label>Bannister et al.(2017)Bannister, Migliorini, Rudd, and
Baker</label><mixed-citation>
Bannister, R. N., Migliorini, S., Rudd, A. C., and Baker, L. H.: Methods of investigating forecast error sensitivity to ensemble size in a limited-area convection-permitting ensemble, Geosci. Model Dev. Discuss. [preprint], <a href="https://doi.org/10.5194/gmd-2017-260" target="_blank">https://doi.org/10.5194/gmd-2017-260</a>, in review, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib10"><label>Bishop and Hodyss(2009a)</label><mixed-citation>
Bishop, C. H. and Hodyss, D.: Ensemble covariances adaptively localized with
ECO-RAP. Part 1: tests on simple error models, Tellus A, 61, 84–96,
<a href="https://doi.org/10.1111/j.1600-0870.2008.00371.x" target="_blank">https://doi.org/10.1111/j.1600-0870.2008.00371.x</a>, 2009a.
</mixed-citation></ref-html>
<ref-html id="bib1.bib11"><label>Bishop and Hodyss(2009b)</label><mixed-citation>
Bishop, C. H. and Hodyss, D.: Ensemble covariances adaptively localized with
ECO-RAP. Part 2: a strategy for the atmosphere, Tellus A, 61, 97–111,
<a href="https://doi.org/10.1111/j.1600-0870.2008.00372.x" target="_blank">https://doi.org/10.1111/j.1600-0870.2008.00372.x</a>, 2009b.
</mixed-citation></ref-html>
<ref-html id="bib1.bib12"><label>Bolin and Wallin(2016)</label><mixed-citation>
Bolin, D. and Wallin, J.: Spatially adaptive covariance tapering, Spatial
Statistics, 18, 163–178, <a href="https://doi.org/10.1016/j.spasta.2016.03.003" target="_blank">https://doi.org/10.1016/j.spasta.2016.03.003</a>,
2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib13"><label>Bonavita et al.(2016)Bonavita, Hólm, Isaksen, and
Fisher</label><mixed-citation>
Bonavita, M., Hólm, E., Isaksen, L., and Fisher, M.: The evolution of the
ECMWF hybrid data assimilation system, Q. J. Roy. Meteor. Soc., 142, 287–303, <a href="https://doi.org/10.1002/qj.2652" target="_blank">https://doi.org/10.1002/qj.2652</a>,
2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib14"><label>Bouttier et al.(2016)Bouttier, Raynaud, Nuissier, and
Menetrier</label><mixed-citation>
Bouttier, F., Raynaud, L., Nuissier, O., and Menetrier, B.: Sensitivity of the
AROME ensemble to initial and surface perturbations during HyMeX, Q. J. Roy. Meteor. Soc., 142, 390–403,
<a href="https://doi.org/10.1002/qj.2622" target="_blank">https://doi.org/10.1002/qj.2622</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib15"><label>Craig et al.(2022)Craig, Puh, Keil, Tempest, Necker, Ruiz, Weissmann,
and Miyoshi</label><mixed-citation>
Craig, G. C., Puh, M., Keil, C., Tempest, K., Necker, T., Ruiz, J., Weissmann,
M., and Miyoshi, T.: Distributions and convergence of forecast variables in
a 1000 member convection-permitting ensemble, Q. J. Roy.
Meteor. Soc., 148, 2325–2343, <a href="https://doi.org/10.1002/qj.4305" target="_blank">https://doi.org/10.1002/qj.4305</a>, 2022.
</mixed-citation></ref-html>
<ref-html id="bib1.bib16"><label>Daley et al.(2015)Daley, Porcu, and Bevilacqua</label><mixed-citation>
Daley, D., Porcu, E., and Bevilacqua, M.: Classes of compactly supported
covariance functions for multivariate random fields,
Stoch. Env. Res. Risk A., 29, 1249–1263,
<a href="https://doi.org/10.1007/s00477-014-0996-y" target="_blank">https://doi.org/10.1007/s00477-014-0996-y</a>, 2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib17"><label>Destouches et al.(2021)Destouches, Montmerle, Michel, and
Ménétrier</label><mixed-citation>
Destouches, M., Montmerle, T., Michel, Y., and Ménétrier, B.: Estimating
optimal localization for sampled background-error covariances of hydrometeor
variables, Q. J. Roy. Meteor. Soc., 147,
74–93, <a href="https://doi.org/10.1002/qj.3906" target="_blank">https://doi.org/10.1002/qj.3906</a>, 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib18"><label>Evensen(1994)</label><mixed-citation>
Evensen, G.: Sequential data assimilation with a nonlinear quasi-geostrophic
model using Monte Carlo methods to forecast error statistics, J.
Geophys. Res.-Oceans, 99, 10143–10162, <a href="https://doi.org/10.1029/94JC00572" target="_blank">https://doi.org/10.1029/94JC00572</a>,
1994.
</mixed-citation></ref-html>
<ref-html id="bib1.bib19"><label>Flowerdew(2015)</label><mixed-citation>
Flowerdew, J.: Towards a theory of optimal localisation, Tellus A, 67, 25257, <a href="https://doi.org/10.3402/tellusa.v67.25257" target="_blank">https://doi.org/10.3402/tellusa.v67.25257</a>,
2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib20"><label>Gaspari and Cohn(1999)</label><mixed-citation>
Gaspari, G. and Cohn, S. E.: Construction of correlation functions in two and
three dimensions, Q. J. Roy. Meteor. Soc., 125,
723–757, <a href="https://doi.org/10.1002/qj.49712555417" target="_blank">https://doi.org/10.1002/qj.49712555417</a>, 1999.
</mixed-citation></ref-html>
<ref-html id="bib1.bib21"><label>Gaspari et al.(2006)Gaspari, Cohn, Guo, and Pawson</label><mixed-citation>
Gaspari, G., Cohn, S. E., Guo, J., and Pawson, S.: Construction and application
of covariance functions with variable length-fields, Q. J.
Roy. Meteor. Soc., 132, 1815–1838,
<a href="https://doi.org/10.1256/qj.05.08" target="_blank">https://doi.org/10.1256/qj.05.08</a>, 2006.
</mixed-citation></ref-html>
<ref-html id="bib1.bib22"><label>Greybush et al.(2011)Greybush, Kalnay, Miyoshi, Ide, and
Hunt</label><mixed-citation>
Greybush, S. J., Kalnay, E., Miyoshi, T., Ide, K., and Hunt, B. R.: Balance and
Ensemble Kalman Filter Localization Techniques, Mon. Weather Rev., 139,
511–522, <a href="https://doi.org/10.1175/2010MWR3328.1" target="_blank">https://doi.org/10.1175/2010MWR3328.1</a>, 2011.
</mixed-citation></ref-html>
<ref-html id="bib1.bib23"><label>Gustafsson et al.(2018)Gustafsson, Janjić, Schraff,
Leuenberger, Weissman, Reich, Brousseau, Montmerle, Wattrelot,
Bučánek, Mile, Hamdi, Lindskog, Barkmeijer, Dahlbom, Macpherson,
Ballard, Inverarity, Carley, Alexander, Dowell, Liu, Ikuta, and
Fujita</label><mixed-citation>
Gustafsson, N., Janjić, T., Schraff, C., Leuenberger, D., Weissman, M.,
Reich, H., Brousseau, P., Montmerle, T., Wattrelot, E., Bučánek,
A., Mile, M., Hamdi, R., Lindskog, M., Barkmeijer, J., Dahlbom, M.,
Macpherson, B., Ballard, S., Inverarity, G., Carley, J., Alexander, C.,
Dowell, D., Liu, S., Ikuta, Y., and Fujita, T.: Survey of data assimilation
methods for convective-scale numerical weather prediction at operational
centres, Q. J. Roy. Meteor. Soc., 144,
1218–1256, <a href="https://doi.org/10.1002/qj.3179" target="_blank">https://doi.org/10.1002/qj.3179</a>, 2018.
</mixed-citation></ref-html>
<ref-html id="bib1.bib24"><label>Hagelin et al.(2017)Hagelin, Son, Swinbank, McCabe, Roberts, and
Tennant</label><mixed-citation>
Hagelin, S., Son, J., Swinbank, R., McCabe, A., Roberts, N., and Tennant, W.:
The Met Office convective-scale ensemble, MOGREPS-UK, Q. J. Roy. Meteor. Soc., 143, 2846–2861, <a href="https://doi.org/10.1002/qj.3135" target="_blank">https://doi.org/10.1002/qj.3135</a>, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib25"><label>Hamill et al.(2001)Hamill, Whitaker, and Snyder</label><mixed-citation>
Hamill, T., Whitaker, J., and Snyder, C.: Distance-Dependent Filtering of
Background Error Covariance Estimates in an Ensemble Kalman Filter, Mon. Weather Rev., 129, 2776–2790,
<a href="https://doi.org/10.1175/1520-0493(2001)129&lt;2776:DDFOBE&gt;2.0.CO;2" target="_blank">https://doi.org/10.1175/1520-0493(2001)129&lt;2776:DDFOBE&gt;2.0.CO;2</a>, 2001.
</mixed-citation></ref-html>
<ref-html id="bib1.bib26"><label>Higham(2002)</label><mixed-citation>
Higham, N. J.: Computing the nearest correlation matrix–a problem from
finance, IMA J. Numer. Anal., 22, 329–343,
<a href="https://doi.org/10.1093/imanum/22.3.329" target="_blank">https://doi.org/10.1093/imanum/22.3.329</a>, 2002.
</mixed-citation></ref-html>
<ref-html id="bib1.bib27"><label>Hohenegger and Schaer(2007)</label><mixed-citation>
Hohenegger, C. and Schaer, C.: Predictability and Error Growth Dynamics in
Cloud-Resolving Models, J. Atmos. Sci., 64, 4467–4478,
<a href="https://doi.org/10.1175/2007JAS2143.1" target="_blank">https://doi.org/10.1175/2007JAS2143.1</a>, 2007.
</mixed-citation></ref-html>
<ref-html id="bib1.bib28"><label>Horn and Johnson(2012)</label><mixed-citation>
Horn, R. A. and Johnson, C. R.: Matrix Analysis, Cambridge University Press, 2
edn., <a href="https://doi.org/10.1017/CBO9781139020411" target="_blank">https://doi.org/10.1017/CBO9781139020411</a>, 2012.
</mixed-citation></ref-html>
<ref-html id="bib1.bib29"><label>Hotta and Ota(2021)</label><mixed-citation>
Hotta, D. and Ota, Y.: Why does EnKF suffer from analysis overconfidence? An
insight into exploiting the ever-increasing volume of observations, Q. J. Roy. Meteor. Soc., 147, 1258–1277,
<a href="https://doi.org/10.1002/qj.3970" target="_blank">https://doi.org/10.1002/qj.3970</a>, 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib30"><label>Houtekamer and Mitchell(1998)</label><mixed-citation>
Houtekamer, P. L. and Mitchell, H. L.: Data Assimilation Using an Ensemble
Kalman Filter Technique, Mon. Weather Rev., 126, 796–811,
<a href="https://doi.org/10.1175/1520-0493(1998)126&lt;0796:DAUAEK&gt;2.0.CO;2" target="_blank">https://doi.org/10.1175/1520-0493(1998)126&lt;0796:DAUAEK&gt;2.0.CO;2</a>, 1998.
</mixed-citation></ref-html>
<ref-html id="bib1.bib31"><label>Houtekamer and Mitchell(2001)</label><mixed-citation>
Houtekamer, P. L. and Mitchell, H. L.: A Sequential Ensemble Kalman Filter for
Atmospheric Data Assimilation, Mon. Weather Rev., 129, 123–137,
<a href="https://doi.org/10.1175/1520-0493(2001)129&lt;0123:ASEKFF&gt;2.0.CO;2" target="_blank">https://doi.org/10.1175/1520-0493(2001)129&lt;0123:ASEKFF&gt;2.0.CO;2</a>, 2001.
</mixed-citation></ref-html>
<ref-html id="bib1.bib32"><label>Hoyer and Hamman(2017)</label><mixed-citation>
Hoyer, S. and Hamman, J. J.: xarray: N-D labeled Arrays and Datasets in
Python, Journal of Open Research Software, 5, 1–6, <a href="https://doi.org/10.5334/jors.148" target="_blank">https://doi.org/10.5334/jors.148</a>,
2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib33"><label>Hu et al.(2023)Hu, Dance, Bannister, Chipilski, Guillet, Macpherson,
Weissmann, and Yussouf</label><mixed-citation>
Hu, G., Dance, S. L., Bannister, R. N., Chipilski, H., Guillet, O., Macpherson,
B., Weissmann, M., and Yussouf, N.: Progress, challenges, and future steps in data assimilation for convection-permitting numerical weather prediction: Report on the virtual meeting held on 10 and 12 November 2021, Atmos. Sci. Lett., 24, e1130, <a href="https://doi.org/10.1002/asl.1130" target="_blank">https://doi.org/10.1002/asl.1130</a>, 2023.
</mixed-citation></ref-html>
<ref-html id="bib1.bib34"><label>Hunt et al.(2007)Hunt, Kostelich, and Szunyogh</label><mixed-citation>
Hunt, B. R., Kostelich, E. J., and Szunyogh, I.: Efficient data assimilation
for spatiotemporal chaos: A local ensemble transform Kalman filter, Physica
D, 230, 112–126, <a href="https://doi.org/10.1016/j.physd.2006.11.008" target="_blank">https://doi.org/10.1016/j.physd.2006.11.008</a>,
2007.
</mixed-citation></ref-html>
<ref-html id="bib1.bib35"><label>Kepert(2009)</label><mixed-citation>
Kepert, J. D.: Covariance localisation and balance in an Ensemble Kalman
Filter, Q. J. Roy. Meteor. Soc., 135,
1157–1176, <a href="https://doi.org/10.1002/qj.443" target="_blank">https://doi.org/10.1002/qj.443</a>, 2009.
</mixed-citation></ref-html>
<ref-html id="bib1.bib36"><label>Kirchgessner et al.(2014)Kirchgessner, Nerger, and
Bunse-Gerstner</label><mixed-citation>
Kirchgessner, P., Nerger, L., and Bunse-Gerstner, A.: On the Choice of an
Optimal Localization Radius in Ensemble Kalman Filter Methods, Mon. Weather Rev., 142, 2165–2175, <a href="https://doi.org/10.1175/MWR-D-13-00246.1" target="_blank">https://doi.org/10.1175/MWR-D-13-00246.1</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib37"><label>Kondo and Miyoshi(2016)</label><mixed-citation>
Kondo, K. and Miyoshi, T.: Impact of Removing Covariance Localization in an
Ensemble Kalman Filter: Experiments with 10 240 Members Using an Intermediate
AGCM, Mon. Weather Rev., 144, 4849–4865, <a href="https://doi.org/10.1175/MWR-D-15-0388.1" target="_blank">https://doi.org/10.1175/MWR-D-15-0388.1</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib38"><label>Kunii(2014)</label><mixed-citation>
Kunii, M.: The 1000-Member Ensemble Kalman Filtering with the JMA
Nonhydrostatic Mesoscale Model on the K Computer,
J. Meteorol. Soc. Jpn., 92, 623–633,
<a href="https://doi.org/10.2151/jmsj.2014-607" target="_blank">https://doi.org/10.2151/jmsj.2014-607</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib39"><label>Lei and Anderson(2014)</label><mixed-citation>
Lei, L. and Anderson, J. L.: Comparisons of Empirical Localization Techniques
for Serial Ensemble Kalman Filters in a Simple Atmospheric General
Circulation Model, Mon. Weather Rev., 142, 739–754,
<a href="https://doi.org/10.1175/MWR-D-13-00152.1" target="_blank">https://doi.org/10.1175/MWR-D-13-00152.1</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib40"><label>Lei et al.(2015)Lei, Anderson, and Romine</label><mixed-citation>
Lei, L., Anderson, J. L., and Romine, G. S.: Empirical Localization Functions
for Ensemble Kalman Filter Data Assimilation in Regions with and without
Precipitation, Mon. Weather Rev., 143, 3664–3679,
<a href="https://doi.org/10.1175/MWR-D-14-00415.1" target="_blank">https://doi.org/10.1175/MWR-D-14-00415.1</a>, 2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib41"><label>Lei et al.(2016)Lei, Anderson, and Whitaker</label><mixed-citation>
Lei, L., Anderson, J. L., and Whitaker, J. S.: Localizing the impact of
satellite radiance observations using a global group ensemble filter, J. Adv. Model. Earth Sy., 8, 719–734,
<a href="https://doi.org/10.1002/2016MS000627" target="_blank">https://doi.org/10.1002/2016MS000627</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib42"><label>Lei et al.(2018)Lei, Whitaker, and Bishop</label><mixed-citation>
Lei, L., Whitaker, J. S., and Bishop, C.: Improving Assimilation of Radiance
Observations by Implementing Model Space Localization in an Ensemble Kalman
Filter, J. Adv. Model. Earth Sy., 10, 3221–3232,
<a href="https://doi.org/10.1029/2018MS001468" target="_blank">https://doi.org/10.1029/2018MS001468</a>, 2018.
</mixed-citation></ref-html>
<ref-html id="bib1.bib43"><label>Lei et al.(2020)Lei, Whitaker, Anderson, and Tan</label><mixed-citation>
Lei, L., Whitaker, J. S., Anderson, J. L., and Tan, Z.: Adaptive Localization
for Satellite Radiance Observations in an Ensemble Kalman Filter, J.
Adv. Model. Earth Sy., 12, e2019MS001693,
<a href="https://doi.org/10.1029/2019MS001693" target="_blank">https://doi.org/10.1029/2019MS001693</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib44"><label>Lien et al.(2017)Lien, Miyoshi, Nishizawa, Yoshida, Yashiro, Adachi,
Yamaura, and Tomita</label><mixed-citation>
Lien, G.-Y., Miyoshi, T., Nishizawa, S., Yoshida, R., Yashiro, H., Adachi,
S. A., Yamaura, T., and Tomita, H.: The Near-Real-Time SCALE-LETKF System: A
Case of the September 2015 Kanto-Tohoku Heavy Rainfall, SOLA, 13, 1–6,
<a href="https://doi.org/10.2151/sola.2017-001" target="_blank">https://doi.org/10.2151/sola.2017-001</a>, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib45"><label>Michel et al.(2011)Michel, Auligné, and Montmerle</label><mixed-citation>
Michel, Y., Auligné, T., and Montmerle, T.: Heterogeneous Convective-Scale
Background Error Covariances with the Inclusion of Hydrometeor Variables,
Mon. Weather Rev., 139, 2994–3015, <a href="https://doi.org/10.1175/2011MWR3632.1" target="_blank">https://doi.org/10.1175/2011MWR3632.1</a>, 2011.
</mixed-citation></ref-html>
<ref-html id="bib1.bib46"><label>Miyoshi and Yamane(2007)</label><mixed-citation>
Miyoshi, T. and Yamane, S.: Local Ensemble Transform Kalman Filtering with an
AGCM at a T159/L48 Resolution, Mon. Weather Rev., 135, 3841–3861,
<a href="https://doi.org/10.1175/2007MWR1873.1" target="_blank">https://doi.org/10.1175/2007MWR1873.1</a>, 2007.
</mixed-citation></ref-html>
<ref-html id="bib1.bib47"><label>Miyoshi et al.(2014)Miyoshi, Kondo, and Imamura</label><mixed-citation>
Miyoshi, T., Kondo, K., and Imamura, T.: The 10,240-member ensemble Kalman
filtering with an intermediate AGCM, Geophys. Res. Lett., 41,
5264–5271, <a href="https://doi.org/10.1002/2014GL060863" target="_blank">https://doi.org/10.1002/2014GL060863</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib48"><label>Miyoshi et al.(2015)Miyoshi, Kondo, and Terasaki</label><mixed-citation>
Miyoshi, T., Kondo, K., and Terasaki, K.: Big Ensemble Data Assimilation in
Numerical Weather Prediction, Computer, 48, 15–21,
<a href="https://doi.org/10.1109/MC.2015.332" target="_blank">https://doi.org/10.1109/MC.2015.332</a>, 2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib49"><label>Miyoshi et al.(2016a)Miyoshi, Kunii, Ruiz, Lien, Satoh,
Ushio, Bessho, Seko, Tomita, and Ishikawa</label><mixed-citation>
Miyoshi, T., Kunii, M., Ruiz, J., Lien, G.-Y., Satoh, S., Ushio, T., Bessho,
K., Seko, H., Tomita, H., and Ishikawa, Y.: Big Data Assimilation
Revolutionizing Severe Weather Prediction, B. Am. Meteor. Soc., 97, 1347–1354, <a href="https://doi.org/10.1175/BAMS-D-15-00144.1" target="_blank">https://doi.org/10.1175/BAMS-D-15-00144.1</a>,
2016a.
</mixed-citation></ref-html>
<ref-html id="bib1.bib50"><label>Miyoshi et al.(2016b)Miyoshi, Lien, Satoh, Ushio,
Bessho, Tomita, Nishizawa, Yoshida, Adachi, Liao, Gerofi, Ishikawa, Kunii,
Ruiz, Maejima, Otsuka, Otsuka, Okamoto, and Seko</label><mixed-citation>
Miyoshi, T., Lien, G., Satoh, S., Ushio, T., Bessho, K., Tomita, H., Nishizawa,
S., Yoshida, R., Adachi, S. A., Liao, J., Gerofi, B., Ishikawa, Y., Kunii,
M., Ruiz, J., Maejima, Y., Otsuka, S., Otsuka, M., Okamoto, K., and Seko, H.:
“Big Data Assimilation” Toward Post-Petascale Severe Weather Prediction:
An Overview and Progress, P. IEEE, 104, 2155–2179,
<a href="https://doi.org/10.1109/JPROC.2016.2602560" target="_blank">https://doi.org/10.1109/JPROC.2016.2602560</a>, 2016b.
</mixed-citation></ref-html>
<ref-html id="bib1.bib51"><label>Ménétrier et al.(2014)Ménétrier, Montmerle, Berre, and
Michel</label><mixed-citation>
Ménétrier, B., Montmerle, T., Berre, L., and Michel, Y.: Estimation and
diagnosis of heterogeneous flow-dependent background-error covariances at the
convective scale using either large or small ensembles, Q. J.
Roy. Meteor. Soc., 140, 2050–2061, <a href="https://doi.org/10.1002/qj.2267" target="_blank">https://doi.org/10.1002/qj.2267</a>,
2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib52"><label>Ménétrier et al.(2015a)Ménétrier, Montmerle, Michel,
and Berre</label><mixed-citation>
Ménétrier, B., Montmerle, T., Michel, Y., and Berre, L.: Linear Filtering of
Sample Covariances for Ensemble-Based Data Assimilation. Part I: Optimality
Criteria and Application to Variance Filtering and Covariance Localization,
Mon. Weather Rev., 143, 1622–1643, <a href="https://doi.org/10.1175/MWR-D-14-00157.1" target="_blank">https://doi.org/10.1175/MWR-D-14-00157.1</a>,
2015a.
</mixed-citation></ref-html>
<ref-html id="bib1.bib53"><label>Ménétrier et al.(2015b)Ménétrier, Montmerle, Michel,
and Berre</label><mixed-citation>
Ménétrier, B., Montmerle, T., Michel, Y., and Berre, L.: Linear Filtering of
Sample Covariances for Ensemble-Based Data Assimilation. Part II: Application
to a Convective-Scale NWP Model, Mon. Weather Rev., 143, 1644–1664,
<a href="https://doi.org/10.1175/MWR-D-14-00156.1" target="_blank">https://doi.org/10.1175/MWR-D-14-00156.1</a>, 2015b.
</mixed-citation></ref-html>
<ref-html id="bib1.bib54"><label>Necker(2022)</label><mixed-citation>
Necker, T.: Empirical optimal localization/Vertical covariance
localization, Zenodo [data set, code], <a href="https://doi.org/10.5281/zenodo.7254119" target="_blank">https://doi.org/10.5281/zenodo.7254119</a>, 2022.
</mixed-citation></ref-html>
<ref-html id="bib1.bib55"><label>Necker et al.(2020a)Necker, Geiss, Weissmann, Ruiz,
Miyoshi, and Lien</label><mixed-citation>
Necker, T., Geiss, S., Weissmann, M., Ruiz, J., Miyoshi, T., and Lien, G.-Y.: A
Convective-scale 1,000-member Ensemble Simulation and Potential Applications,
Q. J. Roy. Meteor. Soc., 146, 1423–1442, <a href="https://doi.org/10.1002/qj.3744" target="_blank">https://doi.org/10.1002/qj.3744</a>,
2020a.
</mixed-citation></ref-html>
<ref-html id="bib1.bib56"><label>Necker et al.(2020b)Necker, Weissmann, Ruckstuhl,
Anderson, and Miyoshi</label><mixed-citation>
Necker, T., Weissmann, M., Ruckstuhl, Y., Anderson, J., and Miyoshi, T.:
Sampling Error Correction Evaluated Using a Convective-Scale
1000-Member Ensemble, Mon. Weather Rev., 148, 1229–1249,
<a href="https://doi.org/10.1175/MWR-D-19-0154.1" target="_blank">https://doi.org/10.1175/MWR-D-19-0154.1</a>, 2020b.
</mixed-citation></ref-html>
<ref-html id="bib1.bib57"><label>Nomokonova et al.(2022)Nomokonova, Griewank, Löhnert, Miyoshi,
Necker, and Weissmann</label><mixed-citation>
Nomokonova, T., Griewank, P. J., Löhnert, U., Miyoshi, T., Necker, T., and
Weissmann, M.: Estimating the benefit of Doppler wind lidars for short-term low-level wind ensemble forecasts, Q. J. Roy. Meteor. Soc., 1–19, <a href="https://doi.org/10.1002/qj.4402" target="_blank">https://doi.org/10.1002/qj.4402</a>, 2022.
</mixed-citation></ref-html>
<ref-html id="bib1.bib58"><label>Perianez et al.(2014)Perianez, Reich, and Potthast</label><mixed-citation>
Perianez, A., Reich, H., and Potthast, R.: Optimal Localization for Ensemble
Kalman Filter Systems, J. Meteor. Soc. Jpn., 92, 585–597, <a href="https://doi.org/10.2151/jmsj.2014-605" target="_blank">https://doi.org/10.2151/jmsj.2014-605</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib59"><label>Piper et al.(2016)Piper, Kunz, Ehmele, Mohr, Mühr, Kron, and
Daniell</label><mixed-citation>
Piper, D., Kunz, M., Ehmele, F., Mohr, S., Mühr, B., Kron, A., and Daniell, J.: Exceptional sequence of severe thunderstorms and related flash floods in May and June 2016 in Germany – Part 1: Meteorological background, Nat. Hazards Earth Syst. Sci., 16, 2835–2850, <a href="https://doi.org/10.5194/nhess-16-2835-2016" target="_blank">https://doi.org/10.5194/nhess-16-2835-2016</a>, 2016.

</mixed-citation></ref-html>
<ref-html id="bib1.bib60"><label>Poterjoy et al.(2014)</label><mixed-citation>
Poterjoy, J., Zhang, F., and Weng, Y.: The Effects of Sampling Errors on the
EnKF Assimilation of Inner-Core Hurricane Observations, Mon. Weather Rev., 142, 1609–1630, <a href="https://doi.org/10.1175/MWR-D-13-00305.1" target="_blank">https://doi.org/10.1175/MWR-D-13-00305.1</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib61"><label>Scheck et al.(2020)Scheck, Weissmann, and Bach</label><mixed-citation>
Scheck, L., Weissmann, M., and Bach, L.: Assimilating visible satellite images
for convective-scale numerical weather prediction: A case-study, Q. J. Roy. Meteor. Soc., 146, 3165–3186,
<a href="https://doi.org/10.1002/qj.3840" target="_blank">https://doi.org/10.1002/qj.3840</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib62"><label>Schraff et al.(2016)Schraff, Reich, Rhodin, Schomburg, Stephan,
Periáñez, and Potthast</label><mixed-citation>
Schraff, C., Reich, H., Rhodin, A., Schomburg, A., Stephan, K., Periáñez, A.,
and Potthast, R.: Kilometre-scale ensemble data assimilation for the COSMO
model (KENDA), Q. J. Roy. Meteor. Soc., 142,
1453–1472, <a href="https://doi.org/10.1002/qj.2748" target="_blank">https://doi.org/10.1002/qj.2748</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib63"><label>Stanley et al.(2021)Stanley, Grooms, and Kleiber</label><mixed-citation>
Stanley, Z., Grooms, I., and Kleiber, W.: Multivariate localization functions for strongly coupled data assimilation in the bivariate Lorenz 96 system, Nonlin. Processes Geophys., 28, 565–583, <a href="https://doi.org/10.5194/npg-28-565-2021" target="_blank">https://doi.org/10.5194/npg-28-565-2021</a>, 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib64"><label>Tabeart et al.(2019)Tabeart, Dance, Lawless, Nichols, and
Waller</label><mixed-citation>
Tabeart, J. M., Dance, S. L., Lawless, A. S., Nichols, N. K., and Waller,
J. A.: Improving the condition number of estimated covariance matrices,
Tellus, 72, 1–19, <a href="https://doi.org/10.1080/16000870.2019.1696646" target="_blank">https://doi.org/10.1080/16000870.2019.1696646</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib65"><label>Virtanen et al.(2020)Virtanen, Gommers, Oliphant, Haberland, Reddy,
Cournapeau, Burovski, Peterson, Weckesser, Bright, van der Walt, Brett,
Wilson, Millman, Mayorov, Nelson, Jones, Kern, Larson, Carey, Polat, Feng,
Moore, VanderPlas, Laxalde, Perktold, Cimrman, Henriksen, Quintero, Harris,
Archibald, Ribeiro, Pedregosa, van Mulbregt, and SciPy 1.0
Contributors</label><mixed-citation>
Virtanen, P., Gommers, R., Oliphant, T. E., Haberland, M., Reddy, T.,
Cournapeau, D., Burovski, E., Peterson, P., Weckesser, W., Bright, J., van
der Walt, S. J., Brett, M., Wilson, J., Millman, K. J., Mayorov, N., Nelson,
A. R. J., Jones, E., Kern, R., Larson, E., Carey, C. J., Polat, I., Feng, Y.,
Moore, E. W., VanderPlas, J., Laxalde, D., Perktold, J., Cimrman, R.,
Henriksen, I., Quintero, E. A., Harris, C. R., Archibald, A. M., Ribeiro,
A. H., Pedregosa, F., van Mulbregt, P., and SciPy 1.0 Contributors:
SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python,
Nat. Methods, 17, 261–272, <a href="https://doi.org/10.1038/s41592-019-0686-2" target="_blank">https://doi.org/10.1038/s41592-019-0686-2</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib66"><label>Whitaker and Hamill(2012)</label><mixed-citation>
Whitaker, J. S. and Hamill, T. M.: Evaluating Methods to Account for System
Errors in Ensemble Data Assimilation, Mon. Weather Rev., 140, 3078–3089, <a href="https://doi.org/10.1175/MWR-D-11-00276.1" target="_blank">https://doi.org/10.1175/MWR-D-11-00276.1</a>, 2012.
</mixed-citation></ref-html>
<ref-html id="bib1.bib67"><label>Wu et al.(2020)Wu, Yang, Tsai, and Cheng</label><mixed-citation>
Wu, P.-Y., Yang, S.-C., Tsai, C.-C., and Cheng, H.-W.: Convective-Scale
Sampling Error and Its Impact on the Ensemble Radar Data Assimilation System:
A Case Study of a Heavy Rainfall Event on 16 June 2008 in Taiwan, Mon. Weather Rev., 148, 3631–3652, <a href="https://doi.org/10.1175/MWR-D-19-0319.1" target="_blank">https://doi.org/10.1175/MWR-D-19-0319.1</a>, 2020.
</mixed-citation></ref-html>--></article>
