Articles | Volume 33, issue 1
https://doi.org/10.5194/npg-33-33-2026
https://doi.org/10.5194/npg-33-33-2026
Research article
 | 
26 Jan 2026
Research article |  | 26 Jan 2026

Localization in the mapping particle filter

Juan M. Guerrieri, Manuel Pulido, Takemasa Miyoshi, Arata Amemiya, and Juan J. Ruiz
Abstract

Data assimilation involves sequential inference in geophysical systems with nonlinear dynamics and observational operators. Non-parametric filters are a promising approach for data assimilation because they are able to represent non-Gaussian densities. The mapping particle filter is an iterative ensemble method that incorporates the Stein Variational Gradient Descent (SVGD) to produce a particle flow transforming state vectors from prior to posterior densities. At every pseudo-time step, the Kullback-Leibler divergence between the intermediate density and the target posterior is evaluated and minimized. However, for applications in geophysical systems, challenges persist in high dimensions, where sample covariance underestimation leads to filter divergence. This work proposes two localization methods, one in which a local kernel function is defined and the particle flow is global. The second method, given a localization radius, physically partitions the state vector and performs local mappings at each grid point. The performance of the proposed Local Mapping Particle Filters (LMPFs) is assessed in synthetic experiments. Observations are produced with a two-scale Lorenz system, while a one-scale Lorenz model is used as surrogate, introducing model error in the inference. The methods are evaluated with both full and partial observations, as well as with different linear and non-linear observational operators. The LMPFs with Gaussian mixtures in the prior density perform similarly to Gaussian filters such as the Ensemble Transform Kalman Filter (ETKF) and the Local Ensemble Transform Kalman Filter (LETKF) in most cases, and in some scenarios, they provide competitive performance in terms of analysis accuracy.

Share
1 Introduction

Particle filters have emerged as a valuable approach for addressing non-linear data assimilation challenges, especially in the context of geophysical systems, with particular promise for improving short-term meteorological forecasting. This potential derives from the inherently non-Gaussian nature of convective instabilities – which dominate short-term weather patterns – and their rapid growth rates compared to synoptic-scale phenomena (Hohenegger and Schar2007). As model resolution increases and observation operators become more complex, including non-linear relationships with the model state, the challenge of accurately representing these growing non-linear and non-Gaussian features becomes more pronounced. Gaussian data assimilation techniques, such as Kalman filter-based methods, encounter limitations when confronted with non-linearity. These methods assume a Gaussian prior probability density function for the state. Variational methods struggle under strong non-Gaussianity resulting in multimodal cost functions or when the observational errors deviate from being Gaussian as well. Ensemble Kalman filters (EnKFs) explicitly assume that the prior density function and the observation likelihood follow a Gaussian distribution. Notably, Ruiz et al. (2021) show that even when drastically reducing the assimilation window of the Local Ensemble Transform Kalman Filter (LETKF), first introduced by Hunt et al. (2007), from 5 min to 30 s in 1 km-resolution experiments, residual non-Gaussianity persists at 40 % levels.

In contrast, particle filters are non-parametric and offer distinct advantages in handling non-Gaussian error statistics (van Leeuwen et al.2019). However, particle filters face challenges when dealing with high dimensionality, which is particularly prominent in geophysical applications characterized by a large number of variables. The standard Sequential Importance Resampling filter (SIR, Doucet et al.2001) preserves and statistically replicates only the particles near observations, leading to sample impoverishment and weight degeneracy. To address this issue, a proposal density can incorporate information from both model dynamics and current observations, guiding particles toward high-probability regions and improving particle diversity by updating weights based on the ratio between the proposal density and the actual posterior density (van Leeuwen et al.2019).

The problem of high dimensionality has also led to the development of several methods. Localization was first introduced for particle filters independently in Bengtsson et al. (2003) and van Leeuwen (2003). Other implementations can be found in Poterjoy (2016); Robert and Künsch (2017). Further methods are based on tempering (Neal1996), which mitigate the computational burden, instability and inaccuracy associated with high-dimensional problems, and jittering (Cotter el al.2020), also referred to as regularisation, used to rejuvenate particles before or after resampling, as well as after tempering steps. An alternative approach to overcome these limitations is provided by particle flow filters (PFFs, Daum et al.2010; van Leeuwen et al.2019). Instead of relying on the two-step process of weighting and resampling, PFFs move particles continuously through state space via a differential equation over a pseudo-time – drawing on ideas from Markov chain Monte Carlo (MCMC) methods such as those in Gallego and Ríos Insua (2018) – transforming the prior into the posterior distribution without modifying particle weights and thus avoiding the resampling and jittering steps required in traditional particle filters.

This work is concerned with developing a localization scheme for the variational mapping particle filter (MPF) proposed by Pulido and van Leeuwen (2019). The MPF is a particle flow filter that holds potential for non-linear applications in meteorology and oceanography. It is a sequential Monte Carlo algorithm that uses the Stein Variational Gradient Descent (SVGD) method, proposed by Liu and Wang (2016). In the MPF, state vectors, also known as particles, are propagated from the state predicted by the model (referred to as the background or forecast state) to states whose probability density function matches the posterior density, through a series of mappings. These gradient descent mappings aim to minimize the Kullback-Leibler divergence between the posterior density, which is obtained by applying Bayes' formula, and the sequence of intermediate densities.

The SVGD is a deterministic inference algorithm that converges in the limit of many particles (Del Moral2013), but it still faces the commonly referred problem known as “the curse of dimensionality”, for representing densities in high-dimensional spaces. This is a common problem in particle filters (Snyder et al.2008). One of its manifestations is the underestimation of the sample covariance and the subsequent divergence of the filter. Zhuo et al. (2018) demonstrated that SVGD often collapses into the modes of the target distribution, and this drawback becomes more severe with higher dimensions. Additionally, Ba et al. (2022) have demonstrated that SVGD-based algorithms offer few convergence guarantees. This issue persists even when the number of particles (or ensemble members) is larger than the dimension of the state. Among these limited cases, convergence is achievable in the mean-field regime, which occurs when the number of particles tends to infinity. To improve its convergence properties, Ba et al. (2022) proposed alternative formulations of SVGD. Furthermore, the SVGD produces biased samples for finite ensemble size and only produces unbiased samples for infinite ensemble size. Gallego and Ríos Insua (2018) solved the bias issue by introducing stochastic noise that makes the method unbiased for any ensemble size. Leviyev et al. (2022) provided an efficient methodology to incorporate the stochastic noise, and also accelerated the convergence through a Newton method that incorporates information from the Hessian. Finally, Ma et al. (2015) offered a complete and general framework for designing samplers based on the SVGD that guarantee a correct stationary distribution and facilitate the exploration of the space.

In the field of geophysical modeling, ensemble-based methods and particle filters are recognized as key frameworks for data assimilation. Both can incorporate localization methods to enhance their performance. Localization is a well-founded assumption considering that the state-dependent correlation between physical variables decreases with the distance between them. In the context of these frameworks, localization techniques serve the purpose of reducing the dimensionality of the assimilation process, ensuring accurate integration of observed data into the model state. For the EnKF, localization is typically achieved by adjusting the influence of observations and the prior error covariances based on their spatial proximity to the estimation point (Houtekamer and Mitchell2001; Hamill and Whitaker2001; Whitaker and Hamill2002). Developing and implementing these localization techniques within the EnKF and particle filters are critical for optimizing their effectiveness in real-world scenarios with spatio-temporal dynamics.

In particle filters, localization can be implemented in many ways (Farchi and Bocquet2018), including resampling-based approaches. For instance, Penny and Miyoshi (2016) proposed a local particle filter (LPF) that uses observation-space localization to compute independent analyses at each grid point. By applying deterministic resampling and smoothing the analysis weights across neighboring points, the LPF effectively mitigates particle degeneracy and enhances performance in highly non-linear and non-Gaussian scenarios. While that work demonstrates the advantages of resampling-based localization, alternative particle flow-based methods avoid resampling and apply continuous transformations to particles from the prior density to the posterior.

Hu and van Leeuwen (2021) addressed the application of the Particle Flow Filter (algorithm based on what we call MPF) in high-dimensional systems, evaluating its performance in a Lorenz-96 system with 1000 variables and 20 particles, observing 25 % of the state variables and using three different observation operators. To avoid the problem of marginal distribution collapse in sparsely observed, high-dimensional settings, they proposed the use of a matrix-valued kernel, noting that a scalar kernel failed in these scenarios. They implemented a preconditioning matrix within the particle flow formulation to accelerate convergence. This matrix was chosen as the localized prior covariance matrix, which was localized using a Schur product with a distance-decaying matrix. This approach resulted in the cancellation of the prior covariance matrix in the particle flow expression. Its performance was comparable to the LETKF and did not require explicit covariance inflation. This method was also applied to a full atmospheric model in Hu et al. (2024).

In this work, two localization schemes in the MPF are introduced to reduce dimensionality and mitigate the problem of the curse of dimensionality in the MPF. These schemes are evaluated in the two-scale Lorenz model using both total and partial observations and nonlinear observation operators.

The work is structured as follows: Sect. 2 introduces two LMPFs methodologies, Sect. 3 describes the experimental design, Sect. 4 presents the results of the experiments and Sect. 5 draws the conclusions of the work.

2 Methodology

2.1 Mapping particle filter's review

The Mapping Particle Filter (MPF), introduced by Pulido and van Leeuwen (2019), is a non-parametric deterministic data assimilation method based on sample points, i.e., particles. It involves the transformation of the sample states from a prior density function to a posterior density by passing through intermediate states. These intermediate states are driven by an interacting particle flow designed to minimize the Kullback-Leibler divergence between a kernelized distribution of the sample states and the target posterior distribution. Subrahmanya et al. (2025) presents a general formulation to minimize the KL divergence. It formulates the flow field using the Fokker-Planck equation to evolve particles and sample the posterior distribution without using a reproducing kernel Hilbert space.

In the MPF, based on a hidden Markov model, a state vector evolves over time using a dynamical model and is observed using an observational model simultaneously,

(1)xk=M(xk-1,ηk),(2)yk=H(xk,νk),

where xkRNx represents the state at time k, is the model operator, ηk denotes the random model error, ykRNy are the observations, is the observation operator, and νk represents the observational error. Here, a general framework is presented in which both model and observational errors can be non-additive.

The target density function of the particle flow corresponds to the posterior probability density using Bayes' formula during the assimilation stage,

(3) p ( x k | y 1 : k ) = p ( y k | x k ) p ( x k | y 1 : k - 1 ) p ( y k | y 1 : k - 1 ) ,

This probability density function delineates the analysis states by capturing the likelihood of the forecast given a particular set of observations and a specified prior density.

Consider a set of Np particles {xk-1(1:Np)} that samples the posterior density at time k−1. To obtain a state that matches the posterior density at time k, the MPF iteratively computes intermediate states from the prior to the target. The particles that sample the prior density at time k are states that undergo dynamical evolution from the particles that sample the posterior density at time k−1, denoted as {xkf(j)=xk,0(j)=M(xk-1a(j),ηk(j))}j=1Np, where the second subscript represents the pseudo-time of the mapping iteration. The superscript f(j) indicates the jth particle of the forecasted states, and a(j) indicates the jth particle of the analysis states (previous estimates). At each iteration, the particles are transformed by

(4) x k , i ( j ) = T ( x k , i - 1 ( j ) ) = x k , i - 1 ( j ) + ϵ v ( x k , i - 1 ( j ) ) ,

where T represents the iteration mapping, v represents the velocity of the particle flow in pseudo-time, ϵ represents the step size of the mapping. It may be considered fixed or estimated adaptively by means of stochastic optimization algorithms (Kingma and Ba2014).

The velocity seeks to minimize the Kullback-Leibler divergence between the target posterior density function, and the density of the intermediate states. Therefore, the sample from the prior density is transformed towards a sample from the posterior density through a set of discrete transformations, which in the infinitesimal limit may be interpreted as a flow in the state space.

MPF is inspired by the Stein Variational Gradient Descent method (Liu and Wang2016) which is kernel-based. These methods are algorithms that rely on kernel functions to measure similarities between state vectors from different particles. The MPF selects a space of functions known as the unit ball of a reproducing kernel Hilbert space (RKHS), denoted as 𝔽. The optimization task is to find v∈𝔽 that indicates the steepest descent direction of the Kullback-Leibler Divergence DKL between the target posterior density and the intermediate density.

By choosing an isotropic kernel K and given a set of particles {xk,i-1(1:Np)} representing a sample of the intermediate density at pseudotime i−1, the gradient of the Monte Carlo integration of the KL divergence is computed as:

(5) v ( x ) = 1 N p j = 1 N p K ( x k , i - 1 ( j ) , x ) x k , i - 1 ( j ) log p ( x k , i - 1 ( j ) ) + x k , i - 1 ( j ) log K ( x k , i - 1 ( j ) , x ) .

The first term in the parenthesis of Eq. (5), called the kernel-smoothed gradient of the posterior density, acts as a central force guiding the samples from an initial distribution density function towards the modes of the posterior density. The second term acts as the repulsive force and prevents the particles from collapsing into modes of the posterior. Note that the variables in Eqs. (4) and (5) are nondimensionalized by proper scaling as in previous works (Pulido and van Leeuwen2019; Liu and Wang2016; Lu et al.2019). To consider dimensional variables one may rewrite ϵ=Dδt, where D is a diffusion coefficient and δt is the pseudo-time step. In that case, the diffusion coefficient controls the optimization convergence rate as in gradient flows (Jordan et al.1998) and must be incorporated into the velocity term. In this work, however, we keep ϵ as the single effective parameter controlling the convergence rate and adapt it during gradient descent using low-order momentum estimates (Kingma and Ba2014).

Radial basis functions are used as kernels in this work,

(6) K ( x , x ) = e - 1 2 x - x Σ 2 ,

where x-xΣ2=(x-x)Σ-1(x-x) denotes the square of the Mahalanobis distance and Σ is referred to as the kernel covariance matrix. This matrix needs to be specified at the beginning of the process. In this work, it is assumed to be proportional to the forecast covariance matrix, though other approaches for defining it are possible.

The gradient of the logarithm of the posterior density requires the analytical forms of the prior density and the likelihood function. In this work, observational errors are assumed to be additive and Gaussian, but the framework is general and other observational error distributions may be considered. The resulting gradient of the log posterior density, evaluated at xk,i-1(j) is:

(7) x k , i - 1 ( j ) log p x k , i - 1 ( j ) = H R k - 1 y k - H x k , i - 1 ( j ) + x k , i - 1 ( j ) log p ( x k , i - 1 ( j ) | y 1 : k - 1 ) .

The first term is the observation likelihood function, in which is the observational operator, (RNy×Nx), and H denotes the tangent linear observation operator, H=dHdx(x), while R stands for the observational error covariance matrix (RNy×Ny). The second term in Eq. (7) is the gradient of the logarithm of the prior density.

In the case of a Gaussian prior density, where

(8) p ( x k ( j ) | y 1 : k - 1 ) = Z e - 1 2 x k , i - 1 ( j ) - x k , 0 B k 2 ,

the second term is reduced to

(9) x k , i - 1 ( j ) log p ( x k ( j ) | y 1 : k - 1 ) = - B k - 1 x k , i - 1 ( j ) - x k , 0 ,

where Z is the normalizing constant, Bk is the prior or background covariance matrix and xk,0 is the prior mean. For sequential Monte Carlo, the prior density in Eq. (9) is given by the forecast density, such that Bk=P^kf.

Alternatively, if we assume the prior density is a Gaussian mixture based on the forecast particles, the prior is

(10) p ( x k ( j ) | y 1 : k - 1 ) = Z exp - 1 2 x k , i - 1 ( j ) - μ k , i , j Q k 2 ,

where μk,i,j=m=1NpψkijmM(xk-1(m))m=1Npψkijm are the Gaussian centroids, ψk,i,j,m=exp-12xk,i-1(j)-M(xk-1(m))Qk2 that represents the adaptive weights, and Qk represents the covariance matrix of the Gaussian mixture. The second term in Eq. (7) results in

(11) x k , i - 1 ( j ) log p ( x k ( j ) | y 1 : k - 1 ) = - Q k - 1 x k , i - 1 ( j ) - m = 1 N p ψ k i j m M ( x k - 1 ( m ) ) m = 1 N p ψ k i j m ,

If the model is stochastic with additive Gaussian errors, Eq. (11) is exact (Pulido and van Leeuwen2019).

To illustrate the practical implementation of this approach, we evaluate Eq. (5) under the assumption of a Gaussian prior distribution with a radial basis function kernel,

(12) v ( x ) = 1 N p j = 1 N p { e - 1 2 x k , i - 1 ( j ) - x Σ 2 H R k - 1 y k - H x k , i - 1 ( j ) - B k - 1 x k , i - 1 ( j ) - x k , 0 - Σ k - 1 x k , i - 1 ( j ) - x } .

The computational cost of a single pseudo-time iteration in the Gaussian-mixture prior case is

(13) O N x 2 N p 2 + N y 2 N p + N x N y N p ,

where the first term corresponds to the kernel and its gradient calculation, while the second and third terms correspond to the computational cost of the likelihood. Assuming that the matrix inversion can be performed in 𝒪inv(Nx), the overall computational cost becomes

(14) O O inv ( N x ) + N it N x 2 N p 2 + N y 2 N p + N x N y N p ,

where Nit denotes the number of pseudo-time iterations.

2.2 Localization methods

The underlying assumption in the two developed localization methods is that on average, error correlations decay with the physical distance so that when the distance between two variables is larger than a given threshold known as the localization radius, the correlation is assumed to be negligible. The correlations of these far points are neglected so that it becomes feasible to produce an inference using only the points of the background state and the observations within the localization radius. This reduction in algorithmic complexity allows to reduce sampling noise and enhance the quality of the analysis for high-dimensional state spaces.

Both Hu and van Leeuwen (2021) and the present work address the challenge of applying MPF in high-dimensional systems, but through different approaches. Hu and van Leeuwen focus on an intrinsic modification of the particle interaction mechanism by transitioning from a scalar kernel to a matrix kernel; within the matrix kernel, they assume the distance between particles is independent for each component of the state vector. In addition to the kernel modification, their work also applies localization to the prior covariance matrix through a Schur product with a distance-decaying correlation matrix. In contrast, this work starts from a localization assumption, and applies it coherently to both the posterior distribution and the kernel. This results in explicit localization schemes that restructure how the optimization process is applied in state space, thereby modifying the sequencing of the optimization process.

The α-localization algorithm assumes that the kernels are localized around each variable, so that distances between particles are measured in a low dimensional space. Furthermore, it uses a localized prior covariance matrix, for instance, by keeping blocks of the global sample covariance via the Schur product. The state updates are determined globally.

On the other hand, the β-localization algorithm assumes the full local variational mapping process is localized around each variable. In terms of the localization assumption, this method is similar to the localization in Hunt et al. (2007) and has also some resemblance to the methodology implemented in Hu et al. (2024). For each variable, the optimization is conducted separately considering the observations and the prior state variables within the localization radius.

2.2.1α-Localization

Given a variable xl of the state x, we consider a neighborhood 𝒞l of xl and denote the variables within this neighborhood as x̃l={xl;lCl}. We assume that the variables located outside of 𝒞l are statistically independent of xl,

(15) p ( x l | x ) = p ( x l | x ̃ l ) .

For simplicity, we assume a single physical type of variable in the state space. In this approach, the local state x̃k,i,l(j) is defined with four indices: k (time index), i (pseudo-time iteration index), l (space index) and j (particle index). To avoid overclutter, time and iteration indices are omitted. In a one dimensional space for a localization radius , the vector of neighbor variables is x̃l={xl;l-ll+} with dimension Nx̃=2+1.

The global update of variable xl from Eq. (5) is

(16) v l ( x ) = 1 N p j = 1 N p K ( x ( j ) , x ) x l ( j ) log p ( x ( j ) ) + x l ( j ) log K ( x ( j ) , x ) .

where xl(j) is the partial derivative with respect to the l variable. Since the variables beyond the localization radius are assumed to be statistically independent, we approximate Eq. (16) by considering a local kernel following Wang et al. (2018) in which only the variables within the localization radius around l are considered. This local kernel is denoted as Kl(x̃l(j),x̃l).

The local kernel is specific to each grid point xl. It calculates the Mahalanobis distance between particles using a state vector that is centered at xl and includes only the neighboring points that fall within the defined localization radius.

The local kernel is defined with a radial basis function as the global one, but with a kernel covariance matrix defined as the Schur (element-wise) product of a localization matrix and the global covariance matrix, Λl=ΓlΣ, where the localization matrix Γl could be a block matrix around l with one's and zeros, as in Eq. (17), or some decaying coefficient with the distance of the rest of the points to the lth grid point (e.g. Gaspari and Cohn1999 factor). The neighborhood variables x̃l are the ones where Γl is not null.

(17) ( Γ l ) m n = 1  if l - m , n l + 0  otherwise 

The resulting local kernel is:

(18) K l ( x ̃ l , x ̃ l ) = e - 1 2 x ̃ l - x ̃ l Λ l 2

The crucial feature of the local kernel is that the Mahalanobis distance calculation only takes into account low-dimensional states. The local flow in the l variable is therefore approximated by

(19) v l ( x ̃ l ) 1 N p j = 1 N p K l ( x ̃ l ( j ) , x ̃ l ) x l ( j ) log p ( x ̃ l ( j ) ) + x l ( j ) log K l ( x ̃ l ( j ) , x ̃ l )

The gradient of the posterior density will be calculated following Eq. (7) with the following modifications. For the gradient of the likelihood term, it is calculated globally using the first term of Eq. (7), resulting in a matrix in RNx×Np. The term used in Eq. (19) corresponds to the lth row of that global matrix. For the prior density term, we calculate it according to Eq. (9) or Eq. (11) (depending on our hypothesis), but we use the localized vector x̃l and apply the localized covariance matrix ΓlBk or ΓlQk. This approach is the same as in the local kernel calculation in Eq. (18).

We could also derive the local velocity, Eq. (19), under the assumption that state-space covariance matrices are l-block diagonal while keeping the state vectors global. In preliminary experiments, we evaluated a hybrid methodology in which the kernel was local, with a block Λl matrix, but the prior density was global with l-banded background covariance matrices. However the performance of this hybrid methodology was suboptimal.

Once the complete velocity vector is reconstructed with each component computed separately, the global states in the next pseudo-time are determined by Eq. (4). Therefore, covariance inversion and the mappings are global. In Algorithm 1 below, a pseudocode of the α-localization algorithm is presented. In this case, the time index is omitted, while the pseudo-time, space, and particle indices are retained. The computational complexity of the LMPF-α using a Gaussian-mixture prior is OOinv(Nx,)+NitNx(2+1)2Np2+Ny2Np+NxNyNp, where 𝒪inv(Nx,ℓ) is the cost of inverting an -banded matrix.

2.2.2β-Localization

The β-localization involves a physical partitioning of the state space centred around each variable based on distance between variables. Subsequently, it leverages the same principles of the global MPF to each partition.

This methodology is based on Zhuo et al. (2018) in which the KL divergence is decomposed as,

(20) D KL q | | p = D KL q ( x l | x ¬ l ) q ( x ¬ l ) | | p ( x l | x ̃ l ) p ( x ¬ l ) + D KL q ( x ¬ l ) | | p ( x ¬ l )

where x¬l is composed by all the state variables except xl. Therefore, we can solve a local minimization problem for xl to find q(xl|x̃l) and by keeping fixed the rest, q(x¬l).

This approach guarantees that the analysis is performed independently for each state variable, with no dependency on intermediate updates of other grid points. However, the neighborhood variables are considered to define the map for each state variable. This means that while the local analysis at a given grid point depends on nearby observations, the convergence at each point remains independent. This reminds the application of normalizing flows with transformations in each direction (Tabak and Turner2012). These local minimizations are iterated along l.

The β-localization algorithm consists of applying the global MPF to the neighborhood of xl. For a given localization radius , we use the neighborhood vector as in the α-localization: x̃l. The local velocity is defined as the global velocity in Eq. (5), but calculated only over the localized state vector x̃l. Therefore, it considers a kernel as in Eq. (6) calculated in the physically partitioned state. A localized posterior density is also used, in which only the forecast states in the local domain xl̃f(j)Nxl̃ are considered. Observations within the localization radius are selected. This localization algorithm can only be applied for observations that have a well-defined location in physical space. For that purpose we define l as the set of observation indices corresponding to the observations that are relevant to the localized state vector x̃l. Specifically, for a one-dimensional domain this is

(21) I l = { m the position of observation  y m lies within the interval  [ l - , l + ] } .

The localized observation vector ỹl is then defined as the subset of observations whose indices belong to l:

(22) y ̃ l = { y m m I l } .

The blocks of H̃lNyl̃×Nxl̃ and RlNyl̃×Nyl̃ are also coherently selected,

(23)H̃l=(H̃)mn with mIl,n=l-,,l+,(24)Rl=(R)mn with m,nIl.

Thus, we are using observations from a subspace and their associated error covariance related to that subset of observations. Additionally, for large localization radii where distant spurious covariances might still occur, a length-decaying factor could be useful.

For each grid point in the domain, i.e. state variable, the following iterative transformation is applied:

(25) x ̃ i , l ( j ) = x ̃ i - 1 , l ( j ) + ϵ v l ( x ̃ i - 1 , l ( j ) )

The convergence is independent for each grid point. To obtain the global analysis vector, only the element at position l from this local analysis vector is kept. This process is repeated for every spatial point on the grid, i.e. variable of the state vector. When updating a given grid point in the β approach, the values of all other grid points are taken from the original prior state of the particle, not from previously updated points in the same cycle. This avoids any dependence on the order in which the domain is processed.

The order of pseudo-time iterations and localized step iterations is reversed between the two methodologies. In the α approach, for each pseudo-time step, the entire domain is updated, resulting in a global state for each pseudo-time step. In contrast, in the β approach, for each local point in the domain, all pseudo-time steps are iterated independently before moving to the next state variable, leading to localized convergence without a global state. The exchange of iterations is easier to observe by looking at the algorithms of LMPF-α in Algorithm 1 and LMPF-β in Algorithm 2.

Algorithm 1LMPF-α: Global update.

Compute global Σ
# Number of pseudo time step iterations is denoted as Nit
for  i=1 to Nit do
for l=1 to Nx do
x̃i-1,l(j)xi-1,m(j) , with m=l-,,l+ and j=1,,Np
Compute Λl-1 and localized log posterior
Compute vl(j)x̃i-1,l(1:Np) as in Eq. (19)
xi,l(j)xi-1,l(j)+ϵvl(j)x̃i-1,l(1:Np)
end for
# Global state updated at pseudo-time step i
end for

Algorithm 2LMPF-β: Local update.

for l=1 to Nx do
x̃0,l(j)x0,m(j), with m=l-,,l+ and j=1,,Np
for  i=1 to Nit do
Compute Σ−1 in the local set {x̃i-1,l(1:Np)}
Compute vl(j)x̃i-1,l(1:Np)
x̃i,l(j)x̃i-1,l(j)+ϵvl(j)x̃i-1,l(1:Np)
end for
Retain center value from x̃Nit,l(j)
# Local convergence at point l
end for

The computational cost of the LMPF-β is ONxOinv(2+1)+NxNit(2+1)2Np2+Ny2Np +(2+1)NyNp, where Ny denotes the maximum number of observations within any localized domain. This is equivalent to O(Nx)OMPF(2+1,Ny,Np), which represents the computational complexity of the global MPF algorithm.

3 Numerical setup

The global MPF and the two variants of the localized MPF are assessed in experiments with synthetic observations. In these experiments, observations are generated based on a known dynamical model. The true state is the solution of the known model, referred to here as the nature model. In contrast, the forecast model is a surrogate for the nature model, so we consider the assimilation experiments in the presence of model error. The surrogate model is used to produce forecasts within the assimilation system, xkf(j)=Msu(xk-1a(j)). After evolving the previous analysis ensemble states, xk-1a(j) with this surrogate model, the data assimilation step is conducted, and so on. This approach allows us to examine the assimilation scheme with a known true state in the presence of model errors. In these proof-of-concept experiments, the nature model is the two-scale Lorenz system (Sect. 3.1), while the surrogate model is the one-scale Lorenz (Sect. 3.2) so that the source of model errors is the lack of the explicit representation of small-scale dynamics. Both models are deterministic, with no explicit additive stochastic error terms.

3.1 Description of the true model

The nature model is defined by the two-scale Lorenz system equations (Lorenz2005):

(26) X ˙ n = - X n - 1 ( X n - 2 - X n + 1 ) - X n + F - h c b j = J ( n - 1 ) + 1 n J Y j n = 1 , , N LS Y ˙ m = - c b Y m + 1 ( Y m + 2 - Y m - 1 ) - c Y m + h c b X int [ ( m - 1 ) / J ] + 1 m = 1 , , N SS

within a cyclic domain, i.e, XNLS+1=X1,X0=XNLS, and X-1=XNLS-1; YNSS+1=Y1, YNSS+2=Y2 and Y0=YNSS. NLS is the number of large-scale (LS) variables, and NSS the number of the small-scale (SS) variables. The equations are solved using a fourth-order Runge-Kutta scheme. The parameters of the nature model are specified in Table 1. They correspond to the standard configuration of the two-scale Lorenz system following Wilks (2005).

3.2 Description of the surrogate model

The forecast model employed in the data assimilation system is the corresponding one-scale Lorenz system (Lorenz and Emanuel1998). This model exclusively replicates the large-scale equations so that the influence of the small-scale variables must be parameterized. As the true model, the equations are solved using a fourth-order Runge-Kutta scheme. The equations for the one-scale case are

(27) X ˙ n = - X n - 1 ( X n - 2 - X n + 1 ) - X n + f ( X n ) n = 1 , , N SU

within a cyclic domain. NSU is the number of variables of the surrogate model. In order to be consistent, NSU must be equal to NLS. The external forcing, f(Xn), is defined as f(Xn)=F+f*(Xn) and consists of a linear parameterization of the effects of small-scale dynamics. The parameterization coefficients, F and f*, are estimated using the methodology proposed by Pulido et al. (2016). The parameters of the forecast model are specified in Table 2.

Table 1True model parameters.

Download Print Version | Download XLSX

Table 2Surrogate model parameters.

Download Print Version | Download XLSX

3.3 Experimental setup

3.3.1 Initial state and observations

To generate the synthetic observations, an initial true state x0t is obtained after integrating the nature model from a random initial condition over a long period. The nature model is then evolved from this initial true state for Nt=10 000 cycle times, where one cycle corresponds to the observation interval of 0.05 time units and consists of 40 integrations of the nature model. Observations are then generated from the large-scale part (LS) of the true states xkt|LS at each cycle,

(28) y k = H x k t | LS + ν k ,

where observational errors are unbiased with variance Rk, i.e. νkN(0,Rk) and xkt represents the evolution of the nature model xkt=Mt(xk-1t). The observation operator is assumed to be constant over time. We assume that the observational covariance matrix is also fixed, and diagonal, i.e.

(29) R k = σ R 2 I N y × N y

Three different observational operators are used: A linear operator , where ℋ(x)=x and σR2=0.5. A square operator , where ℋ(x)=x2 and σR2=0.5. A logarithmic operator , where H(x)=log(|x|+1) and σR2=0.05. The logarithmic operator log(|x|+1) was chosen instead of log(|x|) because, for values of x close to zero, the observation operator may diverge and worsen the performance of the assimilation, making it necessary to apply a quality control routine. Also, following Kurosawa and Poterjoy (2021), a smaller observation error is used to avoid filter divergence.

Experiments for each observation operator were conducted with full observations (that is, Ny=40) and with partial observations (that is, Ny=20 with observations at every other grid point). In addition, each combination of observation operator and observation network was run with Np=20 and Np=50 particles.

To set the Np initial states of the particles, we use randomly chosen times from a long simulation of the surrogate model. This selection is used to create the first ensemble, whose particles are independent of the initial true state.

3.3.2 Specifications of the MPF

As mentioned, a Gaussian radial basis function is used as the kernel in Eq. (6), with its covariance matrix taken to be proportional to the forecast covariance estimated from the sample,

(30) Σ = γ P ^ f = γ N p - 1 j = 1 N p ( x f ( j ) - x f ) ( x f ( j ) - x f )

where γ is a bandwidth hyperparameter and xf denotes the sample mean of the forecasts across the particles. In this work, we tune this γ hyperparameter for the experiments using a brute-force search. The step size of the mapping ϵ is determined adaptively using the Adam optimization method (Kingma and Ba2014) with up to 500 iterations of pseudo-time in each cycle.

For two key experimental setups – the fully observed linear case and the partially observed logarithmic case – we first performed sensitivity analysis by varying the localization radius, which led us to establish a default value of ℓ=3 for all subsequent experiments. We then conducted additional sensitivity tests for these same two scenarios to assess performance dependence on particle number. Finally, for these two scenarios, to evaluate algorithm behavior under large model error conditions, we conducted experiments where the linear parameterization was omitted from the surrogate model, significantly increasing the model error.

A non-Gaussian posterior density may be the result of a non-linear observation operator or a non-Gaussian prior density distribution resulting from non-linear forecasts. One of the objectives of this work is to evaluate the performance of the MPF in experiments with two prior density distributions: a Gaussian and a Gaussian mixture. In the Gaussian experiments, the resulting gradient of the logarithm of the prior density function is given by Eq. (9), in which we take B=P^f. In the global and α-localization cases, this matrix is scaled by a Gaspari-Cohn decaying factor. However, in β-localization, scaling of the prior covariance matrix is not required for small localization radii and thus will not be applied.

In the Gaussian mixture experiments, we use the expression given in Eq. (11) for the density. The matrix Qk is defined as Qk=ξPf where ξ is a bandwidth hyperparameter of the mixtures. Tuning this hyperparameter contributes to enhancing the performance of the MPF. The number of Gaussians corresponds to the number of particles.

In preliminary experiments, we found that a multiplicative or additive inflation factor is not required in the MPF even when applied over an extended period. In fact, adding an inflation factor degraded the performance of the filter.

4 Results

In each experiment, a comparison is made between the global MPF, both localization schemes, the Ensemble Transform Kalman Filter (ETKF, Bishop et al.2001) and the LETKF with ℓ=3. We compare the LMPFs against the LETKF and ETKF as these represent classical, computationally efficient ensemble filters that provide good baseline performance and are widely used in operational data assimilation systems.

The Root Mean Square Error (RMSE) between the true state and the analysis ensemble mean, and the spread of the analysis ensemble are the primary metrics used to compare the performance of each experiment. The time series consists of 10 000 cycles, with the initial 1000 cycles designated as the spin-up period and excluded from the analysis. The temporal averages of the RMSE and spread are then calculated over the subsequent 9000 cycles. Besides, to evaluate the dependence on the initial conditions, for each experiment, 10 realizations were conducted. The results show the mean of these realizations and the error bar represent the sample standard deviation.

Before the experiments, we conducted a hyperparameter optimization. In the case of Kalman filters, this involves a multiplicative inflation factor that minimizes the RMSE of the analyses.

For the MPF and its local versions, one of the key hyperparameters is the proportionality factor of the kernel sample covariance γ. For experiments assuming a Gaussian mixture, another hyperparameter is the width of the Gaussian mixtures, ξ. Thus, the optimization is performed in the 2D space defined by γ and ξ for the Gaussian mixture case by brute force. The hyperparameters are selected to minimize RMSE in order to examine how the methodologies represent ensemble spread under optimal RMSE conditions. The goal is to evaluate whether the methodologies produce a reasonable spread representation at their lowest RMSE without explicitly tuning for it. As an example of the hyperparameter tuning, Fig. 1 shows the optimization of the global MPF with Gaussian mixture prior in a fully observed linear case using 20 particles. As illustrated in Fig. 1, the dependence of RMSE on γ and ξ is nontrivial and non-intuitive, which prevents the definition of a simple rule of thumb. Optimization for the different variants of the MPF and the ensemble Kalman filters is performed for each particle size and observation network. In Appendix A we present the optimal parameters for some of the experiments.

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f01

Figure 1Time and variable averaged RMSE for the MPF experiment as a function of the bandwidth of the Gaussian mixtures ξ and the bandwidth of the kernel γ.

Download

For the experiments shown in this work with the two-scale Lorenz system and its surrogate one-scale Lorenz model, a model integration without data assimilation achieves an RMSE of 6.78 and a spread of 6.55. The RMSE of 6.78 represents the maximum error of the forecast model without assimilation, providing a top value for evaluating the impact of incorporating observational data in the assimilation process. Hereafter, we refer to this value as the NoDA-RMSE.

4.1 Linear observational operator

The first experiment evaluates the performance of the local mapping particle flow filters under a linear observation operator, as in Eq. (28). Figure 2 shows the results for the fully observed (left panels) and partially observed (right panels) scenarios, employing 20 in Fig. 2a and 50 particles in Fig. 2b. Black dots and crosses represent Gaussian filters or MPFs that assume Gaussian priors. Red dots and crosses represent particle flow filters with Gaussian mixture priors.

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f02

Figure 2RMSE and spread in the linear observation operator for 20 and 50 particles, under both fully and partially observed scenarios.

Download

All MPF experiments exhibit better performance than ETKF for the 20-particle experiments with a full observed state, as in Fig. 2a, except for the global MPF using a pure Gaussian prior PDF. This last case converges to an RMSE of 0.644 ± 0.001 but with an extremely high spread value (6.70 ± 0.04). On the other hand, when a Gaussian mixture prior density is utilized, represented by the red dots, all three MPF experiments demonstrate performances similar to LETKF.

Regarding the spread, MPF and LMPF with pure Gaussian priors tend to have large dispersion. In the case of LMPF-β, the spread is 0.92 and is not shown. However, this pattern changes in Gaussian mixture experiments, where the spread is much closer to the RMSE. The global case provides the spread that is closest to the RMSE. It is important to note that these spread results come from experiments using optimal hyperparameters in terms of RMSE.

Despite the linearity of the observational operator, the model dynamics is non-linear. Consequently, it is expected that Gaussian mixtures capture non-linearities more effectively compared to experiments utilizing pure Gaussian priors. This could explain the better performance of the Gaussian mixture experiments.

The right panel of Fig. 2a shows results for partially observed experiments. In the Gaussian prior case, the global MPF converged to a very high RMSE (1.85 ± 0.03) and is therefore not shown. The localized filters achieve a lower RMSE than ETKF, and perform similarly than LETKF. As in the fully observed scenario, the Gaussian mixture experiments show a significant improvement across all MPFs. The resulting RMSE is comparable to that of LETKF.

Figure 2b presents the results for the 50-particle experiments. The performance relationships among the experiments are similar to the previous case, with the notable exception of ETKF, which shows the most significant improvement. In this case, the Gaussian-prior global MPF achieves convergence, although its RMSE remains higher than that of the Kalman filters. Similarly, the Gaussian mixture experiments demonstrate a significant improvement in RMSE.

In the partially observed scenario, the ETKF demonstrates the most significant improvement, and the Gaussian-prior MPF successfully converges. Additionally, the spread of the Gaussian-mixture prior MPF is closer to the RMSE in the global case. The localized particle filters exhibit a similar behavior to that observed in the fully observed case.

We note that these experiments use a localization radius of ℓ=3 which is only optimal for the LETKF with Np=20. The localization radius of ℓ=3 was fixed in all the experiments to ensure fair comparison across all methods and ensemble sizes, recognizing that this choice may not be individually optimal for each configuration. Notably, the LMPFs exhibit better performance for longer radii, so that the ℓ=3 fixed radius choice does not systematically favor the proposed localization methods. This setup ensures that differences in performance can be attributed solely to the algorithms themselves rather than variations in the localization radius.

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f03

Figure 3RMSE and spread for the experiments with a square observation operator for 20 (a) and 50 (b) particles, under both fully (left panels) and partially observed (right panels) scenarios.

Download

4.2 Square observational operator

A square observational operator presents a challenge for data assimilation schemes, as it treats negative and positive true states with the same absolute value as equivalent, so that the error distributions in the hidden state space are likely to be a bimodal distribution.

Figure 3a presents the square- results for the 20-particle experiments. Overall, the RMSE values are smaller compared to the linear case. This difference is linked to the choice of model error variance. While σR2=0.5 in both cases, the magnitude of nonlinear observations is typically much greater than that of linear observations, resulting in a relatively smaller error in the nonlinear case.

In the fully observed case, both the ETKF and the global MPF with a Gaussian prior converged to very high RMSE values. The Gaussian-prior MPF achieved an RMSE > NoDA-RMSE. However, the LETKF and Gaussian-prior LMPFs achieve good RMSE performance, though with a significant underestimation of the spread in the LMPF-α. On the other hand, the three experiments employing Gaussian-mixture priors demonstrate very good performance, similar to the LETKF. The impact of localization is pronounced in the ensemble Kalman filters (as seen in the ETKF vs. LETKF performance) but has only a minor effect on the Gaussian-mixture MPFs.

In the partially observed case, the ETKF diverged for all inflation parameters tested and the Gaussian-prior MPF achieved an RMSE > NoDA-RMSE. LETKF shows a similar RMSE than the Gaussian-mixture particle filters. Figure 3b presents the results for the 50-particle experiments. In the fully observed scenario, the Gaussian-prior MPF successfully converges, unlike in the 20-particle case, but with a high RMSE for this observational operator (0.367 ± 0.007) and a significant low spread. The ETKF demonstrates a notable improvement in accuracy, outperforming its localized version. A similar effect is observed in the Gaussian-mixture experiments, where the MPF achieves similar accuracy than its localized counterparts. In this case, the Gaussian-mixture particle filters provide a higher spread compared to the Gaussian-prior filters, with the exception of the Gaussian-prior LMPF-β.

In the partially observed scenario, the ETKF achieves convergence with an RMSE similar to that of LETKF. Once again, the Gaussian-mixture filters demonstrate the best performance, comparable to the Kalman filters, with the exception of the global MPF, which showed a very high RMSE value. In this case, the effect of localization is very positive. However, as in the 20-particle case, all filters significantly underestimate the spread.

4.3 Logarithmic observational operator

Figure 4 shows the performance of the filters in the logarithmic observation operator case, assessing a highly non-Gaussian regime.

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f04

Figure 4RMSE and spread in the logarithm operator for 20 (a) and 50 (b) particles, under both fully and partially observed scenarios.

Download

For the 20-particle experiments in Fig. 4a, both the ETKF and the Gaussian-prior global MPF achieved very high RMSE values in fully observed cases. In the partially observed case, the ETKF reached an RMSE > NoDA-RMSE. Meanwhile, the LETKF achieves excellent RMSE values with a closely matching spread in the fully observed case. In contrast, the Gaussian-priors filters exhibit the worst performance, while the Gaussian-mixture prior localized filters show good performance, comparable to the LETKF. While the LETKF shows the best mean RMSE, it exhibits large error bars due to its sensitivity to initial conditions. The Gaussian-mixture localized filters show slightly higher mean RMSE values, but these fall within the LETKF's error band and have much smaller error bars, resulting in comparable overall performance.

For 50 particles, Fig. 4b, the ETKF successfully converges in the fully observed case, showing performance comparable to that of the LETKF. In this scenario, the Gaussian-mixture particle filters also demonstrate competitive results. In the partially observed scenario, ETKF and LETKF again show good results, but with large error bar in the case of ETKF. Meanwhile, the localized Gaussian-mixture filters have performance comparable to LETKF. In the case of LMPF-α, the RMSE falls within the LETKF's error band. The MPF's in all its versions show smaller sensitivity to initial conditions, particularly in the partially observed case.

4.4 Sensitivity to the localization radius

The performance of localized particle filters is assessed by varying the radius of localization. This study is made on the linear fully observed case, and on the logarithm and partially observed case, the most non-Gaussian scenario. The number of particles used is 20 and only Gaussian-mixture prior densities are used in the MPFs.

Figure 5 shows the results of the linear experiment. The LETKF achieves a minimum RMSE at a localization radius of ℓ=3. This is the main reason why we selected this localization radius to conduct all localized experiments.

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f05

Figure 5RMSE as a function of localization radius for the LMPFs and the LETKF for the linear and fully observed case with Np=20.

Download

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f06

Figure 6RMSE as a function of localization radius for the LMPFs and the LETKF for the linear and fully observed case with Np=10.

Download

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f07

Figure 7RMSE as a function of localization radius for the LMPFs and the LETKF for the logarithmic and partially observed case with Np=20.

Download

For radii greater than 4, the LETKF degrades more rapidly than LMPFs. The LMPFs tend to converge to the same RMSE performance as the global MPF when using a localization radius of 18. This suggests that for Np=20 the local algorithm benefits from incorporating distant covariances, even with reduced weights, to improve the estimation at each grid point.

LMPF-α exhibits a behavior similar to the β-case but results in slightly higher RMSE values and reaches a minimum around ℓ=12. However, the difference is small considering that the RMSE as a function of the localization radius is almost flat for that range of localization scales.

These results reflect the relationship between localization needs and the system's effective dimensionality relative to the ensemble size. For the 40-dimensional single Lorenz dynamics, the number of positive Lyapunov exponents is smaller than the ensemble size used in these experiments (20 particles). In the case of this surrogate model, using a forcing f=26, the number of positive Lyapunov exponents calculated were around 16–17 with the parameterized forcing, f*, and around 18–19 without it. With 20 particles exceeding the number of unstable directions, the ensemble in principle provides sufficient rank to capture the system's dynamics without requiring strong localization, explaining the optimal performance at larger radii.

To test this hypothesis, we conducted additional experiments with reduced ensemble size (10 particles), where the ensemble rank falls below the number of positive Lyapunov exponents.

The results of this experiment are shown in Fig. 6. The LETKF exhibits behavior similar to the 20-particle case. For the localized particle filters, a minimum value appears around ℓ=9. In the case of the LMPF-α, the filter does not converge for localization radii greater than 14.

We remind that the experiments are under the presence of model error. This affects the optimal localization radius; in particular, the LETKF has a longer optimal localization radius for twin perfect-model experiments. In realistic applications, the presence of model errors is also expected to affect long-range correlations. The MPF appears to behave more robustly to this effect.

Figure 7 shows the performance of the filters for the logarithmic and partially observed experiments for 20 particles. In this scenario, all the filters achieve a minimum RMSE around =2,3. The LETKF shows more sensitivity to initial conditions than the localized filters; however, it achieves slightly lower RMSEs.

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f08

Figure 8RMSE as a function of the number of particles for the fully observed linear case and the partially observed logarithmic case, with a localization radius of 3 for the localized filters.

Download

4.5 Sensitivity to the particle number

The two extreme experimental setups – fully observed with a linear observation operator, and partially observed with a logarithmic observation operator – are used to evaluate the sensitivity of performance to the number of particles. As in the previous subsection, only Gaussian-mixture prior densities are considered. The localization radius is fixed at ℓ=3, and the experiments are conducted with particle numbers of 5, 10, 20, 50, and 100. The results are displayed in Fig. 8.

Global MPF and LMPF-α demonstrate very good performance for small particle numbers in the linear experiment. For larger particle numbers, both localized particle filters achieve excellent performance, comparable to that of the LETKF.

RMSE and spread for the large model error experiments using linear and fully observed cases and logarithm and partial observations with 20 particles and a localization radius of 3.

In contrast, the results for the partially observed logarithmic case are unexpected. For a small number of particles, only Gaussian-mixture MPFs achieved RMSE less than NoDA-RMSE, although with a high RMSE. At larger particle numbers, LMPF-α achieves convergence with an accuracy greater than that of the Kalman filters. The performance of LMPF-β is similar to Kalman filters.

The performance of the LETKF in this non-Gaussian experiment deteriorates for ensembles of 50 and 100 particles. A plausible explanation is that certain ensemble members diverge and fail to return to the Lorenz attractor, an effect that is found in deterministic filters (Amezcua et al.2012). The underlying reason for this behavior is that the data assimilation in deterministic Ensemble Kalman filters only scales the prior density without changing its shape. Consequently, if the prior ensemble contains outliers, they persist in the posterior density and can grow towards the next data assimilation cycle. LMPFs do not degrade in performance with increasing numbers of particles and appear to be unaffected by this issue, as the particle communication during pseudo-time iterations modifies the prior's shape, effectively removing outliers. This issue in the LETKF could be mitigated through techniques such as applying random rotations to the analysis perturbations. However, for the purpose of this comparison, we implement a standard LETKF without additional enhancements to provide a consistent baseline against which to evaluate the proposed MPF and LMPFs methods. Interestingly, neither the MPF nor LMPFs exhibit this performance degradation with increasing ensemble size.

https://npg.copernicus.org/articles/33/33/2026/npg-33-33-2026-f09

Figure 9RMSE and spread for the large model error regime for the fully observed linear case and the partially observed logarithmic case with 20 particles and a localization radius of 3.

Download

4.6 Sensitivity to large model error

In all previous experiments, a linear parameterization of small-scale effects was used. This results in a relatively small model error. To evaluate a large model error scenario, we neglect the linear term of the parameterization f* (Table 2) and only use the external forcing f(Xn)=26 in the surrogate model. As said before, NoDA-RMSE for the parametrized surrogate model is 6.78 and the spread is 6.55. For this large model error environment, the RMSE is 6.86 and the spread reaches 9.01.

Both Kalman and particle filters are tested for the linear and logarithmic observation operators, using ℓ=3 for the local filters and Np=20 particles. Again, we use the localization radius that is optimal for the LETKF, and the hyperparameters γ, ξ and inflation are tuned for these cases. The results are displayed in Fig. 9. In these large model error conditions, the ETKF achieved an RMSE > NoDA-RMSE.

In the linear case, the Gaussian-mixture particle filters have a similar performance in terms of RMSE. Nevertheless, the spread of the global MPF is strongly underestimated. Localized particle filters show higher spreads.

In the logarithmic scenario, the LETKF converged to an RMSE>NoDA-RMSE. In contrast, the MPF and its localized variants handle model error more effectively and perform better in this challenging case. However, all three filters exhibit rather high RMSE values. Among the tested methods, LMPF-α achieves the best performance in this setup.

5 Conclusions

In this work, two localization schemes for the mapping particle filter were proposed. Both schemes are based on the hypothesis that distant observations do not impact the analysis, but their approaches differ. LMPF-α first calculates a global kernel covariance matrix and inverts it. Then, it performs local transformations at each pseudo-time step to obtain a global intermediate state vector in each step. Therefore, convergence is achieved globally. On the other hand, LMPF-β applies the global algorithm in small regions, retaining the center value of each local analysis to obtain a smooth solution. Kernel covariance matrices are calculated in each small domain. Hence, each local analysis achieves convergence independently.

Both frameworks were tested in different setups and compared with the ETKF, LETKF, and the global MPF. In general, there is a clear positive impact when taking the prior probability density as a Gaussian mixture compared to a Gaussian prior density.

For both linear and non-linear operators, LMPF’s improve estimation compared to their global version when a Gaussian prior is used and provide slightly better estimations when Gaussian mixtures are used. Furthermore, LMPF’s provide better estimates compared to the ETKF and competitive performances against the LETKF.

In the linear case, LMPF’s show very good estimations in terms of RMSE. In the squared case, Gaussian-mixture filters show very good estimations. Both Gaussian and non-Gaussian filters show poor spread representation, especially in partially observed scenarios. In the logarithmic case, Gaussian-mixture LMPF’s provide competitive solutions against the LETKF. Again, the partially observed scenario degrades the performance of particle filters while Kalman filters are less affected. LMPFs present a very good performance in the logarithmic operator case under weak model error similar to LETKF.

When the number of particles varies, Gaussian-mixture MPF and LMPF-α show better estimates at low particle numbers. For the experiments with large model error the MPF and LMPF exhibit robust performances and successfully converge while ensemble based Kalman filters did not deal well with large model errors in the logarithmic experiment. However, it is important to highlight that all these experiments required brute-force optimization of two hyperparameters in Gaussian mixtures experiments which is computationally expensive.

The implementation of the particle filter for data assimilation in one-scale Lorenz model experiments represents an essential first step in validating our newly developed methodology. Working with simplified models provides a crucial foundation before advancing to more complex atmospheric forecast models, a direction which has already been explored by Hu et al. (2024), suggesting that applying the proposed LMPF methodologies in large atmospheric models would also be feasible.

Appendix A

The optimal hyperparameters for some of the the experiments are presented, where Kalman filters report the inflation factor, Gaussian-prior particle filters show the γ parameter, and Gaussian-mixture filters display the γ and ξ parameters.

Table A1Optimal parameters for the linear case.

Download Print Version | Download XLSX

Table A2Optimal parameters for the log case.

Download Print Version | Download XLSX

Code availability

The code used in this study is available from the corresponding author upon reasonable request.

Data availability

All data used in this study are synthetic, generated as described in Sect. 3.3.1.

Author contributions

JG and MP participated in the conception of the ideas and in the design of the experiments. JG conducted the experiments. All the authors reviewed the results. JG and MP wrote the draft, all the authors made corrections and comments to the subsequent versions of the manuscript and approved the final version of the paper.

Competing interests

At least one of the (co-)authors is a member of the editorial board of Nonlinear Processes in Geophysics. The peer-review process was guided by an independent editor, and the authors also have no other competing interests to declare.

Disclaimer

Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. The authors bear the ultimate responsibility for providing appropriate place names. Views expressed in the text are those of the authors and do not necessarily reflect the views of the publisher.

Acknowledgements

The authors are grateful to the reviewers, Dr. Farchi and Dr. van Leeuwen, and to the editor, Dr. Talagrand, for their thoughtful comments and significant contributions to this work.

Financial support

This research has been supported by the National Agency for the Promotion of Science and Technology of Argentina (ANPCYT, grant no. PICT 2019/3095). This research has been partially supported by the PREVENIR project implemented by the Japan International Cooperation Agency and the Japan Science and Technology Agency under the Science and Technology Research Partnership for Sustainable Development Program.

Review statement

This paper was edited by Olivier Talagrand and reviewed by Peter Jan van Leeuwen and Alban Farchi.

References

Amezcua, J., Ide, K., Bishop, C. H., and Kalnay, E.: Ensemble clustering in deterministic ensemble Kalman filters, Tellus A: Dynamic Meteorology and Oceanography, 64, 18039, https://doi.org/10.3402/tellusa.v64i0.18039, 2012. a

Ba, J., Erdogdu, M. A., Ghassemi, M., Sun, S., Suzuki, T., Wu, D., and Zhang, T.: Understanding the variance collapse of SVGD in high dimensions, International Conference on Learning Representations, https://openreview.net/forum?id=Qycd9j5Qp9J (last access: 20 January 2026), 2022. a, b

Bengtsson, T., Snyder, C., and Nychka,D.: Toward a nonlinear ensemble filter for high-dimensional systems, J. Geophys. Res., 108, 8775, https://doi.org/10.1029/2002JD002900, 2003. a

Bishop, C. H., Etherton, B. J., and Majumdar, S. J.: Adaptive Sampling with the Ensemble Transform Kalman Filter. Part I: Theoretical Aspects, Monthly Weather Review, 129, 420–436, https://doi.org/10.1175/1520-0493(2001)129<0420:aswtet>2.0.co;2, 2001. a

Cotter, C. and Crisan, D., Holm, D., Pan, W., and Shevchenko, I.: A Particle Filter for Stochastic Advection by Lie Transport: A Case Study for the Damped and Forced Incompressible Two-Dimensional Euler Equation, SIAM/ASA Journal on Uncertainty Quantification, 8, 1446–1492, https://doi.org/10.1137/19M1277606, 2020. a

Daum, F., Huang, J., and Noushin, A.: Exact particle flow for nonlinear filters, in: Signal processing, sensor fusion, and target recognition XIX, Vol. 7697, SPIE, 92–110, https://doi.org/10.1117/12.839590, 2010. a

Del Moral, P.: Mean field simulation for Monte Carlo integration, Monographs on Statistics and Applied Probability, 126, 26, ISBN 978-1-4665-0417-2, 2013. a

Doucet, A., de Freitas, N., and Gordon, N.: Sequential Monte Carlo Methods in Practice, Springer, https://doi.org/10.1007/978-1-4757-3437-9, 2001. a

Farchi, A. and Bocquet, M.: Review article: Comparison of local particle filters and new implementations, Nonlinear Processes in Geophysics, 25, 765–807, https://doi.org/10.5194/npg-25-765-2018, 2018. a

Gallego, V. and Rios Insua, D.: Stochastic gradient MCMC with repulsive forces, arXiv [preprint], https://doi.org/10.48550/arXiv.1812.00071, 2018. a, b

Gaspari, G. and Cohn, S. E.: Construction of correlation functions in two and three dimensions, Quarterly Journal of the Royal Meteorological Society, 125, 723–757, https://doi.org/10.1002/qj.49712555417, 1999. a

Hamill, T. M., Whitaker, J. S., and Snyder, C.: Distance-dependent filtering of background error covariance estimates in an ensemble Kalman filter, Monthly Weather Review, 129, 2776–2790, 2001. a

Hohenegger, C. and Schar, C.: Atmospheric Predictability at Synoptic Versus Cloud-Resolving Scales, Bulletin of the American Meteorological Society, 88, 1783–1794, https://doi.org/10.1175/bams-88-11-1783, 2007. a

Houtekamer, P. L. and Mitchell, H. L.: A sequential ensemble Kalman filter for atmospheric data assimilation, Monthly Weather Review, 129, 123–137, 2001. a

Hu, C. and van Leeuwen, P. J.: A particle flow filter for high‐dimensional system applications, Quarterly Journal of the Royal Meteorological Society, 147, 2352–2374, https://doi.org/10.1002/qj.4028, 2021. a, b

Hu, C., van Leeuwen, P. J., and Anderson, J. L.: An implementation of the particle flow filter in an atmospheric model, Monthly Weather Review, 152, 2247–2264, 2024. a, b, c

Hunt, B. R., Kostelich, E. J., and Szunyogh, I.: Efficient data assimilation for spatiotemporal chaos: A local ensemble transform Kalman filter, Physica D: Nonlinear Phenomena, 230, 112–126, https://doi.org/10.1016/j.physd.2006.11.008, 2007. a, b

Jordan, R., Kinderlehrer, D., and Otto, F.: The variational formulation of the Fokker–Planck equation, SIAM Journal on Mathematical Analysis, 29, 1–17, 1998. a

Kingma, D. P. and Ba, J.: Adam: A method for stochastic optimization, International Conference for Learning Representations, arXiv [preprint] https://doi.org/10.48550/arXiv.1412.6980, 2014. a, b, c

Kurosawa K. and Poterjoy, J.: Data Assimilation Challenges Posed by Nonlinear Operators: A Comparative Study of Ensemble and Variational Filters and Smoothers, Monthly Weather Review, 149, 2369–2389, https://doi.org/10.1175/MWR-D-20-0368.1, 2021. a

Leviyev, A., Chen, J., Wang, Y., Ghattas, O., and Zimmerman, A.: A stochastic stein variational newton method, arXiv [preprint], https://doi.org/10.48550/arXiv.2204.09039, 2022. a

Liu, Q. and Wang, D.: Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm, Advances in Neural Information Processing Systems, 29, https://proceedings.neurips.cc/paper_files/paper/2016/file/b3ba8f1bee1238a2f37603d90b58898d-Paper.pdf (last access: 20 January 2026), 2016. a, b, c

Lorenz, E. N. and Emanuel, K. A.: Optimal Sites for Supplementary Weather Observations: Simulation with a Small Model, Journal of the Atmospheric Sciences, 55, 399–414, https://doi.org/10.1175/1520-0469(1998)055<0399:OSFSWO>2.0.CO;2, 1998. a

Lorenz, E. N.: Designing Chaotic Models, Journal of the Atmospheric Sciences, 62, 1574–1587, https://doi.org/10.1175/jas3430.1, 2005. a

Lu, J., Lu, Y., and Nolen, J.: Scaling limit of the Stein variational gradient descent: The mean field regime, SIAM Journal on Mathematical Analysis, 51, 648–671, 2019. a

Ma, Y.-A., Chen, T., and Fox, E.: A complete recipe for stochastic gradient MCMC, Advances in Neural Information Processing Systems, 28, https://proceedings.neurips.cc/paper_files/paper/2015/file/9a4400501febb2a95e79248486a5f6d3-Paper.pdf (last access: 20 January 2026), 2015. a

Neal, R. M.: Sampling from multimodal distributions using tempered transitions, Statistics and Computing, 6, 353–366, https://doi.org/10.1007/bf00143556, 1996. a

Penny, S. G. and Miyoshi, T.: A local particle filter for high-dimensional geophysical systems, Nonlinear Processes in Geophysics , 23, 391–405, https://doi.org/10.5194/npg-23-391-2016, 2016. a

Poterjoy, J.: A localized particle filter for high-dimensional nonlinear systems, Monthly Weather Review, 144, 59–76, https://doi.org/10.1175/mwr-d-15-0163.1, 2016. a

Pulido, M., Scheffler, G., Ruiz, J. J., Lucini, M. M., and Tandeo, P.: Estimation of the functional form of subgrid-scale parametrizations using ensemble-based data assimilation: a simple model experiment, Quarterly Journal of the Royal Meteorological Society, 142, 2974–2984, https://doi.org/10.1002/qj.2879, 2016. a

Pulido, M. and van Leeuwen, P. J.: Sequential Monte Carlo with kernel embedded mappings: The mapping particle filter, Journal of Computational Physics, 396, 400–415, https://doi.org/10.1016/j.jcp.2019.06.060, 2019. a, b, c, d

Robert, S. and Künsch, H. R.: Localizing the Ensemble Kalman Particle Filter, Tellus A: Dynamic Meteorology and Oceanography, 69, 1282016, https://doi.org/10.1080/16000870.2017.1282016, 2017. a

Ruiz, J., Lien, G.-Y., Kondo, K., Otsuka, S., and Miyoshi, T.: Reduced non-Gaussianity by 30 s rapid update in convective-scale numerical weather prediction, Nonlinear Processes in Geophysics, 28, 615–626, https://doi.org/10.5194/npg-28-615-2021, 2021. a

Snyder, C., Bengtsson, T., Bickel, P., and Anderson, J.: Obstacles to high-dimensional particle filtering, Monthly Weather Review, 136, 4629–4640, https://doi.org/10.1175/2008mwr2529.1, 2008. a

Subrahmanya, A. N., Popov, A. A., and Sandu, A.: Ensemble variational Fokker-Planck methods for data assimilation, Journal of Computational Physics, 523, 113681, https://doi.org/10.1016/j.jcp.2024.113681, 2025. a

Tabak, E. G. and Turner, C. V.: A Family of Nonparametric Density Estimation Algorithms, Communications on Pure and Applied Mathematics, 66, 145–164, https://doi.org/10.1002/cpa.21423, 2012. a

van Leeuwen, P. J.: Nonlinear ensemble data assimilation for the ocean. In seminar Recent Developments in Data Assimilation for Atmosphere and Ocean, ECMWF, Reading, UK, 8–12 September, https://www.ecmwf.int/en/elibrary/77010-nonlinear-ensemble-data-assimilation-ocean (last access: 20 January 2026), 2003. a

van Leeuwen, P. J., Künsch, H. R., Nerger, L., Potthast, R., and Reich, S.: Particle filters for high-dimensional geoscience applications: A review, Quarterly Journal of the Royal Meteorological Society, 145, 2335–2365, https://doi.org/10.1002/qj.3551, 2019. a, b, c

Wang, D., Zeng, Z., and Liu, Q.: Stein variational message passing for continuous graphical models, International Conference on Machine Learning, 5219–5227, https://proceedings.mlr.press/v80/wang18l.html (last access: 20 January 2026), 2018. a

Wilks, D. S.: Effects of stochastic parameterizations in the Lorenz 96 system, Quarterly Journal of the Royal Meteorological Society, 131, 389–407, https://doi.org/10.1256/qj.04.03, 2005. a

Whitaker, J. S. and Hamill, T. M.: Ensemble data assimilation without perturbed observations, Monthly Weather Review, 130, 1913–1924, 2002. a

Zhuo, J., Liu, C., Shi, J., Zhu, J., Chen, N., and Zhang, B.: Message passing Stein variational gradient descent, Proceedings of the 35th International Conference on Machine Learning, 80, 6018–6027, https://proceedings.mlr.press/v80/zhuo18a.html (last access: 20 January 2026), 2018. a, b

Download
Short summary

This work extends the Mapping Particle Filter to account for local dependencies. Two localization methods are tested: a global particle flow with local kernels, and iterative local mappings based on correlation radius. Using a two-scale Lorenz-96 truth and a one-scale forecast model, experiments with full/partial and linear/nonlinear observations show Root Mean Square Error reductions using localized Gaussian mixture priors, achieving competitive performance against Gaussian filters.

Share