Frédéric Cérou’s research while affiliated with French National Centre for Scientific Research and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (48)


Adaptive reduced tempering For Bayesian inverse problems and rare event simulation
  • Preprint

October 2024

·

2 Reads

Frederic Cerou

·

·

Mathias Rousset

This work proposes an adaptive sequential Monte Carlo sampling algorithm for solving inverse Bayesian problems in a context where a (costly) likelihood evaluation can be approximated by a surrogate, constructed from previous evaluations of the true likelihood. A rough error estimation of the obtained surrogates is required. The method is based on an adaptive sequential Monte-Carlo (SMC) simulation that jointly adapts the likelihood approximations and a standard tempering scheme of the target posterior distribution. This algorithm is well-suited to cases where the posterior is concentrated in a rare and unknown region of the prior. It is also suitable for solving low-temperature and rare-event simulation problems. The main contribution is to propose an entropy criteria that associates to the accuracy of the current surrogate a maximum inverse temperature for the likelihood approximation. The latter is used to sample a so-called snapshot, perform an exact likelihood evaluation, and update the surrogate and its error quantification. Some consistency results are presented in an idealized framework of the proposed algorithm. Our numerical experiments use in particular a reduced basis approach to construct approximate parametric solutions of a partially observed solution of an elliptic Partial Differential Equation. They demonstrate the convergence of the algorithm and show a significant cost reduction (close to a factor 10) for comparable accuracy.


Experiment #1. Left: true AMVs superimposed on the pair of incomplete image observations (black pixels correspond to missing data). Right: comparison to the MAP estimate.
Experiment #1. Comparison of the various endpoint error criteria (2.7)–(2.9) with respect to the sample size (N × L) for a chilled HMC simulation.
Experiment #1. Comparison of the different methods in terms of the endpoint error criteria (2.7) with (2.8) for p = 2 (left) or with (2.9) (right) with respect to the sample size N × L = 1×103 (H = 0.5, ζ = 1×10−6 ).
Experiment #1. Influence of the temperature ζ (left) and of the preconditioning parameter H (right) on the evolution of the endpoint error criteria (2.7) with (2.8) for p = 2.
Experiment #1. True versus expected errors obtained by the Laplace method or with chilled HMC ( N=1×102 , L = 10, H = 0.5, ζ = 1×10−6 ). The gray level values range in [0,lmax] , lmax being equal to the empirical mean plus standard deviation over Ωm of the true or expected errors (the values of ∥d⋆(s)−dˆ(s)∥2 are thresholded to lmax≈1.3 ).

+3

Chilled sampling for uncertainty quantification: a motivation from a meteorological inverse problem
  • Article
  • Publisher preview available

December 2023

·

7 Reads

·

2 Citations

Atmospheric motion vectors (AMVs) extracted from satellite imagery are the only wind observations with good global coverage. They are important features for feeding numerical weather prediction (NWP) models. Several Bayesian models have been proposed to estimate AMVs. Although critical for correct assimilation into NWP models, very few methods provide a thorough characterization of the estimation errors. The difficulty of estimating errors stems from the specificity of the posterior distribution, which is both very high dimensional, and highly ill-conditioned due to a singular likelihood, which becomes critical in particular in the case of missing data (unobserved pixels). Motivated by this difficult inverse problem, this work studies the evaluation of the (expected) estimation errors using gradient-based Markov chain Monte Carlo (MCMC) algorithms. The main contribution is to propose a general strategy, called here ‘chilling’, which amounts to sampling a local approximation of the posterior distribution in the neighborhood of a point estimate. From a theoretical point of view, we show that under regularity assumptions, the family of chilled posterior distributions converges in distribution as temperature decreases to an optimal Gaussian approximation at a point estimate given by the maximum a posteriori, also known as the Laplace approximation. Chilled sampling therefore provides access to this approximation generally out of reach in such high-dimensional nonlinear contexts. From an empirical perspective, we evaluate the proposed approach based on some quantitative Bayesian criteria. Our numerical simulations are performed on synthetic and real meteorological data. They reveal that not only the proposed chilling exhibits a significant gain in terms of accuracy of the AMV point estimates and of their associated expected error estimates, but also a substantial acceleration in the convergence speed of the MCMC algorithms.

View access options

Fluctuations of Rare Event Simulation with Monte Carlo Splitting in the Small Noise Asymptotics

December 2022

·

10 Reads

Diffusion processes with small noise conditioned to reach a target set are considered. The AMS algorithm is a Monte Carlo method that is used to sample such rare events by iteratively simulating clones of the process and selecting trajectories that have reached the highest value of a so-called importance function. In this paper, the large sample size relative variance of the AMS small probability estimator is considered. The main result is a large deviations logarithmic equivalent of the latter in the small noise asymptotics, which is rigorously derived. It is given as a maximisation problem explicit in terms of the quasi-potential cost function associated with the underlying small noise large deviations. Necessary and sufficient geometric conditions ensuring the vanishing of the obtained quantity ('weak' asymptotic efficiency) are provided. Interpretations and practical consequences are discussed.


Entropy minimizing distributions are worst-case optimal importance proposals

December 2022

·

8 Reads

Importance sampling of target probability distributions belonging to a given convex class is considered. Motivated by previous results, the cost of importance sampling is quantified using the relative entropy of the target with respect to proposal distributions. Using a reference measure as a reference for cost, we prove under some general conditions that the worst-case optimal proposal is precisely given by the distribution minimizing entropy with respect to the reference within the considered convex class of distributions. The latter conditions are in particular satisfied when the convex class is defined using a push-forward map defining atomless conditional measures. Applications in which the optimal proposal is Gibbsian and can be practically sampled using Monte Carlo methods are discussed.


Uncertainty of Atmospheric Motion Vectors by Sampling Tempered Posterior Distributions

July 2022

·

12 Reads

Atmospheric motion vectors (AMVs) extracted from satellite imagery are the only wind observations with good global coverage. They are important features for feeding numerical weather prediction (NWP) models. Several Bayesian models have been proposed to estimate AMVs. Although critical for correct assimilation into NWP models, very few methods provide a thorough characterization of the estimation errors. The difficulty of estimating errors stems from the specificity of the posterior distribution, which is both very high dimensional, and highly ill-conditioned due to a singular likelihood, which becomes critical in particular in the case of missing data (unobserved pixels). This work studies the evaluation of the expected error of AMVs using gradient-based Markov Chain Monte Carlo (MCMC) algorithms. Our main contribution is to propose a tempering strategy, which amounts to sampling a local approximation of the joint posterior distribution of AMVs and image variables in the neighborhood of a point estimate. In addition, we provide efficient preconditioning with the covariance related to the prior family itself (fractional Brownian motion), with possibly different hyper-parameters. From a theoretical point of view, we show that under regularity assumptions, the family of tempered posterior distributions converges in distribution as temperature decreases to an {optimal} Gaussian approximation at a point estimate given by the Maximum A Posteriori (MAP) log-density. From an empirical perspective, we evaluate the proposed approach based on some quantitative Bayesian evaluation criteria. Our numerical simulations performed on synthetic and real meteorological data reveal a significant gain in terms of accuracy of the AMV point estimates and of their associated expected error estimates, but also a substantial acceleration in the convergence speed of the MCMC algorithms.



On Synchronized Fleming-Viot Particle Systems

November 2019

·

12 Reads

This article presents a variant of Fleming-Viot particle systems, which are a standard way to approximate the law of a Markov process with killing as well as related quantities. Classical Fleming-Viot particle systems proceed by simulating N trajectories, or particles, according to the dynamics of the underlying process, until one of them is killed. At this killing time, the particle is instantaneously branched on one of the (N1)(N-1) other ones, and so on until a fixed and finite final time T. In our variant, we propose to wait until K particles are killed and then rebranch them independently on the (NK)(N-K) alive ones. Specifically, we focus our attention on the large population limit and the regime where K/N has a given limit when N goes to infinity. In this context, we establish consistency and asymptotic normality results. The variant we propose is motivated by applications in rare event estimation problems.


Adaptive Multilevel Splitting: Historical Perspective and Recent Results

April 2019

·

76 Reads

·

63 Citations

This article first presents a short historical perpective of the importance splitting approach to simulate and estimate rare events, with a detailed description of several variants. We then give an account of recent theoretical results on these algorithms, including a central limit theorem for Adaptive Multilevel Splitting (AMS). Considering the asymptotic variance in the latter, the choice of the importance function, called the reaction coordinate in molecular dynamics, is also discussed. Finally, we briefly mention some worthwhile applications of AMS in various domains.


On the Asymptotic Normality of Adaptive Multilevel Splitting

April 2018

·

32 Reads

·

22 Citations

SIAM/ASA Journal on Uncertainty Quantification

Adaptive Multilevel Splitting (AMS for short) is a generic Monte Carlo method for Markov processes that simulates rare events and estimates associated probabilities. Despite its practical efficiency, there are almost no theoretical results on the convergence of this algorithm. The purpose of this paper is to prove both consistency and asymptotic normality results in a general setting. This is done by associating to the original Markov process a level-indexed process, also called a stochastic wave, and by showing that AMS can then be seen as a Fleming-Viot type particle system. This being done, we can finally apply general results on Fleming-Viot particle systems that we have recently obtained.


On the Asymptotic Normality of Adaptive Multilevel Splitting

April 2018

·

1 Read

Adaptive Multilevel Splitting (AMS for short) is a generic Monte Carlo method for Markov processes that simulates rare events and estimates associated probabilities. Despite its practical efficiency, there are almost no theoretical results on the convergence of this algorithm. The purpose of this paper is to prove both consistency and asymptotic normality results in a general setting. This is done by associating to the original Markov process a level-indexed process, also called a stochastic wave, and by showing that AMS can then be seen as a Fleming-Viot type particle system. This being done, we can finally apply general results on Fleming-Viot particle systems that we have recently obtained.


Citations (35)


... Synergistic use of both methods is adopted by NOAA's operational algorithm for AMV production. It is worth noting that sampling approaches (Héas et al. 2023a) for estimating AMVs together with their errors are important for quantitative applications such as improving forecasts through assimilating these winds into NWP models, and such approaches should be further explored and applied to wind estimation. ...

Reference:

Tracking Atmospheric Motions for Obtaining Wind Estimates Using Satellite Observations—From 2D to 3D
Chilled sampling for uncertainty quantification: a motivation from a meteorological inverse problem

... In [7], a similar formula is given for the large sample size variance of all estimators, see Corollary 2.8 and Theorem 2.13. The extension to the case k > 1 under the same assumptions, where k is fixed and N → +∞ can be obtained using the results of [10]. ...

On synchronized Fleming–Viot particle systems
  • Citing Article
  • March 2021

Theory of Probability and Mathematical Statistics

... Estimating the probabilities of rare but impactful events, an important problem throughout science and engineering, is usually done through -generally expensive -Monte Carlo methods such as importance sampling [1] or importance splitting methods [2,3]. A simple alternative approach, which is principled and sampling-free but only asymptotically exact under certain assumptions, consists of using a Laplace approximation, see e.g. ...

Adaptive Multilevel Splitting: Historical Perspective and Recent Results
  • Citing Article
  • April 2019

... = τ {ξ l} (x), the estimator of the rare event probability associated with level l, where I N l is the random number of iterations required so that all clones have reached the target set {ξ l}. The estimator p N l,ams (as well as other nonnormalized estimators) is unbiased E p N l,ams = p ε l (see [7,2]). The empirical distribution of clones at iteration I = Law(X ε | τ l (X ε ) < τ A (X ε )). ...

On the Asymptotic Normality of Adaptive Multilevel Splitting
  • Citing Article
  • April 2018

SIAM/ASA Journal on Uncertainty Quantification

... Moreover, we only consider soft killing at some continuous rate, and no hard killing which would correspond to the case where T is the escape time from some sub-domain (see e.g. [4,21]). Finally, as will be seen below, as far as the long-time behaviour of the process is concerned we will work in a perturbative regime, namely we will assume that the variations of λ are small with respect to the mixing time of the diffusion (1.1) (while λ ∞ itself is not required to be small). ...

A Central Limit Theorem for Fleming-Viot Particle Systems with Hard Killing
  • Citing Article
  • September 2017

... For both standard and reduced estimates (ARMS and ART's IS estimate), we observe approximately a unitary negative slope (on a logarithmic scale) of the variance as a function of cost, implying that the product of variance and cost is constant. Such a behavior suggests asymptotic normality of the proposed IS estimate, as it has been demonstrated for AMS [9]. ...

Fluctuation analysis of adaptive multilevel splitting
  • Citing Article
  • December 2016

... The ensemble Kalman filter has been developed in geophysics (Evensen 1994(Evensen , 2007 and is used frequently in atmospheric physics, oceanography and reservoir modeling. The propagation step is the same as in the bootstrap filter. ...

A mean field theory of sequential Monte Carlo methods P. Del

·

F Cérou

·

·

[...]

·

S Rubenthaler

... Notably, besides [IZ21], we are not aware of any other theoretical work on MCMC methods for Bernoulli group testing. On the other hand, multiple applied papers have used MCMC methods for group testing [STR03, KST96,FGC12] and it is the general understanding that their "…empirical performance appears strong in simulations " [AJS19, Section 3.3.1]. To buttress these claims, it is essential to pursue an improved theoretical understanding of MCMC methods for informationtheoretic optimal designs such as Bernoulli group testing, which is the central focus of this work. ...

Decoding Fingerprinting Using the Markov Chain Monte Carlo Method