Pau Batlle’s research while affiliated with California Institute of Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (12)


Figure 6.2: (Left) Four realizations of the data-generating process where the observations are shown in red. For each realization, the blue points are uniformly distributed samples from its Berger-Boos set, sampled using the VGS sampler. (Center) For a realization of the data-generating process, we plot the distribution of γ-quantiles for the points sampled by the VGS sampler. Notably, a non-trivial percent of these are above χ 2 1,α defining the OSB interval. (Right) For the same realization, we plot the estimated sliced maxquantile function, m γ (µ) in orange alongside χ 2 1,α in red. The blue points correspond to sampled parameter values, each of which has a functional and quantile value, while the solid blue line shows the LLR over the functional varies. All intervals can be read immediately from this image by inspecting where the blue LLR curve intersects the sampled points.
Figure 6.5: Parameter values for the smooth and adversarial settings for x * used to illustrate our interval construction versus the OSB interval. The adversarial setting is made more difficult by the sharp jumps in parameter values.
Confidence intervals for functionals in constrained inverse problems via data-adaptive sampling-based calibration
  • Preprint
  • File available

February 2025

·

14 Reads

Michael Stanley

·

Pau Batlle

·

·

[...]

·

We address functional uncertainty quantification for ill-posed inverse problems where it is possible to evaluate a possibly rank-deficient forward model, the observation noise distribution is known, and there are known parameter constraints. We present four constraint-aware confidence intervals extending the work of Batlle et al. (2023) by making the intervals both computationally feasible and less conservative. Our approach first shrinks the potentially unbounded constraint set compact in a data-adaptive way, obtains samples of the relevant test statistic inside this set to estimate a quantile function, and then uses these computed quantities to produce the intervals. Our data-adaptive bounding approach is based on the approach by Berger and Boos (1994), and involves defining a subset of the constraint set where the true parameter exists with high probability. This probabilistic guarantee is then incorporated into the final coverage guarantee in the form of an uncertainty budget. We then propose custom sampling algorithms to efficiently sample from this subset, even when the parameter space is high-dimensional. Optimization-based interval methods formulate confidence interval computation as two endpoint optimizations, where the optimization constraints can be set to achieve different types of interval calibration while seamlessly incorporating parameter constraints. However, choosing valid optimization constraints has been elusive. We show that all four proposed intervals achieve nominal coverage for a particular functional both theoretically and in practice, with numerical examples demonstrating superior performance of our intervals over the OSB interval in terms of both coverage and expected length. In particular, we show the superior performance in a realistic unfolding simulation from high-energy physics that is severely ill-posed and involves a rank-deficient forward model.

Download


Codiscovering graphical structure and functional relationships within data: A Gaussian Process framework for connecting the dots

August 2024

·

30 Reads

·

4 Citations

Proceedings of the National Academy of Sciences

Most problems within and beyond the scientific domain can be framed into one of the following three levels of complexity of function approximation. Type 1: Approximate an unknown function given input/output data. Type 2: Consider a collection of variables and functions, some of which are unknown, indexed by the nodes and hyperedges of a hypergraph (a generalized graph where edges can connect more than two vertices). Given partial observations of the variables of the hypergraph (satisfying the functional dependencies imposed by its structure), approximate all the unobserved variables and unknown functions. Type 3: Expanding on Type 2, if the hypergraph structure itself is unknown, use partial observations of the variables of the hypergraph to discover its structure and approximate its unknown functions. These hypergraphs offer a natural platform for organizing, communicating, and processing computational knowledge. While most scientific problems can be framed as the data-driven discovery of unknown functions in a computational hypergraph whose structure is known (Type 2), many require the data-driven discovery of the structure (connectivity) of the hypergraph itself (Type 3). We introduce an interpretable Gaussian Process (GP) framework for such (Type 3) problems that does not require randomization of the data, access to or control over its sampling, or sparsity of the unknown functions in a known or learned basis. Its polynomial complexity, which contrasts sharply with the super-exponential complexity of causal inference methods, is enabled by the nonlinear ANOVA capabilities of GPs used as a sensing mechanism.



Induced Seismicity Forecasting with Uncertainty Quantification: Application to the Groningen Gas Field

December 2023

·

110 Reads

·

10 Citations

Seismological Research Letters

Reservoir operations for gas extraction, fluid disposal, carbon dioxide storage, or geothermal energy production are capable of inducing seismicity. Modeling tools exist for seismicity forecasting using operational data, but the computational costs and uncertainty quantification (UQ) pose challenges. We address this issue in the context of seismicity induced by gas production from the Groningen gas field using an integrated modeling framework, which combines reservoir modeling, geomechanical modeling, and stress-based earthquake forecasting. The framework is computationally efficient thanks to a 2D finite-element reservoir model, which assumes vertical flow equilibrium, and the use of semianalytical solutions to calculate poroelastic stress changes and predict seismicity rate. The earthquake nucleation model is based on rate-and-state friction and allows for an initial strength excess so that the faults are not assumed initially critically stressed. We estimate uncertainties in the predicted number of earthquakes and magnitudes. To reduce the computational costs, we assume that the stress model is true, but our UQ algorithm is general enough that the uncertainties in reservoir and stress models could be incorporated. We explore how the selection of either a Poisson or a Gaussian likelihood influences the forecast. We also use a synthetic catalog to estimate the improved forecasting performance that would have resulted from a better seismicity detection threshold. Finally, we use tapered and nontapered Gutenberg–Richter distributions to evaluate the most probable maximum magnitude over time and account for uncertainties in its estimation. Although we did not formally account for uncertainties in the stress model, we tested several alternative stress models, and found negligible impact on the predicted temporal evolution of seismicity and forecast uncertainties. Our study shows that the proposed approach yields realistic estimates of the uncertainties of temporal seismicity and is applicable for operational forecasting or induced seismicity monitoring. It can also be used in probabilistic traffic light systems.



Induced Seismicity Forecasting with Uncertainty Quantification: Application to the Groningen Gas Field

June 2023

·

69 Reads

·

1 Citation

Reservoir operations related to natural gas extraction, fluid disposal, carbon diox- ide storage, or geothermal energy production, are capable of inducing seismicity. Mod- eling tools have been developed that allow for quantitative forecasting of seismicity based on operations data, but the computational cost of such models and the difficulty in rep- resenting various sources of uncertainties make uncertainty quantification challenging. We address this issue in the context of an integrated modeling framework, which com- bines reservoir modeling, geomechanical modeling, and stress-based earthquake forecast- ing. We use the Groningen gas field as a case example of application. The modeling frame- work is computationally efficient thanks to a 2-D finite-element reservoir model which assumes vertical flow equilibrium, and the use of semi-analytical solutions to calculate poroelastic stress changes and predict seismicity rate. The earthquake nucleation model is based on rate-and-state friction and allows for an initial strength excess so that the faults are not assumed initially critically stressed. The model parameters and their un- certainties are estimated using either a Poisson or a Gaussian likelihood. We investigate the effect of the likelihood choice on the forecast performance and we estimate uncer- tainties in the predicted number of earthquakes as well as in the expected magnitudes. We use a synthetic catalog to estimate the improved forecasting performance that would have resulted from a better seismicity detection threshold. Finally, we use tapered and non-tapered Gutenberg-Richter distributions to evaluate the most probable maximum magnitude over time and account for uncertainties in its estimation. We show that the framework yields realistic estimates of the seismicity model uncertainties and is appli- cable for operational forecasting or to design induced seismicity monitoring. It could also serve as a basis for probabilistic traffic-light systems.



Error Analysis of Kernel/GP Methods for Nonlinear and Parametric PDEs

May 2023

·

123 Reads

We introduce a priori Sobolev-space error estimates for the solution of nonlinear, and possibly parametric, PDEs using Gaussian process and kernel based methods. The primary assumptions are: (1) a continuous embedding of the reproducing kernel Hilbert space of the kernel into a Sobolev space of sufficient regularity; and (2) the stability of the differential operator and the solution map of the PDE between corresponding Sobolev spaces. The proof is articulated around Sobolev norm error estimates for kernel interpolants and relies on the minimizing norm property of the solution. The error estimates demonstrate dimension-benign convergence rates if the solution space of the PDE is smooth enough. We illustrate these points with applications to high-dimensional nonlinear elliptic PDEs and parametric PDEs. Although some recent machine learning methods have been presented as breaking the curse of dimensionality in solving high-dimensional PDEs, our analysis suggests a more nuanced picture: there is a trade-off between the regularity of the solution and the presence of the curse of dimensionality. Therefore, our results are in line with the understanding that the curse is absent when the solution is regular enough.


Kernel Methods are Competitive for Operator Learning

April 2023

·

67 Reads

·

3 Citations

We present a general kernel-based framework for learning operators between Banach spaces along with a priori error analysis and comprehensive numerical comparisons with popular neural net (NN) approaches such as Deep Operator Net (DeepONet) [Lu et al.] and Fourier Neural Operator (FNO) [Li et al.]. We consider the setting where the input/output spaces of target operator G:UV\mathcal{G}^\dagger\,:\, \mathcal{U}\to \mathcal{V} are reproducing kernel Hilbert spaces (RKHS), the data comes in the form of partial observations ϕ(ui),φ(vi)\phi(u_i), \varphi(v_i) of input/output functions vi=G(ui)v_i=\mathcal{G}^\dagger(u_i) (i=1,,Ni=1,\ldots,N), and the measurement operators ϕ:URn\phi\,:\, \mathcal{U}\to \mathbb{R}^n and φ:VRm\varphi\,:\, \mathcal{V} \to \mathbb{R}^m are linear. Writing ψ:RnU\psi\,:\, \mathbb{R}^n \to \mathcal{U} and χ:RmV\chi\,:\, \mathbb{R}^m \to \mathcal{V} for the optimal recovery maps associated with ϕ\phi and φ\varphi, we approximate G\mathcal{G}^\dagger with Gˉ=χfˉϕ\bar{\mathcal{G}}=\chi \circ \bar{f} \circ \phi where fˉ\bar{f} is an optimal recovery approximation of f:=φGψ:RnRmf^\dagger:=\varphi \circ \mathcal{G}^\dagger \circ \psi\,:\,\mathbb{R}^n \to \mathbb{R}^m. We show that, even when using vanilla kernels (e.g., linear or Mat\'{e}rn), our approach is competitive in terms of cost-accuracy trade-off and either matches or beats the performance of NN methods on a majority of benchmarks. Additionally, our framework offers several advantages inherited from kernel methods: simplicity, interpretability, convergence guarantees, a priori error estimates, and Bayesian uncertainty quantification. As such, it can serve as a natural benchmark for operator learning.


Citations (6)


... However, neural network based approaches, such as physics-informed neural networks (PINNs) [47], present significant challenges in training, due to their sensitivity to hyperparameter tuning and the substantial computational cost associated with large, often over-parameterized architectures that may include many redundant features. A notable recent development is the Gaussian process (GP) approaches [5,15,16], which extend kernel-based methods to solving nonlinear PDEs. The central idea of GP typically involves solving a linear system as an inner step combined with an outer optimization step. ...

Reference:

Solving Nonlinear PDEs with Sparse Radial Basis Function Networks
Error Analysis of Kernel/GP Methods for Nonlinear and Parametric PDEs
  • Citing Article
  • October 2024

Journal of Computational Physics

... Another connection between numerical approximation and statistical inference arises in the framework of optimal recovery introduced by Micchelli and Rivlin (1977), and its connection to Bayesian inference and GP regression through decision and game theory (Wald 1945;von Neumann 1928). Optimal recovery was initially used for solving linear PDEs (Harder and Desmarais 1972;Duchon 1977;Owhadi 2015), but was extended to nonlinear PDEs in Chen et al. (2021) and to general computational graph completion/discovery problems in Owhadi (2022) and Bourdais et al. (2024) where the connection between optimal recovery and the GP perspective on solving PDEs is made explicit. Finally, we mention the recent papers (Long et al. 2022) and Vadeboncoeur et al. (2023) where numerical errors are analyzed as Bayesian posterior measures. ...

Codiscovering graphical structure and functional relationships within data: A Gaussian Process framework for connecting the dots
  • Citing Article
  • August 2024

Proceedings of the National Academy of Sciences

... The risk of earthquakes induced by subsurface fluid injection or extraction due to oil and gas production, geothermal energy development, and CO 2 sequestration has been a growing concern (Zoback and Gorelick, 2012;Ellsworth, 2013;Lee et al., 2019). Much progress has been made in the understanding of the mechanisms at play and in the development of stressbased methods to forecast the rate of induced earthquakes (e.g., Dieterich et al., 2015;Guglielmi et al., 2015;Segall and Lu, 2015;Langenbruch and Zoback, 2016;Bourne and Oates., 2017;Kim et al., 2018;Cappa et al., 2019;Kwiatek et al., 2019;Schultz et al., 2020;Acosta et al., 2023;Kaveh et al., 2024). However, the factors determining the magnitude of induced earthquakes remain poorly understood. ...

Induced Seismicity Forecasting with Uncertainty Quantification: Application to the Groningen Gas Field

Seismological Research Letters

... FNOs have shown comparable performances for equispaced domains with respect to the vanilla Deep Operator Networks [26]. Another promising approach is Kernel Operator Learning (KOL) [4], which builds on standard kernel regression arguments to approximate the mapping between function spaces. Compared to other neural operator methodologies, the key advantage of the Kernel Operator Learning (KOL) approach lies in its non-iterative formulation; the operator is obtained by solving a (potentially large) symmetric and positive definite linear system, thereby eliminating the need for iterative training procedures typically required in neural networkbased frameworks. ...

Kernel Methods are Competitive for Operator Learning
  • Citing Article
  • October 2023

Journal of Computational Physics

... The upper bound implies that complex output patterns must have low probabilities, while high-probability outputs must be simple. Example systems in which simplicity bias has been observed include RNA structures [1,6], differential equation solutions [1], finite state transducers [2], time series patterns in natural data [7], and natural protein structures [8], among others. Which systems will and will not show simplicity bias has yet to be determined, but the phenomenon is expected to appear in a wide class of input-output maps, under fairly general conditions. ...

Multiclass classification utilising an estimated algorithmic probability prior
  • Citing Article
  • June 2023

Physica D Nonlinear Phenomena

... However, if one wishes to obtain a point estimator, it is worth noting that the midpoint of the interval can be justified from a decision-theoretic perspective. This idea has been discussed in previous works by Micchelli and Rivlin (1977); Bajgiran et al. (2022), among others. ...

Uncertainty quantification of the 4th kind; optimal posterior accuracy-uncertainty tradeoff with the minimum enclosing ball
  • Citing Article
  • December 2022

Journal of Computational Physics