Anton Rask Lundborg’s research while affiliated with IT University of Copenhagen and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (9)


The projected covariance measure for assumption-lean variable significance testing
  • Article

December 2024

·

2 Reads

·

14 Citations

The Annals of Statistics

Anton Rask Lundborg

·

Ilmun Kim

·

·

Richard J. Samworth

Algorithm-agnostic significance testing in supervised learning with multimodal data

September 2024

·

3 Reads

·

3 Citations

Briefings in Bioinformatics

Motivation Valid statistical inference is crucial for decision-making but difficult to obtain in supervised learning with multimodal data, e.g. combinations of clinical features, genomic data, and medical images. Multimodal data often warrants the use of black-box algorithms, for instance, random forests or neural networks, which impede the use of traditional variable significance tests. Results We address this problem by proposing the use of COvariance MEasure Tests (COMETs), which are calibrated and powerful tests that can be combined with any sufficiently predictive supervised learning algorithm. We apply COMETs to several high-dimensional, multimodal data sets to illustrate (i) variable significance testing for finding relevant mutations modulating drug-activity, (ii) modality selection for predicting survival in liver cancer patients with multiomics data, and (iii) modality selection with clinical features and medical imaging data. In all applications, COMETs yield results consistent with domain knowledge without requiring data-driven pre-processing, which may invalidate type I error control. These novel applications with high-dimensional multimodal data corroborate prior results on the power and robustness of COMETs for significance testing. Availability and implementation COMETs are implemented in the cometsR package available on CRAN and pycometsPython library available on GitHub. Source code for reproducing all results is available at https://github.com/LucasKook/comets. All data sets used in this work are openly available.


Model-Based Causal Feature Selection for General Response Types

August 2024

·

22 Reads

·

4 Citations

Lucas Kook

·

Sorawit Saengkyongam

·

Anton Rask Lundborg

·

[...]

·

Jonas Peters

Flow chart showing the number of children for whom propensity scores were calculated in the primary matched analyses and reasons for covariate and outcome exclusion. PIM, Primeira Infância Melhor.
Towards a better understanding of real-world home-visiting programs: a large-scale effectiveness study of parenting mechanisms in Brazil
  • Article
  • Full-text available

February 2024

·

38 Reads

·

1 Citation

Background The scale-up of parenting programmes to support early childhood development (ECD) is poorly understood. Little is known about how and when early interventions are most effective. Sustainability of ECD programming requires a better understanding of the mechanisms of real-world interventions. We examined the effects on caregiving practices of Primeira Infância Melhor (PIM), a state-wide home-visiting programme in Brazil. Methods This propensity score matched, longitudinal, quasiexperimental study uses data from the 2015 Pelotas Birth Cohort. We matched children who received PIM at any age with other cohort children on 25 key covariates. Sensitivity, guidance and responsiveness were assessed using video-recorded play tasks. Coerciveness and the parent–child relationship were assessed using the Parenting and Family Adjustment Scales. All parenting outcomes were examined at age 4 years. Separate moderation analyses were conducted for each effect modifier: family income, child age and duration of participation. Results Out of 4275 children in the cohort, 797 were enrolled in PIM up to age 4 years. 3018 children (70.6%) were included in the analytic sample, of whom 587 received PIM and 2431 were potential controls. We found a positive effect of PIM on responsiveness (β=0.08, 95% CIs 0.002 to 0.16) and sensitivity (β=0.10, 95% CIs 0.02 to 0.19). No effect was found for any secondary outcomes. Moderation analyses revealed a stronger positive effect on sensitivity for low-income parents (β=0.18, 95% CIs 0.03 to 0.34). Conclusion A state-wide, home-visiting programme in Brazil improved aspects of responsive caregiving. Effects were more pronounced for low-income families, suggesting benefits of purposeful targeting.

Download

Figure 4: Power in the alternative settings considered in Section E.1 for nominal 5%-level tests.
The Projected Covariance Measure for assumption-lean variable significance testing

November 2022

·

70 Reads

·

1 Citation

Testing the significance of a variable or group of variables X for predicting a response Y, given additional covariates Z, is a ubiquitous task in statistics. A simple but common approach is to specify a linear model, and then test whether the regression coefficient for X is non-zero. However, when the model is misspecified, the test may have poor power, for example when X is involved in complex interactions, or lead to many false rejections. In this work we study the problem of testing the model-free null of conditional mean independence, i.e. that the conditional mean of Y given X and Z does not depend on X. We propose a simple and general framework that can leverage flexible nonparametric or machine learning methods, such as additive models or random forests, to yield both robust error control and high power. The procedure involves using these methods to perform regressions, first to estimate a form of projection of Y on X and Z using one half of the data, and then to estimate the expected conditional covariance between this projection and Y on the remaining half of the data. While the approach is general, we show that a version of our procedure using spline regression achieves what we show is the minimax optimal rate in this nonparametric testing problem. Numerical experiments demonstrate the effectiveness of our approach both in terms of maintaining Type I error control, and power, compared to several existing approaches.


Conditional Independence Testing in Hilbert Spaces with Applications to Functional Data Analysis

November 2022

·

47 Reads

·

21 Citations

Journal of the Royal Statistical Society Series B (Statistical Methodology)

We study the problem of testing the null hypothesis that X and Y are conditionally independent given Z, where each of X, Y and Z may be functional random variables. This generalises testing the significance of X in a regression model of scalar response Y on functional regressors X and Z. We show, however, that even in the idealised setting where additionally (X, Y, Z) has a Gaussian distribution, the power of any test cannot exceed its size. Further modelling assumptions are needed and we argue that a convenient way of specifying these assumptions is based on choosing methods for regressing each of X and Y on Z. We propose a test statistic involving inner products of the resulting residuals that is simple to compute and calibrate: type I error is controlled uniformly when the in‐sample prediction errors are sufficiently small. We show this requirement is met by ridge regression in functional linear model settings without requiring any eigen‐spacing conditions or lower bounds on the eigenvalues of the covariance of the functional regressor. We apply our test in constructing confidence intervals for truncation points in truncated functional linear models and testing for edges in a functional graphical model for EEG data.


Modern Methods for Variable Significance Testing

October 2022

·

25 Reads

·

3 Citations

This thesis concerns the ubiquitous statistical problem of variable significance testing. The first chapter contains an account of classical approaches to variable significance testing including different perspectives on how to formalise the notion of `variable significance'. The historical development is contrasted with more recent methods that are adapted to both the scale of modern datasets but also the power of advanced machine learning techniques. This chapter also includes a description of and motivation for the theoretical framework that permeates the rest of the thesis: providing theoretical guarantees that hold uniformly over large classes of distributions. The second chapter deals with testing the null that Y ⊥ X | Z where X and Y take values in separable Hilbert spaces with a focus on applications to functional data. The first main result of the chapter shows that for functional data it is impossible to construct a non-trivial test for conditional independence even when assuming that the data are jointly Gaussian. A novel regression-based test, called the Generalised Hilbertian Covariance Measure (GHCM), is presented and theoretical guarantees for uniform asymptotic Type I error control are provided with the key assumption requiring that the product of the mean squared errors of regressing Y on Z and X on Z converges faster than n1^{-1}, where n is the sample size. A power analysis is conducted under the same assumptions to illustrate that the test has uniform power over local alternatives where the expected conditional covariance operator has a Hilbert--Schmidt norm going to 0 at a nn\sqrt[n]{n}-rate. The chapter also contains extensive empirical evidence in the form of simulations demonstrating the validity and power properties of the test. The usefulness of the test is demonstrated by using the GHCM to construct confidence intervals for the boundary point in a truncated functional linear model and to detect edges in a graphical model for an EEG dataset. The third and final chapter analyses the problem of nonparametric variable significance testing by testing for conditional mean independence, that is, testing the null that E(Y | X, Z) = E(Y | Z) for real-valued Y. A test, called the Projected Covariance Measure (PCM), is derived by considering a family of studentised test statistics and choosing a member of this family in a data-driven way that balances robustness and power properties of the resulting test. The test is regression-based and is computed by splitting a set of observations of (X, Y, Z) into two sets of equal size, where one half is used to learn a projection of Y onto X and Z (nonparametrically) and the second half is used to test for vanishing expected conditional correlation given Z between the projection and Y. The chapter contains general conditions that ensure uniform asymptotic Type I control of the resulting test by imposing conditions on the mean-squared error of the involved regressions. A modification of the PCM using additional sample splitting and employing spline regression is shown to achieve the minimax optimal separation rate between null and alternative under Hölder smoothness assumptions on the regression functions and the conditional density of X given Z=z. The chapter also shows through simulation studies that the test maintains the strong type I error control of methods like the Generalised Covariance Measure (GCM) but has power against a broader class of alternatives.


Fig 1. (a) M-bias: U 1 , U 2 are unobserved and S = {L}. It holds that A ⊥ ⊥ Y (a) for a = 0, 1 unconditionally. However, controlling for L opens the back-door path A ← U 1 → L ← U 2 → Y and hence would introduce a bias. (b) Suppose S = {X 1 , X 2 } and we have R A (S) = {X 1 }, R Y |A (S) = {X 2 }. Using the definitions from Section 3.3, we note that the conjunctive C∩(S) = R A (S) ∩ R Y |A (S) = ∅ fails to control for confounding. Meanwhile, C AY (S) = {X 1 }, C YA (S) = {X 2 } and C∪(S) = {X 1 , X 2 } all suffice to control for confounding.
Confounder Selection: Objectives and Approaches

August 2022

·

59 Reads

Confounder selection is perhaps the most important step in the design of observational studies. A number of criteria, often with different objectives and approaches, have been proposed, and their validity and practical value have been debated in the literature. Here, we provide a unified review of these criteria and the assumptions behind them. We list several objectives that confounder selection methods aim to achieve and discuss the amount of structural knowledge required by different approaches. Finally, we discuss limitations of the existing approaches and implications for practitioners.


Figure 2: As Figure 1 but for the alternative settings, see (21).
Figure 3: Rejection rates against significance level for the pfr (red) and GHCM (green) tests under null (light) and alternative (dark) settings when n = 500.
Figure 5: Histograms of the left endpoints of 95% confidence intervals for truncation points θ = 0.275 (left) and θ = 0.675 (right), given by red vertical lines, in model (22) across 500 simulations.
Conditional Independence Testing in Hilbert Spaces with Applications to Functional Data Analysis

January 2021

·

216 Reads

We study the problem of testing the null hypothesis that X and Y are conditionally independent given Z, where each of X, Y and Z may be functional random variables. This generalises, for example, testing the significance of X in a scalar-on-function linear regression model of response Y on functional regressors X and Z. We show however that even in the idealised setting where additionally (X, Y, Z) have a non-singular Gaussian distribution, the power of any test cannot exceed its size. Further modelling assumptions are needed to restrict the null and we argue that a convenient way of specifying these is based on choosing methods for regressing each of X and Y on Z. We thus propose as a test statistic, the Hilbert-Schmidt norm of the outer product of the resulting residuals, and prove that type I error control is guaranteed when the in-sample prediction errors are sufficiently small. We show this requirement is met by ridge regression in functional linear model settings without requiring any eigen-spacing conditions or lower bounds on the eigenvalues of the covariance of the functional regressor. We apply our test in constructing confidence intervals for truncation points in truncated functional linear models.

Citations (6)


... This "sample splitting plus self normalization" is a fundamental principle of the DA approach. After its introduction in Kim and Ramdas (2024), this technique for DA inference (as opposed to just inference) has been successfully applied to various high-dimensional inference problems (e.g., Liu et al., 2022;Shekhar et al., 2022Shekhar et al., , 2023Gao et al., 2023;Zhang and Shao, 2024;Lundborg et al., 2024;Liu et al., 2024;Zhang et al., 2025;Takatsu and Kuchibhotla, 2025). By extending this framework to the discrete argmin inference problem, our work ensures asymptotic validity under mild moment conditions and achieves minimax-optimal power across both low-and high-dimensional regimes, even for heavy-tailed data. ...

Reference:

Locally minimax optimal and dimension-agnostic discrete argmin inference
The projected covariance measure for assumption-lean variable significance testing
  • Citing Article
  • December 2024

The Annals of Statistics

... Possible downstream tasks where such embeddings are useful can be of statistical nature, e.g. conditional independence testing for high-dimensional data such as images (Simnacher et al., 2024, Kook andLundborg, 2024). Alternatively, discrete extraction of features could apply if high-dimensional data are structured in clusters, and where the variability inside a cluster might not be relevant to the data processing task. ...

Algorithm-agnostic significance testing in supervised learning with multimodal data
  • Citing Article
  • September 2024

Briefings in Bioinformatics

... What differentiated ICP from other methods for causal discovery using both interventional and observational data, is that ICP does not require the analyst to know where the interventions are performed, but instead only to know from which setting a particular data point comes. This powerful idea was later extended to non-linear models [22], dynamical systems [36], time series data [37], spatio-temporal data [9] and different outcome models [26]. Moreover, the invariant principle has been also used together with adjacent machine learning methods: active learning [12], to find the causal features with experimental data efficiently; and reinforcement learning [40], for policy learning. ...

Model-Based Causal Feature Selection for General Response Types
  • Citing Article
  • August 2024

... A recent systematic review identified the PCF home-visiting program as an effective strategy for strengthening social support, developing parenting skills, and enhancing family bonds among vulnerable families in Brazil [27]. Another public home-visiting initiative, the Better Early Childhood Program (Programa Primeira Infância Melhor), has demonstrated improvements in responsive caregiving, particularly among low-income families [47]. Evaluations of large-scale parenting programs in Latin America have shown promising results in enhancing caregiving practices in countries such as Colombia [48], Mexico [49], and Peru [50], suggesting that home-visiting programs are able to improve low-income families' quality of life. ...

Towards a better understanding of real-world home-visiting programs: a large-scale effectiveness study of parenting mechanisms in Brazil

... To the best of our knowledge, most existing CMI tests focus on univariate Y and face one or more of the following three major issues: (1) finite-sample performance deteriorates when some or all of (X, Y, Z) are high-dimensional; (2) the tests lack theoretical size guarantees in general nonparametric settings; and (3) they exhibit weak power in detecting local alternatives. For a recent survey, see Lundborg (2022). We discuss how these challenges arise and how existing CMI tests have partially addressed them. ...

Modern Methods for Variable Significance Testing
  • Citing Thesis
  • October 2022

... This would typically be a uniformly bounded p1`δq th or p2`δq th moment for some δ ą 0 in the cases of laws of large numbers and central limit theorems, respectively. See [8,61,52,35] for an incomplete list of examples. Let us now provide a distribution-uniform bound on the expected rejection time. ...

Conditional Independence Testing in Hilbert Spaces with Applications to Functional Data Analysis

Journal of the Royal Statistical Society Series B (Statistical Methodology)