Moritz Grosse-Wentrup’s research while affiliated with University of Vienna and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (122)


Figure 1. Overall model performance of the EEGNet, shallow ConvNet and deep ConvNet after 100, 200 and 300 training iterations. Each data point represents the mean of the ROC and accuracy scores over all three paradigms. The error bars show the bootstrapped 95% confidence interval.
Figure 2. Overall model performance of the EEGNet compared to the shallow & deep ConvNets, the Riemannian minimum distance to mean (RMDM) and tangent space mapping (TGSP) approaches. Each point represents an individual classification score. The lower axis shows the bootstrapped 95% confidence interval of the mean difference between the EEGNet and the other pipelines. A negative mean difference indicates the respective pipeline performing worse than the EEGNet.
Figure 3. Comparative analysis of classification performances between the EEGNet and the respectively best-performing Riemannian TGSP method. Each point represents an individual classification score. The distinct colors indicate the evaluation method.
Figure A1. Gardner-Altman estimation plot for Motor Imagery within-session evaluation. Each point represents an individual classification score. The shadings represent individual sessions within a dataset. The lower axis shows the bootstrapped 95% confidence interval of the mean difference between the EEGNet and the other pipelines.
Benchmarking brain-computer interface algorithms: Riemannian approaches vs convolutional neural networks
  • Article
  • Full-text available

July 2024

·

18 Reads

Manuel Eder

·

Jiachen Xu

·

Moritz Grosse-Wentrup

Objective. To date, a comprehensive comparison of Riemannian decoding methods with deep convolutional neural networks for EEG-based Brain-Computer Interfaces remains absent from published work. We address this research gap by using MOABB, The Mother Of All BCI Benchmarks, to compare novel convolutional neural networks to state-of-the-art Riemannian approaches across a broad range of EEG datasets, including motor imagery, P300, and steady-state visual evoked potentials paradigms. Approach. We systematically evaluated the performance of convolutional neural networks, specifically EEGNet, shallow ConvNet, and deep ConvNet, against well-established Riemannian decoding methods using MOABB processing pipelines. This evaluation included within- session, cross-session, and cross-subject methods, to provide a practical analysis of model effectiveness and to find an overall solution that performs well across different experimental settings. Main results. We find no significant differences in decoding performance between convolutional neural networks and Riemannian methods for within-session, cross-session, and cross-subject analyses. Significance. The results show that, when using traditional Brain-Computer Interface paradigms, the choice between CNNs and Riemannian methods may not heavily impact decoding performances in many experimental settings. These findings provide researchers with flexibility in choosing decoding approaches based on factors such as ease of implementation, computational efficiency or individual preferences.

Download

Neuro-Cognitive Multilevel Causal Modeling: A Framework that Bridges the Explanatory Gap between Neuronal Activity and Cognition

October 2023

·

10 Reads

Explaining how neuronal activity gives rise to cognition arguably remains the most significant challenge in cognitive neuroscience. We introduce neuro-cognitive multilevel causal modeling (NC-MCM), a framework that bridges the explanatory gap between neuronal activity and cognition by construing cognitive states as (behaviorally and dynamically) causally consistent abstractions of neuronal states. Multilevel causal modeling allows us to interchangeably reason about the neuronal- and cognitive causes of behavior while maintaining a physicalist (in contrast to a strong dualist) position. We introduce an algorithm for learning cognitive-level causal models from neuronal activation patterns and demonstrate its ability to learn cognitive states of the nematode C. elegans from calcium imaging data. We show that the cognitive-level model of the NC-MCM framework provides a concise representation of the neuronal manifold of C. elegans and its relation to behavior as a graph, which, in contrast to other neuronal manifold learning algorithms, supports causal reasoning. We conclude the article by arguing that the ability of the NC-MCM framework to learn causally interpretable abstractions of neuronal dynamics and their relation to behavior in a purely data-driven fashion is essential for understanding more biological systems whose complexity prohibits the development of hand-crafted computational models.



BundDLe-Net: Neuronal Manifold Learning Meets Behaviour

August 2023

·

111 Reads

Neuronal manifold learning techniques represent high-dimensional neuronal dynamics in low-dimensional embeddings to reveal the intrinsic structure of neuronal manifolds. Common to these techniques is their goal to learn low-dimensional embeddings that preserve all dynamic information in the high-dimensional neuronal data, i.e., embeddings that allow for reconstructing the original data. We introduce a novel neuronal manifold learning technique, BunDLe-Net, that learns a low-dimensional Markovian embedding of the neuronal dynamics which preserves only those aspects of the neuronal dynamics that are relevant for a given behavioural context. In this way, BunDLe-Net eliminates neuronal dynamics that are irrelevant to decoding behaviour, effectively de-noising the data to reveal better the intricate relationships between neuronal dynamics and behaviour. We demonstrate the quantitative superiority of BunDLe-Net over commonly used and state-of-the-art neuronal manifold learning techniques in terms of dynamic and behavioural information in the learned manifold on calcium imaging data recorded in the nematode C. elegans. Qualitatively, we show that BunDLe-Net learns highly consistent manifolds across multiple worms that reveal the neuronal and behavioural motifs that form the building blocks of the neuronal manifold.



Improvement-Focused Causal Recourse (ICR)

June 2023

·

10 Reads

·

10 Citations

Proceedings of the AAAI Conference on Artificial Intelligence

Algorithmic recourse recommendations inform stakeholders of how to act to revert unfavorable decisions. However, existing methods may recommend actions that lead to acceptance (i.e., revert the model's decision) but do not lead to improvement (i.e., may not revert the underlying real-world state). To recommend such actions is to recommend fooling the predictor. We introduce a novel method, Improvement-Focused Causal Recourse (ICR), which involves a conceptual shift: Firstly, we require ICR recommendations to guide toward improvement. Secondly, we do not tailor the recommendations to be accepted by a specific predictor. Instead, we leverage causal knowledge to design decision systems that predict accurately pre- and post-recourse, such that improvement guarantees translate into acceptance guarantees. Curiously, optimal pre-recourse classifiers are robust to ICR actions and thus suitable post-recourse. In semi-synthetic experiments, we demonstrate that given correct causal knowledge ICR, in contrast to existing approaches, guides toward both acceptance and improvement.


Efficient SAGE Estimation via Causal Structure Learning

April 2023

·

40 Reads

The Shapley Additive Global Importance (SAGE) value is a theoretically appealing interpretability method that fairly attributes global importance to a model's features. However, its exact calculation requires the computation of the feature's surplus performance contributions over an exponential number of feature sets. This is computationally expensive, particularly because estimating the surplus contributions requires sampling from conditional distributions. Thus, SAGE approximation algorithms only take a fraction of the feature sets into account. We propose d-SAGE, a method that accelerates SAGE approximation. d-SAGE is motivated by the observation that conditional independencies (CIs) between a feature and the model target imply zero surplus contributions, such that their computation can be skipped. To identify CIs, we leverage causal structure learning (CSL) to infer a graph that encodes (conditional) independencies in the data as d-separations. This is computationally more efficient because the expense of the one-time graph inference and the d-separation queries is negligible compared to the expense of surplus contribution evaluations. Empirically we demonstrate that d-SAGE enables the efficient and accurate estimation of SAGE values.



Improvement-Focused Causal Recourse (ICR)

October 2022

·

32 Reads

Algorithmic recourse recommendations, such as Karimi et al.'s (2021) causal recourse (CR), inform stakeholders of how to act to revert unfavourable decisions. However, some actions lead to acceptance (i.e., revert the model's decision) but do not lead to improvement (i.e., may not revert the underlying real-world state). To recommend such actions is to recommend fooling the predictor. We introduce a novel method, Improvement-Focused Causal Recourse (ICR), which involves a conceptual shift: Firstly, we require ICR recommendations to guide towards improvement. Secondly, we do not tailor the recommendations to be accepted by a specific predictor. Instead, we leverage causal knowledge to design decision systems that predict accurately pre- and post-recourse. As a result, improvement guarantees translate into acceptance guarantees. We demonstrate that given correct causal knowledge, ICR, in contrast to existing approaches, guides towards both acceptance and improvement.


General Pitfalls of Model-Agnostic Interpretation Methods for Machine Learning Models

April 2022

·

206 Reads

·

115 Citations

Lecture Notes in Computer Science

An increasing number of model-agnostic interpretation techniques for machine learning (ML) models such as partial dependence plots (PDP), permutation feature importance (PFI) and Shapley values provide insightful model interpretations, but can lead to wrong conclusions if applied incorrectly. We highlight many general pitfalls of ML model interpretation, such as using interpretation techniques in the wrong context, interpreting models that do not generalize well, ignoring feature dependencies, interactions, uncertainty estimates and issues in high-dimensional settings, or making unjustified causal interpretations, and illustrate them with examples. We focus on pitfalls for global methods that describe the average model behavior, but many pitfalls also apply to local methods that explain individual predictions. Our paper addresses ML practitioners by raising awareness of pitfalls and identifying solutions for correct model interpretation, but also addresses ML researchers by discussing open issues for further research.


Citations (70)


... Abid is no more creditworthy than he was before, but the decisioncriteria will classify him as creditworthy. To remedy this, some authors have proposed that explanations have to provide causal information so that individuals can improve outcomes (30,39). Finally, one might doubt the ability of explainable AI to promote recourse because it only approximates the underlying model. ...

Reference:

Transparency and Explainability for Public Policy
Improvement-Focused Causal Recourse (ICR)
  • Citing Article
  • June 2023

Proceedings of the AAAI Conference on Artificial Intelligence

... The second part of the third stage describes the main classification metrics used for evaluating the model performance and ML classification performance metrics-accuracy, precision, recall, F1-score and area under receiver operating characteristic (AUROC). Finally, the last part of the stage includes the extraction of the most relevant features using Permutation Feature Importance (PFI) [50] that builds upon the Mean Dropout Loss (MDL) metric [51]. The second stage explains the Multiple Correspondence Analysis (MCA) procedure for reducing the dimensionality of a dataset into two-dimensional space. ...

General Pitfalls of Model-Agnostic Interpretation Methods for Machine Learning Models

Lecture Notes in Computer Science

... Notably, the results of the CNN pipelines show a more pronounced bimodal distribution compared to the Riemannian pipelines. The observation that deep convolutional networks may be particularly useful in transfer learning settings is in line with the results of a recent BCI decoding competition that was also won by a CNN approach [37]. However, we remark that the methods benchmarked here were not explicitly designed for cross-subject decoding. ...

2021 BEETL Competition: Advancing Transfer Learning for Subject Independence & Heterogenous EEG Data Sets
  • Citing Preprint
  • February 2022

... The MOABB -the Mother Of All BCI Benchmarks -is an open-source package for benchmarking BCIs written in Python that allows such a comparison across a broad range of publicly available datasets [9]. While MOABB has been used to compare various machine learning pipelines, two of the most successful machine learning frameworks for BCI decoding, Riemannian decoding [10] and deep convolutional networks [11,12], have not yet been rigorously compared with each other [13]. ...

Workshops of the eighth international brain-computer interface meeting: BCIs: the next frontier Workshops of the eighth international brain-computer interface meeting: BCIs: the next frontier

Brain-Computer Interfaces

... /fnhum. . that act in an Euclidean space, the most commonly employed in the BCI community being the linear discriminant analysis (LDA; Barachant et al., 2012), support vector classifier (SVC; Xu et al., 2021) and Lasso logistic regression (LR) (Tomioka et al., 2006). In this study, we use the SVC. ...

Distance Covariance: A Nonlinear Extension of Riemannian Geometry for EEG-based Brain-Computer Interfacing
  • Citing Conference Paper
  • October 2021

... Immersive platforms provide more realistic results pertaining to human interaction, as compared to visualizations on flat screens [39]. Virtual Reality (VR) systems such as Arm-Sym [27] and the Virtual Integration Environment or VIE [25] take advantage of this. The HoloPHAM platform [28] uses AR for direct EMG control training. ...

ArmSym: A Virtual Human–Robot Interaction Laboratory for Assistive Robotics
  • Citing Article
  • September 2021

IEEE Transactions on Human-Machine Systems

... If fine motor tracking is performed during upper limb rehabilitation, and if wearable or nonwearable equipment are used, virtual rehabilitation research can be characterised (Stetz et al. 2011;Shum et al. 2019). Hand tracking is possible using a VR HMD, which has already been released as a commercial product (e.g., Oculus Quest) in the case of camera approaches among nonwearable devices (McDermott et al. 2021). Table 2 shows the key characteristics of the meta-analyses that were included. ...

Predicting motor behavior: an efficient EEG signal processing pipeline to detect brain states with potential therapeutic relevance for VR-based neurorehabilitation

... Conducting feature importance analysis allows for the identification of the most significant features or nodes contributing to the model's predictions [40]. By understanding which inputs have the greatest impact on the model's decisions, stakeholders can better comprehend the decision-making process, thereby enhancing the model's interpretability. ...

Relative Feature Importance
  • Citing Conference Paper
  • January 2021

... Research on INTs has revealed that the brain exhibits a temporal hierarchical organization: unimodal regions accumulate and process information over shorter timescales, while transmodal regions integrate and process stimuli across longer time durations (Golesorkhi et al., 2021a). INTs have recently been posited as a key mechanism for consciousness in the Temporo-spatial Theory of Consciousness (TTC) (Northoff and Zilio, 2022;Zilio et al., 2023Zilio et al., , 2021, according to which they allow the brain to encode contents within their respective contexts. In this framework, TTC posits INTs as giving rise to the structure of consciousness, representing the background in the figurebackground dyad. ...

Are intrinsic neural timescales related to sensory processing? Evidence from abnormal behavioral states

NeuroImage

... Electroencephalography (EEG) enables investigations into brain function and health in an economical manner and for a wide array of purposes, including sleep monitoring, pathology screening, neurofeedback, brain-computer interfacing and anaesthesia monitoring [1,2,3,4,5,6]. Thanks to recent advances in mobile EEG technology, these applications can now be more easily translated from the lab and clinic to contexts such as at-home or ambulatory assessments. ...

MYND: Unsupervised Evaluation of Novel BCI Control Strategies on Consumer Hardware
  • Citing Conference Paper
  • October 2020