Michaël Defferrard’s research while affiliated with Swiss Federal Institute of Technology in Lausanne and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (21)


Deep learning for classifying neuronal morphologies: combining topological data analysis and graph neural networks
  • Preprint

September 2024

·

32 Reads

·

1 Citation

·

Stanislav Schmidt

·

Francesco Casalegno

·

[...]

·

The shape of neuronal morphologies plays a critical role in determining their dynamical properties and the functionality of the brain. With an abundance of neuronal morphology reconstructions, a robust definition of cell types is important to understand their role in brain functionality. However, an objective morphology classification scheme is hard to establish due to disagreements on the definition of cell types, on which subjective views of field experts show significant differences. The robust grouping of neurons based on their morphological shapes is important for generative models and for establishing a link between anatomical properties and other modalities, such as biophysical and transcriptomic information. We combine deep learning techniques with a variety of mathematical descriptions of neurons and evaluate the classification accuracy of different methods. We demonstrate that various methodologies, including graph neural networks, topological morphology descriptors, and morphometrics, consistently perform with the highest accuracy for a variety of datasets. Based on these methods, we present a robust classification of both inhibitory and excitatory cell types in the rodent cortex and propose a generalized scheme for a consistent classification of neurons into classes.


Computation of the surface similarity score (SurfS)
Protein surface is converted into a point cloud where each point is used to compute shape and electrostatic features. To compute the SurfS score, all individual points of the point clouds are compared and the shape similarity value is derived from closest points of the two surfaces in space while electrostatic similarity is evaluated by correlation analysis of the electrostatic potentials of both surfaces.
Single mutant discrimination using surface similarity score in protein-protein complexes
A) Surface similarity evaluation protocol for single amino acids. B) Recovery for all 19 considered amino acid types in bound (top) and unbound (bottom) complex states, evaluated with four different metrics: SS (shape similarity), REU (Rosetta energy unit), ES (electrostatic similarity), and SurfS (surface similarity). C) Average surface similarity score when performing all-against-all amino acid comparison for bound (top) and unbound (bottom) complex states. The highest mean SurfS score for every amino acid is highlighted.
Sequence recovery of protein interfaces
A) Sequence recovery benchmark pipeline. Sequences in the interfaces of protein-protein complexes are evaluated in the presence and absence of the binding partner. The tested complexes were grouped into interfaces with low and high shape complementarity, and antibody-antigen complexes. Surface-centric design (RosettaSurf and RosettaSurf-site) is compared to a standard structural protein design protocol (FixBB). B) Interface sequence recovery of the complete dataset. C) Sequence recovery of low shape complementarity interfaces. D) Sequence recovery of high shape complementarity interfaces. E) Sequence recovery of antigen-antibody complexes. Dashed lines represent median and triangles represent mean recovery values.
Comparison of SSM data obtained for the designed interleukin-2/15 antagonists in comparison to RosettaSurf-site predictions
The structure highlights the five selected positions of the interleukin design that were computationally and experimentally sampled. Different sampling results of the experimental SSM, RosettaSurf-site with SC, and Rosetta’s energy function are reported in the table. Mutations resulting in the experimentally reported best binding design are highlighted in dark green. RosettaSurf-site was able to recover four out of five key binding mutations (dark green) while evaluating mutations with Rosetta’s energy function could only retrieve one binding mutation at position 98. Additionally, RosettaSurf-site was able to recover four affinity improving mutations not present in the best binding design, whereas Rosetta’s energy function could identify only two of these mutations (light green).
Surface centric design of a viral antigenic site present the in RSVF
A) Design process of site 0-mimicking protein scaffolds. Starting scaffolds are selected from the PDB based on structural alignments with the epitope helix. The surface mimicking designs are generated by grafting the side chains of the helix segment of the epitope onto the scaffold and surface-centric design is employed to optimize the loop region. Before and after design of the surface compared to native site 0. Blue areas indicate high similarity. B) Mimicry of surface geometry of WT scaffold, RSV_FixBB, and Surf_03 designs compared to native site 0. C) Representative SPR measurements of Surf_03 and RSV_FixBB against site 0-specific antibodies D25 and ADI14496. D) Binding profiles of Surf_03, a FixBB designed protein, and a helix-only design against a panel of site-specific antibodies with green indicating binding and red cells corresponding to non-binding. A knockout mutant of Surf_03 and the WT protein binding profiles are listed as reference.
RosettaSurf—A surface-centric computational design approach
  • Article
  • Full-text available

March 2022

·

167 Reads

·

4 Citations

Proteins are typically represented by discrete atomic coordinates providing an accessible framework to describe different conformations. However, in some fields proteins are more accurately represented as near-continuous surfaces, as these are imprinted with geometric (shape) and chemical (electrostatics) features of the underlying protein structure. Protein surfaces are dependent on their chemical composition and, ultimately determine protein function, acting as the interface that engages in interactions with other molecules. In the past, such representations were utilized to compare protein structures on global and local scales and have shed light on functional properties of proteins. Here we describe RosettaSurf, a surface-centric computational design protocol, that focuses on the molecular surface shape and electrostatic properties as means for protein engineering, offering a unique approach for the design of proteins and their functions. The RosettaSurf protocol combines the explicit optimization of molecular surface features with a global scoring function during the sequence design process, diverging from the typical design approaches that rely solely on an energy scoring function. With this computational approach, we attempt to address a fundamental problem in protein design related to the design of functional sites in proteins, even when structurally similar templates are absent in the characterized structural repertoire. Surface-centric design exploits the premise that molecular surfaces are, to a certain extent, independent of the underlying sequence and backbone configuration, meaning that different sequences in different proteins may present similar surfaces. We benchmarked RosettaSurf on various sequence recovery datasets and showcased its design capabilities by generating epitope mimics that were biochemically validated. Overall, our results indicate that the explicit optimization of surface features may lead to new routes for the design of functional proteins.

Download

ChebLieNet: Invariant Spectral Graph NNs Turned Equivariant by Riemannian Geometry on Lie Groups

November 2021

·

170 Reads

We introduce ChebLieNet, a group-equivariant method on (anisotropic) manifolds. Surfing on the success of graph- and group-based neural networks, we take advantage of the recent developments in the geometric deep learning field to derive a new approach to exploit any anisotropies in data. Via discrete approximations of Lie groups, we develop a graph neural network made of anisotropic convolutional layers (Chebyshev convolutions), spatial pooling and unpooling layers, and global pooling layers. Group equivariance is achieved via equivariant and invariant operators on graphs with anisotropic left-invariant Riemannian distance-based affinities encoded on the edges. Thanks to its simple form, the Riemannian metric can model any anisotropies, both in the spatial and orientation domains. This control on anisotropies of the Riemannian metrics allows to balance equivariance (anisotropic metric) against invariance (isotropic metric) of the graph convolution layers. Hence we open the doors to a better understanding of anisotropic properties. Furthermore, we empirically prove the existence of (data-dependent) sweet spots for anisotropic parameters on CIFAR10. This crucial result is evidence of the benefice we could get by exploiting anisotropic properties in data. We also evaluate the scalability of this approach on STL10 (image data) and ClimateNet (spherical data), showing its remarkable adaptability to diverse tasks.


Figure 3. Comparison of SSM data obtained for the designed interleukin-2/15 antagonists in comparison to
Figure 6. Computation of the surface similarity score (Surf S ). Protein surface is converted into a point cloud where
RosettaSurf - a surface-centric computational design approach

June 2021

·

171 Reads

Proteins are typically represented by discrete atomic coordinates providing an accessible framework to describe different conformations. However, in some fields proteins are more accurately represented as near-continuous surfaces, as these are imprinted with geometric (shape) and chemical (electrostatics) features of the underlying protein structure. Protein surfaces are dependent on their chemical composition and, ultimately determine protein function, acting as the interface that engages in interactions with other molecules. In the past, such representations were utilized to compare protein structures on global and local scales and have shed light on functional properties of proteins. Here we describe RosettaSurf, a surface-centric computational design protocol, that focuses on the molecular surface shape and electrostatic properties as means for protein engineering, offering a unique approach for the design of proteins and their functions. The RosettaSurf protocol combines the explicit optimization of molecular surface features with a global scoring function during the sequence design process, diverging from the typical design approaches that rely solely on an energy scoring function. With this computational approach, we attempt to address a fundamental problem in protein design related to the design of functional sites in proteins, even when structurally similar templates are absent in the characterized structural repertoire. Surface-centric design exploits the premise that molecular surfaces are, to a certain extent, independent of the underlying sequence and backbone configuration, meaning that different sequences in different proteins may present similar surfaces. We benchmarked RosettaSurf on various sequence recovery datasets and showcased its design capabilities by generating epitope mimics that were biochemically validated. Overall, our results indicate that the explicit optimization of surface features may lead to new routes for the design of functional proteins. Author Summary Finely orchestrated protein-protein interactions are at the heart of virtually all fundamental cellular processes. Altering these processes or encoding new functions in proteins has been a long-standing goal in computational protein design. Protein design methods commonly rely on scoring functions that seek to identify amino acid sequences that optimize structural configurations of atoms while minimizing a variety of physics-based and statistical terms. The objectives of the large majority of computational design protocols have been focused on obtaining a predefined structural conformation. However, routinely introducing a functional aspect on designer proteins has been more challenging. Our results suggest that the molecular surface features can be a useful optimization parameter to guide the design process towards functional surfaces that mimic known protein binding sites and interact with their intended targets. Specifically, we demonstrate that our design method can optimize experimental libraries through computational screening, creating a basis for highly specific protein binders, as well as design a potent immunogen that engages with site-specific antibodies. The ability to create proteins with novel functions will be transformative for biomedical applications, providing many opportunities for the design of novel immunogens, protein components for synthetic biology, and other protein-based biotechnologies.


Learning to recover orientations from projections in single-particle cryo-EM

April 2021

·

54 Reads

A major challenge in single-particle cryo-electron microscopy (cryo-EM) is that the orientations adopted by the 3D particles prior to imaging are unknown; yet, this knowledge is essential for high-resolution reconstruction. We present a method to recover these orientations directly from the acquired set of 2D projections. Our approach consists of two steps: (i) the estimation of distances between pairs of projections, and (ii) the recovery of the orientation of each projection from these distances. In step (i), pairwise distances are estimated by a Siamese neural network trained on synthetic cryo-EM projections from resolved bio-structures. In step (ii), orientations are recovered by minimizing the difference between the distances estimated from the projections and the distances induced by the recovered orientations. We evaluated the method on synthetic cryo-EM datasets. Current results demonstrate that orientations can be accurately recovered from projections that are shifted and corrupted with a high level of noise. The accuracy of the recovery depends on the accuracy of the distance estimator. While not yet deployed in a real experimental setup, the proposed method offers a novel learning-based take on orientation recovery in SPA. Our code is available at https://github.com/JelenaBanjac/protein-reconstruction


DeepSphere: a graph-based spherical CNN

December 2020

·

136 Reads

·

2 Citations

Designing a convolution for a spherical neural network requires a delicate tradeoff between efficiency and rotation equivariance. DeepSphere, a method based on a graph representation of the sampled sphere, strikes a controllable balance between these two desiderata. This contribution is twofold. First, we study both theoretically and empirically how equivariance is affected by the underlying graph with respect to the number of vertices and neighbors. Second, we evaluate DeepSphere on relevant problems. Experiments show state-of-the-art performance and demonstrates the efficiency and flexibility of this formulation. Perhaps surprisingly, comparison with previous work suggests that anisotropic filters might be an unnecessary price to pay. Our code is available at https://github.com/deepsphere


Simplicial Neural Networks

October 2020

·

91 Reads

We present simplicial neural networks (SNNs), a generalization of graph neural networks to data that live on a class of topological spaces called simplicial complexes. These are natural multi-dimensional extensions of graphs that encode not only pairwise relationships but also higher-order interactions between vertices - allowing us to consider richer data, including vector fields and n-fold collaboration networks. We define an appropriate notion of convolution that we leverage to construct the desired convolutional neural networks. We test the SNNs on the task of imputing missing data on coauthorship complexes.


Figure 1. a: Illustration of workflow, from left to right. The structural connectivity matrix encodes a graph in which each node is a brain region defined by a parcellation (see brain surface below matrix; brain regions, nodes, and matrix rows/columns are color-coded) and edges are defined by white matter connectivity. The graph Laplacian is computed from this matrix, and network harmonics are obtained as eigenvectors of this graph Laplacian. The eigenvectors are ordered by ascending eigenvalue. b: The first four network harmonics in vector form (magnified from panel a), projected onto the surface of the brain, and in graph representation. Colors visualize arbitrary units, i.e. the weights in the orthonormal eigenvectors. c: The graph Laplacian can be reconstructed as a weighted sum of rank-1 matrices defined by the outer products of its eigenvectors. Large values of equal sign in the eigenvector (network harmonic) lead to large positive weights in the outer product (red entries in the illustration). d: The correlation between the (upper or lower triangle) of the reconstructed matrices * ^ obtained from the outer products and the original graph Laplacian * is used to quantify how well the eigenvectors capture the SC, both when they are used cumulatively (circles) and on their own (crosses). Open blue circles and black crosses mark non-significant correlations for each case. e: Two brain regions, here: inferior parietal, are close together in network harmonic 2 (x-axis), but far apart in network harmonic 3 (y-axis), illustrating how network harmonics capture integration and segregation in multiple dimensions.
Figure 2. a: EEG signal averaged over all trials and all subjects in its original domain (top panel, rows are brain regions) and in the graph frequency/spectral domain (bottom panel, rows are network harmonics/eigenvalues). The windows which were used to assess the sparsity of the signal are marked as colored rectangles (dashed lines: pre-stimulus interval, solid lines: post-stimulus interval), as well as the interval in which the stimulus was presented (black bar, 0-200 ms) and the mean and standard deviation of the reaction times (solid and dashed black lines, respectively) across all trials and subjects . b: Signal averaged over the time points within the windows marked in panel a in the "original" and in the graph frequency domain. Brain regions are ordered by amplitude, network harmonics are ordered by eigenvalue. GFT: graph Fourier transform. c: Power captured cumulatively as more network harmonics are added, for the SC-derived graph (red line) and for 100 randomized graphs (grey lines). The dotted line marks 90% of the overall power. d: Illustration of how the L1 norm captures sparsity of the signal.The smaller the L1 norm, the more compact the signal, even as the L2 norm (power) remains the same. e: Difference between the L1 norms of the signals shown in panel b, both averaged over subjects, and, for illustration, for each subject. The star marks a significance difference between the means as assessed by a permutation test (see Methods).
Connectome spectral analysis to track EEG task dynamics on a subsecond scale

July 2020

·

205 Reads

·

52 Citations

NeuroImage

We present an approach for tracking fast spatiotemporal cortical dynamics in which we combine white matter connectivity data with source-projected electroencephalographic (EEG) data. We employ the mathematical framework of graph signal processing in order to derive the Fourier modes of the brain structural connectivity graph, or “network harmonics” . These network harmonics are naturally ordered by smoothness. Smoothness in this context can be understood as the amount of variation along the cortex, leading to a multi-scale representation of brain connectivity. We demonstrate that network harmonics provide a sparse representation of the EEG signal, where, at certain times, the smoothest 15 network harmonics capture 90% of the signal power. This suggests that network harmonics are functionally meaningful, which we demonstrate by using them as a basis for the functional EEG data recorded from a face detection task. There, only 13 network harmonics are sufficient to track the large-scale cortical activity during the processing of the stimuli with a 50 ms resolution, reproducing well-known activity in the fusiform face area as well as revealing co-activation patterns in somatosensory/motor and frontal cortices that an unconstrained ROI-by-ROI analysis fails to capture. The proposed approach is simple and fast, provides a means of integration of multimodal datasets, and is tied to a theoretical framework in mathematics and physics. Thus, network harmonics point towards promising research directions both theoretically - for example in exploring the relationship between structure and function in the brain - and practically - for example for network tracking in different tasks and groups of individuals, such as patients.


Figure 1. ​ a: ​ Illustration of workflow, from left to right. The structural connectivity matrix encodes a graph in which each node is a brain region defined by a parcellation (see brain surface below matrix; brain regions, nodes, and matrix rows/columns are color-coded) and edges are defined by white matter connectivity. The graph Laplacian is computed from this matrix, and network harmonics are obtained as eigenvectors of this graph Laplacian. The eigenvectors are ordered by ascending eigenvalue. ​ b:​ The first four network harmonics in vector form (magnified from panel ​ a​ ), projected onto the surface of the brain, and in graph representation. Colors visualize arbitrary units, i.e. the weights in the orthonormal eigenvectors. ​ c:​ The graph Laplacian can be reconstructed as a weighted sum of rank-1 matrices defined by the outer products of its eigenvectors. Large values of equal sign in the eigenvector (network harmonic) lead to large positive weights in the outer product (red entries in the illustration). ​ d​ : The correlation between the (upper or lower triangle) of the reconstructed matrices obtained from the outer products and the Y ˆ original graph Laplacian is used to quantify how well the eigenvectors capture the SC, both when they Y are used cumulatively (circles) and on their own (crosses). Open blue circles and black crosses mark
Connectome spectral analysis to track EEG task dynamics on a subsecond scale

June 2020

·

187 Reads

We present an approach for tracking fast spatiotemporal cortical dynamics in which we combine white matter connectivity data with source-projected electroencephalographic (EEG) data. We employ the mathematical framework of graph signal processing; in order to derive the Fourier modes of the brain structural connectivity graph, or "network harmonics". These network harmonics are naturally ordered by smoothness. Smoothness in this context can be understood as the amount of variation along the cortex, leading to a multi-scale representation of brain connectivity. We demonstrate that network harmonics provide a sparse representation of the EEG signal, where, at certain times, the smoothest 15 network harmonics capture 90% of the signal power. This suggests that network harmonics are functionally meaningful, which we demonstrate by using them as a basis for the functional EEG data recorded from a face detection task. There, only 13 network harmonics are sufficient to track the large-scale cortical activity during the processing of the stimuli with a 50 ms resolution, reproducing well-known activity in the fusiform face area as well as revealing co-activation patterns in somatosensory/motor and frontal cortices that an unconstrained ROI-by-ROI analysis fails to capture. The proposed approach is simple and fast, provides a means of integration of multimodal datasets, and is tied to a theoretical framework in mathematics and physics. Thus, network harmonics point towards promising research directions both theoretically - for example in exploring the relationship between structure and function in the brain - and practically - for example for network tracking in different tasks and groups of individuals, such as patients.


Figure 3: Accuracy on test sets.
Summary of data features and representation.
Results for link prediction task
Bilateral Trade Modeling with Graph Neural Networks

February 2020

·

1,668 Reads

Bilateral trade agreements confer preferred trading status between participating countries, enabling increased trade and potential economic growth. Predicting such trade flows often serve as important economic indicators used by economists and policy makers with impactful ramifications in economic policies adopted by respective countries. However, traditional approaches to predicting potential trade partners is through the use of gravity methods which are cumbersome to define due to the exponentially growing number of constants that need to be considered. In this work, we present a framework for directly predicting bilateral trade partners from observed trade records using graph representation learning. Furthermore, we show as a downstream task that modelling bilateral trade as a graph allows for the classification of countries into various income levels. Empirically, we observe accuracies of up to 98% for predicting trading partners and 68% on income level classification.


Citations (13)


... Recently, motivated by the development of deep learning techniques, many studies utilize neural networks to extract deep features from 3D neuron data Zhao et al., 2022;Kanari et al., 2024) or 2D images Sun et al., 2023). The sparsity of 3D neuron data makes training 3D neural networks challenging. ...

Reference:

Multi-level feature fusion network for neuronal morphology classification
Deep learning for classifying neuronal morphologies: combining topological data analysis and graph neural networks
  • Citing Preprint
  • September 2024

... Most methods based on protein surface analysis exploit so-called surface complementarity [1,2], as interacting surfaces tend to have opposite curvature and electrostatics. This feature is used in protein-protein docking [3], protein design [4] and so on. ...

RosettaSurf—A surface-centric computational design approach

... Overall, Graph-based neural networks are commonly used for signals with irregular spatial distributions. In this approach, the nodes on a sphere are represented as nodes in an established graph, enabling fast implementation and good performance 28 . Following the emergence of GNNs, spatiotemporal forecasting models have predominantly relied on graph-based neural networks because of their capacity to learn representations of spatially irregular distributed signals. ...

DeepSphere: a graph-based spherical CNN

... However, it remains unclear how functional networks at different time scales relate to one another and their common structural substrate. Previous studies focused primarily on resting-state activity as inferred from EEG, MEG, and fMRI data, at both global and local levels [48][49][50][51][52][52][53][54][55] . However, the extent to which this relationship varies across resting and task-specific periods has not been fully explored. ...

Connectome spectral analysis to track EEG task dynamics on a subsecond scale

NeuroImage

... For spherical data, SO(3)-equivariant convolutions are often achieved via spherical harmonicsparameterized convolutions [16,21,42] or isotropic convolutions on spherical graphs [53]. These models have been extended to voxel-wise fODF estimation [19,62] but do not use the spatial correlation between fODFs. ...

DeepSphere: Efficient spherical convolutional neural network with HEALPix sampling for cosmological applications
  • Citing Article
  • April 2019

Astronomy and Computing

... The teleconnections of climate events (Tsonis et al., 2006), through atmospheric and oceanic circulation or large-scale waves, are increasingly considered as an important factor for developing DL methods (Cachay et al., 2021;Taylor & Feng, 2022) and can be incorporated using the graph structure. Besides, the use of grids and CNNs still have some inherent problems: handling of missing values, and the lack of rotation equivariance (Defferrard et al., 2019) and the issue of computational receptive fields (Luo et al., 2016) for Earth data. These inherent drawbacks might affect the performance of DL-based climatological forecasts and will be discussed in the Methods section. ...

DeepSphere: towards an equivariant graph-based spherical CNN
  • Citing Preprint
  • April 2019

... Fortunately, graph structures can represent such complex road network structures. Benefiting from the powerful structure capture ability of Graph Neural Network (GNNs) [7][8][9][10], a series of GNN-based traffic flow forecasting methods have been proposed [11,12]. They generally integrate graph neural networks into Recurrent Neural Networks (RNN) or Convolutional Neural Networks (CNN) to capture the complex spatial-temporal dependencies of traffic data [13,14]. ...

Structured Sequence Modeling with Graph Convolutional Recurrent Networks: 25th International Conference, ICONIP 2018, Siem Reap, Cambodia, December 13-16, 2018, Proceedings, Part I
  • Citing Chapter
  • November 2018

Lecture Notes in Computer Science

... These samples can be mixtures for the first model, vocals for the second, and both for the third. The models are then evaluated on real singers using novel splits of two open datasets, the Free Music Archive (FMA) [4,5] and MTG-Jamendo (MTG) [6], and a closed dataset consisting of 176,141 songs that span 7500 popular singers. We use this dataset due to its scale and the fact that its singers are often the target of music voice deepfakes, some of which we use in this paper. ...

Learning to Recognize Musical Genre from Audio