Alexios Koutsoukas’s research while affiliated with Bristol-Myers Squibb and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (32)


An overview of the AP-Net architecture where (A) is the atomic property module and (B) is the interaction energy module. AP-Net predicts the four physically meaningful components of a protein–ligand interaction: electrostatics (Eelst), exchange (Eexch), induction/polarization (Eind), and London dispersion (Edisp)
(A) Depiction of the Splinter dimer dataset. This dataset was constructed by exhaustively pairing small protein and ligand fragments. Between 50 and 500 dimer configurations were generated for each pair of fragments. (B) Distribution of interaction energies (left) and AP-Net errors with respect to interaction energies (right) over 150 000 validation dimers of the Splinter dataset, in kcal mol⁻¹. The respective mean absolute interaction energies and mean absolute errors of the two sets of distributions are labeled
An example dimer from the SAPT-PDB-13K dataset. A small molecule inhibitor interacts with the nearest amino acid, a tyrosine, of an Escherichia coli sliding clamp protein. This dimer was extracted from PDB entry 4PNU
Correlation between AP-Net predicted interaction energies and computed SAPT0/aDZ interaction energies on the 13 216 dimers in the SAPT-PDB-13K dataset
Example of an alchemical ΔΔEint experiment. The chlorine group of the P1 substructure of the Factor Xa inhibitor, BAY 59-7939, is mutated to a methyl. The structure is extracted from PDB entry 2W26

+2

A physics-aware neural network for protein–ligand interactions with quantum chemical accuracy
  • Article
  • Full-text available

July 2024

·

108 Reads

·

1 Citation

Zachary L. Glick

·

Derek P. Metcalf

·

Caroline S. Glick

·

[...]

·

Quantifying intermolecular interactions with quantum chemistry (QC) is useful for many chemical problems, including understanding the nature of protein–ligand interactions. Unfortunately, QC computations on protein–ligand systems are too computationally expensive for most use cases. The flourishing field of machine-learned (ML) potentials is a promising solution, but it is limited by an inability to easily capture long range, non-local interactions. In this work we develop an atomic-pairwise neural network (AP-Net) specialized for modeling intermolecular interactions. This model benefits from a number of physical constraints, including a two-component equivariant message passing neural network architecture that predicts interaction energies via an intermediate prediction of monomer electron densities. The AP-Net model is trained on a comprehensive dataset composed of paired ligand and protein fragments. This model accurately predicts QC-quality interaction energies of protein–ligand systems at a computational cost reduced by orders of magnitude. Applications of the AP-Net model to molecular crystal structure prediction are explored, as well as limitations in modeling highly polarizable systems.

Download

A Physics-Aware Neural Network for Protein-Ligand Interactions with Quantum Chemical Accuracy

February 2024

·

40 Reads

·

1 Citation

Quantifying intermolecular interactions with quantum chemistry (QC) is useful for many chemical problems, including understanding the nature of protein-ligand interactions. Unfortunately, QC computations on protein-ligand systems are too computationally expensive for most use cases. The flourishing field of machine-learned (ML) potentials is a promising solution, but it is limited by an in- ability to easily capture long range, non-local interactions. In this work we develop an atomic-pairwise neural network (AP-Net) specialized for modeling intermolecular interactions. This model benefits from a number of physical constraints, including a two-component equivariant message passing neural network architecture that predicts interaction energies via an intermediate prediction of monomer electron densities. The AP-Net model also benefits from a comprehensive training dataset com- posed of paired ligand and protein fragments. This model accurately predicts QC-quality interaction energies of protein-ligand systems at a computational cost reduced by orders of magnitude.


Cartesian message passing neural networks for directional properties: Fast and transferable atomic multipoles

June 2021

·

39 Reads

·

24 Citations

The message passing neural network (MPNN) framework is a promising tool for modeling atomic properties but is, until recently, incompatible with directional properties, such as Cartesian tensors. We propose a modified Cartesian MPNN (CMPNN) suitable for predicting atom-centered multipoles, an essential component of ab initio force fields. The efficacy of this model is demonstrated on a newly developed dataset consisting of 46 623 chemical structures and corresponding high-quality atomic multipoles, which was deposited into the publicly available Molecular Sciences Software Institute QCArchive server. We show that the CMPNN accurately predicts atom-centered charges, dipoles, and quadrupoles and that errors in the predicted atomic multipoles have a negligible effect on multipole–multipole electrostatic energies. The CMPNN is accurate enough to model conformational dependencies of a molecule’s electronic structure. This opens up the possibility of recomputing atomic multipoles on the fly throughout a simulation in which they might exhibit strong conformational dependence.


CLIFF: A component-based, machine-learned, intermolecular force field

May 2021

·

71 Reads

·

35 Citations

Computation of intermolecular interactions is a challenge in drug discovery because accurate ab initio techniques are too computationally expensive to be routinely applied to drug–protein models. Classical force fields are more computationally feasible, and force fields designed to match symmetry adapted perturbation theory (SAPT) interaction energies can remain accurate in this context. Unfortunately, the application of such force fields is complicated by the laborious parameterization required for computations on new molecules. Here, we introduce the component-based machine-learned intermolecular force field (CLIFF), which combines accurate, physics-based equations for intermolecular interaction energies with machine-learning models to enable automatic parameterization. The CLIFF uses functional forms corresponding to electrostatic, exchange-repulsion, induction/polarization, and London dispersion components in SAPT. Molecule-independent parameters are fit with respect to SAPT2+(3)δMP2/aug-cc-pVTZ, and molecule-dependent atomic parameters (atomic widths, atomic multipoles, and Hirshfeld ratios) are obtained from machine learning models developed for C, N, O, H, S, F, Cl, and Br. The CLIFF achieves mean absolute errors (MAEs) no worse than 0.70 kcal mol⁻¹ in both total and component energies across a diverse dimer test set. For the side chain–side chain interaction database derived from protein fragments, the CLIFF produces total interaction energies with an MAE of 0.27 kcal mol⁻¹ with respect to reference data, outperforming similar and even more expensive methods. In applications to a set of model drug–protein interactions, the CLIFF is able to accurately rank-order ligand binding strengths and achieves less than 10% error with respect to SAPT reference values for most complexes.


AP-Net: An atomic-pairwise neural network for smooth and transferable interaction potentials

July 2020

·

74 Reads

·

66 Citations

Intermolecular interactions are critical to many chemical phenomena, but their accurate computation using ab initio methods is often limited by computational cost. The recent emergence of machine learning (ML) potentials may be a promising alternative. Useful ML models should not only estimate accurate interaction energies but also predict smooth and asymptotically correct potential energy surfaces. However, existing ML models are not guaranteed to obey these constraints. Indeed, systemic deficiencies are apparent in the predictions of our previous hydrogen-bond model as well as the popular ANI-1X model, which we attribute to the use of an atomic energy partition. As a solution, we propose an alternative atomic-pairwise framework specifically for intermolecular ML potentials, and we introduce AP-Net—a neural network model for interaction energies. The AP-Net model is developed using this physically motivated atomic-pairwise paradigm and also exploits the interpretability of symmetry adapted perturbation theory (SAPT). We show that in contrast to other models, AP-Net produces smooth, physically meaningful intermolecular potentials exhibiting correct asymptotic behavior. Initially trained on only a limited number of mostly hydrogen-bonded dimers, AP-Net makes accurate predictions across the chemically diverse S66x8 dataset, demonstrating significant transferability. On a test set including experimental hydrogen-bonded dimers, AP-Net predicts total interaction energies with a mean absolute error of 0.37 kcal mol⁻¹, reducing errors by a factor of 2–5 across SAPT components from previous neural network potentials. The pairwise interaction energies of the model are physically interpretable, and an investigation of predicted electrostatic energies suggests that the model “learns” the physics of hydrogen-bonded interactions.


AP-Net: An Atomic-Pairwise Neural Network for Smooth and Transferable Interaction Potentials

May 2020

·

8 Reads

div> Intermolecular interactions are critical to many chemical phenomena, but their accurate computation using ab initio methods is often limited by computational cost. The recent emergence of machine learning (ML) potentials may be a promising alternative. Useful ML models should not only estimate accurate interaction energies, but also predict smooth and asymptotically correct potential energy surfaces. However, existing ML models are not guaranteed to obey these constraints. Indeed, systemic deficiencies are apparent in the predictions of our previous hydrogen-bond model as well as the popular ANI-1X model, which we attribute to the use of an atomic energy partition. As a solution, we propose an alternative atomic-pairwise framework specifically for intermolecular ML potentials, and we introduce AP-Net—a neural network model for interaction energies. The AP-Net model is developed using this physically motivated atomic-pairwise paradigm and also exploits the interpretability of symmetry adapted perturbation theory (SAPT). We show that in contrast to other models, AP-Net produces smooth, physically meaningful intermolecular potentials exhibiting correct asymptotic behavior. Initially trained on only a limited number of mostly hydrogen-bonded dimers, AP-Net makes accurate predictions across the chemically diverse S66x8 dataset, demonstrating significant transferability. On a test set including experimental hydrogen-bonded dimers, AP-Net predicts total interaction energies with a mean absolute error of 0.37 kcal mol−1, reducing errors by a factor of 2-5 across SAPT components from previous neural network potentials. The pairwise interaction energies of the model are physically interpretable, and an investigation of predicted electrostatic energies suggests that the model ‘learns’ the physics of hydrogen-bonded interactions. </div


Approaches for machine learning intermolecular interaction energies and application to energy components from symmetry adapted perturbation theory

February 2020

·

222 Reads

·

38 Citations

Accurate prediction of intermolecular interaction energies is a fundamental challenge in electronic structure theory due to their subtle character and small magnitudes relative to total molecular energies. Symmetry adapted perturbation theory (SAPT) provides rigorous quantum mechanical means for computing such quantities directly and accurately, but for a computational cost of at least O(N5), where N is the number of atoms. Here, we report machine learned models of SAPT components with a computational cost that scales asymptotically linearly, O(N). We use modified multi-target Behler–Parrinello neural networks and specialized intermolecular symmetry functions to address the idiosyncrasies of the intermolecular problem, achieving 1.2 kcal mol⁻¹ mean absolute errors on a test set of hydrogen bound complexes including structural data extracted from the Cambridge Structural Database and Protein Data Bank, spanning an interaction energy range of 20 kcal mol⁻¹. Additionally, we recover accurate predictions of the physically meaningful SAPT component energies, of which dispersion and induction/polarization were the easiest to predict and electrostatics and exchange–repulsion are the most difficult.




a A feed-forward deep neural network with two hidden layers, each layer consists of multiple neurons, which are fully connected with neurons of the previous and following layers. b Each artificial neuron receives one or more input signals x1, x2,…, xm and outputs a value y to neurons of the next layer. The output y is a nonlinear weighted sum of input signals. Nonlinearity is achieved by passing the linear sum through non-linear functions known as activation functions. c Popular neurons activation functions: the rectified linear unit (ReLU) (red), Sigmoid (Sigm) (green) and Tanh (blue)
Comparison of activation functions rectified linear units (ReLU), Tanh and Sigmoid (Sigm) on the performance of DNNs. DNNs with a single hidden layer and variable number of neurons were trained and tested using the activation functions ReLU (red), Sigm (green) and Tanh (blue) over fivefold cross validation. The performance was measured using MCC as evaluation metric
Effect of the hyper-parameters (i) number of hidden layers, (ii) number of neurons and (iii) dropout regularization on the performance of DNNs measured by MCC as evaluation metric. DNN configuration A shows results obtained by DNN with a single hidden layer and 10 neurons, ReLU activation function and no regularization averaged over the seven activity datasets, B a two hidden layered DNN with 500 neurons in each layer, ReLU activation function and no regularization, C two hidden layers with 3000 neurons per hidden layer and dropout regularization (0% for the input and 50% for hidden layers), D two hidden layers with 3000 neurons per hidden layer and dropout regularization (20% for the input and 50% for hidden layers), E two hidden layers with 3000 neurons per hidden layer and dropout regularization (50% for both the input and hidden layers), F three hidden layers with 3000 neurons per layer and dropout regularization (50% for both the input and hidden layers) and G four hidden layers with 3500 neurons per layer and dropout regularization (50% for the input and hidden layers)
Boxplot of differences between performances achieved by tuned DNN and the rest algorithms measured using MCC as evaluation metric on the validation sets over the seven activity classes. Results are ranked by decreased mean differences. The differences ranged on average from 0.149 MCC units between DNN and NB, 0.092 DNN and kNN, 0.052 DNN and SVM with linear kernel, 0.021 DNN and RF and 0.009 DNN and SVM with “rbf” kernel
Robustness of machine learning methods to different levels of noise for 4 out of 7 activity classes. At low levels of noise, lower that 20%, non-linear methods performed well achieving performance higher than 0.7 MCC units for most of the tested datasets. Instead, at higher level of noise, equal to or higher than 30%, performance for most algorithms dropped below 0.7 MCC and in several occasions even lower than 0.6 at 50% of noise. Naïve Bayes method was found to be the least affected method achieving in several tested datasets performance higher than 0.6 MCC even at the highest level of noise tested 50% and outperforming more complex methods
Deep-learning: Investigating deep neural networks hyper-parameters and comparison of performance to shallow methods for modeling bioactivity data

June 2017

·

1,280 Reads

·

286 Citations

Journal of Cheminformatics

Abstract Background In recent years, research in artificial neural networks has resurged, now under the deep-learning umbrella, and grown extremely popular. Recently reported success of DL techniques in crowd-sourced QSAR and predictive toxicology competitions has showcased these methods as powerful tools in drug-discovery and toxicology research. The aim of this work was dual, first large number of hyper-parameter configurations were explored to investigate how they affect the performance of DNNs and could act as starting points when tuning DNNs and second their performance was compared to popular methods widely employed in the field of cheminformatics namely Naïve Bayes, k-nearest neighbor, random forest and support vector machines. Moreover, robustness of machine learning methods to different levels of artificially introduced noise was assessed. The open-source Caffe deep-learning framework and modern NVidia GPU units were utilized to carry out this study, allowing large number of DNN configurations to be explored. Results We show that feed-forward deep neural networks are capable of achieving strong classification performance and outperform shallow methods across diverse activity classes when optimized. Hyper-parameters that were found to play critical role are the activation function, dropout regularization, number hidden layers and number of neurons. When compared to the rest methods, tuned DNNs were found to statistically outperform, with p value


Citations (25)


... 48 For the purpose of obtaining training data for ML, direct correlation with experiment is not the most important consideration; sampling, solvation, and entropic corrections can be added later, using a low-cost ML force field trained on ∆E int values from electronic structure calculations. 124,125 What is more important is obtaining high-quality quantumchemical benchmark data. ...

Reference:

Convergent Protocols for Computing Protein–Ligand Interaction Energies Using Fragment-Based Quantum Chemistry
A physics-aware neural network for protein–ligand interactions with quantum chemical accuracy

... Recently, machine learning (ML) methods have been employed to express ab initio potential energy surfaces as a function of nuclear coordinates [58][59][60][61][62] to preserve the accuracy of AIMD and enhance the computational efficiency. The ordinary ML potentials are designed to predict the energies and forces of atoms [58,59], while predicting charges and multipoles requires additional mechanisms that optimize these physical properties [63,64]. The latter is also essential for calculating dielectric properties. ...

Cartesian message passing neural networks for directional properties: Fast and transferable atomic multipoles
  • Citing Article
  • June 2021

... Various approaches have sought to extend the applicability of SAPT to large systems, including (a) the application of efficient numerical techniques, such as density fitting 12 and local approximations; 13,14 (b) the development of alternative SAPTbased methods, such as extended SAPT (XSAPT), 15 SAPT with empirical [16][17][18] and semiempirical 19,20 dispersion, and density functional theory (DFT)-based SAPT [SAPT(DFT) or DFT-SAPT]; 21,22 and (c) the creation of SAPT-based force fields 23,24 and machine learning potentials. [25][26][27] We show here that electrostatic embedding can expand the applicability of SAPT to very large systems, even up to a complete protein, thus facilitating the computation of SAPT energies in complex molecular environments. The method remains fundamentally a SAPT approach: it computes the quantum mechanical interaction energy and its physical components, between two sets of atoms, which we might label A and B. However, with the extensions presented here, distant or less important atoms in A and/or B can be replaced with point charge representations. ...

CLIFF: A component-based, machine-learned, intermolecular force field
  • Citing Article
  • May 2021

... 48 For the purpose of obtaining training data for ML, direct correlation with experiment is not the most important consideration; sampling, solvation, and entropic corrections can be added later, using a low-cost ML force field trained on ∆E int values from electronic structure calculations. 124,125 What is more important is obtaining high-quality quantumchemical benchmark data. ...

AP-Net: An atomic-pairwise neural network for smooth and transferable interaction potentials
  • Citing Article
  • July 2020

... Metcalf [113] features hydrogen-bonded dimers involving N-methylacetamide (NMA) paired with 126 different molecules (46 donors and 80 acceptors). Optimized geometries for each monomer were obtained and paired with NMA in various spatial configurations to generate thousands of complexes. ...

Approaches for machine learning intermolecular interaction energies and application to energy components from symmetry adapted perturbation theory
  • Citing Article
  • February 2020

... In the process of developing neurons from a low level to a high level, the substructure of the characteristic-coding toxicological carrier in which neurons are located gradually becomes larger until it occupies the whole toxicological carrier. Therefore, deep neural networks (DNNs) can learn from complex toxicological characteristic data, which leads to a high prediction ability of toxicity [42]. ...

Deep-learning: Investigating deep neural networks hyper-parameters and comparison of performance to shallow methods for modeling bioactivity data

Journal of Cheminformatics

... However, drug repositioning applying this method requires drug target-miRNA and miRNA-disease associations, which is limited in number. Chandrasekaran et al. proposed to apply and combine multi-perspective and multi-approach learning to study the association between drugs and diseases [2]. However, the approach they propose needs to incorporate a lot of multi-source information. ...

Investigating Multiview and Multitask Learning Frameworks for Predicting Drug-Disease Associations
  • Citing Conference Paper
  • October 2016

... This showed that fusing models built on two different feature spaces that provide complementary information was able to improve the prediction of bioactivity endpoints. Previous work has also shown that combinations of descriptors can significantly improve prediction for MOA classification [15,25,26] (using gene expression and cell morphology data), cytotoxicity [16], mitochondria toxicity [18] and anonymised assay activity [27] (using chemical, gene expression, cell morphology and predicted bioactivity data), prediction of sigma 1 (σ1) receptor antagonist [28] (using cell morphology data and thermal proteome profiling), and even organism-level toxicity [29] (using chemical, protein target and cytotoxicity qHTS data). Thus, the combination of models built from complementary feature spaces can expand a model's applicability domain by allowing predictions in new structural space [30]. ...

Improving the Prediction of Organism-level Toxicity through Integration of Chemical, Protein Target and Cytotoxicity qHTS Data

Toxicology Research

... In toxicology, RF models can be trained on large sets of chemical data (e.g., physicochemical properties, structural features, etc.) and their corresponding toxicity outcomes (Koutsoukas et al., 2016). These models can then be used to predict the toxicity of new chemicals based on their features. ...

Predictive Toxicology: Modeling Chemical Induced Toxicological Response Combining Circular Fingerprints with Random Forest and Support Vector Machine

... The spectra are a result from binary (yes/no) predictions on 477 protein targets from a Naïve Bayes classifier developed by Koutsoukas et al 22,23 using class-specific score thresholds. 24 It can be seen that Table 1 contains both balanced and unbalanced class data. Since classification algorithms -and by extension, decision trees-work better with balanced class data, Exterior Releasing and Heat clearing were selected as candidate class variables, which are the most balanced medicinal classes in terms of instances in our dataset. ...

Comparing Global and Local Likelihood Score Thresholds in Multiclass Laplacian-Modified Naive Bayes Protein Target Prediction
  • Citing Article
  • March 2015

Combinatorial Chemistry & High Throughput Screening