<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Mol. Biosci.</journal-id>
<journal-title>Frontiers in Molecular Biosciences</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Mol. Biosci.</abbrev-journal-title>
<issn pub-type="epub">2296-889X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">867241</article-id>
<article-id pub-id-type="doi">10.3389/fmolb.2022.867241</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Molecular Biosciences</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>NeuralDock: Rapid and Conformation-Agnostic Docking of Small Molecules</article-title>
<alt-title alt-title-type="left-running-head">Sha et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">Neural Network Drug Docking</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Sha</surname>
<given-names>Congzhou M.</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="fn" rid="fn1">
<sup>&#x2020;</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Jian</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="fn" rid="fn1">
<sup>&#x2020;</sup>
</xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Dokholyan</surname>
<given-names>Nikolay V.</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<xref ref-type="fn" rid="fn1">
<sup>&#x2020;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1659855/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Department of Engineering Science and Mechanics</institution>, <institution>Pennsylvania State University</institution>, <addr-line>University Park</addr-line>, <addr-line>PA</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Department of Pharmacology</institution>, <institution>Penn State College of Medicine</institution>, <addr-line>Hershey</addr-line>, <addr-line>PA</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Department of Biochemistry and Molecular Biology</institution>, <institution>Penn State College of Medicine</institution>, <addr-line>Hershey</addr-line>, <addr-line>PA</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>Departments of Chemistry and Biomedical Engineering</institution>, <institution>Penn State University</institution>, <addr-line>University Park</addr-line>, <addr-line>PA</addr-line>, <country>United&#x20;States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/182057/overview">Feng Ding</ext-link>, Clemson University, United&#x20;States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/724445/overview">Binquan Luan</ext-link>, IBM Research, United&#x20;States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/828638/overview">Yanting Xing</ext-link>, Clemson University, United&#x20;States</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Nikolay V. Dokholyan, <email>dokh@psu.edu</email>
</corresp>
<fn fn-type="equal" id="fn1">
<label>
<bold>
<sup>&#x2020;</sup>
</bold>
</label>
<p>
<bold>ORCID ID:</bold> Congzhou M. Sha, <ext-link ext-link-type="uri" xlink:href="http://orcid.org/0000-0001-5301-9459">orcid.org/0000-0001-5301-9459</ext-link>; Jian Wang, <ext-link ext-link-type="uri" xlink:href="http://orcid.org/0000-0001-7768-2802">orcid.org/0000-0001-7768-2802</ext-link>; Nikolay V. Dokholyan, <ext-link ext-link-type="uri" xlink:href="http://orcid.org/0000-0002-8225-4025">orcid.org/0000-0002-8225-4025</ext-link>
</p>
</fn>
<fn fn-type="other">
<p>This article was submitted to Molecular Diagnostics and Therapeutics, a section of the journal Frontiers in Molecular Biosciences</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>22</day>
<month>03</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>9</volume>
<elocation-id>867241</elocation-id>
<history>
<date date-type="received">
<day>31</day>
<month>01</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>22</day>
<month>02</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2022 Sha, Wang and Dokholyan.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Sha, Wang and Dokholyan</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Virtual screening is a cost- and time-effective alternative to traditional high-throughput screening in the drug discovery process. Both virtual screening approaches, structure-based molecular docking and ligand-based cheminformatics, suffer from computational cost, low accuracy, and/or reliance on prior knowledge of a ligand that binds to a given target. Here, we propose a neural network framework, NeuralDock, which accelerates the process of high-quality computational docking by a factor of 10<sup>6</sup>, and does not require prior knowledge of a ligand that binds to a given target. By approximating both protein-small molecule conformational sampling and energy-based scoring, NeuralDock accurately predicts the binding energy, and affinity of a protein-small molecule pair, based on protein pocket 3D structure and small molecule topology. We use NeuralDock and 25 GPUs to dock 937 million molecules from the ZINC database against superoxide dismutase-1 in 21&#xa0;h, which we validate with physical docking using MedusaDock. Due to its speed and accuracy, NeuralDock may be useful in brute-force virtual screening of massive chemical libraries and training of generative drug models.</p>
</abstract>
<kwd-group>
<kwd>virtual docking</kwd>
<kwd>small molecule screening</kwd>
<kwd>drug screening</kwd>
<kwd>machine learning</kwd>
<kwd>binding affinity</kwd>
</kwd-group>
<contract-sponsor id="cn001">National Institutes of Health<named-content content-type="fundref-id">10.13039/100000002</named-content>
</contract-sponsor>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>Drug discovery as carried out by pharmaceutical companies requires an investment of years of research effort and billions of dollars (<xref ref-type="bibr" rid="B10">DiMasi et&#x20;al., 2016</xref>). The preclinical pipeline for identifying a small molecule ligand for a protein target is: (1) biochemical screening of small molecules against a protein target or cellular assay, (2) medicinal chemistry optimization of candidate small molecules, and (3) validation of promising molecules in animals (<xref ref-type="bibr" rid="B12">Eder and Herrling, 2016</xref>). Step (1) is critical in identifying small molecules which bind tightly to the target (hits) and for their subsequent optimization in Step (2) (leads). Step (1) is expensive and time-consuming, taking several months to screen a small library of &#x3e;10<sup>5</sup> compounds. Insufficient binding affinities of leads and hits from Steps (1) and (2) often lead to drug attrition in Step (3) and subsequent clinical trials, with attrition rates as high as 95% (<xref ref-type="bibr" rid="B25">Hutchinson and Kirk, 2011</xref>; <xref ref-type="bibr" rid="B47">Waring et&#x20;al., 2015</xref>). Molecular dynamics and rational drug design can explore a larger part of the chemical space and potentially increase binding affinity of hits, but typical docking tools still only dock one compound every few minutes at moderate sampling accuracy [AutoDock Vina: 1.2&#xa0;min (<xref ref-type="bibr" rid="B42">Trott and Olson, 2010</xref>), DOCK 6: 4.8&#xa0;min (<xref ref-type="bibr" rid="B2">Allen et&#x20;al., 2015</xref>), Glide 1.7&#xa0;min (<xref ref-type="bibr" rid="B37">Schr&#xf6;dinger, 2020</xref>), MedusaDock: seconds to minutes (<xref ref-type="bibr" rid="B14">Fan et&#x20;al., 2021</xref>)], while the chemical space of potential drugs may be as large as 10<sup>60</sup> small molecules (<xref ref-type="bibr" rid="B7">Bohacek et&#x20;al., 1996</xref>). Neural networks have shown significant promise in structural biology, accurately reproducing gold standard results in a fraction of the time (<xref ref-type="bibr" rid="B28">Jumper, 2020</xref>). Here, we accelerate the virtual docking process by 6 orders of magnitude, enabling docking of 10<sup>9</sup> compounds in a single day at low&#x20;cost.</p>
<p>Modern methods of computational drug docking are implemented by tools such as MedusaDock (<xref ref-type="bibr" rid="B11">Ding et&#x20;al., 2010</xref>; <xref ref-type="bibr" rid="B45">Wang and Dokholyan, 2019</xref>), AutoDock Vina (<xref ref-type="bibr" rid="B42">Trott and Olson, 2010</xref>; <xref ref-type="bibr" rid="B17">Forli et&#x20;al., 2016</xref>; <xref ref-type="bibr" rid="B22">Goodsell et&#x20;al., 2021</xref>), DOCK (<xref ref-type="bibr" rid="B2">Allen et&#x20;al., 2015</xref>), and Glide (<xref ref-type="bibr" rid="B19">Friesner et&#x20;al., 2004</xref>). These tools perform molecular docking using classical force fields to evaluate the binding energy or affinity of a small molecule to a protein pocket of interest. Here, we focus on MedusaDock because it performs fully flexible conformational sampling of both the protein and ligand, which mimics the induced fit model of protein-small molecule binding, whereas other tools generate ensembles of the protein which are then rigidly fixed and docked to the small molecule [AutoDock Vina (<xref ref-type="bibr" rid="B13">Evangelista et&#x20;al., 2016</xref>), DOCK (<xref ref-type="bibr" rid="B2">Allen et&#x20;al., 2015</xref>), Glide (<xref ref-type="bibr" rid="B39">Schr&#xf6;dinger, 2016</xref>)]. MedusaDock consists of two independent tasks: conformational sampling and scoring. Here, we show that both tasks can be well-approximated by a deep neural network at a fraction of the computational cost of traditional docking. Although we used MedusaDock to generate our data, the framework we have developed can be applied to the results of other docking&#x20;tools.</p>
<p>The principal advantages of a neural network over traditional docking tools include differentiability (propagation of gradients in model training through automatic differentiation) and speed. Neural networks are also valuable for their composability, in which they can be used as subnetworks of larger neural networks while providing gradients for training (<xref ref-type="bibr" rid="B35">Rockt&#xe4;schel and Riedel, 2017</xref>; <xref ref-type="bibr" rid="B15">Feng et&#x20;al., 2020</xref>). Neural network inference is highly optimized on modern processors, and particularly on GPUs. One can achieve many orders of magnitude higher performance with neural network approximation than with traditional algorithms based on exact calculation (<xref ref-type="bibr" rid="B3">Basu et&#x20;al., 2010</xref>). As molecular dynamics is already a significant approximation to a quantum mechanical and statistical reality, inaccuracies in neural network predictions may be acceptable for virtual docking purposes (<xref ref-type="bibr" rid="B24">He et&#x20;al., 2014</xref>).</p>
<p>Deep neural networks for predicting binding affinities have been successful; however, there are drawbacks to specific approaches (<xref ref-type="bibr" rid="B8">Cang et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B27">Jim&#xe9;nez et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B18">Francoeur et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B20">Gentile et&#x20;al., 2020</xref>). For example, <xref ref-type="bibr" rid="B18">Francoeur et&#x20;al. (2020)</xref> proposed to approximate force fields using neural network-based scoring while still relying on extensive conformational sampling, hence retaining the major computational bottleneck of virtual docking. <xref ref-type="bibr" rid="B20">Gentile et&#x20;al. (2020)</xref> used neural networks to aid in chemical screening, but the predictions were nonspecific: information about the protein pocket was not used in the screening (<xref ref-type="bibr" rid="B20">Gentile et&#x20;al., 2020</xref>). K<sub>DEEP</sub> by <xref ref-type="bibr" rid="B27">Jim&#xe9;nez et&#x20;al. (2018)</xref> uses computationally expensive convolutional architectures which limit inference speed, and was trained directly on protein-small molecule 3D crystal structures. Due to the inclusion of protein-small molecule crystal structures, K<sub>DEEP</sub> is biased and has limited generalizability to proteins not bound to small molecules. This issue of bias has been discussed in <xref ref-type="bibr" rid="B18">Francoeur et&#x20;al. (2020)</xref> and arises from self-docking, in which the neural network is provided with a low conformational energy crystal structure as input, and therefore does not perform conformational sampling. <xref ref-type="bibr" rid="B8">Cang et&#x20;al. (2018)</xref> use convolutional networks with manually constructed ligand features based on ligand topology but provide no forward validation with docking tools; instead, only binding affinity is predicted, increasing the risk of overfitting and self-docking&#x20;bias.</p>
<p>In stark contrast to work such as in <xref ref-type="bibr" rid="B18">Francoeur et&#x20;al. (2020)</xref>, we do not explicitly train our neural network to predict the energy of a specific conformation. Rather, we train our network in a conformation-invariant manner by withholding conformational information in the inputs and by predicting population parameters like minimum energy and binding affinity. Unlike in work from <xref ref-type="bibr" rid="B27">Jim&#xe9;nez et&#x20;al. (2018)</xref> and <xref ref-type="bibr" rid="B8">Cang et&#x20;al. (2018)</xref>, we limit bias from self-docking and overfitting. We train the neural network to directly predict the minimum binding energy evaluated by MedusaDock, based on a coarse 3D representation of the protein and a graph representation of the small molecule. The direct prediction of binding energy makes conformational sampling implicit in the neural network. The coarse protein representation and graph representation of the small molecule withholds the optimal orientation, alignment, and conformation of the small molecule in the protein pocket of interest from the neural network. Since we augment our training data with MedusaDock energies, we also reduce the likelihood of overfitting. With these design elements in our neural network NeuralDock, we achieve class-leading performance, as tested on the PDBbind 2013 core set. We perform a proof-of-concept docking for benchmarking and external validation purposes, using NeuralDock to dock nearly 10<sup>9</sup> molecules from the ZINC database against the enzyme superoxide dismutase-1 (SOD1), which is not present in the training, validation, or test sets. Finally, we validate the predicted energies using MedusaDock.</p>
</sec>
<sec sec-type="materials|methods" id="s2">
<title>Materials and Methods</title>
<sec id="s2-1">
<title>Data</title>
<p>For NeuralDock inputs, we used crystal structures of 3875 known protein-ligand pairs from the PDBbind 2017 refined set (<xref ref-type="bibr" rid="B46">Wang et&#x20;al., 2004</xref>), which were shuffled into a training set (N &#x3d; 2,712, 70% of structures), and a validation set (N &#x3d; 1,163, 30% of structures). One of these structures was not processable by MedusaDock 2.0, and 127 structures exceeded our computational resources. A further 496 ligands were not processable by rdkit 2020.09.3<sup>27</sup>. This left a training set of N &#x3d; 2,279 (84.0% of original) and a validation set of N &#x3d; 972 (83.6% of original). The core set of PDBbind 2013 was used as the test set (N &#x3d; 195); 154 of these were processable by rdkit. There is no overlap of proteins among the training, validation, and test&#x20;sets.</p>
<p>We extracted the atoms in the protein which were within a cube of side length 20 angstroms, centered at the ligand. These protein pockets were encoded as 10&#x20;&#xd7; 10&#x20;&#xd7; 10, 2-angstrom resolution images with 8 channels corresponding to a one-hot encoding of (no atom, C, O, N, S, P, H, and other). If multiple atoms were contained in the same 2-angstrom cube, we took the maximum of each channel of the one-hot encodings. Using rdkit, we encoded the ligand as a length 36 atom type vectors with 7 channels (no atom, C, N, O, F, S, and other) and a 36&#x20;&#xd7; 36 bond adjacency matrix with 5 channels (no bond, single, double, triple, aromatic/conjugated). For ligands with greater than 36 heavy atoms (hydrogens were excluded), we removed atoms with the least bond order until we were left with 36 total heavy atoms, thus attempting to preserve the important topologies present in the ligand. These input dimensions were chosen so that a comparable number of parameters in NeuralDock would be devoted to processing the protein and the ligand each (<xref ref-type="fig" rid="F1">Figure&#x20;1A</xref>), as well as to prevent the ligand representation from becoming too sparse. We also chose a limit of 36 heavy atoms since molecules with greater than 36 heavy atoms are likely to exceed the 500 Dalton cutoff for Lipinski&#x2019;s rule of&#x20;five.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The neural network architecture on the left and performance comparison with MedusaDock on the right. <bold>(A)</bold> Inputs, hidden layers, and outputs are shown for the architecture. The protein pocket is flattened and fed into a subnetwork, and the ligand is processed similarly. The outputs of the two subnetworks are concatenated and fed into another subnetwork, which outputs 13 &#xd7; 7 &#x2b; 1 values representing the 7 summary statistics of the 13 energies output by MedusaDock, as well as the pK of the protein-ligand pair. The structure of each FC layer is shown at the bottom. <bold>(B)</bold> The 45 million parameter NeuralDock network achieves class-leading performance on the PDBbind 2013 core set (<italic>r</italic>&#x20;&#x3d; 0.85, <italic>p</italic>&#x20;&#x3c; 0.0001). <bold>(C)</bold> The 45 million parameter NeuralDock network achieves good agreement with experimentally determined pK on the validation set (<italic>r</italic>&#x20;&#x3d; 0.62, <italic>p</italic>&#x20;&#x3c; 0.0001).</p>
</caption>
<graphic xlink:href="fmolb-09-867241-g001.tif"/>
</fig>
<p>For each protein-ligand pair, we ran MedusaDock for 24&#xa0;h or 1000 iterations (whichever came first) on a single core of an Intel Xeon E5-2,680 v3 processor with 6&#xa0;GB RAM. We collected summary statistics (mean, median, standard deviation, minimum, maximum, skew, and kurtosis) on the 13 interaction energies computed in MedusaDock&#x2019;s force field, MedusaScore (<xref ref-type="bibr" rid="B49">Yin et&#x20;al., 2008</xref>). Five of these energies were zero for most or all structures (see <xref ref-type="sec" rid="s11">Supplementary Information</xref>). An interaction energy is defined as the total energy of the protein-ligand complex <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> minus the contributions from the protein <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>P</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and the ligand <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>L</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> when they are isolated.<disp-formula id="equ1">
<mml:math id="m4">
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">int</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>P</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>L</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</disp-formula>
</p>
<p>In developing our models, we chose to focus on E<sub>without</sub> <sub>VDWR</sub>, the interaction energy excluding repulsive van der Waals forces. E<sub>without</sub> <sub>VDWR</sub> is the output of MedusaScore which is known to be the most highly correlated with experimental binding affinity (<xref ref-type="bibr" rid="B49">Yin et&#x20;al., 2008</xref>).</p>
<p>The compounds (N &#x3d; 997,402,117) were downloaded from the ZINC database (<xref ref-type="bibr" rid="B26">Irwin and Shoichet, 2005</xref>) (available in Tranches) and processed into the tensor input format described above. Of these compounds, 936,054,166 (94%) were processable by rdkit. We chose the 1UXM SOD1 PDB structure as a protein target (<xref ref-type="bibr" rid="B5">Berman et&#x20;al., 2000</xref>). The interface between the A and B chains (<xref ref-type="fig" rid="F2">Figure&#x20;2</xref>) was chosen as the binding pocket for&#x20;1UXM.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>1UXM A4V SOD1 dimer chains A (gold) and B (lilac), with the protein pocket of interest (green) in billion-molecule docking. A cartoon, stick, and water-accessible surface representation of 1UXM, an A4V mutant SOD1 dimer structure. The image was generated using PyMOL 2.4.0 (<xref ref-type="bibr" rid="B38">Schr&#xf6;dinger, 2021</xref>).</p>
</caption>
<graphic xlink:href="fmolb-09-867241-g002.tif"/>
</fig>
</sec>
<sec id="s2-2">
<title>Neural Network</title>
<p>For NeuralDock, we chose a fully connected architecture, with spectral normalization (<xref ref-type="bibr" rid="B32">Miyato et&#x20;al., 2018</xref>), dropout (<xref ref-type="bibr" rid="B40">Srivastava et&#x20;al., 2014</xref>) with rate 0.2, LeakyReLU activation (<xref ref-type="bibr" rid="B31">Maas et&#x20;al., 2013</xref>), and skip connections (<xref ref-type="bibr" rid="B23">He et&#x20;al., 2016</xref>). These characteristics were chosen as standard methods of model regularization to prevent overfitting and vanishing or exploding gradients during training.</p>
<p>The NeuralDock architecture (<xref ref-type="fig" rid="F1">Figure&#x20;1A</xref>) was implemented in TensorFlow 2.4.0 (<xref ref-type="bibr" rid="B1">Abadi et&#x20;al., 2016</xref>) and Python 3.7. The inputs were the protein pocket and ligand topology, and the outputs were the 7 summary statistics of the 13 interaction energies computed by MedusaDock as well as the binding affinity pK &#x3d; log<sub>10</sub>&#xa0;K (dissociation/inhibitor constant K<sub>D/I</sub> with units of molar). We considered K<sub>D</sub> and K<sub>I</sub> to be the same for the training purposes. We chose 10 fully connected (FC) hidden layers for each of the three parts (protein encoder, ligand encoder, affinity predictor) of the network (<xref ref-type="fig" rid="F1">Figure&#x20;1A</xref>), resulting in approximately 45 million trainable parameters. We varied the number of hidden layers as well as their widths during hyperparameter optimization (<xref ref-type="table" rid="T1">Table&#x20;1</xref>). The loss function was the L<sup>2</sup> (squared difference) loss between the NeuralDock output energies and the MedusaDock output energies, as well as pKs. The Adam optimizer was used (<xref ref-type="bibr" rid="B29">Kingma and Ba, 2015</xref>), with a learning rate of 10<sup>&#x2212;6</sup>. Training of each model took place on one NVIDIA Tesla T4 GPU, and the models were trained to convergence within a week. We trained a convolutional architecture, in which the FC blocks in the protein encoder (top left of <xref ref-type="fig" rid="F1">Figure&#x20;1A</xref>) were replaced by spectrally normalized 3D inception modules (<xref ref-type="bibr" rid="B41">Szegedy et&#x20;al., 2015</xref>) with a comparable number of trainable parameters (<xref ref-type="table" rid="T1">Table&#x20;1</xref>).</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Correlation coefficients of NeuralDock predicted minimum energy and MedusaDock output for a variety of architectures.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Number of hidden layers per subnetwork (total)</th>
<th align="center">Dimension of hidden layer</th>
<th align="center">Number of trainable parameters</th>
<th align="center">Correlation coefficient for E<sub>without</sub> <sub>VDWR</sub>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">10 FC blocks (30 total)</td>
<td align="center">2048</td>
<td align="center">83,221,686</td>
<td align="char" char=".">0.794</td>
</tr>
<tr>
<td align="left">10 FC blocks (30 total)</td>
<td align="center">1024</td>
<td align="center">45,618,267</td>
<td align="char" char=".">0.838&#x2a;</td>
</tr>
<tr>
<td align="left">6 FC blocks (18 total)</td>
<td align="center">512</td>
<td align="center">12,055,131</td>
<td align="char" char=".">0.800</td>
</tr>
<tr>
<td align="left">6 FC blocks (18 total)</td>
<td align="center">256</td>
<td align="center">4,913,499</td>
<td align="char" char=".">0.758</td>
</tr>
<tr>
<td align="left">6 inception blocks (18 total)</td>
<td align="center">N/A</td>
<td align="center">55,937,979</td>
<td align="char" char=".">0.775</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2-3">
<title>Chemical Scoring</title>
<p>We used Lipinski&#x2019;s rule of five (<xref ref-type="bibr" rid="B30">Lipinski et&#x20;al., 2012</xref>) to evaluate small molecule lead quality and drug likeness, as well as the Quantitative Estimate of Drug likeness (QED) (<xref ref-type="bibr" rid="B6">Bickerton et&#x20;al., 2012</xref>). Octanol-water partition coefficients (log P) were extracted from HTML files of the ZINC database, while all other quantities were computed using rdkit (<xref ref-type="bibr" rid="B34">RDKit, 2016</xref>).</p>
</sec>
<sec id="s2-4">
<title>Statistics</title>
<p>Least squares regression was performed in Python 3.7 using SciPy 1.6.0 (<xref ref-type="bibr" rid="B44">Virtanen et&#x20;al., 2020</xref>). Analysis of covariance (ANCOVA) was performed in Python 3.7 using Pingouin 0.3.12 (<xref ref-type="bibr" rid="B43">Vallat, 2018</xref>).</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec id="s3-1">
<title>Validation of the Ability of NeuralDock to Predict MedusaDock Energy and Experimental pK</title>
<p>NeuralDock training on MedusaDock energies and experimental pKs converged, achieving agreement with experimental binding affinities for the test set, the PDBbind 2013 core set (<xref ref-type="fig" rid="F1">Figure&#x20;1B</xref>), and the validation set taken from PDBbind 2017 (<xref ref-type="fig" rid="F1">Figure&#x20;1C</xref>). As discussed by <xref ref-type="bibr" rid="B18">Francoeur et&#x20;al. (2020)</xref>, testing on the core set may offer a biased evaluation of binding affinity prediction performance. We provide the core set correlation for comparison with other methods which solely report that data (<xref ref-type="table" rid="T2">Table&#x20;2</xref>). Given our relatively small training set (2,331 structures) and the massive diversity of potential protein pockets and ligands, NeuralDock was able to learn the binding affinities accurately. We believe that the key to NeuralDock&#x2019;s success is using high quality data produced by MedusaDock, which sampled thousands of conformations for each protein-ligand pair, while supplying coarse 3D protein information and only the small molecule&#x2019;s topology. By using only the topology of the small molecule as input, we forced NeuralDock to approximate the effects of conformational sampling.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Correlation coefficients for binding affinity prediction of a variety of neural networks.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Model</th>
<th align="center">PDBbind core set binding affinity correlation</th>
<th align="center">Number of test set structures</th>
<th align="center">Number of training set structures</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Def2018 General Ensemble [<xref ref-type="bibr" rid="B18">Francoeur et&#x20;al. (2020)</xref>]</td>
<td align="char" char=".">0.80</td>
<td align="char" char=".">280</td>
<td align="left">18,450&#x20;protein-ligand complexes and 22,584,102 poses</td>
</tr>
<tr>
<td align="left">K<sub>DEEP</sub> [<xref ref-type="bibr" rid="B27">Jim&#xe9;nez et&#x20;al. (2018)</xref>]</td>
<td align="char" char=".">0.82</td>
<td align="char" char=".">195</td>
<td align="left">13,308&#x20;protein-ligand complexes</td>
</tr>
<tr>
<td align="left">TopBP-ML [<xref ref-type="bibr" rid="B8">Cang et&#x20;al. (2018)</xref>]</td>
<td align="char" char=".">0.85</td>
<td align="char" char=".">195</td>
<td align="left">22,886 compounds against each of 102 protein targets</td>
</tr>
<tr>
<td align="left">NeuralDock</td>
<td align="char" char=".">0.85</td>
<td align="char" char=".">154</td>
<td align="left">2331&#x20;protein-ligand complexes</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To test the robustness, generalizability, and speed of NeuralDock, we performed virtual screening of a massive library of ligands (N &#x3d; 936,054,166) against the pocket at the dimeric interface of SOD1 (<xref ref-type="fig" rid="F2">Figure&#x20;2</xref>). We compared relationships among MedusaDock E<sub>without</sub> <sub>VDWR</sub>, NeuralDock E<sub>without</sub> <sub>VDWR</sub>, and experimental binding affinities (pK), and found that they agree (<xref ref-type="fig" rid="F3">Figure&#x20;3</xref>). E<sub>without</sub> <sub>VDWR</sub> is the component of MedusaScore which is most highly correlated with experimental pK (<xref ref-type="bibr" rid="B49">Yin et&#x20;al., 2008</xref>). We demonstrate the correlation between NeuralDock and MedusaDock E<sub>without</sub> <sub>VDWR</sub>, first on the validation set drawn from PDBbind 2017 refined set (<italic>r</italic>&#x20;&#x3d; 0.83, <italic>p</italic>&#x20;&#x3c; 0.0001), and then on 100 random small molecules from ZINC docked to 1UXM, an A4V mutant of the human superoxide dismutase-1 enzyme (<italic>r</italic>&#x20;&#x3d; 0.69, <italic>p</italic>&#x20;&#x3c; 0.0001) (<xref ref-type="fig" rid="F3">Figure&#x20;3A</xref>). Therefore, NeuralDock was successful in learning to predict the minimum E<sub>without</sub> <sub>VDWR</sub> in a single shot. Additionally, we performed 2-way ANCOVA, which measures the effects of a categorical variable. In this instance, it showed no statistically significant difference in the trends for the validation set and the external validation on 1UXM (F &#x3d; 0.67, <italic>p</italic>&#x20;&#x3d; 0.41), which provides evidence that NeuralDock successfully generalizes from the training set to other protein-small molecule pairs, as well as achieving cross-docking success. Since the native structure 1UXM was not bound to any small molecule and MedusaDock performs fully flexible conformational sampling of both the protein and the small molecule, our results provide external validation of NeuralDock and demonstrate robustness under cross-docking, i.e.,&#x20;docking of small molecules to a protein as it appears in its native, unbound conformation.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Comparisons among MedusaDock energies, NeuralDock predicted energies, and experimental binding affinity data. <bold>(A)</bold> The correlation between NeuralDock predicted E<sub>without</sub> <sub>VDWR</sub> and MedusaDock E<sub>without</sub> <sub>VDWR</sub> on the validation set (blue circles, <italic>r</italic>&#x20;&#x3d; 0.83, <italic>p</italic>&#x20;&#x3c; 0.0001) and 100 random small molecules docked to 1UXM (orange triangles, <italic>r</italic>&#x20;&#x3d; 0.69, <italic>p</italic>&#x20;&#x3c; 0.0001). NeuralDock performs well on the validation set in predicting MedusaDock energies, and the trend generalizes to 1UXM with no significant difference (2-way ANCOVA F &#x3d; 0.67, <italic>p</italic>&#x20;&#x3d; 0.41). <bold>(B)</bold> The correlations of MedusaDock E<sub>without</sub> <sub>VDWR</sub> (blue circles), NeuralDock predicted E<sub>without</sub> <sub>VDWR</sub> (magenta triangles), and experimental binding affinity (pK) on the validation set (<italic>r</italic>&#x20;&#x3d; &#x2212;0.48 for both data sets, <italic>p</italic>&#x20;&#x3c; 0.0001), with no significant difference (2-way ANCOVA F &#x3d; 1.27, <italic>p</italic>&#x20;&#x3d; 0.26). <bold>(C)</bold> The 100 small molecules with maximum NeuralDock pK (green triangles), from docking of 936,054,166 small molecules from the ZINC library against 1UXM; the corresponding predicted E<sub>without</sub> <sub>VDWR</sub> is plotted (lilac circles). The left is higher binding affinity (higher pK) and lower energy (lower E<sub>without</sub> <sub>VDWR</sub>). <bold>(D)</bold> and <bold>(E)</bold> The relative frequency distributions (300 bins) of NeuralDock predicted pK (mean 4.07, std 0.47) and E<sub>without</sub> <sub>VDWR</sub> (mean &#x2212;36.6, std 4.1), respectively, on 8,099,176 (9% of total) randomly selected small molecules from the docking of 1UXM. The plots are centered at the means, the <italic>x</italic>-axis ranges are&#x20;&#xb1; 5 standard deviations&#x20;from the mean, and colors are repeated from <bold>(C)</bold>. Note that both the E<sub>without</sub> <sub>VDWR</sub> and pKs in <bold>(C)</bold> are drawn from the extreme tails of the distributions shown in <bold>(D)</bold>.</p>
</caption>
<graphic xlink:href="fmolb-09-867241-g003.tif"/>
</fig>
<p>We found agreement of MedusaDock E<sub>without</sub> <sub>VDWR</sub> and NeuralDock E<sub>without</sub> <sub>VDWR</sub> with experimental pK, with similar correlations for both tools (<italic>r</italic>&#x20;&#x3d; &#x2212;0.48 for both, <italic>p</italic>&#x20;&#x3c; 0.0001), and no statistically significant difference in correlations with 2-way ANCOVA (F &#x3d; 1.27, <italic>p</italic>&#x20;&#x3d; 0.26) (<xref ref-type="fig" rid="F3">Figure&#x20;3B</xref>). The correlation of MedusaDock E<sub>without</sub> <sub>VDWR</sub> with experimental pK (<italic>r</italic>&#x20;&#x3d; &#x2212;0.48) is comparable to that of AutoDock Vina scoring (<italic>r</italic>&#x20;&#x3d; 0.41) as reported in <xref ref-type="bibr" rid="B18">Francoeur et&#x20;al. (2020)</xref>, with MedusaDock performing better, likely due to our extensive sampling and computational effort for each protein-ligand pair. NeuralDock predicts experimental pK better than MedusaDock on the validation set (<xref ref-type="fig" rid="F1">Figures 1B</xref>, C, <xref ref-type="fig" rid="F3">3B</xref>), however, experimental validation of the pKs is needed to confirm that this result holds for SOD1 and other proteins. NeuralDock&#x2019;s agreement with MedusaDock predictions demonstrates that it may be useful in replacing traditional docking tools such as MedusaDock and AutoDock&#x20;Vina.</p>
<p>For the billion-molecule docking of 1UXM, we find that the 100 small molecules with the highest NeuralDock-predicted pKs also have low binding energies, with agreement of NeuralDock and MedusaDock E<sub>without</sub> <sub>VDWR</sub> on those small molecules (<xref ref-type="fig" rid="F3">Figure&#x20;3C</xref>). Furthermore, NeuralDock predicts that few other molecules have such high pK and low energy (<xref ref-type="fig" rid="F3">Figure&#x20;3D</xref>). Combining these results with our validation set regression (<xref ref-type="fig" rid="F1">Figure&#x20;1C</xref>), we predict that some of the 100 small molecules chosen have pK of approximately 8 (K &#x2248; 10 nanomolar). Ninety-five of the 100 compounds satisfy Lipinski&#x2019;s rule of five, and all have reasonable QED scores (median 0.50, range 0.33&#x2013;0.66). The drug likeness of these compounds can be explained as bias from the ZINC Tranches, in which only 14,744,513 (1.5% of the over 997 million) compounds have molecular weight exceeding 500 Daltons or log P exceeding&#x20;5.</p>
</sec>
<sec id="s3-2">
<title>Comparison of Various NeuralDock Architectures</title>
<p>Various architectures for NeuralDock yielded similar results (<xref ref-type="table" rid="T1">Table&#x20;1</xref>). We optimized the NeuralDock architecture based on E<sub>without</sub> <sub>VDWR</sub> correlation, as E<sub>without</sub> <sub>VDWR</sub> is most highly correlated with binding affinity (<xref ref-type="bibr" rid="B49">Yin et&#x20;al., 2008</xref>). Optimizing the network based on E<sub>without</sub> <sub>VDWR</sub> prevents overfitting the hyperparameters (i.e.,&#x20;number of layers and total number of trainable parameters), as MedusaDock E<sub>without</sub> <sub>VDWR</sub> is an external source of training information. Including other MedusaDock energies in training also helped to prevent overfitting (<xref ref-type="sec" rid="s11">Supplementary Table S2</xref>). Even a limited network such as the 4.9 million parameter model was able to generalize from the training set to the validation set (<xref ref-type="table" rid="T1">Table&#x20;1</xref>).</p>
</sec>
<sec id="s3-3">
<title>NeuralDock Dramatically Accelerates Large-Scale Docking-Based Virtual Drug Screening</title>
<p>Benchmarking on a single Tesla T4 GPU was able to predict energies and pKs of 96,000&#x20;protein-small molecule pairs in 203.8&#xa0;s, or 2&#x20;ms per protein-small molecule pair. Docking of the 937 million ZINC compounds with NeuralDock took 21&#xa0;h on 25 GPUs, which matches our benchmarking of 2&#x20;ms per protein-small molecule pair per GPU. For 3875 structures (including those that exceeded our computational resources), it took 120 processors 4&#xa0;weeks to perform extensive MedusaDock sampling (1000 iterations), or an average of 20&#xa0;h per structure per processor. Note that the ability of MedusaDock to find low energy binding poses requires repeated conformational sampling, which can be controlled by running many iterations of MedusaDock, with each iteration taking seconds to minutes. Taking 10&#xa0;h as a conservative minimum compute time for 1000 iterations of MedusaDock, NeuralDock performs 1.7 &#xd7; 10<sup>7</sup> times faster. NeuralDock performs 10<sup>5</sup> times faster than DOCK 6<sup>6</sup>, and 100&#x20;times faster than K<sub>DEEP</sub> (<xref ref-type="bibr" rid="B27">Jim&#xe9;nez et&#x20;al., 2018</xref>). NeuralDock is also faster than the Def2018 General Ensemble by <xref ref-type="bibr" rid="B18">Francoeur et&#x20;al. (2020)</xref>, since we directly predict binding affinity in a single forward pass instead of requiring many conformational samples and forward passes through the network (<xref ref-type="bibr" rid="B18">Francoeur et&#x20;al., 2020</xref>).</p>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<sec id="s4-1">
<title>Potential Improvements to NeuralDock</title>
<p>There are several potential improvements to NeuralDock. First, NeuralDock takes a coarse atomic image of the 20&#xa0;&#xc5; protein pocket as input. Including inputs such as protein amino acid composition, secondary structure, and hydrophobicity may improve NeuralDock predictions. This potential change in inputs is supported by the fact that NeuralDock was able to learn local contributions to the energy well (i.e.,&#x20;E<sub>VWDA</sub>, <xref ref-type="sec" rid="s11">Supplementary Tables S1, S2</xref>), while having difficulty with global contributions (e.g., E<sub>solv</sub>, <xref ref-type="sec" rid="s11">Supplementary Tables S1, S2</xref>). Second, NeuralDock is trained on a relatively small training set, and we did not use any small molecules which are off-target, i.e.,&#x20;do not bind to the given protein. However, our external validation with MedusaDock indicates that this potential source of bias was at least partially addressed by the variety of binding affinities present in our training set. This deficiency may be remedied by docking more targets with MedusaDock as a reference, such as the PDBbind general set, which incurs a significant computational cost in data preparation. Additionally, the PDBbind general set experimental pKs may be noisier than in the refined set. Third, the binding energies and pK are invariant under rotations of the protein pocket as well as under permutations in the atom order for the small molecule bond adjacency matrix and atom type vector. Newer neural network architectures which directly leverage these physical symmetries have been proposed (<xref ref-type="bibr" rid="B16">Finzi et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B48">Wu et&#x20;al., 2021</xref>).</p>
</sec>
<sec id="s4-2">
<title>NeuralDock Performance and Training Past the Interpolation Threshold</title>
<p>We posit that our small training set is not necessarily a major impediment to NeuralDock performance. The so-called double descent phenomenon occurs when the number of tunable parameters in a statistical model significantly exceeds the number of training samples (or degrees of freedom) (<xref ref-type="bibr" rid="B4">Belkin et&#x20;al., 2019</xref>). The traditional tradeoff between bias and variance constitutes the first descent in test error, in which the optimal number of model parameters lies below the number necessary to precisely fit the data, a number known as the interpolation threshold. Contrary to conventional statistical wisdom, however, over-parametrizing may provide benefit in the model performance, causing a second descent in test error. In certain cases, additional training data may actually hurt model performance, whereas training past the point of 0 training set error (another instance of the interpolation threshold) provides a similar benefit to over-parametrization (<xref ref-type="bibr" rid="B33">Nakkiran et&#x20;al., 2020</xref>). Although the double descent phenomenon has yet to be formally proven for deep neural networks, we observed the phenomenon in practice while training NeuralDock. With 46 million parameters and roughly 2000 data points for training, NeuralDock takes advantage of the second descent to achieve its high accuracy.</p>
</sec>
<sec id="s4-3">
<title>Applications of NeuralDock</title>
<p>MolGAN is a generative adversarial approach to drug discovery, which does not take into account the protein pocket (<xref ref-type="bibr" rid="B21">Goodfellow et&#x20;al., 2014</xref>; <xref ref-type="bibr" rid="B9">de Cao and Kipf, 2018</xref>). Since NeuralDock is end-to-end differentiable, one can augment MolGAN training with NeuralDock acting as a scoring function to create drugs targeting specific proteins. The addition of such guidance may improve the stability of MolGAN training. While automatically differentiable force fields are currently being developed for molecular dynamics (<xref ref-type="bibr" rid="B36">Schoenholz and Cubuk, 2019</xref>), the advantage of NeuralDock is it can immediately evaluate the quality of a generated small molecule, implicitly performing conformational sampling, whereas conformational sampling must still be done for a differentiable force field. NeuralDock does not predict the explicit minimum energy configuration for the protein-small molecule pair, however, once specific compounds are identified via NeuralDock, the candidates can be docked by MedusaDock or another docking software prior to experimental validation, as the computational cost of doing so is much less than the cost of traditional docking for the entire compound library.</p>
</sec>
</sec>
<sec sec-type="conclusion" id="s5">
<title>Conclusion</title>
<p>NeuralDock is a robust neural network for predicting binding energies and affinities for protein-small molecule pairs. The key design elements that result in its class-leading performance include high quality scoring of thousands of protein-small molecule conformations, coarse 3D image representation of the protein pocket, and topological representation of the small molecule. Because NeuralDock is trained to output predicted energies and pK in a single shot, NeuralDock is faster than competing methods and can be used in billion-molecule virtual screening. NeuralDock has been validated using the fully flexible docking tool MedusaDock and is ready for experimental validation.</p>
</sec>
</body>
<back>
<sec id="s6">
<title>Data Availability Statement</title>
<p>The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/<xref ref-type="sec" rid="s11">Supplementary Material</xref>.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>CMS implemented, trained, and validated the neural networks; performed the MedusaDock simulations; performed statistical and chemical calculations; created <xref ref-type="fig" rid="F1">Figures 1</xref>, <xref ref-type="fig" rid="F3">3</xref>; and wrote the manuscript. JW wrote the protein pocket extraction software, aided with the MedusaDock simulations, and created <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. All three authors contributed to discussions regarding architecture exploration and design, the cost function, and data analysis and interpretation.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>We acknowledge support from the National Institutes for Health (R35 GM134864 and RF1 AG071675 to NVD), the National Science Foundation (2040667), and the Passan Foundation. The content in this article is solely the responsibility of the authors and does not necessarily represent the official views of the&#x20;NIH.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ack>
<p>We are thankful to the NSF Convergence Accelerator (award number 2040667) team for helpful discussions on machine learning.</p>
</ack>
<sec id="s11">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fmolb.2022.867241/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fmolb.2022.867241/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="Table1.DOCX" id="SM1" mimetype="application/DOCX" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Abadi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Barham</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Davis</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Dean</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). &#x201c;<article-title>TensorFlow: A System for Large-Scale Machine Learning</article-title>,&#x201d; in <source>Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation OSDI 2016</source>. <publisher-loc>Berkeley</publisher-loc>: <publisher-name>USENIX Association</publisher-name>. </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Allen</surname>
<given-names>W. J.</given-names>
</name>
<name>
<surname>Balius</surname>
<given-names>T. E.</given-names>
</name>
<name>
<surname>Mukherjee</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Brozell</surname>
<given-names>S. R.</given-names>
</name>
<name>
<surname>Moustakas</surname>
<given-names>D. T.</given-names>
</name>
<name>
<surname>Lang</surname>
<given-names>P. T.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>DOCK 6: Impact of New Features and Current Docking Performance</article-title>. <source>J.&#x20;Comput. Chem.</source> <volume>36</volume>, <fpage>1132</fpage>&#x2013;<lpage>1156</lpage>. <pub-id pub-id-type="doi">10.1002/jcc.23905</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Basu</surname>
<given-names>J.&#x20;K.</given-names>
</name>
<name>
<surname>Bhattacharyya</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Use of Artificial Neural Network in Pattern Recognition</article-title>. <source>Int. J.&#x20;Softw. Eng. Its Appl.</source> <volume>4</volume>. </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Belkin</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Hsu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Mandal</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Reconciling Modern Machine-Learning Practice and the Classical Bias-Variance Trade-Off</article-title>. <source>Proc. Natl. Acad. Sci. U. S. A.</source> <volume>116</volume>, <fpage>15849</fpage>&#x2013;<lpage>15854</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1903070116</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Berman</surname>
<given-names>H. M.</given-names>
</name>
<name>
<surname>Westbrook</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Gilliland</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Bhat</surname>
<given-names>T. N.</given-names>
</name>
<name>
<surname>Weissig</surname>
<given-names>H.</given-names>
</name>
<etal/>
</person-group> (<year>2000</year>). <article-title>The Protein Data Bank</article-title>. <source>Nucleic Acids Res.</source> <volume>28</volume>, <fpage>235</fpage>&#x2013;<lpage>242</lpage>. <pub-id pub-id-type="doi">10.1093/nar/28.1.235</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bickerton</surname>
<given-names>G. R.</given-names>
</name>
<name>
<surname>Paolini</surname>
<given-names>G. V.</given-names>
</name>
<name>
<surname>Besnard</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Muresan</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Hopkins</surname>
<given-names>A. L.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Quantifying the Chemical beauty of Drugs</article-title>. <source>Nat. Chem.</source> <volume>4</volume>, <fpage>90</fpage>&#x2013;<lpage>98</lpage>. <pub-id pub-id-type="doi">10.1038/nchem.1243</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bohacek</surname>
<given-names>R. S.</given-names>
</name>
<name>
<surname>McMartin</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Guida</surname>
<given-names>W. C.</given-names>
</name>
</person-group> (<year>1996</year>). <article-title>The Art and Practice of Structure-Based Drug Design: A Molecular Modeling Perspective</article-title>. <source>Med. Res. Rev.</source> <volume>16</volume>, <fpage>3</fpage>&#x2013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.1002/(SICI)1098-1128(199601)16:1&#x3c;3::AID-MED1&#x3e;3.0.CO;2-6</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Mu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>G. W.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Representability of Algebraic Topology for Biomolecules in Machine Learning Based Scoring and Virtual Screening</article-title>. <source>Plos Comput. Biol.</source> <volume>14</volume>, <fpage>e1005929</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1005929</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>de Cao</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Kipf</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2018</year>). <source>MolGAN: An Implicit Generative Model for Small Molecular Graphs</source>. <publisher-loc>Ithaca</publisher-loc>: <publisher-name>arXiv</publisher-name>. </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>DiMasi</surname>
<given-names>J.&#x20;A.</given-names>
</name>
<name>
<surname>Grabowski</surname>
<given-names>H. G.</given-names>
</name>
<name>
<surname>Hansen</surname>
<given-names>R. W.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Innovation in the Pharmaceutical Industry: New Estimates of R&#x26;D Costs</article-title>. <source>J.&#x20;Health Econ.</source> <volume>47</volume>, <fpage>20</fpage>&#x2013;<lpage>33</lpage>. <pub-id pub-id-type="doi">10.1016/j.jhealeco.2016.01.012</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ding</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Yin</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Dokholyan</surname>
<given-names>N. V.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Rapid Flexible Docking Using a Stochastic Rotamer Library of Ligands</article-title>. <source>J.&#x20;Chem. Inf. Model.</source> <volume>50</volume>, <fpage>1623</fpage>&#x2013;<lpage>1632</lpage>. <pub-id pub-id-type="doi">10.1021/ci100218t</pub-id> </citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Eder</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Herrling</surname>
<given-names>P. L.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Trends in Modern Drug Discovery</article-title>. <source>Handbook Exp. Pharmacol.</source> <volume>232</volume>, <fpage>3</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1007/164_2015_20</pub-id> </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Evangelista</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Weir</surname>
<given-names>R. L.</given-names>
</name>
<name>
<surname>Ellingson</surname>
<given-names>S. R.</given-names>
</name>
<name>
<surname>Harris</surname>
<given-names>J.&#x20;B.</given-names>
</name>
<name>
<surname>Kapoor</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Smith</surname>
<given-names>J.&#x20;C.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>Ensemble-based Docking: From Hit Discovery to Metabolism and Toxicity Predictions</article-title>. <source>Bioorg. Med. Chem.</source> <volume>24</volume>, <fpage>4928</fpage>&#x2013;<lpage>4935</lpage>. <pub-id pub-id-type="doi">10.1016/j.bmc.2016.07.064</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fan</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Mahdavi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Madduri</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>GPU-accelerated Flexible Molecular Docking</article-title>. <source>J.&#x20;Phys. Chem. B</source> <volume>125</volume>, <fpage>1049</fpage>&#x2013;<lpage>1060</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpcb.0c09051</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Feng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>Z. o.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Greenspan</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2020</year>). <source>Exploring End-To-End Differentiable Natural Logic Modeling</source>. <publisher-loc>Ithaca</publisher-loc>: <publisher-name>arXiv</publisher-name>. <pub-id pub-id-type="doi">10.18653/v1/2020.coling-main.101</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Finzi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Stanton</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Izmailov</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Wilson</surname>
<given-names>A. G.</given-names>
</name>
</person-group> (<year>2020</year>). <source>Generalizing Convolutional Neural Networks for Equivariance to Lie Groups on Arbitrary Continuous Data</source>. <publisher-loc>Ithaca</publisher-loc>: <publisher-name>arXiv</publisher-name>. </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Forli</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Huey</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Pique</surname>
<given-names>M. E.</given-names>
</name>
<name>
<surname>Sanner</surname>
<given-names>M. F.</given-names>
</name>
<name>
<surname>Goodsell</surname>
<given-names>D. S.</given-names>
</name>
<name>
<surname>Olson</surname>
<given-names>A. J.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Computational Protein-Ligand Docking and Virtual Drug Screening with the AutoDock Suite</article-title>. <source>Nat. Protoc.</source> <volume>11</volume>, <fpage>905</fpage>&#x2013;<lpage>919</lpage>. <pub-id pub-id-type="doi">10.1038/nprot.2016.051</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Francoeur</surname>
<given-names>P. G.</given-names>
</name>
<name>
<surname>Masuda</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Sunseri</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jia</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Iovanisci</surname>
<given-names>R. B.</given-names>
</name>
<name>
<surname>Snyder</surname>
<given-names>I.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Three-Dimensional Convolutional Neural Networks and a Cross-Docked Data Set for Structure-Based Drug Design</article-title>. <source>J.&#x20;Chem. Inf. Model.</source> <volume>60</volume>, <fpage>4200</fpage>&#x2013;<lpage>4215</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.0c00411</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Friesner</surname>
<given-names>R. A.</given-names>
</name>
<name>
<surname>Banks</surname>
<given-names>J.&#x20;L.</given-names>
</name>
<name>
<surname>Murphy</surname>
<given-names>R. B.</given-names>
</name>
<name>
<surname>Halgren</surname>
<given-names>T. A.</given-names>
</name>
<name>
<surname>Klicic</surname>
<given-names>J.&#x20;J.</given-names>
</name>
<name>
<surname>Mainz</surname>
<given-names>D. T.</given-names>
</name>
<etal/>
</person-group> (<year>2004</year>). <article-title>Glide: A New Approach for Rapid, Accurate Docking and Scoring. 1. Method and Assessment of Docking Accuracy</article-title>. <source>J.&#x20;Med. Chem.</source> <volume>47</volume>, <fpage>1739</fpage>&#x2013;<lpage>1749</lpage>. <pub-id pub-id-type="doi">10.1021/jm0306430</pub-id> </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gentile</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Agrawal</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Hsing</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ton</surname>
<given-names>A. T.</given-names>
</name>
<name>
<surname>Ban</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Norinder</surname>
<given-names>U.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Deep Docking: A Deep Learning Platform for Augmentation of Structure Based Drug Discovery</article-title>. <source>ACS Cent. Sci.</source> <volume>6</volume>, <fpage>939</fpage>&#x2013;<lpage>949</lpage>. <pub-id pub-id-type="doi">10.1021/acscentsci.0c00229</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Goodfellow</surname>
<given-names>I. J.</given-names>
</name>
<name>
<surname>Pouget-Abadie</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Mirza</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Warde-Farley</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Ozair</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). &#x201c;<article-title>Generative Adversarial Nets</article-title>,&#x201d; in <source>Advances in Neural Information Processing Systems</source>. <publisher-loc>Montreal</publisher-loc>: <publisher-name>NeurIPS</publisher-name>, <volume>3</volume>. </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Goodsell</surname>
<given-names>D. S.</given-names>
</name>
<name>
<surname>Sanner</surname>
<given-names>M. F.</given-names>
</name>
<name>
<surname>Olson</surname>
<given-names>A. J.</given-names>
</name>
<name>
<surname>Forli</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>The AutoDock Suite at 30</article-title>. <source>Protein Sci.</source> <volume>30</volume>, <fpage>31</fpage>&#x2013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.1002/pro.3934</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Ren</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Deep Residual Learning for Image Recognition</article-title>,&#x201d; in <source>Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition</source>. <publisher-loc>Piscataway</publisher-loc>: <publisher-name>IEEE</publisher-name>. <comment>vols 2016-December</comment>. <pub-id pub-id-type="doi">10.1109/cvpr.2016.90</pub-id> </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.&#x20;Z.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Fragment Quantum Mechanical Calculation of Proteins and its Applications</article-title>. <source>Acc. Chem. Res.</source> <volume>47</volume>, <fpage>2748</fpage>&#x2013;<lpage>2757</lpage>. <pub-id pub-id-type="doi">10.1021/ar500077t</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hutchinson</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Kirk</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>High Drug Attrition Rates-Wwhere Are We Going Wrong?</article-title> <source>Nat. Rev. Clin. Oncol.</source> <volume>8</volume>, <fpage>189</fpage>&#x2013;<lpage>190</lpage>. <pub-id pub-id-type="doi">10.1038/nrclinonc.2011.34</pub-id> </citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Irwin</surname>
<given-names>J.&#x20;J.</given-names>
</name>
<name>
<surname>Shoichet</surname>
<given-names>B. K.</given-names>
</name>
</person-group> (<year>2005</year>). <article-title>ZINC--a Free Database of Commercially Available Compounds for Virtual Screening</article-title>. <source>J.&#x20;Chem. Inf. Model.</source> <volume>45</volume>, <fpage>177</fpage>&#x2013;<lpage>182</lpage>. <pub-id pub-id-type="doi">10.1021/ci049714&#x2b;</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jim&#xe9;nez</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>&#x160;kali&#x10d;</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Mart&#xed;nez-Rosell</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>De Fabritiis</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>KDEEP: Protein-Ligand Absolute Binding Affinity Prediction via 3D-Convolutional Neural Networks</article-title>. <source>J.&#x20;Chem. Inf. Model.</source> <volume>58</volume>, <fpage>287</fpage>&#x2013;<lpage>296</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.7b00650</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Jumper</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Evans</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Pritzel</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Green</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Figurnov</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ronneberger</surname>
<given-names>O.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Highly accurate protein structure prediction with AlphaFold</article-title>. <source>Nature</source> <volume>596</volume>, <fpage>583</fpage>&#x2013;<lpage>589</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-021-03819-2</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kingma</surname>
<given-names>D. P.</given-names>
</name>
<name>
<surname>Ba</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>A Method for Stochastic Optimization</article-title>,&#x201d; in <source>3rd International Conference on Learning Representations</source> (<publisher-loc>San Diego</publisher-loc>: <publisher-name>ICLR 2015 - Conference Track Proceedings</publisher-name>). </citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lipinski</surname>
<given-names>C. A.</given-names>
</name>
<name>
<surname>Lombardo</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Dominy</surname>
<given-names>B. W.</given-names>
</name>
<name>
<surname>Feeney</surname>
<given-names>P. J.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Experimental and Computational Approaches to Estimate Solubility and Permeability in Drug Discovery and Development Settings</article-title>. <source>Adv. Drug Deliv. Rev.</source> <volume>64</volume>. <pub-id pub-id-type="doi">10.1016/j.addr.2012.09.019</pub-id> </citation>
</ref>
<ref id="B31">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Maas</surname>
<given-names>A. L.</given-names>
</name>
<name>
<surname>Hannun</surname>
<given-names>A. Y.</given-names>
</name>
<name>
<surname>Ng</surname>
<given-names>A. Y.</given-names>
</name>
</person-group> (<year>2013</year>). &#x201c;<article-title>Rectifier Nonlinearities Improve Neural Network Acoustic Models</article-title>,&#x201d; in <source>ICML Workshop on Deep Learning for Audio, Speech and Language Processing</source>. <publisher-loc>Atlanta</publisher-loc>: <publisher-name>Proceedings of Machine Learning Research</publisher-name>. </citation>
</ref>
<ref id="B32">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Miyato</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kataoka</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Koyama</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Yoshida</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Spectral Normalization for Generative Adversarial Networks</article-title>,&#x201d; in <source>6th International Conference on Learning Representations</source> (<publisher-loc>Vancouver</publisher-loc>: <publisher-name>ICLR 2018 - Conference Track Proceedings</publisher-name>). </citation>
</ref>
<ref id="B33">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nakkiran</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Kaplun</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Bansal</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Barak</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Sutskever</surname>
<given-names>I.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). &#x201c;<article-title>Deep Double Descent: Where Bigger Models and More Data Hurt</article-title>,&#x201d; in <source>8th International Conference on Learning Representations</source>. <comment>
<italic>$\\{$ICLR$\\}$ 2020</italic>
</comment>. </citation>
</ref>
<ref id="B34">
<citation citation-type="book">
<collab>RDKit</collab> (<year>2016</year>). <source>RDKit: Open-Source Cheminformatics Software (RRID:SCR_014274)</source>. </citation>
</ref>
<ref id="B35">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Rockt&#xe4;schel</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Riedel</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>End-to-end Differentiable Proving</article-title>,&#x201d; in <source>Advances in Neural Information Processing Systems</source>. <publisher-loc>Long Beach</publisher-loc>: <publisher-name>NeurIPS</publisher-name>. <comment>vols 2017-December</comment>. </citation>
</ref>
<ref id="B36">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Schoenholz</surname>
<given-names>S. S.</given-names>
</name>
<name>
<surname>Cubuk</surname>
<given-names>E. D.</given-names>
</name>
</person-group> (<year>2019</year>). <source>End-to-end Differentiable, Hardware Accelerated, Molecular Dynamics in Pure Python</source>. <publisher-loc>Ithaca</publisher-loc>: <publisher-name>arXiv</publisher-name>. </citation>
</ref>
<ref id="B37">
<citation citation-type="book">
<collab>Schr&#xf6;dinger</collab> (<year>2020</year>). <source>How Long Does it Take to Screen 10,000 Compounds with Glide?</source> <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Schr&#xf6;dinger LLC</publisher-name>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="https://www.schrodinger.com/kb/1012">https://www.schrodinger.com/kb/1012</ext-link>
</comment>. </citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<collab>Schr&#xf6;dinger</collab> (<year>2021</year>). <source>The PyMOL Molecular Graphics System</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Schr&#xf6;dingerLLC</publisher-name>. <comment>Version 2.4</comment> </citation>
</ref>
<ref id="B39">
<citation citation-type="book">
<collab>Schr&#xf6;dinger</collab> (<year>2016</year>). <source>What Is Ensemble Docking and How Can I Use it?</source> . <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Schr&#xf6;dinger LLC</publisher-name>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="https://www.schrodinger.com/kb/28">https://www.schrodinger.com/kb/28</ext-link>
</comment>. </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Srivastava</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Hinton</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Krizhevsky</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sutskever</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Salakhutdinov</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Dropout: A Simple Way to Prevent Neural Networks from Overfitting</article-title>. <source>J.&#x20;Mach. Learn. Res.</source> <volume>15</volume>, <fpage>1929</fpage>&#x2013;<lpage>1958</lpage>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="https://jmlr.org/papers/v15/srivastava14a.html">https://jmlr.org/papers/v15/srivastava14a.html</ext-link>
</comment>. </citation>
</ref>
<ref id="B41">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Jia</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Sermanet</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Reed</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Anguelov</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). &#x201c;<article-title>Going Deeper with Convolutions</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Boston, MA</conf-loc>, <conf-date>7-12 June 2015</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x2013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr.2015.7298594</pub-id> </citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Trott</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Olson</surname>
<given-names>A. J.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>AutoDock Vina: Improving the Speed and Accuracy of Docking with a New Scoring Function, Efficient Optimization, and Multithreading</article-title>. <source>J.&#x20;Comput. Chem.</source> <volume>31</volume> (<issue>2</issue>), <fpage>455</fpage>&#x2013;<lpage>461</lpage>. <pub-id pub-id-type="doi">10.1002/jcc.21334</pub-id> </citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Vallat</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Pingouin: Statistics in Python</article-title>. <source>J.&#x20;Open Source Softw.</source> <volume>3</volume>. <pub-id pub-id-type="doi">10.21105/joss.01026</pub-id> </citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Virtanen</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Gommers</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Oliphant</surname>
<given-names>T. E.</given-names>
</name>
<name>
<surname>Haberland</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Reddy</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Cournapeau</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python</article-title>. <source>Nat. Methods</source> <volume>17</volume>, <fpage>261</fpage>&#x2013;<lpage>272</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-019-0686-2</pub-id> </citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dokholyan</surname>
<given-names>N. V.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>MedusaDock 2.0: Efficient and Accurate Protein-Ligand Docking with Constraints</article-title>. <source>J.&#x20;Chem. Inf. Model.</source> <volume>59</volume>, <fpage>2509</fpage>&#x2013;<lpage>2515</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.8b00905</pub-id> </citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Fang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>The PDBbind Database: Collection of Binding Affinities for Protein-Ligand Complexes with Known Three-Dimensional Structures</article-title>. <source>J.&#x20;Med. Chem.</source> <volume>47</volume>, <fpage>2977</fpage>&#x2013;<lpage>2980</lpage>. <pub-id pub-id-type="doi">10.1021/jm030580l</pub-id> </citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Waring</surname>
<given-names>M. J.</given-names>
</name>
<name>
<surname>Arrowsmith</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Leach</surname>
<given-names>A. R.</given-names>
</name>
<name>
<surname>Leeson</surname>
<given-names>P. D.</given-names>
</name>
<name>
<surname>Mandrell</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Owen</surname>
<given-names>R. M.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>An Analysis of the Attrition of Drug Candidates from Four Major Pharmaceutical Companies</article-title>. <source>Nat. Rev. Drug Discov.</source> <volume>14</volume>, <fpage>475</fpage>&#x2013;<lpage>486</lpage>. <pub-id pub-id-type="doi">10.1038/nrd4609</pub-id> </citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Long</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>P. S</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>A Comprehensive Survey on Graph Neural Networks</article-title>. <source>IEEE Trans. Neural Networks Learn. Syst.</source> <volume>32</volume>. <pub-id pub-id-type="doi">10.1109/tnnls.2020.2978386</pub-id> </citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yin</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Biedermannova</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Vondrasek</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dokholyan</surname>
<given-names>N. V.</given-names>
</name>
</person-group> (<year>2008</year>). <article-title>MedusaScore: an Accurate Force Field-Based Scoring Function for Virtual Drug Screening</article-title>. <source>J.&#x20;Chem. Inf. Model.</source> <volume>48</volume>, <fpage>1656</fpage>&#x2013;<lpage>1662</lpage>. <pub-id pub-id-type="doi">10.1021/ci8001167</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>