<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Archiving and Interchange DTD v2.3 20070202//EN" "archivearticle.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="methods-article" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Microbiol.</journal-id>
<journal-title>Frontiers in Microbiology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Microbiol.</abbrev-journal-title>
<issn pub-type="epub">1664-302X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmicb.2024.1394302</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Microbiology</subject>
<subj-group>
<subject>Methods</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Prediction of microbe&#x2013;drug associations based on a modified graph attention variational autoencoder and random forest</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Wang</surname> <given-names>Bo</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2668960/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Ma</surname> <given-names>Fangjian</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Du</surname> <given-names>Xiaoxin</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Zhang</surname> <given-names>Guangda</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Li</surname> <given-names>Jingyou</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>College of Computer and Control Engineering, Qiqihar University</institution>, <addr-line>Qiqihar</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>Heilongjiang Key Laboratory of Big Data Network Security Detection and Analysis, Qiqihar University</institution>, <addr-line>Qiqihar</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by" id="fn0001">
<p>Edited by: Kunal R. Jain, Sardar Patel University, India</p>
</fn>
<fn fn-type="edited-by" id="fn0002">
<p>Reviewed by: Anjun Ma, The Ohio State University, United States</p>
<p>Fabr&#x00ED;cio Motteran, Federal University of Pernambuco, Brazil</p>
</fn>
<corresp id="c001">&#x002A;Correspondence: Bo Wang, <email>bowangdr@qqhru.edu.cn</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>31</day>
<month>05</month>
<year>2024</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>15</volume>
<elocation-id>1394302</elocation-id>
<history>
<date date-type="received">
<day>01</day>
<month>03</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>10</day>
<month>05</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2024 Wang, Ma, Du, Zhang and Li.</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Wang, Ma, Du, Zhang and Li</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>The identification of microbe&#x2013;drug associations can greatly facilitate drug research and development. Traditional methods for screening microbe-drug associations are time-consuming, manpower-intensive, and costly to conduct, so computational methods are a good alternative. However, most of them ignore the combination of abundant sequence, structural information, and microbe-drug network topology.</p>
</sec>
<sec>
<title>Methods</title>
<p>In this study, we developed a computational framework based on a modified graph attention variational autoencoder (MGAVAEMDA) to infer potential microbedrug associations by combining biological information with the variational autoencoder. In MGAVAEMDA, we first used multiple databases, which include microbial sequences, drug structures, and microbe-drug association databases, to establish two comprehensive feature matrices of microbes and drugs after multiple similarity computations, fusion, smoothing, and thresholding. Then, we employed a combination of variational autoencoder and graph attention to extract low-dimensional feature representations of microbes and drugs. Finally, the lowdimensional feature representation and graphical adjacency matrix were input into the random forest classifier to obtain the microbe&#x2013;drug association score to identify the potential microbe-drug association. Moreover, in order to correct the model complexity and redundant calculation to improve efficiency, we introduced a modified graph convolutional neural network embedded into the variational autoencoder for computing low dimensional features.</p>
</sec>
<sec>
<title>Results</title>
<p>The experiment results demonstrate that the prediction performance of MGAVAEMDA is better than the five state-of-the-art methods. For the major measurements (AUC =0.9357, AUPR =0.9378), the relative improvements of MGAVAEMDA compared to the suboptimal methods are 1.76 and 1.47%, respectively.</p>
</sec>
<sec>
<title>Discussion</title>
<p>We conducted case studies on two drugs and found that more than 85% of the predicted associations have been reported in PubMed. The comprehensive experimental results validated the reliability of our models in accurately inferring potential microbe-drug associations.</p>
</sec>
</abstract>
<kwd-group>
<kwd>microbe&#x2013;drug association prediction</kwd>
<kwd>variational autoencoder</kwd>
<kwd>modified graph convolutional neural network</kwd>
<kwd>graph attention network</kwd>
<kwd>random forest classifier</kwd>
<kwd>computational prediction model</kwd>
</kwd-group>
<counts>
<fig-count count="7"/>
<table-count count="2"/>
<equation-count count="29"/>
<ref-count count="38"/>
<page-count count="14"/>
<word-count count="9069"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Microbiotechnology</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1</label>
<title>Introduction</title>
<p>Microbes, including bacteria, viruses, archaea, fungi, and protists, are dynamic, diverse, and complex gene pools. These microbes form different microbiomes and inhabit different parts of the human body, such as the gut, mouth, vagina, and uterus (<xref ref-type="bibr" rid="ref11">Huttenhower et al., 2012</xref>). Microbes are considered to be &#x201C;forgotten&#x201D; organs that are beneficial to humans, such as assisting the human body in regulating and promoting metabolism by providing protection against pathogens (<xref ref-type="bibr" rid="ref8">Gill et al., 2006</xref>). In addition, microbes play a crucial role in maintaining the ecological environment within organisms (<xref ref-type="bibr" rid="ref7">ElRakaiby et al., 2014</xref>). Abnormal growth or decline of microbes can affect human health, likely leading to obesity (<xref ref-type="bibr" rid="ref4">de la Cuesta-zuluaga et al., 2023</xref>), diabetes (<xref ref-type="bibr" rid="ref33">Wen et al., 2008</xref>), inflammatory bowel disease (<xref ref-type="bibr" rid="ref6">Durack and Lynch, 2019</xref>), and even cancer (<xref ref-type="bibr" rid="ref26">Schwabe and Jobin, 2013</xref>). In recent years, many features of the microbiome and its potential roles in human health have been widely reported. For example, <xref ref-type="bibr" rid="ref28">Sprockett et al. (2018)</xref> explored how preferential effects affect the microbial community of the gastrointestinal tract in early childhood, and <xref ref-type="bibr" rid="ref35">Ximenez and Torres (2017)</xref> discussed the development of the microbiome in early life, spanning from pregnancy to birth and extending into the first years of life. In addition, gut microbial communities have been shown to play a key role in cardiometabolic disorders, neuropsychiatric disorders, and cancer. Moreover, some bacteria or viruses can cause very serious infectious diseases, such as COVID-19 (<xref ref-type="bibr" rid="ref34">Xiang et al., 2020</xref>). Therefore, microbes are considered new therapeutic targets for precision medicine.</p>
<p>Currently, with the rapid increase in drug-resistant pathogenic microbes, it is urgent to determine the association between microbes and drugs to promote subsequent drug development (<xref ref-type="bibr" rid="ref25">Ramirez et al., 2016</xref>). Recent studies have demonstrated that microbes have an important role in modulating drug activity and toxicity, and drugs can, in turn, influence the diversity and function of microbial communities. There is increasing reporting on the relationship between microbes and drugs. For example, <xref ref-type="bibr" rid="ref9">Haiser et al. (2013)</xref> noted that the intestinal actinomycete <italic>Eggerthella lenta</italic> is responsible for the inactivation of the cardiac drug digoxin. <xref ref-type="bibr" rid="ref37">Yoon et al. (2019)</xref> noted that <italic>Enterococcus faecalis</italic> is highly sensitive to imipenem, amikacin, and piperacillin.</p>
<p>Although these microbe&#x2013;drug associations are obtained from experimental methods, it is practically impossible to identify target microbes, which leads to the slow development of new drugs. In order to overcome this problem, most studies have been devoted to the reuse of known drugs and drug combinations. However, the emergence of drug-resistant microbes poses insightful challenges to drug development. Therefore, there is an urgent need to develop an effective method to infer target microbes with new drug associations. Since traditional wet-lab experiments are time-consuming, labor-intensive, and expensive, computer-based methods can be an effective complement to provide accurate predictions of microbe&#x2013;drug associations through computation.</p>
<p>At present, existing methods for predicting microbe&#x2013;drug associations can be classified into three categories: based on network propagation, based on machine learning, and based on deep learning.</p>
<list list-type="simple">
<list-item>
<p>(1) Methods based on network propagation</p>
</list-item>
</list>
<p>Microbe&#x2013;drug associations were predicted by constructing a heterogeneous network based on known microbe&#x2013;drug associations, microbial similarity, and drug similarity. <xref ref-type="bibr" rid="ref38">Zhu et al. (2021)</xref> designed an HMDAKATZ computational model based on KATZ measurements (<xref ref-type="bibr" rid="ref14">Lei and Zhang, 2019</xref>) using the chemical structure of the drug to identify potential human microbe&#x2013;drug associations. The drug was extracted by obtaining the chemical structure of the drug to calculate its similarity with other drugs. The Gaussian interaction profile kernel (<xref ref-type="bibr" rid="ref31">van Laarhoven et al., 2011</xref>) was then used to calculate the similarity of microbes. Finally, the microbe&#x2013;microbe similarity network, drug&#x2013;drug similarity network, and microbe&#x2013;drug similarity network were combined to construct a microbe&#x2013;drug similarity network. The potential relationship between microbes and drugs was predicted using the KATZ algorithm. This model can correctly predict microbe&#x2013;drug association relationships by using a simple metric for heterogeneous networks, but it is not applicable to predict new drugs without known microbial associations or isolated microbes without any known disease associations. <xref ref-type="bibr" rid="ref18">Long et al. (2020)</xref> used rich biological information to construct a heterogeneous network of microbes and drugs and then utilized a framework based on graph convolutional networks to predict human microbe&#x2013;drug associations. However, the similarity calculation of microbes (drugs) is still highly dependent on known microbe&#x2013;drug association information, and the prediction accuracy is not high. <xref ref-type="bibr" rid="ref30">Tan et al. (2022)</xref> first constructed a heterogeneous network by integrating known microbe&#x2013;drug associations, microbial similarity, and drug similarity. Then, autoencoder modules based on GAE (<xref ref-type="bibr" rid="ref17">Liu et al., 2022</xref>) and sparse autoencoder (<xref ref-type="bibr" rid="ref12">Jiang et al., 2020</xref>) modules are used to learn the topological representation and attribute representation of nodes in the newly constructed heterogeneous network, respectively. Finally, based on these two node representations, two feature matrices for microbes and drugs are constructed separately, and they are used to calculate the possible association scores of microbe&#x2013;drug pairs. A novel computational model, GSAMDA, based on graph-attentive networks and sparse autoencoder is proposed to infer potential microbe&#x2013;drug associations, but this model is still not perfect for predicting matrices with sparse data.</p>
<list list-type="simple">
<list-item>
<p>(2) Methods based on machine learning</p>
</list-item>
</list>
<p>Constructing microbial (drug) profiles uses microbial (drug) similarities and known microbe&#x2013;disease associations, as well as designing classifiers to identify microbes associated with diseases. <xref ref-type="bibr" rid="ref12">Jiang et al. (2020)</xref> proposed a computational model based on neighborhood inference and restricted Boltzmann machine (<xref ref-type="bibr" rid="ref13">Kirubahari and Amali, 2023</xref>). Neighborhood inference can leverage abundant similar information about microbes (drugs), while the restricted Boltzmann machine can learn the latent probability distributions hidden in known microbe&#x2013;drug associations. Finally, integrated learning is used to combine the individual learners into a stronger predictor. However, this method is not sufficient to reveal the association between drugs and new microbes, or the association between microbes and new diseases, or without any associations. <xref ref-type="bibr" rid="ref20">Ma and Liu (2022)</xref> integrated multiple sets of data to calculate the functional and semantic similarity of microbes, the structural similarity of drugs, and the information on microbe&#x2013;drug associations. The hypergraph is constructed using strong neighborhood information. In order to improve the performance of the hypergraph, a simple volume is used to calculate the hyperedge weights. Hypergraph regularization is introduced to the generalized matrix decomposition model, and the higher-order structural information is used to improve the representation of the low-dimensional features. A kind of generalized matrix decomposition based on weighted hypergraph learning (WHGMF) is proposed for predicting potential microbe&#x2013;drug associations, but using a fixed number of neighbors to construct the hypergraph may limit the adaptability of the model to some extent. In addition, WHGMF uses the generalized matrix decomposition framework, which is not efficient computationally. <xref ref-type="bibr" rid="ref36">Yang et al. (2022)</xref> proposed a multi-kernel fusion model based on a graph convolutional network, which uses graph convolution to extract multi-layer features, calculates kernel matrices by embedding features on each layer, and fuses multiple kernel matrices based on the average weighting method. Dual Laplacian regularized least squares (<xref ref-type="bibr" rid="ref32">Wang et al., 2017</xref>) are then used to combine kernels in the microbial and drug spaces to infer new microbe&#x2013;drug associations.</p>
<list list-type="simple">
<list-item>
<p>(3) Methods based on deep learning</p>
</list-item>
</list>
<p>The complex heterogeneous network graph of microbe&#x2013;drug associations was constructed by integrating multi-source bioinformatics data of microbes, drugs, and diseases to extract the non-linear features of microbes and drugs to achieve the prediction of microbe&#x2013;drug associations. <xref ref-type="bibr" rid="ref15">Li et al. (2023)</xref> constructed a heterogeneous network of drugs, microbes, and diseases by using multiple sources of biomedical data, then developed a model with a matrix decomposition and a three-layer heterogeneous network to predict potential microbe&#x2013;drug associations. However, the similarity networks of drugs, microbes, and diseases still have some noise, which leads to this model not being good enough, so there is still more room for improvement. <xref ref-type="bibr" rid="ref21">Ma et al. (2023)</xref> constructed two heterogeneous microbe&#x2013;drug networks using multiple similarity metrics of microbes and drugs, as well as known microbe&#x2013;drug associations or known microbe&#x2013;disease&#x2013;drug associations, respectively, and then obtained the feature matrices of microbes and drugs. A computational model, GACNNMDA, based on a graph convolutional neural network was designed to predict the possible scores of microbe&#x2013;drug pairs. However, this model did not take into account other relevant biological information (e.g., microbial sequences and drug similarity based on side effects), and the prediction accuracy is not high. <xref ref-type="bibr" rid="ref18">Long et al. (2020)</xref> utilized a variety of sources of biomedical information to construct microbial and multiple networks of drugs. A new integrated framework of graph attention networks with hierarchical attention mechanism and node-level attention was developed for predicting microbe&#x2013;drug associations from the constructed multiple microbe&#x2013;drug graphs, but still, there is noise in the features extracted from these similarities. So, this model needs to improve the prediction results.</p>
<p>In order to overcome the inherent defects of the above three types of methods, the investigation set up a new microbe&#x2013;drug association prediction model named MGAVAEMDA. This model uses a variational autoencoder and incorporates the modified graph convolutional network and graph attention network to improve prediction accuracy in a three-stage process.</p>
<p>The construction process of the MGAVAEMDA model is divided into three steps:</p>
<list list-type="simple">
<list-item>
<p>(1) In this part, based on the downloaded microbe&#x2013;drug associations, the structure information of drugs, and the sequence information of microbes. The microbe&#x2013;microbe similarity matrix and the drug&#x2013;drug similarity matrix are obtained through multiple similarity measures and data processing.</p>
</list-item>
<list-item>
<p>(2) In this part, the microbe&#x2013;microbe similarity matrix, drug&#x2013;drug similarity matrix, and microbe&#x2013;drug similarity matrix are inputted into the modified graph variational autoencoder to learn the low-dimensional feature representations of microbes and drugs. The graph attention network is then introduced to extract important features using the attention mechanism to reduce dependence on other biological information.</p>
</list-item>
<list-item>
<p>(3) In this last section, the random forest-based classifier was introduced to calculate the possible scores of microbe&#x2013;drug associations. Those newly learned important feature representations are combined to form the inputs for the classifier. The final result of the model is voted on by each base classifier.</p>
</list-item>
</list>
</sec>
<sec id="sec2">
<label>2</label>
<title>Materials and equipment</title>
<sec id="sec3">
<label>2.1</label>
<title>Microbe&#x2013;drug associations</title>
<p>In this article, three different datasets, namely MDAD (<xref ref-type="bibr" rid="ref29">Sun et al., 2018</xref>), aBiofilm (<xref ref-type="bibr" rid="ref24">Rajput et al., 2018</xref>), and DrugVirus (<xref ref-type="bibr" rid="ref2">Andersen et al., 2020</xref>), are used to test the predictive power of the MGAVAEMDA model. The MDAD dataset used in the model contains 2,470 associations between 1,373 drugs and 173 microbes. The aBiofilm dataset used in the model includes 2,884 associations between 1,720 drugs and 140 microbes. The DrugVirus database summarizes experimentally confirmed microbe&#x2013;drug correlations, including 933 associations between 175 drugs and 95 viruses.</p>
<p>Here, an adjacency matrix <inline-formula>
<mml:math id="M1">
<mml:mrow>
<mml:mi mathvariant="normal">MD</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mo>&#x00D7;</mml:mo>
<mml:mi mathvariant="normal">j</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> was built to preserve the microbe&#x2013;drug association information. The <inline-formula>
<mml:math id="M2">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> represents the number of microbes and <inline-formula>
<mml:math id="M3">
<mml:mi>j</mml:mi>
</mml:math>
</inline-formula> represents the number of drugs. If the microbes <inline-formula>
<mml:math id="M4">
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and drugs <inline-formula>
<mml:math id="M5">
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are related, the entity <inline-formula>
<mml:math id="M6">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> has a value of 1, otherwise it is 0.</p>
<disp-formula id="EQ1">
<label>(1)</label>
<mml:math id="M7">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr columnalign="left">
<mml:mtd columnalign="left">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mrow>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mspace width="thickmathspace"/>
<mml:mi mathvariant="normal">microbe</mml:mi>
<mml:mspace width="thickmathspace"/>
<mml:msub>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mspace width="thickmathspace"/>
<mml:mi>a</mml:mi>
<mml:mi mathvariant="normal">ssociated with durg</mml:mi>
<mml:mspace width="thickmathspace"/>
<mml:msub>
<mml:mi mathvariant="normal">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mrow></mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr columnalign="left">
<mml:mtd columnalign="left">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mrow>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mi>o</mml:mi>
<mml:mi mathvariant="normal">therwise</mml:mi>
</mml:mrow>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mrow></mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
</sec>
<sec id="sec4">
<label>2.2</label>
<title>Sequence similarity of microbes</title>
<p>To calculate the sequence similarity of microbial genomes, BLAST (<xref ref-type="bibr" rid="ref1">Altschul et al., 1990</xref>) can be used to perform a pairwise sequence alignment of microbial genomes. Briefly, the main function of BLAST is to discover regions of local similarity between sequences and then calculate the similarity using a local comparison algorithm (<xref ref-type="bibr" rid="ref27">Smith and Waterman, 1981</xref>). For example, <inline-formula>
<mml:math id="M8">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi mathvariant="normal">A</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mn>1</mml:mn>
</mml:msubsup>
<mml:msubsup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x22EF;</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math id="M9">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi mathvariant="normal">B</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mi mathvariant="normal">B</mml:mi>
<mml:mn>1</mml:mn>
</mml:msubsup>
<mml:msubsup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mi mathvariant="normal">B</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x22EF;</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mi mathvariant="normal">B</mml:mi>
<mml:mi mathvariant="normal">b</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> are asked about the genome sequences of microbe A and microbe B, where a and b are the lengths of sequences <inline-formula>
<mml:math id="M10">
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math id="M11">
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>B</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. BLAST creates the scoring matrix <inline-formula>
<mml:math id="M12">
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x00D7;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and sets the elements of the first row and column to zero. The element <inline-formula>
<mml:math id="M13">
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced>
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mfenced>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x00D7;</mml:mo>
<mml:mfenced>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>&#x2026;</mml:mo>
<mml:mi>b</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> of the formula in this scoring matrix are:</p>
<disp-formula id="EQ2">
<label>(2)</label>
<mml:math id="M14">
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mfenced close="" open="{">
<mml:mrow>
<mml:mtable equalrows="true" equalcolumns="true">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
<mml:msup>
<mml:mfenced>
<mml:mrow>
<mml:msubsup>
<mml:mi>m</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>i</mml:mi>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mi>m</mml:mi>
<mml:mi mathvariant="normal">B</mml:mi>
<mml:mi>i</mml:mi>
</mml:msubsup>
<mml:mi mathvariant="normal">, </mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">, </mml:mi>
<mml:msubsup>
<mml:mi>m</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>i</mml:mi>
</mml:msubsup>
<mml:mo>&#x2260;</mml:mo>
<mml:msubsup>
<mml:mi>m</mml:mi>
<mml:mi mathvariant="normal">B</mml:mi>
<mml:mi>i</mml:mi>
</mml:msubsup>
<mml:mi mathvariant="normal">, </mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2190;</mml:mo>
</mml:msup>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M15">
<mml:mrow>
<mml:msubsup>
<mml:mi>m</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>i</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> represents the <italic>i</italic>th sequence value of microbe A, take the matrix <inline-formula>
<mml:math id="M16">
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x00D7;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the maximum value is <inline-formula>
<mml:math id="M17">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
</mml:msub>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>B</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. The similarity between microbe A and microbe B is shown in the following equation:</p>
<disp-formula id="EQ3">
<label>(3)</label>
<mml:math id="M18">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>b</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
</mml:msub>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>B</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
</mml:msub>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x00D7;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>B</mml:mi>
</mml:msub>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>B</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>According to the similarity between two or more microbes, the microbial similarity matrix <inline-formula>
<mml:math id="M19">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> can be constructed.</p>
</sec>
<sec id="sec5">
<label>2.3</label>
<title>Structural similarity of drugs</title>
<p>In this study, SIMCOMP2 was used to search (<xref ref-type="bibr" rid="ref10">Hattori et al., 2010</xref>) for drug structure similarity. SIMCOMP2 search is a chemical structure search server that provides links to the KEGG PATHWAY database, which contains hand-drawn pathway maps with information on molecular interactions, reactions, and relationships. In the SIMCOMP2 search, by mapping the drugs in the dataset to the drugs in KEGG, the drug structure similarity can be obtained with a cut-off score of 0.01 to filter out drugs with a structure similarity score of 0.01 or higher. Then, matrix DSS is defined to preserve the structural similarity of drugs, where the element <inline-formula>
<mml:math id="M20">
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the drug <inline-formula>
<mml:math id="M21">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and drug <inline-formula>
<mml:math id="M22">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mfenced>
<mml:mi>j</mml:mi>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> similarity value.</p>
</sec>
<sec id="sec6">
<label>2.4</label>
<title>Gaussian interaction profile kernel similarity of microbes</title>
<p>Gaussian interaction profile kernel similarity (<xref ref-type="bibr" rid="ref15">Li et al., 2023</xref>) has been widely used in previous studies for the similarity of biological entities. Given the sparse nature of the similarity matrices of microbes and drugs obtained by the above methods, Gaussian interaction profile kernel similarity was constructed based on known microbe&#x2013;drug associations to obtain a more comprehensive microbial similarity. The matrix MGS represents the microbial Gaussian interaction profile kernel similarity, the matrix element <inline-formula>
<mml:math id="M23">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>G</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> represents the Gaussian interaction profile kernel similarity of microbes <inline-formula>
<mml:math id="M24">
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math id="M25">
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, which is calculated as follows:</p>
<disp-formula id="EQ4">
<label>(4)</label>
<mml:math id="M26">
<mml:mrow>
<mml:mi mathvariant="normal">MGS</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">,m</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">j</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>&#x03B3;</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:msup>
<mml:mo>&#x2225;</mml:mo>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M27">
<mml:mrow>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the <italic>i</italic>th column vector of the adjacency matrix <inline-formula>
<mml:math id="M28">
<mml:mrow>
<mml:mi mathvariant="normal">MD</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> as the spectrum kernel for the interactions of microbe <inline-formula>
<mml:math id="M29">
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M30">
<mml:mrow>
<mml:msub>
<mml:mi>&#x03B3;</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the normalized kernel bandwidth of microbe, which can be normalized by the parameter <inline-formula>
<mml:math id="M31">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x03B3;</mml:mi>
<mml:mi>m</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>. It is calculated as follows:</p>
<disp-formula id="EQ5">
<label>(5)</label>
<mml:math id="M32">
<mml:mrow>
<mml:msub>
<mml:mi>&#x03B3;</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x03B3;</mml:mi>
<mml:mi>m</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
</mml:mrow>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:msup>
<mml:mo>&#x2225;</mml:mo>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
</sec>
<sec id="sec7">
<label>2.5</label>
<title>Gaussian interaction profile kernel similarity of drugs</title>
<p>Similar to microbes, the Gaussian interaction profile kernel similarity of drugs was calculated. The matrix DGS represents the drug Gaussian interaction profile kernel similarity, and the matrix element <inline-formula>
<mml:math id="M33">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>G</mml:mi>
<mml:mi>S</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mfenced>
<mml:mi>i</mml:mi>
</mml:mfenced>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi>d</mml:mi>
<mml:mfenced>
<mml:mi>j</mml:mi>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> represents the Gaussian interaction profile kernel similarity of drugs <inline-formula>
<mml:math id="M34">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math id="M35">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. It is calculated as follows:</p>
<disp-formula id="EQ6">
<label>(6)</label>
<mml:math id="M36">
<mml:mrow>
<mml:mi mathvariant="normal">DGS</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">j</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>&#x03B3;</mml:mi>
<mml:mi>d</mml:mi>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:msup>
<mml:mo>&#x2225;</mml:mo>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M37">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the <italic>i</italic>th column vector of the adjacency matrix <inline-formula>
<mml:math id="M38">
<mml:mrow>
<mml:mi mathvariant="normal">MD</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> as the spectrum kernel for the interactions of drug <inline-formula>
<mml:math id="M39">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula>
<mml:math id="M40">
<mml:mrow>
<mml:msub>
<mml:mi>&#x03B3;</mml:mi>
<mml:mi>d</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the normalized kernel bandwidth of drug.</p>
</sec>
<sec id="sec8">
<label>2.6</label>
<title>Similarity fusion</title>
<p>As mentioned above, the similarity of microbes and drugs in different aspects is calculated separately. In order to obtain their comprehensive similarity matrix, the similarities from different perspectives need to be fused. The integrated similarity matrix of microbes is constructed as follows:</p>
<disp-formula id="EQ7">
<label>(7)</label>
<mml:math id="M41">
<mml:mrow>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi>G</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:mfrac>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mi mathvariant="normal">if</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2260;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi>G</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi mathvariant="normal">otherwise</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The integrated similarity of microbes is the similarity of the Gaussian interaction profile of microbes <inline-formula>
<mml:math id="M42">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>G</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> if the sequence similarity of microbes <inline-formula>
<mml:math id="M43">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> otherwise, it is half of the sum of the two.</p>
<p>Similarly, the integrated similarity matrix of drugs is calculated as follows:</p>
<disp-formula id="EQ8">
<label>(8)</label>
<mml:math id="M44">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>G</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:mfrac>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mi mathvariant="normal">if</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi>D</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2260;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>G</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi mathvariant="normal">otherwise</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
</sec>
<sec id="sec9">
<label>2.7</label>
<title>Data processing</title>
<p>In order to reduce noise or fluctuations in the data and make it easier to subsequently analyze the data trends, the combined similarity matrices of microbes and drugs obtained above were smoothed, and for microbes and drugs, the smoothing matrices were calculated using the following formula:</p>
<disp-formula id="EQ9">
<label>(9)</label>
<mml:math id="M45">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mo>.</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M46">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to <inline-formula>
<mml:math id="M47">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the data points within the window, usually from the past <inline-formula>
<mml:math id="M48">
<mml:mi>n</mml:mi>
</mml:math>
</inline-formula> time points. <inline-formula>
<mml:math id="M49">
<mml:mi>n</mml:mi>
</mml:math>
</inline-formula> is the size of the window, which determines the number of data points to be computed within the window. <inline-formula>
<mml:math id="M50">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to <inline-formula>
<mml:math id="M51">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> correspond to the data within the microbial similarity matrix. Similarly, <inline-formula>
<mml:math id="M52">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> to <inline-formula>
<mml:math id="M53">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, the same for drugs. Subsequently, several experiments were conducted for the choice of window, and the optimal window value was obtained. By summing the data points within the window and dividing by the window size <italic>n</italic>, a simple moving average at time point <inline-formula>
<mml:math id="M54">
<mml:mi>j</mml:mi>
</mml:math>
</inline-formula> can be calculated. This smoothing technique helps to observe trends, reducing noise and sudden fluctuations for a better understanding of the long-term trend of the data.</p>
<p>The composite matrix was obtained after the curve of the microbe and drug was smoothed. The data were binarized to reduce its complexity, where continuous-type data were transformed into a process containing only two values of data (0 and 1). In binarization, a threshold value is selected, and then each value in the data is compared with that threshold value. If the data are equal to or greater than the threshold, it is mapped to 1. If the data are less than the threshold, it is mapped to 0. In this way, the data are converted into binary form, which is analyzed in the next step. In this study, different thresholds are selected for microbes and drugs, <italic>m</italic>th represents the microbe threshold, and <italic>d</italic>th represents the drug threshold. The binarization matrix of microbes is calculated as follows:</p>
<disp-formula id="EQ10">
<label>(10)</label>
<mml:math id="M55">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>=</mml:mo>
<mml:mfenced close="" open="{">
<mml:mrow>
<mml:mtable equalrows="true" equalcolumns="true">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi mathvariant="normal">if</mml:mi>
<mml:mspace width="thickmathspace"/>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mi>m</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mi mathvariant="normal">otherwise</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The binarized matrix for the same drug is calculated as follows:</p>
<disp-formula id="EQ11">
<label>(11)</label>
<mml:math id="M56">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>S</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>=</mml:mo>
<mml:mfenced close="" open="{">
<mml:mrow>
<mml:mtable equalrows="true" equalcolumns="true">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi mathvariant="normal">if</mml:mi>
<mml:mspace width="thickmathspace"/>
<mml:mi>D</mml:mi>
<mml:mi>S</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mi>d</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal"></mml:mi>
<mml:mi mathvariant="normal">otherwise</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>
<p>Finally, the binarized similarity matrix is transformed into an adjacency matrix and feature matrix, where the adjacency matrix represents the connectivity of the graph and the feature matrix represents the features of each node. In the next step, both the adjacency matrix and feature matrix are put into MGAVAEMDA for low-dimensional feature extraction.</p>
</sec>
</sec>
<sec sec-type="methods" id="sec10">
<label>3</label>
<title>Methods</title>
<sec id="sec11">
<label>3.1</label>
<title>MGAVAEMDA framework</title>
<p>The flowchart of the MGAVAEMDA model is shown in <xref ref-type="fig" rid="fig1">Figure 1</xref>, based on a graph-attentive variational autoencoder. The model is divided into three main steps: (1) construct the similarity networks of microbes and drugs, respectively; (2) extract the feature representations of microbes and drugs using the modified graph attention variational autoencoder (MGAVAE); and (3) embed the combination of microbe and drug representations into the random forest classifier to obtain the final prediction scores.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>The flowchart of the MGAVAEMDA model.</p>
</caption>
<graphic xlink:href="fmicb-15-1394302-g001.tif"/>
</fig>
</sec>
<sec id="sec12">
<label>3.2</label>
<title>Variational autoencoder</title>
<p>MGAVAE can reconstruct the node attributes and graph structure of structured graph data through an encoder and decoder. MGAVAE can be used to extract the low-dimensional features of microbes and drugs. The MGAVAE model consists of multiple encoding and decoding layers, which have the same number of layers, and the multiple encoders can improve the learning ability of the model.</p>
<p>The input of MGAVAE is the feature matrix <inline-formula>
<mml:math id="M57">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> or <inline-formula>
<mml:math id="M58">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and the adjacency matrix <inline-formula>
<mml:math id="M59">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> or<inline-formula>
<mml:math id="M60">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> calculated form the comprehensive microbial or drug similarity matrix <inline-formula>
<mml:math id="M61">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> or <inline-formula>
<mml:math id="M62">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and the microbe&#x2013;drug similarity matrix <inline-formula>
<mml:math id="M63">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. The key point of the variational autoencoder is the application of a two-layer graphical neural network structure for generating low-dimensional representations. The first layer of the neural network is used to compute a low-dimensional feature matrix <inline-formula>
<mml:math id="M64">
<mml:mi>X</mml:mi>
</mml:math>
</inline-formula>:</p>
<disp-formula id="EQ12">
<label>(12)</label>
<mml:math id="M65">
<mml:mrow>
<mml:mi mathvariant="normal">X</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="normal">GCN</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi mathvariant="normal">X,</mml:mi>
<mml:mover accent="true">
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mo>&#x02DC;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mfenced>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="normal">LeakyReLU</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mo>&#x02DC;</mml:mo>
</mml:mover>
<mml:mi mathvariant="normal">XW</mml:mi>
</mml:mrow>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ13">
<label>(13)</label>
<mml:math id="M66">
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mo>&#x02DC;</mml:mo>
</mml:mover>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mn>2</mml:mn>
</mml:mfrac>
</mml:mrow>
</mml:msup>
<mml:mi>A</mml:mi>
<mml:msup>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mn>2</mml:mn>
</mml:mfrac>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M67">
<mml:mover accent="true">
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mo>&#x02DC;</mml:mo>
</mml:mover>
</mml:math>
</inline-formula> is the symmetric normalized adjacency matrix, <inline-formula>
<mml:math id="M68">
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>y</mml:mi>
<mml:mi>Re</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the activation function, and <inline-formula>
<mml:math id="M69">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the weight parameter of this layer of graph neural net.</p>
<p>In the second layer of the graph neural network, the mean and variance vectors of the feature matrix are computed using the weight parameter <inline-formula>
<mml:math id="M70">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>:</p>
<disp-formula id="EQ14">
<label>(14)</label>
<mml:math id="M71">
<mml:mrow>
<mml:mi>&#x03BC;</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">GCN</mml:mi>
</mml:mrow>
<mml:mi>&#x03BC;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>X</mml:mi>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ15">
<label>(15)</label>
<mml:math id="M72">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:msup>
<mml:mi>&#x03C3;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mi>G</mml:mi>
<mml:mi>C</mml:mi>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>&#x03C3;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>X</mml:mi>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where the mean and variance share the same class of weight coefficients. A reparameterization method is used to compute the obtained low-dimensional features:</p>
<disp-formula id="EQ16">
<label>(16)</label>
<mml:math id="M73">
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi>&#x03BC;</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>&#x03C3;</mml:mi>
<mml:mo>&#x2299;</mml:mo>
<mml:mi>&#x03B5;</mml:mi>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M74">
<mml:mrow>
<mml:mi>&#x03B5;</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> represents the standard normal distribution. Here, the decoder is implemented in terms of the matrix inner product, and hence the adjacency matrix is reconstructed as follows:</p>
<disp-formula id="EQ17">
<label>(17)</label>
<mml:math id="M75">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi mathvariant="normal">|</mml:mi>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mi>&#x03C3;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mo>&#x22C5;</mml:mo>
<mml:msup>
<mml:mi>Z</mml:mi>
<mml:mi mathvariant="normal">T</mml:mi>
</mml:msup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>Finally, the loss function contains two types of errors. The first type is the reconstruction error, which measures the direct similarity of the input and output adjacency matrices. The second type of error is to make the initial label q and the predicted label p as close as possible. The mathematical expression for the loss function is as follows:</p>
<disp-formula id="EQ18">
<label>(18)</label>
<mml:math id="M76">
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi mathvariant="fraktur">q</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mi mathvariant="normal">|,X|,A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:msub>
<mml:mfenced close="]" open="[">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mi>p</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi mathvariant="normal">|Z</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>K</mml:mi>
<mml:mi>L</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mi mathvariant="normal">|,</mml:mi>
<mml:mi>X</mml:mi>
<mml:mi mathvariant="normal">|,</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2225;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mfenced>
<mml:mi>Z</mml:mi>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M77">
<mml:mrow>
<mml:mi>K</mml:mi>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mo>.</mml:mo>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> represents the Kullback&#x2013;Leibler divergence between two probability distributions. Finally, the generated low-dimensional features are integrated to obtain the final input for the prediction model.</p>
</sec>
<sec id="sec13">
<label>3.3</label>
<title>Graph convolutional neural networks</title>
<p>In this article, modified graph convolution (MGC) is embedded into a variational autoencoder for computing low-dimensional features. Modified graph convolution is modified on the original graph convolution neural network (GCN) to correct the model complexity and redundant computation. The calculation is simplified, and the efficiency is improved by reducing the collapsed weight matrices and non-linearities between successive layers, where the convolution kernel is modified as follows:</p>
<disp-formula id="EQ19">
<label>(19)</label>
<mml:math id="M78">
<mml:mrow>
<mml:msub>
<mml:mi>Y</mml:mi>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>G</mml:mi>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>x</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mo>&#x2026;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>X</mml:mi>
<mml:msup>
<mml:mi>M</mml:mi>
<mml:mrow>
<mml:mfenced>
<mml:mn>1</mml:mn>
</mml:mfenced>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2026;</mml:mo>
<mml:msup>
<mml:mi>M</mml:mi>
<mml:mrow>
<mml:mfenced>
<mml:mi>K</mml:mi>
</mml:mfenced>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
<mml:mo>=</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>x</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:msup>
<mml:mi>S</mml:mi>
<mml:mi>K</mml:mi>
</mml:msup>
<mml:mi>X</mml:mi>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M79">
<mml:mi>S</mml:mi>
</mml:math>
</inline-formula> is the normalized adjacency matrix, <inline-formula>
<mml:math id="M80">
<mml:mi>X</mml:mi>
</mml:math>
</inline-formula> is the feature matrix, <inline-formula>
<mml:math id="M81">
<mml:mi>M</mml:mi>
</mml:math>
</inline-formula> is the weight matrix, and <inline-formula>
<mml:math id="M82">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>max</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the normalized exponential function.</p>
</sec>
<sec id="sec14">
<label>3.4</label>
<title>Graph attention networks</title>
<p>The graph attention network learns the representation of nodes on the graph through the attention mechanism, assigning different learning weights to different neighboring nodes so that the correlation between node features is better integrated into the model and better prediction performance is obtained (<xref ref-type="bibr" rid="ref3">Bian et al., 2021</xref>). It uses the graph as an input, including the structural information of the graph and the graph node features. For a two-part graph consisting of microbe and drug associations, the node features are constructed as <inline-formula>
<mml:math id="M83">
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>d</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, and the adjacency matrix of the graph is constructed as follows:</p>
<disp-formula id="EQ20">
<label>(20)</label>
<mml:math id="M84">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mn>0</mml:mn>
</mml:mtd>
<mml:mtd>
<mml:mi>A</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msup>
<mml:mi>A</mml:mi>
<mml:mi mathvariant="normal">T</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M85">
<mml:mi>A</mml:mi>
</mml:math>
</inline-formula> is the association matrix of microbes with drugs.</p>
<p>A linear transformation of the input features to enhance their expressiveness is defined as follows:</p>
<disp-formula id="EQ21">
<label>(21)</label>
<mml:math id="M86">
<mml:mrow>
<mml:msup>
<mml:mi>Z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mi>Z</mml:mi>
<mml:mo>&#x22C5;</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>Z</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M87">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>Z</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is a learnable weight matrix.</p>
<p>The core idea of graph convolutional networks is to aggregate domain information to update node features. Considering that the importance of different nodes is not the same, a self-attention mechanism is used on the nodes to compute the non-normalized attention coefficient <inline-formula>
<mml:math id="M88">
<mml:mrow>
<mml:msub>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> using the current node <inline-formula>
<mml:math id="M89">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> and its first-order neighboring node <inline-formula>
<mml:math id="M90">
<mml:mi>j</mml:mi>
</mml:math>
</inline-formula>, defined as follows:</p>
<disp-formula id="EQ22">
<label>(22)</label>
<mml:math id="M91">
<mml:mrow>
<mml:msub>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi>L</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>y</mml:mi>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced>
<mml:mrow>
<mml:mfenced>
<mml:mrow>
<mml:msubsup>
<mml:mi>Z</mml:mi>
<mml:mi>i</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
<mml:mo>&#x2225;</mml:mo>
<mml:msubsup>
<mml:mi>Z</mml:mi>
<mml:mi>j</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x22C5;</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M92">
<mml:mo>&#x2225;</mml:mo>
</mml:math>
</inline-formula> represents the concatenation operation, <inline-formula>
<mml:math id="M93">
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>y</mml:mi>
<mml:mi>Re</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the activation function, and<inline-formula>
<mml:math id="M94">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the learnable matrix.</p>
<p>The attention coefficient after using the softmax normalized exponential function is calculated as follows:</p>
<disp-formula id="EQ23">
<label>(23)</label>
<mml:math id="M95">
<mml:mrow>
<mml:msub>
<mml:mi>a</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mi>&#x03F5;</mml:mi>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munder>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M96">
<mml:mrow>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the node<inline-formula>
<mml:math id="M97">
<mml:mrow>
<mml:mspace width="thickmathspace"/>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of the first-order neighbor node.</p>
<p>The first-order neighborhood features are updated by aggregating the <inline-formula>
<mml:math id="M98">
<mml:mi>l</mml:mi>
</mml:math>
</inline-formula> aggregation of first-order neighborhood features with attention coefficients in the layer, updating the <inline-formula>
<mml:math id="M99">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> node features of the layer, which are defined as follows:</p>
<disp-formula id="EQ24">
<label>(24)</label>
<mml:math id="M100">
<mml:mrow>
<mml:msubsup>
<mml:mi>Z</mml:mi>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:mi>L</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>y</mml:mi>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mi>&#x03F5;</mml:mi>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munder>
<mml:msubsup>
<mml:mi>a</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mi>l</mml:mi>
</mml:msubsup>
<mml:msubsup>
<mml:msup>
<mml:mi>Z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mi>l</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>Utilizing multi-head attention to expand model capabilities and stabilize the learning process. In this process, the initial node features have a dimension of <inline-formula>
<mml:math id="M101">
<mml:mrow>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, first, the initial nodes are replicated to obtain <inline-formula>
<mml:math id="M102">
<mml:mi>M</mml:mi>
</mml:math>
</inline-formula> feature matrices of size <inline-formula>
<mml:math id="M103">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula>
<mml:math id="M104">
<mml:mi>N</mml:mi>
</mml:math>
</inline-formula> represents the number of nodes. Each replicated feature matrix is then processed with different weight matrices <inline-formula>
<mml:math id="M105">
<mml:mrow>
<mml:msubsup>
<mml:mi>W</mml:mi>
<mml:mi>Z</mml:mi>
<mml:mi>h</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> to compute the outputs of <inline-formula>
<mml:math id="M106">
<mml:mi>M</mml:mi>
</mml:math>
</inline-formula> attention heads. The dimension of each attention head&#x2019;s output is the same as the initial node feature dimension. Subsequently, the outputs are concatenated to obtain <inline-formula>
<mml:math id="M107">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> layer output feature of dimension <inline-formula>
<mml:math id="M108">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mi>M</mml:mi>
<mml:mo>&#x00D7;</mml:mo>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. The output feature of each node <inline-formula>
<mml:math id="M109">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> can be calculated as follows:</p>
<disp-formula id="EQ25">
<label>(25)</label>
<mml:math id="M110">
<mml:mrow>
<mml:msubsup>
<mml:mi>Z</mml:mi>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mo>&#x2225;</mml:mo>
<mml:mrow>
<mml:mi>h</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>M</mml:mi>
</mml:msubsup>
<mml:mi>L</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>y</mml:mi>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mo stretchy="false">(</mml:mo>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mi>&#x03F5;</mml:mi>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munder>
<mml:msubsup>
<mml:mi>a</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mi>l</mml:mi>
</mml:msubsup>
<mml:msubsup>
<mml:msup>
<mml:mi>Z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mi>l</mml:mi>
</mml:msubsup>
<mml:mo>&#x22C5;</mml:mo>
<mml:msubsup>
<mml:mi>W</mml:mi>
<mml:mi>Z</mml:mi>
<mml:mi>h</mml:mi>
</mml:msubsup>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M111">
<mml:mrow>
<mml:msubsup>
<mml:mi>W</mml:mi>
<mml:mi>Z</mml:mi>
<mml:mi>h</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> represents the weight matrices; each attention head has its own weight matrix. During model training, weight matrices are randomly initialized and continuously adjusted through optimization algorithms to minimize the model&#x2019;s loss function and obtain suitable weight matrices.</p>
</sec>
<sec id="sec15">
<label>3.5</label>
<title>Random forest</title>
<p>The random forest algorithm is a well-known integrated learning method. The core idea is to build a forest in a random way, which consists of many decision trees, and the decision trees are used as the base classifiers to form a large multi-classifier. When test data were inputted into the model, the output categories of multiple decision trees were voted to get the final prediction. A decision tree is actually a process of node split, which starts from the root node and continuously splits downward. Knowing that the dataset can no longer be split, the decision tree stops growing.</p>
<p>The core idea is to select <inline-formula>
<mml:math id="M112">
<mml:mi>n</mml:mi>
</mml:math>
</inline-formula> samples from the training set as a training subset and then generate a decision tree, which is a base classifier, the above process is repeated a total of <inline-formula>
<mml:math id="M113">
<mml:mi>n</mml:mi>
</mml:math>
</inline-formula> times, generating <inline-formula>
<mml:math id="M114">
<mml:mi>n</mml:mi>
</mml:math>
</inline-formula> decision trees to form the final random forest. Each base classifier can participate in decision-making and for classification. The final result of the model is decided by the voting of each base classifier, and the class label with the largest number of votes for the classification result is selected.</p>
</sec>
</sec>
<sec sec-type="results" id="sec16">
<label>4</label>
<title>Results</title>
<sec id="sec17">
<label>4.1</label>
<title>Evaluation indicators</title>
<p>In this article, five-fold cross-validation and ten-fold cross-validation will be used to evaluate the prediction performance of the computational model. In the case of the five-fold cross-validation, the specific steps are as follows: first, all the microbe&#x2013;drug association pairs were divided into five subsets. Each subset was saved individually as a test set; in turn, the remaining four subsets were used as a training set instead of selecting a microbe&#x2013;drug association pair from them as a test sample. After cross-validation of the model, the receiver operating characteristic (ROC) curve is usually plotted, and the AUC (area under the ROC curve) is calculated to visually assess the predictive performance of the computational model. For a comprehensive evaluation of this computational model, we also evaluate its performance using accuracy (Acc), precision (Pre), recall (R), F1 score, and area under the accuracy-recall curve (AUPR).</p>
<p>The formula for Acc, Pre, R, and F1 score is as follows:</p>
<disp-formula id="EQ26">
<label>(26)</label>
<mml:math id="M115">
<mml:mrow>
<mml:mi mathvariant="normal">Acc</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">TN</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">TN</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ27">
<label>(27)</label>
<mml:math id="M116">
<mml:mrow>
<mml:mi mathvariant="normal">Pre</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ28">
<label>(28)</label>
<mml:math id="M117">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ29">
<label>(29)</label>
<mml:math id="M118">
<mml:mrow>
<mml:mi mathvariant="normal">F</mml:mi>
<mml:mn>1</mml:mn>
<mml:mspace width="thickmathspace"/>
<mml:mi mathvariant="normal">score</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#x00D7;</mml:mo>
<mml:mi mathvariant="normal">Pre</mml:mi>
<mml:mo>&#x00D7;</mml:mo>
<mml:mi mathvariant="normal">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">Pre</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">R</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where TN is true negative, indicating the number of microbe&#x2013;drug non-associations correctly identified by the model in the negative samples. TP is true positive, indicating the number of microbe&#x2013;drug associations correctly identified by the model in the positive samples. FN is false positive, indicating the number of microbe&#x2013;drug non-associations incorrectly predicted as microbe&#x2013;drug non-associations by the model in the positive samples. FP is false positive, indicating the number of microbe&#x2013;drug associations incorrectly predicted as microbe&#x2013;drug associations by the model in the negative samples and the number of microbe&#x2013;drug associations in the negative sample.</p>
<p>The curves of ROC and AUPR for MGAVAEMDA obtained after five-fold and ten-fold cross-validation are shown in <xref ref-type="fig" rid="fig2">Figure 2</xref>.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Five-fold and ten-fold cross-validation results of MGAVAEMDA.</p>
</caption>
<graphic xlink:href="fmicb-15-1394302-g002.tif"/>
</fig>
<p>As can be seen from the results of the figure, the AUC and AUPR values of the MGAVAEMDA model on ten-fold cross-validation are 0.52 and 0.35% points, which is higher than the five-fold cross-validation. These data indicate that the MGAVAEMDA model utilizes more training data using ten-fold cross-validation. It can more accurately evaluate the performance of the MGAVAEMDA model in the prediction of microbe&#x2013;drug association.</p>
</sec>
<sec id="sec18">
<label>4.2</label>
<title>Influence of parameter selection</title>
<p>The analysis of the parameters can quantitatively assess the stability of the model (<xref ref-type="bibr" rid="ref16">Li et al., 2020</xref>). In order to obtain more accurate prediction results, the influence of different parameters on the prediction results was analyzed through experiments. The parameters are divided into three parts: the parameters in MGAVAE, the parameters in the random forest classifier, the parameters of the binarization threshold, and the parameters of the smooth window.</p>
<sec id="sec19">
<label>4.2.1</label>
<title>Parameter selection in MGAVAE</title>
<sec id="sec20">
<label>4.2.1.1</label>
<title>Hidden layer dimension</title>
<p>The fixed learning rate <inline-formula>
<mml:math id="M119">
<mml:mi>k</mml:mi>
</mml:math>
</inline-formula> was set at 0.01 according to the literature to analyze the effect of hidden layer dimension <inline-formula>
<mml:math id="M120">
<mml:mi>d</mml:mi>
</mml:math>
</inline-formula> on the performance of MGAVAEMDA. For each <inline-formula>
<mml:math id="M121">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>&#x03B5;</mml:mi>
<mml:mfenced close="}" open="{">
<mml:mrow>
<mml:mn>32</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>64</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>128</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>256</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula>, five-fold cross-validation, the corresponding AUC and AUPR values are obtained, as shown in <xref ref-type="fig" rid="fig3">Figure 3</xref> Graph A. According to <xref ref-type="fig" rid="fig3">Figure 3</xref>, the higher the hidden layer dimension, the smaller the error, but it will increase the complexity of the model and may also be overfitting. From the data, when the dimensionality increases from 32 to 128, the performance of MGAVAEMDA increases with it. When the dimensionality is 256, the values of AUC and AUPR are 1.17 and 1.07% points lower than the dimensionality of 128, respectively. The values of AUC and AUPR of the model are maximum when the dimension is 128. Therefore, setting the hidden layer dimension <italic>d</italic> to 128 ensures the model prediction performance and saves time cost.</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>Graph A: AUC and AUPR values for different hidden layer dimensions. Graph B: AUC and AUPR values for different learning rates. Graph C: AUC and AUPR values for different n_estimators. Graph D: AUC, AUPR, and Acc values for different windows.</p>
</caption>
<graphic xlink:href="fmicb-15-1394302-g003.tif"/>
</fig>
</sec>
<sec id="sec21">
<label>4.2.1.2</label>
<title>Learning rate</title>
<p>Fix the hidden layer dimension <inline-formula>
<mml:math id="M122">
<mml:mi>d</mml:mi>
</mml:math>
</inline-formula> to 128 and change the learning rate to a common value. For each <inline-formula>
<mml:math id="M123">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mi>&#x03B5;</mml:mi>
<mml:mfenced close="}" open="{">
<mml:mrow>
<mml:mn>0.001</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>0.005</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>0.01</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>0.05</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula>, five-fold cross-validation, the corresponding AUC and AUPR values are obtained, as shown in <xref ref-type="fig" rid="fig3">Figure 3</xref>, Graph B. According to <xref ref-type="fig" rid="fig3">Figure 3</xref>, the values of AUC and AUPR are highest when the learning rate <inline-formula>
<mml:math id="M124">
<mml:mi>k</mml:mi>
</mml:math>
</inline-formula> is 0.01.</p>
</sec>
</sec>
<sec id="sec22">
<label>4.2.2</label>
<title>Parameter selection in the classifier</title>
<sec id="sec23">
<label>4.2.2.1</label>
<title>n_estimators</title>
<p>n_estimators is the number of weak learner. Generally speaking, if n_estimators is too small, it is easy to underfit. If n_estimators is too large, it is easy to overfit, so we usually choose a moderate value. For random forest, increasing the number of &#x201C;sub-models&#x201D; (n_estimators) can significantly reduce the variance of the overall model and will not have any effect on the bias or variance of the sub-models. The accuracy of the model increases with the increase in the number of sub-models, and there is an upper limit to the increase in accuracy because the reduction is the second term of the overall model variance formula. In this article, in order to obtain the optimal n_estimators, for every <inline-formula>
<mml:math id="M125">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>_</mml:mo>
<mml:mi>e</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>&#x03B5;</mml:mi>
<mml:mfenced close="}" open="{">
<mml:mrow>
<mml:mn>10</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>50</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>100</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>150</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>200</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula>, five-fold cross-validation, the corresponding AUC and AUPR values are obtained, as shown in <xref ref-type="fig" rid="fig3">Figure 3</xref> Graph C. According to <xref ref-type="fig" rid="fig3">Figure 3</xref>, when n_estimators increase from 10 to 100, the performance of MGAVAEMDA increases with it from 100 to 200. The performance of MGAVAEMDA decreases with it, and when the value of n_estimators is 100, the accuracy arrives at the highest level and the overall variance is the smallest. Therefore, n_estimators is set as 100.</p>
</sec>
</sec>
<sec id="sec24">
<label>4.2.3</label>
<title>Parameter selection in the smoothing window</title>
<sec id="sec25">
<label>4.2.3.1</label>
<title>Smooth window size n</title>
<p>Using a smoothing window to reduce the feature noise in the similarity matrix may lead to the loss of useful information in the similarity network. Adjusting the size of the smoothing window is the key. Too large a window may lead to loss of detailed information, while too small a window may not be effective for noise reduction. Therefore, several experiments are needed in the smoothing process to find the most suitable window size. For each <inline-formula>
<mml:math id="M126">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mfenced close="}" open="{">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>5</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>7</mml:mn>
<mml:mi mathvariant="normal">,</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mn>9</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula>, five-fold cross-validation, <xref ref-type="fig" rid="fig3">Figure 3</xref> Graph D shows that when the window size is 5, AUC, AUPR, and Acc are the highest, so smooth window is set as 5.</p>
</sec>
</sec>
<sec id="sec26">
<label>4.2.4</label>
<title>Parameter selection in binarization</title>
<sec id="sec27">
<label>4.2.4.1</label>
<title>Thresholds (<italic>m</italic>th) and (<italic>d</italic>th)</title>
<p>The <italic>m</italic>th and <italic>d</italic>th are the threshold points for binarization of microbe-integrated similarity networks and drug-integrated similarity networks, respectively. Higher thresholds can effectively reduce the noise in the similarity network but also eliminate the useful information in the similarity network. As shown in <xref ref-type="fig" rid="fig4">Figure 4</xref>, the performance of the model gradually improves with the increase of <italic>m</italic>th and <italic>d</italic>th, and the AUC and AUPR reach the maximum when <italic>m</italic>th and <italic>d</italic>th reach 0.8 and 0.7, respectively. In order to ensure that there is more useful information in the similar network and to achieve the best performance of the model, this article considers that it is most appropriate to set <italic>m</italic>th and <italic>d</italic>th to 0.8 and 0.7, respectively.</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>Performance with different combinations of the two hyperparameters.</p>
</caption>
<graphic xlink:href="fmicb-15-1394302-g004.tif"/>
</fig>
</sec>
</sec>
</sec>
<sec id="sec28">
<label>4.3</label>
<title>Comparison of different datasets</title>
<p>In order to further verify the prediction ability of the MGAVAEMDA model, this article conducted extension experiments on both aBiofilm and DrugVirus, two microbe&#x2013;drug databases. After the five-fold cross-validation, the ROC curves and AUPR curves of the three datasets were obtained, which are shown in the following figure. As shown in <xref ref-type="fig" rid="fig5">Figure 5</xref>, MGAVAEMDA achieved AUC values of 0.9357, 0.8563, and 0.8490 and AUPR values of 0.9378, 0.8601, and 0.8550 on the MDAD, aBiofilm, and DrugVirus datasets, respectively. The experimental results showed that MGAVAEMDA achieved AUC values and AUPR values that achieved more than 0.8400 and 0.8500 prediction results on different datasets. The results indicate that the model is robust and can be applied to different scales of data.</p>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Comparison of prediction results on different datasets.</p>
</caption>
<graphic xlink:href="fmicb-15-1394302-g005.tif"/>
</fig>
</sec>
<sec id="sec29">
<label>4.4</label>
<title>Comparison with existing methods</title>
<p>Under the same dataset conditions, the performance of MGAVAEMDA is compared with five advanced models: HMDAKATZ, GATECDA (<xref ref-type="bibr" rid="ref5">Deng et al., 2022</xref>), NTSHMDA (<xref ref-type="bibr" rid="ref19">Luo and Long, 2020</xref>), EGATMDA, and GCNMDA using AUC, AUPR, Acc, Pre, R, and F1 score as the evaluation indicators, and the parameters involved are selected to be the optimal ones recommended in the respective models. The ROC curves and AUPR curves after the five-fold cross-validation are shown in <xref ref-type="fig" rid="fig6">Figure 6</xref>.</p>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>PR curve, ROC curve, Pre, ACC, R, and F1 score curves comparison of the proposed model and existing models.</p>
</caption>
<graphic xlink:href="fmicb-15-1394302-g006.tif"/>
</fig>
<p>MGAVAEMDA is superior to the other five groups of comparison experiments, as shown in <xref ref-type="fig" rid="fig6">Figure 6</xref>. The AUC values are higher by 4.6, 5.54, 3, 1.76, and 2.92%, and the AUPR values are higher by 4.42, 5.33, 2.66, 1.47, and 2.54%, respectively. For the other four evaluation indicators, MGAVAEMDA is also better than the other five comparison experiments.</p>
<p>Thus, the MGAVAEMDA model has better predictive performance.</p>
</sec>
<sec id="sec30">
<label>4.5</label>
<title>Ablation experiments</title>
<p>Ablation experiments are performed by progressively removing modules from the model to assess how much these modules contribute to the overall performance.</p>
<p>(1) To verify the effect of the introduced self-attention mechanism, reconstruction loss function, and KL divergence loss function on the three times improvement of the prediction accuracy of the model MGAVAEMDA, this article conducts four sets of comparative experiments: (1) Group 1: introducing the KL divergence loss function; (2) Group 2: introducing the reconstruction loss function; (3) Group 3: introducing the attention mechanism; (4) Group 4: blank experiment.</p>
<p>After five-fold cross-validation, the ROC curves and AUPR curves of the four group comparison experiments were obtained, which are shown in <xref ref-type="fig" rid="fig7">Figure 7</xref>. As can be seen in <xref ref-type="fig" rid="fig7">Figure 7</xref>, both the AUC and AUPR values obtained from the MGAVAEMDA model are better than the results from four groups of comparison experiments. Among them, the AUC value of the first group is 34.81% higher than that of the fourth group. The AUPR value is 35.11%, which is higher than the results of the fourth group. It indicates that the addition of the KL divergence loss function reduces the error and improves the optimization ability of the model. The AUC value of the second group is 38.21% higher than that of the fourth group. The AUPR value is 38.65% higher than that of the fourth group, which indicates that the addition of the reconstruction loss function also reduces the error and improves the optimization ability of the model. The AUC value of the third group is 35.39%, higher than that of the fourth group. The AUPR value is 35.69 higher than that of the fourth group, indicating that the addition of the self-attention mechanism reduces the dependence on other information and significantly helps in the fusion of microbe medicines. The AUC value of the MGAVAEMDA is higher than that of groups 1, 2, and 3 by 6.73, 3.33, and 6.15%, and the AUPR values were 6.24, 2.70, and 5.66%, higher than those of the first, second, and third groups, respectively. Those results indicate that the combination of the KL divergence loss function. The reconstruction loss function and the self-attention mechanism can obtain more useful information, which can improve the prediction accuracy of the model. All the groups have shown that the introduction of the KL divergence loss function, reconstruction loss function, and attention mechanism is crucial for the improvement of the prediction accuracy of MGAVAEMDA.</p>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>Comparison of ablation experimental results.</p>
</caption>
<graphic xlink:href="fmicb-15-1394302-g007.tif"/>
</fig>
<p>(2) To verify the effect of the introduced modified graph convolutional neural network on improving the computational efficiency of the model MGAVAEMDA. This article sets up a comparative experiment with fixed hardware equipment. The running time of the model without modification is 422&#x2009;s, and the running time of the model after modification is 353&#x2009;s, which is significantly improved, indicating that the introduction of the modified graph convolutional neural network improves the computational efficiency of the model.</p>
</sec>
<sec id="sec31">
<label>4.6</label>
<title>Case studies</title>
<p>To evaluate the predictive effectiveness of the MGAVAEMDA model, two case studies were conducted using ceftazidime and curcumin. After calculating the predicted microbes associated with these two drugs, the top 10 microbes were screened after arranging the obtained association prediction scores in descending order, as shown in <xref ref-type="table" rid="tab1">Tables 1</xref>, <xref ref-type="table" rid="tab2">2</xref>.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Top 10 microbes related to ceftazidime.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Ranking</th>
<th align="left" valign="top">Name of microbe</th>
<th align="center" valign="top">PMID number</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">1</td>
<td align="left" valign="middle">
<italic>Stenotrophomonas maltophilia</italic>
</td>
<td align="center" valign="middle">37615040</td>
</tr>
<tr>
<td align="left" valign="middle">2</td>
<td align="left" valign="middle">
<italic>Haemophilus influenzae</italic>
</td>
<td align="center" valign="middle">6376458</td>
</tr>
<tr>
<td align="left" valign="middle">3</td>
<td align="left" valign="middle">
<italic>Shigella flexneri</italic>
</td>
<td align="center" valign="middle">31519769</td>
</tr>
<tr>
<td align="left" valign="middle">4</td>
<td align="left" valign="middle">
<italic>Escherichia coli</italic>
</td>
<td align="center" valign="middle">37574665</td>
</tr>
<tr>
<td align="left" valign="middle">5</td>
<td align="left" valign="middle">
<italic>Pseudomonas aeruginosa</italic>
</td>
<td align="center" valign="middle">34990760</td>
</tr>
<tr>
<td align="left" valign="middle">6</td>
<td align="left" valign="middle">
<italic>Bacillus subtilis</italic>
</td>
<td align="center" valign="middle">31420587</td>
</tr>
<tr>
<td align="left" valign="middle">7</td>
<td align="left" valign="middle">
<italic>Mycobacterium tuberculosis</italic>
</td>
<td align="center" valign="middle">28875168</td>
</tr>
<tr>
<td align="left" valign="middle">8</td>
<td align="left" valign="middle">
<italic>Mycobacterium avium</italic>
</td>
<td align="center" valign="middle">28922808</td>
</tr>
<tr>
<td align="left" valign="middle">9</td>
<td align="left" valign="middle"><italic>Streptococcus pneumoniae</italic> serotype 4</td>
<td align="center" valign="middle">8126192</td>
</tr>
<tr>
<td align="left" valign="middle">10</td>
<td align="left" valign="middle">
<italic>Proteus vulgaris</italic>
</td>
<td align="center" valign="middle">19802966</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Top 10 microbes related to curcumin.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Ranking</th>
<th align="left" valign="top">Name of microbe</th>
<th align="center" valign="top">PMID number</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">1</td>
<td align="left" valign="middle">
<italic>Streptococcus mutans</italic>
</td>
<td align="center" valign="middle">23778072</td>
</tr>
<tr>
<td align="left" valign="middle">2</td>
<td align="left" valign="middle">
<italic>Proteus mirabilis</italic>
</td>
<td align="center" valign="middle">21808656</td>
</tr>
<tr>
<td align="left" valign="middle">3</td>
<td align="left" valign="middle">
<italic>Vibrio anguillarum</italic>
</td>
<td align="center" valign="middle">31930829</td>
</tr>
<tr>
<td align="left" valign="middle">4</td>
<td align="left" valign="middle">
<italic>Pseudomonas aeruginosa</italic>
</td>
<td align="center" valign="middle">32421995</td>
</tr>
<tr>
<td align="left" valign="middle">5</td>
<td align="left" valign="middle">
<italic>Haemophilus influenzae</italic>
</td>
<td align="center" valign="middle">27538525</td>
</tr>
<tr>
<td align="left" valign="middle">6</td>
<td align="left" valign="middle">
<italic>Vibrio cholerae</italic>
</td>
<td align="center" valign="middle">35140698</td>
</tr>
<tr>
<td align="left" valign="middle">7</td>
<td align="left" valign="middle">
<italic>Burkholderia cenocepacia</italic>
</td>
<td align="center" valign="middle">Unverified</td>
</tr>
<tr>
<td align="left" valign="middle">8</td>
<td align="left" valign="middle">
<italic>Enterococcus faecalis</italic>
</td>
<td align="center" valign="middle">34320428</td>
</tr>
<tr>
<td align="left" valign="middle">9</td>
<td align="left" valign="middle">
<italic>Burkholderia multivorans</italic>
</td>
<td align="center" valign="middle">Unverified</td>
</tr>
<tr>
<td align="left" valign="middle">10</td>
<td align="left" valign="middle">
<italic>Eikenella corrodens</italic>
</td>
<td align="center" valign="middle">Unverified</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The prediction data of the above two tables were obtained from the MGAVAEMDA model by searching the PMID database for relevant literature and reports. The data in the table show that among the top 10 microbes predicted in the MGAVAEMDA model with ceftazidime and curcumin, 10 and 7 each have been confirmed in the literature. Among them, curcumin-mediated EDTA blue light PDI has a strong inhibitory effect on <italic>Streptococcus mutans</italic> in planktonic culture. It is expected to be a promising technique for disinfection of oral tissues due to its unclear targeting mechanism (<xref ref-type="bibr" rid="ref22">Nima et al., 2021</xref>). Curcumin has an inhibitory effect on urease activity in <italic>Proteus mirabilis</italic> (<xref ref-type="bibr" rid="ref23">Prywer and Torzewska, 2012</xref>), and the addition of curcumin increases the induction time and decreases the growth efficiency of guano stones as compared to the absence of curcumin. Curcumin has been used in the case of <italic>Proteus mirabilis</italic>-induced growth of guano crystals in association with urinary stone formation, which has been shown to have great potential for further research. The case study further validates that the MGAVAEMDA model has a good performance in identifying microbe-associated drugs and has a certain application value.</p>
</sec>
</sec>
<sec sec-type="conclusions" id="sec32">
<label>5</label>
<title>Conclusion</title>
<p>In this study, the MGAVAEMDA model was set up by using the variational autoencoder, modified graph convolutional neural network, and graph attention network to realize the microbe&#x2013;drug association prediction through two-stage fusion.</p>
<p>The case study indicates that the MGAVAEMDA model can overcome the shortcomings of other models, such as long training times and low prediction accuracy. Moreover, the MGAVAEMDA model has better prediction performance.</p>
<p>The predicted performance of the model will be improved after more biological datasets are integrated in the future, so this model&#x2019;s practical application will increase.</p>
</sec>
<sec sec-type="data-availability" id="sec33">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec sec-type="author-contributions" id="sec34">
<title>Author contributions</title>
<p>BW: Writing &#x2013; review &#x0026; editing. FJM: Writing &#x2013; editing. XXD: Data curation, Writing &#x2013; review &#x0026; editing. GDZ: Formal analysis, Writing &#x2013; review &#x0026; editing. JYL: Investigation, Writing &#x2013; review &#x0026; editing.</p>
</sec>
</body>
<back>
<sec sec-type="funding-information" id="sec35">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This study was supported by the Basic Scientific Research Business of Colleges and Universities in Heilongjiang Province (No. 145209125).</p>
</sec>
<sec sec-type="COI-statement" id="sec36">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="sec37">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Altschul</surname> <given-names>S. F.</given-names></name> <name><surname>Gish</surname> <given-names>W.</given-names></name> <name><surname>Miller</surname> <given-names>W.</given-names></name> <name><surname>Myers</surname> <given-names>E. W.</given-names></name> <name><surname>Lipman</surname> <given-names>D. J.</given-names></name></person-group> (<year>1990</year>). <article-title>Basic local alignment search tool</article-title>. <source>J. Mol. Biol.</source> <volume>215</volume>, <fpage>403</fpage>&#x2013;<lpage>410</lpage>. doi: <pub-id pub-id-type="doi">10.1016/s0022-2836(05)80360-2</pub-id></citation></ref>
<ref id="ref2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Andersen</surname> <given-names>P. I.</given-names></name> <name><surname>Ianevski</surname> <given-names>A.</given-names></name> <name><surname>Lysvand</surname> <given-names>H.</given-names></name> <name><surname>Vitkauskiene</surname> <given-names>A.</given-names></name> <name><surname>Oksenych</surname> <given-names>V.</given-names></name> <name><surname>Bj&#x00F8;r&#x00E5;s</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Discovery and development of safe-in-man broad-spectrum antiviral agents</article-title>. <source>Int. J. Infect. Dis.</source> <volume>93</volume>, <fpage>268</fpage>&#x2013;<lpage>276</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ijid.2020.02.018</pub-id>, PMID: <pub-id pub-id-type="pmid">32081774</pub-id></citation></ref>
<ref id="ref3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bian</surname> <given-names>C.</given-names></name> <name><surname>Lei</surname> <given-names>X. J.</given-names></name> <name><surname>Wu</surname> <given-names>F. X.</given-names></name></person-group> (<year>2021</year>). <article-title>GATCDA: predicting circRNA-disease associations based on graph attention network</article-title>. <source>Cancers</source> <volume>13</volume>:<fpage>2595</fpage>. doi: <pub-id pub-id-type="doi">10.3390/cancers13112595</pub-id>, PMID: <pub-id pub-id-type="pmid">34070678</pub-id></citation></ref>
<ref id="ref4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>de la Cuesta-zuluaga</surname> <given-names>J. K. E.</given-names></name> <name><surname>Huus</surname> <given-names>N. D.</given-names></name> <name><surname>Youngblut</surname> <given-names>J. S.</given-names></name> <name><surname>Escobar</surname></name> <name><surname>Ley</surname> <given-names>R. E.</given-names></name></person-group> (<year>2023</year>). <article-title>Obesity is the main driver of altered gut microbiome functions in the metabolically unhealthy</article-title>. <source>Gut Microbes</source> <volume>15</volume>:<fpage>2246634</fpage>. doi: <pub-id pub-id-type="doi">10.1080/19490976.2023.2246634</pub-id>, PMID: <pub-id pub-id-type="pmid">37680093</pub-id></citation></ref>
<ref id="ref5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Deng</surname> <given-names>L. Z. X.</given-names></name> <name><surname>Liu</surname> <given-names>Y. R.</given-names></name> <name><surname>Qian</surname></name> <name><surname>Zhang</surname> <given-names>J. P.</given-names></name></person-group> (<year>2022</year>). <article-title>Predicting circRNA-drug sensitivity associations via graph attention auto-encoder</article-title>. <source>BMC Bioinformatics</source> <volume>23</volume>:<fpage>160</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s12859-022-04694-y</pub-id>, PMID: <pub-id pub-id-type="pmid">35508967</pub-id></citation></ref>
<ref id="ref6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Durack</surname> <given-names>J.</given-names></name> <name><surname>Lynch</surname> <given-names>S. V.</given-names></name></person-group> (<year>2019</year>). <article-title>The gut microbiome: relationships with disease and opportunities for therapy</article-title>. <source>J. Exp. Med.</source> <volume>216</volume>, <fpage>20</fpage>&#x2013;<lpage>40</lpage>. doi: <pub-id pub-id-type="doi">10.1084/jem.20180448</pub-id>, PMID: <pub-id pub-id-type="pmid">30322864</pub-id></citation></ref>
<ref id="ref7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>ElRakaiby</surname> <given-names>M.</given-names></name> <name><surname>Dutilh</surname> <given-names>B. E.</given-names></name> <name><surname>Rizkallah</surname> <given-names>M. R.</given-names></name> <name><surname>Boleij</surname> <given-names>A.</given-names></name> <name><surname>Cole</surname> <given-names>J. N.</given-names></name> <name><surname>Aziz</surname> <given-names>R. K.</given-names></name></person-group> (<year>2014</year>). <article-title>Pharmacomicrobiomics: the impact of human microbiome variations on systems pharmacology and personalized therapeutics</article-title>. <source>OMICS</source> <volume>18</volume>, <fpage>402</fpage>&#x2013;<lpage>414</lpage>. doi: <pub-id pub-id-type="doi">10.1089/omi.2014.0018</pub-id>, PMID: <pub-id pub-id-type="pmid">24785449</pub-id></citation></ref>
<ref id="ref8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gill</surname> <given-names>S. R. M.</given-names></name> <name><surname>Pop</surname> <given-names>R. </given-names></name> <name><surname>Deboy</surname> <given-names>P. B.</given-names></name> <name><surname>Eckburg</surname> <given-names>P. J.</given-names></name> <name><surname>Turnbaugh</surname> <given-names>B. S.</given-names></name> <name><surname>Samuel</surname></name><etal/></person-group>. (<year>2006</year>). <article-title>Metagenomic analysis of the human distal gut microbiome</article-title>. <source>Science</source> <volume>312</volume>, <fpage>1355</fpage>&#x2013;<lpage>1359</lpage>. doi: <pub-id pub-id-type="doi">10.1126/science.1124234</pub-id>, PMID: <pub-id pub-id-type="pmid">16741115</pub-id></citation></ref>
<ref id="ref9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Haiser</surname> <given-names>H. J.</given-names></name> <name><surname>Gootenberg</surname> <given-names>D. B.</given-names></name> <name><surname>Chatman</surname> <given-names>K.</given-names></name> <name><surname>Sirasani</surname> <given-names>G.</given-names></name> <name><surname>Balskus</surname> <given-names>E. P.</given-names></name> <name><surname>Turnbaugh</surname> <given-names>P. J.</given-names></name></person-group> (<year>2013</year>). <article-title>Predicting and manipulating cardiac drug inactivation by the human gut bacterium <italic>Eggerthella lenta</italic></article-title>. <source>Science</source> <volume>341</volume>, <fpage>295</fpage>&#x2013;<lpage>298</lpage>. doi: <pub-id pub-id-type="doi">10.1126/science.1235872</pub-id>, PMID: <pub-id pub-id-type="pmid">23869020</pub-id></citation></ref>
<ref id="ref10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hattori</surname> <given-names>M.</given-names></name> <name><surname>Tanaka</surname> <given-names>N.</given-names></name> <name><surname>Kanehisa</surname> <given-names>M.</given-names></name> <name><surname>Goto</surname> <given-names>S.</given-names></name></person-group> (<year>2010</year>). <article-title>SIMCOMP/SUBCOMP: chemical structure search servers for network analyses</article-title>. <source>Nucleic Acids Res.</source> <volume>38</volume>, <fpage>W652</fpage>&#x2013;<lpage>W656</lpage>. doi: <pub-id pub-id-type="doi">10.1093/nar/gkq367</pub-id>, PMID: <pub-id pub-id-type="pmid">20460463</pub-id></citation></ref>
<ref id="ref11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huttenhower</surname> <given-names>C. D.</given-names></name> <name><surname>Gevers</surname> <given-names>R.</given-names></name> <name><surname>Knight</surname> <given-names>S.</given-names></name> <name><surname>Abubucker</surname> <given-names>J. H.</given-names></name> <name><surname>Badger</surname> <given-names>A. T.</given-names></name> <name><surname>Chinwalla</surname></name><etal/></person-group>. (<year>2012</year>). <article-title>Structure, function and diversity of the healthy human microbiome</article-title>. <source>Nature</source> <volume>486</volume>, <fpage>207</fpage>&#x2013;<lpage>214</lpage>. doi: <pub-id pub-id-type="doi">10.1038/nature11234</pub-id>, PMID: <pub-id pub-id-type="pmid">22699609</pub-id></citation></ref>
<ref id="ref12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>H. J.</given-names></name> <name><surname>Huang</surname> <given-names>Y. A.</given-names></name> <name><surname>You</surname> <given-names>Z. H.</given-names></name></person-group> (<year>2020</year>). <article-title>SAEROF: an ensemble approach for large-scale drug-disease association prediction by incorporating rotation forest and sparse autoencoder deep neural network</article-title>. <source>Sci. Rep.</source> <volume>10</volume>:<fpage>4972</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-020-61616-9</pub-id>, PMID: <pub-id pub-id-type="pmid">32188871</pub-id></citation></ref>
<ref id="ref13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kirubahari</surname> <given-names>R.</given-names></name> <name><surname>Amali</surname> <given-names>S. M. J.</given-names></name></person-group> (<year>2023</year>). <article-title>An improved restricted Boltzmann machine using Bayesian optimization for recommender systems</article-title>. <source>Evol. Syst</source>. <fpage>1</fpage>&#x2013;<lpage>13</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s12530-023-09520-1</pub-id></citation></ref>
<ref id="ref14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lei</surname> <given-names>X. J.</given-names></name> <name><surname>Zhang</surname> <given-names>C.</given-names></name></person-group> (<year>2019</year>). <article-title>Predicting metabolite-disease associations based on KATZ model</article-title>. <source>Biodata Mining</source> <volume>12</volume>:<fpage>19</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s13040-019-0206-z</pub-id>, PMID: <pub-id pub-id-type="pmid">31673292</pub-id></citation></ref>
<ref id="ref15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>H.</given-names></name> <name><surname>Hou</surname> <given-names>Z. J.</given-names></name> <name><surname>Zhang</surname> <given-names>W. G.</given-names></name> <name><surname>Qu</surname> <given-names>J.</given-names></name> <name><surname>Yao</surname> <given-names>H. B.</given-names></name> <name><surname>Chen</surname> <given-names>Y.</given-names></name></person-group> (<year>2023</year>). <article-title>Prediction of potential drug-microbe associations based on matrix factorization and a three-layer heterogeneous network</article-title>. <source>Comput. Biol. Chem.</source> <volume>104</volume>:<fpage>107857</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiolchem.2023.107857</pub-id>, PMID: <pub-id pub-id-type="pmid">37018909</pub-id></citation></ref>
<ref id="ref16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>S.</given-names></name> <name><surname>Liu</surname> <given-names>T.</given-names></name> <name><surname>Ning</surname> <given-names>C.</given-names></name> <name><surname>Zhang</surname> <given-names>Z.</given-names></name> <name><surname>Zhou</surname> <given-names>W.</given-names></name></person-group> (<year>2020</year>). <article-title>Neural inductive matrix completion with graph convolutional networks for miRNA-disease association prediction</article-title>. <source>Bioinformatics</source> <volume>36</volume>, <fpage>2538</fpage>&#x2013;<lpage>2546</lpage>. doi: <pub-id pub-id-type="doi">10.1093/bioinformatics/btz965</pub-id>, PMID: <pub-id pub-id-type="pmid">31904845</pub-id></citation></ref>
<ref id="ref17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>D. Y. J. Y.</given-names></name> <name><surname>Liu</surname> <given-names>J.</given-names></name> <name><surname>Luo</surname> <given-names>Y.</given-names></name> <name><surname>He</surname></name> <name><surname>Deng</surname> <given-names>L.</given-names></name></person-group> (<year>2022</year>). <article-title>MGATMDA: predicting microbe-disease associations via multi-component graph attention network</article-title>. <source>IEEE/ACM Trans. Comput. Biol. Bioinform.</source> <volume>19</volume>, <fpage>3578</fpage>&#x2013;<lpage>3585</lpage>. doi: <pub-id pub-id-type="doi">10.1109/tcbb.2021.3116318</pub-id>, PMID: <pub-id pub-id-type="pmid">34587092</pub-id></citation></ref>
<ref id="ref18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Long</surname> <given-names>Y. H. M.</given-names></name> <name><surname>Wu</surname> <given-names>Y.</given-names></name> <name><surname>Liu</surname> <given-names>C. K.</given-names></name> <name><surname>Kwoh</surname> <given-names>J. W.</given-names></name> <name><surname>Luo</surname></name> <name><surname>Li</surname> <given-names>X. L.</given-names></name></person-group> (<year>2020</year>). <article-title>Ensembling graph attention networks for human microbe-drug association prediction</article-title>. <source>Bioinformatics</source> <volume>36</volume>, <fpage>I779</fpage>&#x2013;<lpage>I786</lpage>. doi: <pub-id pub-id-type="doi">10.1093/bioinformatics/btaa891</pub-id>, PMID: <pub-id pub-id-type="pmid">33381844</pub-id></citation></ref>
<ref id="ref19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Luo</surname> <given-names>J. W.</given-names></name> <name><surname>Long</surname> <given-names>Y. H.</given-names></name></person-group> (<year>2020</year>). <article-title>NTSHMDA: prediction of human microbe-disease association based on random walk by integrating network topological similarity</article-title>. <source>IEEE/ACM Trans. Comput. Biol. Bioinform.</source> <volume>17</volume>, <fpage>1341</fpage>&#x2013;<lpage>1351</lpage>. doi: <pub-id pub-id-type="doi">10.1109/tcbb.2018.2883041</pub-id>, PMID: <pub-id pub-id-type="pmid">30489271</pub-id></citation></ref>
<ref id="ref20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>Y. J.</given-names></name> <name><surname>Liu</surname> <given-names>Q. Q.</given-names></name></person-group> (<year>2022</year>). <article-title>Generalized matrix factorization based on weighted hypergraph learning for microbe-drug association prediction</article-title>. <source>Comput. Biol. Med.</source> <volume>145</volume>:<fpage>105503</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiomed.2022.105503</pub-id>, PMID: <pub-id pub-id-type="pmid">35427986</pub-id></citation></ref>
<ref id="ref21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>Q.</given-names></name> <name><surname>Tan</surname> <given-names>Y. Q.</given-names></name> <name><surname>Wang</surname> <given-names>L.</given-names></name></person-group> (<year>2023</year>). <article-title>GACNNMDA: a computational model for predicting potential human microbe-drug associations based on graph attention network and CNN-based classifier</article-title>. <source>BMC Bioinformatics</source> <volume>24</volume>:<fpage>35</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s12859-023-05158-7</pub-id>, PMID: <pub-id pub-id-type="pmid">36732704</pub-id></citation></ref>
<ref id="ref22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nima</surname> <given-names>G.</given-names></name> <name><surname>Soto-Montero</surname> <given-names>J.</given-names></name> <name><surname>Alves</surname> <given-names>L. A.</given-names></name> <name><surname>Mattos-Graner</surname> <given-names>R. O.</given-names></name> <name><surname>Giannini</surname> <given-names>M.</given-names></name></person-group> (<year>2021</year>). <article-title>Photodynamic inactivation of <italic>Streptococcus mutans</italic> by curcumin in combination with EDTA</article-title>. <source>Dent. Mater.</source> <volume>37</volume>, <fpage>E1</fpage>&#x2013;<lpage>E14</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.dental.2020.09.015</pub-id>, PMID: <pub-id pub-id-type="pmid">33143940</pub-id></citation></ref>
<ref id="ref23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Prywer</surname> <given-names>J.</given-names></name> <name><surname>Torzewska</surname> <given-names>A.</given-names></name></person-group> (<year>2012</year>). <article-title>Effect of curcumin against <italic>Proteus mirabilis</italic> during crystallization of struvite from artificial urine</article-title>. <source>Evid. Based Complement. Alternat. Med.</source> <volume>2012</volume>, <fpage>1</fpage>&#x2013;<lpage>7</lpage>. doi: <pub-id pub-id-type="doi">10.1155/2012/862794</pub-id>, PMID: <pub-id pub-id-type="pmid">21808656</pub-id></citation></ref>
<ref id="ref24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rajput</surname> <given-names>A.</given-names></name> <name><surname>Thakur</surname> <given-names>A.</given-names></name> <name><surname>Sharma</surname> <given-names>S.</given-names></name> <name><surname>Kumar</surname> <given-names>M.</given-names></name></person-group> (<year>2018</year>). <article-title>aBiofilm: a resource of anti-biofilm agents and their potential implications in targeting antibiotic drug resistance</article-title>. <source>Nucleic Acids Res.</source> <volume>46</volume>, <fpage>D894</fpage>&#x2013;<lpage>D900</lpage>. doi: <pub-id pub-id-type="doi">10.1093/nar/gkx1157</pub-id>, PMID: <pub-id pub-id-type="pmid">29156005</pub-id></citation></ref>
<ref id="ref25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ramirez</surname> <given-names>M.</given-names></name> <name><surname>Rajaram</surname> <given-names>S.</given-names></name> <name><surname>Steininger</surname> <given-names>R. J.</given-names></name> <name><surname>Osipchuk</surname> <given-names>D.</given-names></name> <name><surname>Roth</surname> <given-names>M. A.</given-names></name> <name><surname>Morinishi</surname> <given-names>L. S.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Diverse drug-resistance mechanisms can emerge from drug-tolerant cancer persister cells</article-title>. <source>Nat. Commun.</source> <volume>7</volume>:<fpage>10690</fpage>. doi: <pub-id pub-id-type="doi">10.1038/ncomms10690</pub-id>, PMID: <pub-id pub-id-type="pmid">26891683</pub-id></citation></ref>
<ref id="ref26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schwabe</surname> <given-names>R. F.</given-names></name> <name><surname>Jobin</surname> <given-names>C.</given-names></name></person-group> (<year>2013</year>). <article-title>The microbiome and cancer</article-title>. <source>Nat. Rev. Cancer</source> <volume>13</volume>, <fpage>800</fpage>&#x2013;<lpage>812</lpage>. doi: <pub-id pub-id-type="doi">10.1038/nrc3610</pub-id>, PMID: <pub-id pub-id-type="pmid">24132111</pub-id></citation></ref>
<ref id="ref27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Smith</surname> <given-names>T. F.</given-names></name> <name><surname>Waterman</surname> <given-names>M. S.</given-names></name></person-group> (<year>1981</year>). <article-title>Identification of common molecular subsequences</article-title>. <source>J. Mol. Biol.</source> <volume>147</volume>, <fpage>195</fpage>&#x2013;<lpage>197</lpage>. doi: <pub-id pub-id-type="doi">10.1016/0022-2836(81)90087-5</pub-id>, PMID: <pub-id pub-id-type="pmid">7265238</pub-id></citation></ref>
<ref id="ref28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sprockett</surname> <given-names>D.</given-names></name> <name><surname>Fukami</surname> <given-names>T.</given-names></name> <name><surname>Relman</surname> <given-names>D. A.</given-names></name></person-group> (<year>2018</year>). <article-title>Role of priority effects in the early-life assembly of the gut microbiota</article-title>. <source>Nat. Rev. Gastroenterol. Hepatol.</source> <volume>15</volume>, <fpage>197</fpage>&#x2013;<lpage>205</lpage>. doi: <pub-id pub-id-type="doi">10.1038/nrgastro.2017.173</pub-id>, PMID: <pub-id pub-id-type="pmid">29362469</pub-id></citation></ref>
<ref id="ref29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>Y. Z.</given-names></name> <name><surname>Zhang</surname> <given-names>D. H.</given-names></name> <name><surname>Cai</surname> <given-names>S. B.</given-names></name> <name><surname>Ming</surname> <given-names>Z.</given-names></name> <name><surname>Li</surname> <given-names>J. Q.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name></person-group> (<year>2018</year>). <article-title>MDAD: a special resource for microbe-drug associations</article-title>. <source>Front. Cell. Infect. Microbiol.</source> <volume>8</volume>:<fpage>424</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fcimb.2018.00424</pub-id>, PMID: <pub-id pub-id-type="pmid">30581775</pub-id></citation></ref>
<ref id="ref30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tan</surname> <given-names>Y. Q.</given-names></name> <name><surname>Zou</surname> <given-names>J.</given-names></name> <name><surname>Kuang</surname> <given-names>L. N.</given-names></name> <name><surname>Wang</surname> <given-names>X. Y.</given-names></name> <name><surname>Zeng</surname> <given-names>B.</given-names></name> <name><surname>Zhang</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>GSAMDA: a computational model for predicting potential microbe-drug associations based on graph attention network and sparse autoencoder</article-title>. <source>BMC Bioinformatics</source> <volume>23</volume>:<fpage>492</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s12859-022-05053-7</pub-id>, PMID: <pub-id pub-id-type="pmid">36401174</pub-id></citation></ref>
<ref id="ref31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>van Laarhoven</surname> <given-names>T.</given-names></name> <name><surname>Nabuurs</surname> <given-names>S. B.</given-names></name> <name><surname>Marchiori</surname> <given-names>E.</given-names></name></person-group> (<year>2011</year>). <article-title>Gaussian interaction profile kernels for predicting drug-target interaction</article-title>. <source>Bioinformatics</source> <volume>27</volume>, <fpage>3036</fpage>&#x2013;<lpage>3043</lpage>. doi: <pub-id pub-id-type="doi">10.1093/bioinformatics/btr500</pub-id>, PMID: <pub-id pub-id-type="pmid">21893517</pub-id></citation></ref>
<ref id="ref32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>F.</given-names></name> <name><surname>Huang</surname> <given-names>Z. A.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Zhu</surname> <given-names>Z.</given-names></name> <name><surname>Wen</surname> <given-names>Z.</given-names></name> <name><surname>Zhao</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>LRLSHMDA: Laplacian regularized least squares for human microbe-disease association prediction</article-title>. <source>Sci. Rep.</source> <volume>7</volume>:<fpage>7601</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-017-08127-2</pub-id>, PMID: <pub-id pub-id-type="pmid">28790448</pub-id></citation></ref>
<ref id="ref33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wen</surname> <given-names>L.</given-names></name> <name><surname>Ley</surname> <given-names>R. E.</given-names></name> <name><surname>Volchkov</surname> <given-names>P. Y.</given-names></name> <name><surname>Stranges</surname> <given-names>P. B.</given-names></name> <name><surname>Avanesyan</surname> <given-names>L.</given-names></name> <name><surname>Stonebraker</surname> <given-names>A. C.</given-names></name> <etal/></person-group>. (<year>2008</year>). <article-title>Innate immunity and intestinal microbiota in the development of type 1 diabetes</article-title>. <source>Nature</source> <volume>455</volume>, <fpage>1109</fpage>&#x2013;<lpage>1113</lpage>. doi: <pub-id pub-id-type="doi">10.1038/nature07336</pub-id>, PMID: <pub-id pub-id-type="pmid">18806780</pub-id></citation></ref>
<ref id="ref34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xiang</surname> <given-names>Y. T.</given-names></name> <name><surname>Li</surname> <given-names>W.</given-names></name> <name><surname>Zhang</surname> <given-names>Q. G.</given-names></name> <name><surname>Jin</surname> <given-names>Y.</given-names></name> <name><surname>Rao</surname> <given-names>W. W.</given-names></name> <name><surname>Zeng</surname> <given-names>L. N.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Timely research papers about COVID-19 in China</article-title>. <source>Lancet</source> <volume>395</volume>, <fpage>684</fpage>&#x2013;<lpage>685</lpage>. doi: <pub-id pub-id-type="doi">10.1016/s0140-6736(20)30375-5</pub-id>, PMID: <pub-id pub-id-type="pmid">32078803</pub-id></citation></ref>
<ref id="ref35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ximenez</surname> <given-names>C.</given-names></name> <name><surname>Torres</surname> <given-names>J.</given-names></name></person-group> (<year>2017</year>). <article-title>Development of microbiota in infants and its role in maturation of gut mucosa and immune system</article-title>. <source>Arch. Med. Res.</source> <volume>48</volume>, <fpage>666</fpage>&#x2013;<lpage>680</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.arcmed.2017.11.007</pub-id>, PMID: <pub-id pub-id-type="pmid">29198451</pub-id></citation></ref>
<ref id="ref36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>H. P.</given-names></name> <name><surname>Ding</surname> <given-names>Y. J.</given-names></name> <name><surname>Tang</surname> <given-names>J. J.</given-names></name> <name><surname>Guo</surname> <given-names>F.</given-names></name></person-group> (<year>2022</year>). <article-title>Inferring human microbe-drug associations via multiple kernel fusion on graph neural network</article-title>. <source>Knowl.-Based Syst.</source> <volume>238</volume>:<fpage>107888</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.knosys.2021.107888</pub-id></citation></ref>
<ref id="ref37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yoon</surname> <given-names>Y. K.</given-names></name> <name><surname>Kim</surname> <given-names>J.</given-names></name> <name><surname>Moon</surname> <given-names>C.</given-names></name> <name><surname>Lee</surname> <given-names>M. S.</given-names></name> <name><surname>Hur</surname> <given-names>J.</given-names></name> <name><surname>Lee</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Antimicrobial susceptibility of microorganisms isolated from patients with intraabdominal infection in Korea: a multicenter study</article-title>. <source>J. Korean Med. Sci.</source> <volume>34</volume>:<fpage>e309</fpage>. doi: <pub-id pub-id-type="doi">10.3346/jkms.2019.34.e309</pub-id>, PMID: <pub-id pub-id-type="pmid">31808326</pub-id></citation></ref>
<ref id="ref38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname> <given-names>L. Z.</given-names></name> <name><surname>Duan</surname> <given-names>G. H.</given-names></name> <name><surname>Yan</surname> <given-names>C.</given-names></name> <name><surname>Wang</surname> <given-names>J. X.</given-names></name></person-group>. (<year>2021</year>). <article-title>Prediction of microbe-drug associations based on chemical structures and the KATZ measure</article-title>. <source>Curr. Bioinforma.</source> <volume>16</volume>, <fpage>807</fpage>&#x2013;<lpage>819</lpage>. doi: <pub-id pub-id-type="doi">10.2174/1574893616666210204144721</pub-id></citation></ref>
</ref-list>
</back>
</article>