<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Pharmacol.</journal-id>
<journal-title>Frontiers in Pharmacology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Pharmacol.</abbrev-journal-title>
<issn pub-type="epub">1663-9812</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1529128</article-id>
<article-id pub-id-type="doi">10.3389/fphar.2024.1529128</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Pharmacology</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Accurate identification of snoRNA targets using variational graph autoencoder to advance the redevelopment of traditional medicines</article-title>
<alt-title alt-title-type="left-running-head">Wang et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fphar.2024.1529128">10.3389/fphar.2024.1529128</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Zhina</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Chen</surname>
<given-names>Yangyuan</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2897439/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Ma</surname>
<given-names>Hongming</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Gao</surname>
<given-names>Hong</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Zhu</surname>
<given-names>Yangbin</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2896734/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Hongwu</given-names>
</name>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Zhang</surname>
<given-names>Nan</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1550285/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Department of Pulmonary and Critical Care Medicine II</institution>, <institution>Emergency General Hospital</institution>, <addr-line>Beijing</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Department of Oncology</institution>, <institution>Emergency General Hospital</institution>, <addr-line>Beijing</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>School of Data Science and Artificial Intelligence</institution>, <institution>Wenzhou University of Technology</institution>, <addr-line>Wenzhou</addr-line>, <country>China</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>Respiratory Disease Center</institution>, <institution>Dongzhimen Hospital</institution>, <institution>Beijing University of Chinese Medicine</institution>, <addr-line>Beijing</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1472367/overview">Junlin Xu</ext-link>, Hunan University, China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1058102/overview">Xiangzheng Fu</ext-link>, Hunan University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2267638/overview">Xiong Li</ext-link>, East China Jiaotong University, China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Yangbin Zhu, <email>yangbinzhu1008@163.com</email>; Nan Zhang, <email>zhangnan20191217@sina.com</email>
</corresp>
</author-notes>
<pub-date pub-type="epub">
<day>06</day>
<month>01</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>15</volume>
<elocation-id>1529128</elocation-id>
<history>
<date date-type="received">
<day>16</day>
<month>11</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>11</day>
<month>12</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 Wang, Chen, Ma, Gao, Zhu, Wang and Zhang.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Wang, Chen, Ma, Gao, Zhu, Wang and Zhang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Existing studies indicate that dysregulation or abnormal expression of small nucleolar RNA (snoRNA) is closely associated with various diseases, including lung cancer. Furthermore, these diseases often involve multiple targets, making the redevelopment of traditional medicines highly promising. Accurate prediction of potential snoRNA therapeutic targets is essential for early disease intervention and the redevelopment of traditional medicines. Additionally, researchers have developed artificial intelligence (AI)-based methods to screen and predict potential snoRNA therapeutic targets, thereby advancing traditional drug redevelopment. However, existing methods face challenges such as imbalanced datasets and the dominance of high-degree nodes in graph neural networks (GNNs), which compromise the accuracy of node representations. To address these challenges, we propose an AI model based on variational graph autoencoders (VGAEs) that integrates decoupling and Kolmogorov-Arnold Network (KAN) technologies. The model reconstructs snoRNA-disease graphs by learning snoRNA and disease representations, accurately identifying potential snoRNA therapeutic targets. By decoupling similarity from node degree, the model mitigates the dominance of high-degree nodes, enhances prediction accuracy in scenarios like lung cancer, and leverages KAN technology to improve adaptability and flexibility to new data. Case studies revealed that snoRNA SNORA21 and SNORD33 are abnormally expressed in lung cancer patients and are strong candidates for potential therapeutic targets. These findings validate the proposed model&#x2019;s effectiveness in identifying therapeutic targets for diseases like lung cancer, supporting early screening and treatment, and advancing the redevelopment of traditional medicines. Data and experimental findings are archived in: <ext-link ext-link-type="uri" xlink:href="https://github.com/shmildsj/data">https://github.com/shmildsj/data</ext-link>.</p>
</abstract>
<kwd-group>
<kwd>redevelopment of traditional medicines</kwd>
<kwd>snoRNA therapeutic targets</kwd>
<kwd>lung cancer</kwd>
<kwd>variational graph autoencoder (VGAE)</kwd>
<kwd>artificial intelligence (AI)</kwd>
</kwd-group>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Ethnopharmacology</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>SnoRNA is a type of non-coding RNA, typically 60 to 300 nucleotides long, and predominantly found in eukaryotic cells (<xref ref-type="bibr" rid="B5">Esteller, 2011</xref>). Based on function, snoRNAs are categorized into two types: box C/D and box H/ACA (<xref ref-type="bibr" rid="B17">Liao et al., 2010</xref>). Numerous studies have demonstrated that snoRNA plays a crucial regulatory role. For instance, snoRNA regulates the methylation of rRNA and tRNA (<xref ref-type="bibr" rid="B10">Kiss-L&#xe1;szl&#xf3; et al., 1996</xref>), and is also involved in mRNA splicing (<xref ref-type="bibr" rid="B9">Kishore and Stamm, 2006</xref>). SnoRNA can bind to specific proteins to prevent enzyme cleavage and facilitate related RNA processing (<xref ref-type="bibr" rid="B5">Esteller, 2011</xref>). Furthermore, growing evidence suggests that snoRNA is a key factor in the occurrence and progression of various diseases. For example, dysregulation of snoRNA may promote various diseases, including gastric cancer (<xref ref-type="bibr" rid="B37">Wang et al., 2019</xref>). SnoRNA downregulation may lead to brain cavernous malformations (<xref ref-type="bibr" rid="B28">Qi et al., 2016</xref>), and promote lung cancer (<xref ref-type="bibr" rid="B17">Liao et al., 2010</xref>) or tumor development (<xref ref-type="bibr" rid="B12">Krishnan et al., 2016</xref>). Similarly, a significant increase in snoRNA levels may promote the progression of various cancers (<xref ref-type="bibr" rid="B51">Zheng et al., 2015</xref>). Thus, studying the regulatory mechanisms of snoRNA is crucial for advancing disease treatment strategies. Clinical trials can help accurately uncover the regulatory mechanisms of specific snoRNAs in diseases. However, these methods often depend on long-term experiments and observations, as well as costly equipment.</p>
<p>Fortunately, previous research has accumulated substantial data, enabling data scientists to uncover new regulatory mechanisms of snoRNA in diseases. For instance, X et al. developed the MNDR database, which is based on the regulatory mechanisms of mammalian ncRNA in diseases (<xref ref-type="bibr" rid="B26">Ning et al., 2021</xref>). Chen and Zhang et al. integrated and updated the RNA-disease association databases RNADisease v4.0 and ncRPheno, making them publicly accessible (<xref ref-type="bibr" rid="B2">Chen et al., 2023</xref>; <xref ref-type="bibr" rid="B46">Zhang W. et al., 2020</xref>). These databases contain SDA data and have the potential to advance the development of related computational methods. For example, Sun et al. collected high-quality SDA data from the MNDR database, integrated snoRNA and disease similarity networks, and predicted unknown SDAs using matrix completion (<xref ref-type="bibr" rid="B31">Sun et al., 2022</xref>). Hu et al. gathered known SDAs from RNADisease v4.0 and ncRPheno, constructed snoRNA-disease networks, and applied subgraph extraction and graph collaborative filtering to identify unknown SDAs (<xref ref-type="bibr" rid="B7">Hu et al., 2024</xref>). Overall, relatively few computational methods directly explore SDAs. And these models depend on complex feature extraction and classification processes, which significantly limit their broad applicability. However, numerous studies focus on potential interactions in biological networks, which are fundamentally similar to SDA prediction tasks. Advanced methods primarily include deep learning, GNNs, and graph autoencoder techniques.</p>
<p>The rapid development of deep learning is evident in its widespread application in bioinformatics (<xref ref-type="bibr" rid="B38">Wang Y. et al., 2023</xref>; <xref ref-type="bibr" rid="B32">Wang R. et al., 2024</xref>; <xref ref-type="bibr" rid="B33">Wang R. et al., 2023</xref>; <xref ref-type="bibr" rid="B59">Zhuo et al., 2023</xref>), particularly in exploring potential interactions within biological networks, achieving notable success. For instance, Zhou et al. automatically extracted sequence features of lncRNA and proteins using the Transformer architecture, integrated them to obtain representations of lncRNA-protein pairs, and predicted unknown lncRNA-protein interactions via multi-layer perceptrons (<xref ref-type="bibr" rid="B54">Zhou et al., 2023</xref>). Wei et al. incorporated data augmentation, feature alignment, and other techniques into a self-supervised learning framework, extracting precise node representations to efficiently predict unknown food-drug interactions (<xref ref-type="bibr" rid="B39">Wei et al., 2024a</xref>). Liu et al. applied meta-path technology to extract features of circular RNA and diseases, integrated them to obtain representations of circular RNA-disease pairs, and predicted unknown circular RNA-disease associations using contrastive learning and MLP (<xref ref-type="bibr" rid="B19">Liu et al., 2023a</xref>). Then, Liu et al. proposed a novel method for predicting gene regulatory networks, refining the topological network from both global and local perspectives, and accounting for edge importance to enhance prediction performance (<xref ref-type="bibr" rid="B21">Liu et al., 2023b</xref>). Additionally, Wei et al. employed an integrated deep learning framework, incorporating clustering-based parameter fine-tuning, to significantly enhance the accuracy of drug-target interaction prediction (<xref ref-type="bibr" rid="B41">Wei et al., 2024b</xref>). Moreover, Wei demonstrated that large language models have a significant impact on drug repositioning (<xref ref-type="bibr" rid="B42">Wei et al., 2024c</xref>). Deep learning technology effectively extracts deep node representations, operates independently of manually designed features, and swiftly and accurately infers potential interactions within biological networks. However, a major drawback of these methods is their neglect of the topological information in known biological networks.</p>
<p>Graph neural networks (GNNs) uncover the structure of topological networks via message propagation mechanisms. GNNs have emerged in bioinformatics fields like property prediction (<xref ref-type="bibr" rid="B35">Wang et al., 2024b</xref>; <xref ref-type="bibr" rid="B23">Ma et al., 2024a</xref>; <xref ref-type="bibr" rid="B34">Wang et al., 2024c</xref>) and gene detection (<xref ref-type="bibr" rid="B24">Ma et al., 2024b</xref>; <xref ref-type="bibr" rid="B36">Wang et al., 2024d</xref>), challenging traditional machine learning and deep learning approaches. Similarly, GNN technologies play a crucial role in predicting interactions within biological networks. For example, Zhuo et al. used two graph convolutional networks to extract node representations of lncRNA and proteins, followed by pairwise learning to train predictors for identifying unknown lncRNA-protein interactions (<xref ref-type="bibr" rid="B58">Zhuo et al., 2022</xref>). Wei et al. subsequently applied sampling to enhance the GCN model&#x2019;s performance in handling sparse data (<xref ref-type="bibr" rid="B43">Wei et al., 2023a</xref>). Liao et al. employed an autoencoder to extract representations of miRNA and diseases and used GCN to predict unknown miRNA-disease associations (<xref ref-type="bibr" rid="B18">Liao et al., 2023</xref>). Wei et al. applied a graph collaborative filtering model and multi-perspective contrastive learning to enhance node representations, aiming to accurately predict unknown miRNA-drug sensitivity (<xref ref-type="bibr" rid="B44">Wei et al., 2023b</xref>). Furthermore, Zhou et al. noted that message propagation across the entire graph may lead to &#x201c;over-smoothing&#x201d; and adopted a subgraph enhancement strategy to improve microbial-drug interaction predictions by focusing on local features (<xref ref-type="bibr" rid="B57">Zhou et al., 2024a</xref>). Building on this, Zhou et al. adopted an energy-constrained diffusion mechanism to extract global node representations of drugs and proteins, uncovering potential drug-protein relationships and accurately predicting unknown interactions (<xref ref-type="bibr" rid="B55">Zhou et al., 2024b</xref>). Additionally, Li et al. integrated multi-source similarity networks of miRNA and diseases to enhance model performance in predicting miRNA-disease interactions (<xref ref-type="bibr" rid="B16">Li Z. et al., 2024</xref>). GNN technology explores the topological information of biological networks via message propagation mechanisms, significantly enhancing interaction analysis within these networks. However, methods related to GNN often overlook the intrinsic information of the nodes. Additionally, the dominance of high-degree nodes in message propagation within GNNs can limit model performance.</p>
<p>The integration of self-supervised learning strategies with GNN technology has driven the development of graph autoencoders, widely applied to uncover potential information in biological networks. For instance, Zhou et al. proposed a method utilizing the graph autoencoder framework and edge masking to predict potential interactions between small molecules and miRNAs (<xref ref-type="bibr" rid="B56">Zhou et al., 2024c</xref>). This method first masks edges in the small molecule-miRNA graph using a Bernoulli distribution, then applies a GNN encoder and inner product decoder to reconstruct the masked graph. Zhou et al. subsequently masked paths in the miRNA-drug graph and trained the miRNA-drug association predictor using graph autoencoder technology (<xref ref-type="bibr" rid="B53">Zhou et al., 2024d</xref>). Zhang et al. subsequently integrated multi-source similarity networks and employed graph autoencoders along with polyloss technology to infer unknown lncRNA-protein interactions (<xref ref-type="bibr" rid="B47">Zhang et al., 2024a</xref>). Graph autoencoders are becoming increasingly crucial in biological network research due to their self-supervised nature, simplicity, and efficiency. Like GNN models, GAE-related models also encounter issues with message propagation dominated by high-degree nodes.</p>
<p>In summary, these advanced models have demonstrated success in interaction prediction within biological networks and should be capable of handling snoRNA-disease prediction tasks. However, these models face several significant challenges. First, these models, particularly graph-based models, have an inherent limitation due to their message propagation mechanism. Studies have shown that nodes with high degrees tend to have large embedding norms, which dominate message propagation and hinder GNN models from extracting accurate snoRNA or disease representations. Second, extremely sparse or isolated snoRNA or disease nodes often exist in the snoRNA-disease graph, leading to unpredictable results during gradient backpropagation. To address these issues, we incorporate L2 regularization and decoupling techniques into the variational graph autoencoder framework, proposing a snoRNA-disease association prediction model, named DK-SDA, to mitigate these challenges and enhance prediction performance. Our contributions can be summarized as follows:<list list-type="simple">
<list-item>
<p>(1) We proposed an SDA prediction model within the VGAE framework, yielding reliable results.</p>
</list-item>
<list-item>
<p>(2) We employed graph-regularized convolutional networks to extract snoRNA and disease representations, mitigating the issue of sparse nodes causing unpredictable gradient propagation.</p>
</list-item>
<list-item>
<p>(3) We applied decoupling technology to separate snoRNA-disease similarity from the node degrees of snoRNA or disease, thus mitigating the dominant effect of high-degree nodes on message propagation.</p>
</list-item>
<list-item>
<p>(4) We conducted multiple experiments on public datasets, and the results confirmed the model&#x2019;s high efficiency in lung cancer research, providing strong support for early screening and treatment.</p>
</list-item>
</list>
</p>
</sec>
<sec sec-type="materials|methods" id="s2">
<title>Materials and methods</title>
<sec id="s2-1">
<title>Data preparation</title>
<p>This study utilized two datasets collected by previous research (<xref ref-type="bibr" rid="B7">Hu et al., 2024</xref>) to evaluate the performance of the proposed and comparison models. The first dataset was sourced from the RNADisease v4.0 (<xref ref-type="bibr" rid="B2">Chen et al., 2023</xref>) database, where experimentally verified and predicted SDAs were extracted, with duplicates and missing data removed. The final dataset included 471 snoRNAs, 84 diseases, and 1,095 SDAs. The dataset was then divided into training and test sets in a 4:1 ratio. The second dataset was sourced from the ncRPheo database (<xref ref-type="bibr" rid="B46">Zhang W. et al., 2020</xref>). After data cleaning, it contained 82 diseases, 13 snoRNAs, and 439 SDAs. This dataset was used as an external test set to evaluate model performance under isolation conditions.</p>
</sec>
<sec id="s2-2">
<title>Method</title>
<p>This study proposes a snoRNA-disease association prediction model within the VGAE framework, aiming to efficiently predict unknown snoRNA-disease pairs from observed SDAs. Compared to other advanced interaction prediction models in biological networks, this model presents two key differences. First, the model addresses sparse nodes in the snoRNA-disease network, which can undermine the reliability of the extracted embeddings. Then, L2 regularization is applied to process node representations in both the initial and intermediate layers to alleviate the difficulty in predicting low-degree node embeddings during gradient backpropagation. Second, the model addresses the dominant effect of high-degree nodes on message propagation. Therefore, decoupling technology is introduced to separate snoRNA-disease similarity from node norms, thereby enhancing the reliability of message propagation. The following sections will elaborate on the relevant principles and techniques.</p>
<sec id="s2-2-1">
<title>Preliminary</title>
<p>In this study, the snoRNA-disease network is represented as <bold>
<italic>G &#x3d; &#x3c;V,E,X&#x3e;</italic>
</bold>, where <bold>
<italic>V &#x3d; Vs</italic>&#x222A;<italic>Vd</italic>
</bold> denotes the set of snoRNA and disease nodes, <bold>
<italic>E&#x2208;Vs &#xd7; Vd</italic>
</bold> denotes the observed SDAs, and <bold>
<italic>X &#x3d; Xs</italic>&#x222a;<italic>Xd</italic>
</bold> denotes the initial feature set of snoRNA and disease nodes. A graph convolutional network (GCN) is used as an encoder within the VGAE framework to perform operations such as message propagation and node embedding extraction. Recently, studies have focused on the L2 norm of node embeddings across various fields. For example, in translation, uncommon words are often assigned lower L2 norms (<xref ref-type="bibr" rid="B11">Kobayashi et al., 2020</xref>). In image processing or computer vision, low-quality images tend to have lower L2 norms for their corresponding embeddings (<xref ref-type="bibr" rid="B20">Liu et al., 2017</xref>). Subsequent studies have employed L2 regularization to reduce quantization error between large and small norms (<xref ref-type="bibr" rid="B29">Ranjan et al., 2017</xref>). Zhang et al. noted that L2 regularization can alleviate the instability between large and small norms during gradient backpropagation (<xref ref-type="bibr" rid="B45">Zhang D. et al., 2020</xref>). Additionally, Zheng et al. have applied L2 regularization in GCNs to mitigate the over-smoothing problem during message propagation (<xref ref-type="bibr" rid="B52">Zheng et al., 2018</xref>).</p>
</sec>
<sec id="s2-2-2">
<title>Model framework</title>
<p>
<xref ref-type="fig" rid="F1">Figure 1</xref> depicts the DK-SDA model&#x2019;s architecture, which encompasses (A) snoRNA-disease network construction, (B) sampling process (C) adjacency matrix reconstruction based on snoRNA-disease similarity, and (D) adjacency matrix reconstruction based on node degree. In Module A, data retrieved from the database constructs the snoRNA-disease network. Module C initiates with Bernoulli sampling as per Module B. If the SDAs are established based on snoRNA-disease similarity, the process concludes; otherwise, it transitions to Module D for adjacency matrix reconstruction based on node degree.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>DK-SDA model&#x2019;s architecture, comprising: <bold>(A)</bold> construction of the snoRNA-disease network, <bold>(B)</bold> the sampling process, <bold>(C)</bold> reconstruction of the adjacency matrix based on snoRNA-disease similarity, and <bold>(D)</bold> reconstruction of the adjacency matrix based on node degree.</p>
</caption>
<graphic xlink:href="fphar-15-1529128-g001.tif"/>
</fig>
</sec>
<sec id="s2-2-3">
<title>L2-normalization in GCN</title>
<p>To mitigate the issue of unpredictable gradient propagation caused by sparse nodes, we introduced L2 regularization in GCN based on previous work (<xref ref-type="bibr" rid="B1">Ahn and Kim, 2021</xref>). Specifically, L2 regularization is applied before GCN performs message propagation. For a snoRNA or disease node <bold>
<italic>a</italic>
</bold>, its initial feature is represented as <bold>
<italic>x</italic>
</bold>
<sub>
<bold>
<italic>a</italic>
</bold>
</sub>, which is mapped to a trainable weight matrix as:<disp-formula id="e1">
<mml:math id="m1">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>g</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mi>W</mml:mi>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>
</p>
<p>Next, L2 regularization is applied to <bold>
<italic>h</italic>
</bold>
<sub>
<bold>
<italic>a</italic>
</bold>
</sub> as defined in <xref ref-type="disp-formula" rid="e2">Equation 2</xref>:<disp-formula id="e2">
<mml:math id="m2">
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mfrac>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="&#x2016;" close="&#x2016;" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>where <bold>
<italic>t</italic>
</bold> represents the scaling parameter, set to 1.8 by default. Based on the regularized representation, GCN performs message propagation and outputs the embedding of node <bold>
<italic>a</italic>
</bold>:<disp-formula id="e3">
<mml:math id="m3">
<mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msqrt>
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mi>b</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mi>b</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>where <inline-formula id="inf1">
<mml:math id="m4">
<mml:mrow>
<mml:mi mathvariant="bold-italic">N</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the set of neighbor nodes of node <bold>
<italic>a</italic>
</bold>, and <inline-formula id="inf2">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">d</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the degree of node <bold>
<italic>a</italic>
</bold>. Notably, in the subsequent process, the similarity between snoRNA and disease and the norms of snoRNA or disease nodes are decoupled. The similarity and node norms are sampled from the embeddings extracted by GCN, and the GCN corresponding to the node norm does not execute <xref ref-type="disp-formula" rid="e3">Equation 3</xref>. Then the operation of this GCN can be formalized as <xref ref-type="disp-formula" rid="e4">Equation 4</xref>:<disp-formula id="e4">
<mml:math id="m6">
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>t</mml:mi>
<mml:msup>
<mml:mover accent="true">
<mml:mi>D</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:msup>
<mml:mover accent="true">
<mml:mi>A</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:msup>
<mml:mover accent="true">
<mml:mi>D</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:msup>
<mml:mi>g</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>where <inline-formula id="inf3">
<mml:math id="m7">
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="bold-italic">D</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> represents the degree matrix and <inline-formula id="inf4">
<mml:math id="m8">
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> represents the adjacency matrix with self-loop.</p>
</sec>
<sec id="s2-2-4">
<title>Variational graph autoencoder (VGAE) framework</title>
<p>Generally, the node embedding matrix <bold>
<italic>Z</italic>
</bold> output by GCN can be used with an inner product decoder to reconstruct the snoRNA-disease network <inline-formula id="inf5">
<mml:math id="m9">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="bold-italic">g</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:msup>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:mi mathvariant="bold-italic">T</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. However, studies suggest that variational graph autoencoders (VGAE) (<xref ref-type="bibr" rid="B8">Kipf and Welling, 2016</xref>) offer greater advantages. The implementation of VGAE is straightforward, consisting primarily of two components: a GNN encoder and a decoder. The GCN encoder serves as an inference model, sampling and deriving the variational posterior of node embeddings based on an approximate Gaussian distribution. The snoRNA-disease network is then reconstructed using the decoder. In the GCN encoder, the posterior of node embeddings is difficult to derive directly and is typically approximated using a Gaussian distribution (<xref ref-type="bibr" rid="B8">Kipf and Welling, 2016</xref>) as <xref ref-type="disp-formula" rid="e5">Equation 5</xref>:<disp-formula id="e5">
<mml:math id="m10">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msubsup>
<mml:mo>&#x220f;</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:msubsup>
<mml:mi>q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="|" close="" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>where<disp-formula id="e6">
<mml:math id="m11">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="|" close="" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi>&#x3bc;</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>d</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>g</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>a</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>and<disp-formula id="e7">
<mml:math id="m12">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3bc;</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:msub>
<mml:mi>&#x3bc;</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>X</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>
</p>
<p>In <xref ref-type="disp-formula" rid="e6">Equations 6</xref> and <xref ref-type="disp-formula" rid="e7">7</xref>, <inline-formula id="inf6">
<mml:math id="m13">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3bc;</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf7">
<mml:math id="m14">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3c3;</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> share the first layer parameters. And <inline-formula id="inf8">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3bc;</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> undergoes regularization, while <inline-formula id="inf9">
<mml:math id="m16">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3c3;</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> does not. Both <inline-formula id="inf10">
<mml:math id="m17">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3bc;</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf11">
<mml:math id="m18">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3c3;</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> perform the propagation process independently. <xref ref-type="disp-formula" rid="e7">Equation 7</xref> are used for feature extraction based on snoRNA-disease similarity. This study also incorporates an encoder based on node degree. However, unlike <xref ref-type="disp-formula" rid="e7">Equation 7</xref>, the mean and variance based on node degree are calculated without applying <xref ref-type="disp-formula" rid="e3">Equation 3</xref>.</p>
</sec>
<sec id="s2-2-5">
<title>Kolmogorov-Arnold Network (KAN)</title>
<p>The Multilayer Perceptron (MLP) has been instrumental in the field of deep learning due to its strong ability to approximate nonlinear functions. However, MLP faces unavoidable challenges, including large parameter sizes and limited flexibility. To address these challenges, KAN technology was developed (<xref ref-type="bibr" rid="B22">Liu et al., 2024</xref>). The core concept of the Kolmogorov-Arnold representation theorem is that any multivariate continuous function can be expressed as a composition of continuous univariate functions:<disp-formula id="e8">
<mml:math id="m19">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>v</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi>n</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:msub>
<mml:mi>&#x3c0;</mml:mi>
<mml:mi>v</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>u</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c9;</mml:mi>
<mml:mrow>
<mml:mi>v</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>u</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>u</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>
</p>
<p>In <xref ref-type="disp-formula" rid="e8">Equation 8</xref> theorem establishes the theoretical foundation that any multivariate continuous function can be represented as the sum of continuous univariate functions. KAN technology leverages this theory to efficiently approximate complex functions. In this framework, <inline-formula id="inf12">
<mml:math id="m20">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3c9;</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">v</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">u</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mtext>&#x2009;</mml:mtext>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> functions as a univariate transformation of each input variable <inline-formula id="inf13">
<mml:math id="m21">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mi mathvariant="bold-italic">u</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf14">
<mml:math id="m22">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3c9;</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">v</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">u</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:mn mathvariant="bold">0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2192;</mml:mo>
<mml:mi mathvariant="bold-italic">R</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf15">
<mml:math id="m23">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3c0;</mml:mi>
<mml:mi mathvariant="bold-italic">v</mml:mi>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="bold-italic">R</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi mathvariant="bold-italic">R</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. Given that all functions are univariate, each can be parameterized as b-spline curves with local basis function coefficients that can be optimized.</p>
<p>Building on the structure of the multi-layer perceptron (MLP), a deeper KAN architecture is constructed by stacking <bold>
<italic>L</italic>
</bold> layers:<disp-formula id="e9">
<mml:math id="m24">
<mml:mrow>
<mml:mi>K</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x398;</mml:mi>
<mml:mi>L</mml:mi>
</mml:msub>
<mml:mo>&#x2218;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x398;</mml:mi>
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2218;</mml:mo>
<mml:mo>&#x22ef;</mml:mo>
<mml:mo>&#x2218;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x398;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mi>h</mml:mi>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>where <inline-formula id="inf16">
<mml:math id="m25">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x398;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denotes the parameters of the <bold>
<italic>i</italic>
</bold>-th layer. Therefore, KAN can externally stack multiple layers, similar to MLP, to extract deep features as <xref ref-type="disp-formula" rid="e9">Equation 9</xref>. Internally, the univariate function within KAN offers greater flexibility to the model. This study employs the KAN network architecture in the second layer of the node degree-based encoder.</p>
<p>KANs offer a novel approach that enhances flexibility and interpretability compared to traditional MLPs by using spline functions as edge weights. This design allows KANs to better capture the underlying features driving snoRNA-disease interactions, thereby enhancing the model&#x2019;s interpretability. Researchers can analyze the spline function coefficients to understand how the network processes input data and makes predictions, making KANs a more transparent alternative to other neural network architectures.</p>
</sec>
<sec id="s2-2-6">
<title>Decoupling node similarity and degree</title>
<p>In typical VGAE-based methods, message propagation relies on both node similarity and norm. When these two factors are misaligned, the norm often dominates message propagation. Several studies have shown that in networks or graphs, the embedding norm of low-degree nodes is typically low, while high-degree nodes have higher norms (<xref ref-type="bibr" rid="B3">Cho, 2024</xref>). According to prior studies (<xref ref-type="bibr" rid="B3">Cho, 2024</xref>), there are two conditions under which snoRNA and disease can be considered associated. The first condition is a high degree of association between snoRNA and disease, indicated by high similarity. This is referred to as snoRNA-disease similarity. The second condition is when a snoRNA or disease node has a high degree, indicating its popularity, which increases the likelihood of forming associations with other nodes, though they may remain undiscovered. This is referred to as node degree. However, when establishing an association between snoRNA and disease, it is difficult to determine which condition applies due to the lack of precise prior knowledge. Therefore, for each snoRNA-disease pair, this study employs Bernoulli distribution sampling based on snoRNA-disease similarity to determine which condition applies for establishing SDAs.</p>
<p>Therefore, this study employed the previously established method (<xref ref-type="bibr" rid="B3">Cho, 2024</xref>), extracting node embeddings based on snoRNA-disease similarity and node degree, followed by constructing individual decoders. For snoRNA <bold>
<italic>a</italic>
</bold> or disease <bold>
<italic>b</italic>
</bold>, sampling is conducted based on similarity to derive the potential node embedding as <xref ref-type="disp-formula" rid="e10">Equation 10</xref>:<disp-formula id="e10">
<mml:math id="m26">
<mml:mrow>
<mml:msubsup>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>s</mml:mi>
</mml:msubsup>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi>u</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>z</mml:mi>
<mml:mi>b</mml:mi>
<mml:mi>s</mml:mi>
</mml:msubsup>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi>u</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>where <inline-formula id="inf17">
<mml:math id="m27">
<mml:mrow>
<mml:msup>
<mml:mi>u</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf18">
<mml:math id="m28">
<mml:mrow>
<mml:msup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> are calculated using <xref ref-type="disp-formula" rid="e7">Equation 7</xref>. For snoRNA <bold>
<italic>a</italic>
</bold> or disease <bold>
<italic>b</italic>
</bold>, sampling is conducted based on node degree to derive the potential node embedding as <xref ref-type="disp-formula" rid="e11">Equation 11</xref>:<disp-formula id="e11">
<mml:math id="m29">
<mml:mrow>
<mml:msubsup>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi>u</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>z</mml:mi>
<mml:mi>b</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi>u</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>where <inline-formula id="inf19">
<mml:math id="m30">
<mml:mrow>
<mml:msup>
<mml:mi>u</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf20">
<mml:math id="m31">
<mml:mrow>
<mml:msup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> are calculated using <xref ref-type="disp-formula" rid="e7">Equation 7</xref>. Notably, <xref ref-type="disp-formula" rid="e3">Equation 3</xref> is not applied at this stage.</p>
<p>For any snoRNA-disease pair <bold>
<italic>&#x3c;a,b&#x3e;</italic>
</bold>, a two-stage approach is employed to establish their association:<list list-type="simple">
<list-item>
<p>(1) SnoRNA-disease similarity: Establish SDAs based on the Bernoulli distribution.</p>
</list-item>
</list>
<disp-formula id="e12">
<mml:math id="m32">
<mml:mrow>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x223c;</mml:mo>
<mml:mi>B</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msubsup>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>s</mml:mi>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>z</mml:mi>
<mml:mi>b</mml:mi>
<mml:mi>s</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(12)</label>
</disp-formula>where <inline-formula id="inf21">
<mml:math id="m33">
<mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3c3;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the SIGMOD function, <inline-formula id="inf22">
<mml:math id="m34">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
<bold>()</bold> measures the similarity between snoRNA and disease, and the normalized inner product can be used.<list list-type="simple">
<list-item>
<p>(2) Node Degree: If no association is established after the initial sampling, continue sampling as follows.</p>
</list-item>
</list>
<disp-formula id="e13">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x223c;</mml:mo>
<mml:mi>B</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mi>d</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msubsup>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>z</mml:mi>
<mml:mi>b</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(13)</label>
</disp-formula>where <inline-formula id="inf23">
<mml:math id="m36">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mi mathvariant="bold-italic">d</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
<bold>()</bold> measures node popularity, and the addition operation can be used.</p>
<p>This process achieves the decoupling of snoRNA-disease similarity and node degree. Firstly, <inline-formula id="inf24">
<mml:math id="m37">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
<bold>()</bold> specifically explains the rationale for establishing a connection based on snoRNA-disease similarity. And <inline-formula id="inf25">
<mml:math id="m38">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mi mathvariant="bold-italic">d</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
<bold>()</bold> specifically explains the rationale for establishing a connection based on degree.</p>
</sec>
<sec id="s2-2-7">
<title>Optimization</title>
<p>The goal of this study is to reconstruct the snoRNA-disease network using the decoder under two conditions, represented by the adjacency matrix <bold>
<italic>A</italic>
</bold>, with the corresponding posterior probability formalized as <xref ref-type="disp-formula" rid="e14">Equation 14</xref>:<disp-formula id="e14">
<mml:math id="m39">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>Z</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mo>&#x220f;</mml:mo>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:msub>
<mml:mo>&#x220f;</mml:mo>
<mml:mi>b</mml:mi>
</mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>b</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(14)</label>
</disp-formula>
</p>
<p>By integrating the embedding loss values for snoRNA-disease similarity and node degree, the following can be derived as <xref ref-type="disp-formula" rid="e15">Equation 15</xref>:<disp-formula id="e15">
<mml:math id="m40">
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3c6;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">&#x3c8;</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi mathvariant="normal">A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">q</mml:mi>
<mml:mi mathvariant="normal">&#x3c6;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="normal">Z</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">s</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">d</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mrow>
<mml:mfenced open="|" close="" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">X</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mi mathvariant="normal">&#x3c8;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="" close="|" separators="&#x7c;">
<mml:mrow>
<mml:mi mathvariant="normal">A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msup>
<mml:mi mathvariant="normal">Z</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">s</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">d</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mtext>KL</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">q</mml:mi>
<mml:mi mathvariant="normal">&#x3c6;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="normal">Z</mml:mi>
<mml:mi mathvariant="normal">d</mml:mi>
</mml:msup>
<mml:mrow>
<mml:mfenced open="|" close="" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">I</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">A</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="normal">Z</mml:mi>
<mml:mi mathvariant="normal">d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mtext>KL</mml:mtext>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">q</mml:mi>
<mml:mi mathvariant="normal">&#x3c6;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="normal">Z</mml:mi>
<mml:mi mathvariant="normal">s</mml:mi>
</mml:msup>
<mml:mrow>
<mml:mfenced open="|" close="" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">X</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">A</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="normal">Z</mml:mi>
<mml:mi mathvariant="normal">s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(15)</label>
</disp-formula>were <inline-formula id="inf26">
<mml:math id="m41">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">s</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">d</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> represents the node embedding matrix derived for snoRNA-disease similarity and node degree. These are SDAs reconstructed under different conditions. <bold>
<italic>KL()</italic>
</bold> denotes KL divergence, a distance metric. Both <inline-formula id="inf27">
<mml:math id="m42">
<mml:mrow>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf28">
<mml:math id="m43">
<mml:mrow>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:mi mathvariant="bold-italic">d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> follow Gaussian priors <inline-formula id="inf29">
<mml:math id="m44">
<mml:mrow>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">z</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="bold-italic">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">z</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="|" close="" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mn mathvariant="bold">0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">I</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where <bold>
<italic>I</italic>
</bold> denotes identity matrix. And <inline-formula id="inf30">
<mml:math id="m45">
<mml:mrow>
<mml:mi mathvariant="bold-italic">q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf31">
<mml:math id="m46">
<mml:mrow>
<mml:mi mathvariant="bold-italic">q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:mi mathvariant="bold-italic">d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> follow the posterior estimates of the Gaussian distribution, with <inline-formula id="inf32">
<mml:math id="m47">
<mml:mrow>
<mml:mi mathvariant="bold-italic">q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">z</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> following. Notably, In the case of node degree, the focus is on the node itself rather than the feature vector <bold>
<italic>X</italic>
</bold>.</p>
<p>The first term of the above optimization objective can be further decomposed as follows:<disp-formula id="e16">
<mml:math id="m48">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>q</mml:mi>
<mml:mi>&#x3c6;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi>Z</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mrow>
<mml:mfenced open="|" close="" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:mi mathvariant="italic">log</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:msubsup>
<mml:mi>p</mml:mi>
<mml:mi>&#x3c8;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="" close="|" separators="&#x7c;">
<mml:mrow>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msup>
<mml:mi>Z</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msup>
<mml:mi>p</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>q</mml:mi>
<mml:mi>&#x3c6;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msup>
<mml:mi>Z</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mrow>
<mml:mfenced open="|" close="" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:mi mathvariant="italic">log</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:msubsup>
<mml:mi>p</mml:mi>
<mml:mi>&#x3c8;</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="" close="|" separators="&#x7c;">
<mml:mrow>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msup>
<mml:mi>Z</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msup>
<mml:mi>q</mml:mi>
<mml:mi>s</mml:mi>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(16)</label>
</disp-formula>where <inline-formula id="inf33">
<mml:math id="m49">
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">&#x3c8;</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="" close="|" separators="&#x7c;">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf34">
<mml:math id="m50">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">&#x3c8;</mml:mi>
<mml:mi mathvariant="bold-italic">d</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="" close="|" separators="&#x7c;">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">Z</mml:mi>
<mml:mi mathvariant="bold-italic">d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> follow <xref ref-type="disp-formula" rid="e12">Equations 12</xref>, <xref ref-type="disp-formula" rid="e13">13</xref>, respectively, representing similarity and node degree. And <inline-formula id="inf35">
<mml:math id="m51">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf36">
<mml:math id="m52">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">q</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> are Bernoulli parameters, representing the probabilities of establishing SDAs based on snoRNA-disease similarity and node degree, respectively. Finally, the corresponding parameters are learned using the EM algorithm.</p>
</sec>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec id="s3-1">
<title>Experimental setup</title>
<p>This study evaluated the performance of the proposed DK-SDA model and comparison models using the RNADisease (<xref ref-type="bibr" rid="B2">Chen et al., 2023</xref>) and ncRPheo (<xref ref-type="bibr" rid="B46">Zhang W. et al., 2020</xref>) datasets. Six GNN-based comparison models were selected: NIMCGCN (<xref ref-type="bibr" rid="B13">Li et al., 2020</xref>), AMHMDA (<xref ref-type="bibr" rid="B27">Ning et al., 2023</xref>), NSAMDA (<xref ref-type="bibr" rid="B50">Zhao et al., 2022</xref>), iPiDA-GCN (<xref ref-type="bibr" rid="B6">Hou et al., 2022</xref>), VGAMF (<xref ref-type="bibr" rid="B4">Ding et al., 2021</xref>), and IGCNSDA (<xref ref-type="bibr" rid="B7">Hu et al., 2024</xref>). The NIMCGCN model was originally developed for miRNA-disease association prediction. Two GCNs were used to extract features from miRNA and disease similarity networks, followed by matrix completion to identify potential MDAs. Similarly, the AMHMDA model used two GCNs to extract features from miRNA and disease similarity networks, predicting potential MDAs via hypergraphs and hierarchical attention mechanisms. The NSAMDA model employed nearest neighbor and graph attention network (GAT) techniques to extract features from the miRNA-disease heterogeneous graph, predicting unknown MDAs via an inner product decoder. The VGAMF model employs VGAE technology to extract features from miRNA and disease similarity networks, using matrix decomposition to predict potential MDAs. The iPiDA-GCN model uses GCN to identify unknown piRNA-disease associations. The IGCNSDA model employs graph collaborative filtering to predict potential SDAs. NIMCGCN, AMHMDA, NSAMDA, and VGAMF were originally designed for MDA prediction, while iPiDA-GCN was developed for piRNA-disease association prediction. In this study, the inputs for these methods were changed to the RNADisease and ncRPheo datasets. All models used the same training, test, and external test sets. As in previous studies (<xref ref-type="bibr" rid="B40">Wei et al., 2024d</xref>; <xref ref-type="bibr" rid="B49">Zhang et al., 2024b</xref>), we primarily used AUC, AUPR, Accuracy (ACC), Precision (PRE), Sensitivity (SEN), and F1-Score (F1) metrics to evaluate model performance.</p>
<p>Our strategy manages the imbalance between positive and negative samples during model training. Specifically, we use a technique to randomly select an equal number of negative samples to match the positive samples in each training batch. This method ensures the model trains on a balanced dataset, crucial for preventing bias towards predicting non-interacting classes. Additionally, to further mitigate class imbalance risk, we employ stratified sampling during the cross-validation process. This strategy maintains a balanced distribution of positive and negative classes across different folds, enhancing the model&#x2019;s generalizability and robustness.</p>
</sec>
<sec id="s3-2">
<title>Performance comparison</title>
<p>
<xref ref-type="table" rid="T1">Table 1</xref> presents the results of all models on the RNADisease database. It is evident that the AMHMDA, NIMCGCN, NSAMDA, and VGAMF models perform poorly, likely due to their reliance on feature extraction from similarity networks. Determining an appropriate threshold for similarity networks is challenging, often resulting in the extraction of biased features. Comparatively, the performance of the NSAMDA and VGAMF models shows slight improvement. This improvement may be attributed to the use of nearest neighbor and GAT technologies in the NSAMDA model. The VGAMF model employs VGAE technology, which is better suited for SDA prediction tasks. The iPiDA_GCN and IGCNSDA models directly perform SDA prediction on the snoRNA-disease graph, leading to further performance improvements. Notably, the proposed model achieved 95.84% AUC, 97.23% AUPR, and 92.01% ACC, significantly outperforming all competing models. This superior performance may result from the proposed model&#x2019;s use of the VGAE framework with decoupling technology, which mitigates the dominance of high-degree nodes in message propagation.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Results of all models on the RNADisease dataset (%).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Models/Metrics</th>
<th align="center">AUC</th>
<th align="center">AUPR</th>
<th align="center">ACC</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">AMHMDA</td>
<td align="center">65.72</td>
<td align="center">72.44</td>
<td align="center">65.06</td>
</tr>
<tr>
<td align="center">NIMCGCN</td>
<td align="center">68.03</td>
<td align="center">66.55</td>
<td align="center">61.41</td>
</tr>
<tr>
<td align="center">NSAMDA</td>
<td align="center">70.23</td>
<td align="center">75.88</td>
<td align="center">68.26</td>
</tr>
<tr>
<td align="center">VGAMF</td>
<td align="center">73.04</td>
<td align="center">76.65</td>
<td align="center">68.09</td>
</tr>
<tr>
<td align="center">iPiDA_GCN</td>
<td align="center">81.01</td>
<td align="center">81.73</td>
<td align="center">70.54</td>
</tr>
<tr>
<td align="center">IGCNSDA</td>
<td align="center">84.38</td>
<td align="center">87.44</td>
<td align="center">78.31</td>
</tr>
<tr>
<td align="center">Ours</td>
<td align="center">95.84</td>
<td align="center">97.23</td>
<td align="center">92.01</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Additionally, the proposed model incorporates KAN technology to enhance its flexibility and adaptability. Additionally, we tested the models&#x2019; adaptability to new data and evaluated their performance on external datasets. Specifically, all models were trained on the complete RNADisease dataset and subsequently tested on the external ncRPheo dataset. <xref ref-type="table" rid="T2">Table 2</xref> presents the independent testing results of all models on external datasets. The results show that the proposed model significantly outperforms all competing models, except in the AUPR metric. These results demonstrate that the proposed DK-SDA model exhibits strong adaptability to new data.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Results of all models on external testing set (%).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Models/Metrics</th>
<th align="center">AUC</th>
<th align="center">AUPR</th>
<th align="center">ACC</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">AMNMDA</td>
<td align="center">50.10</td>
<td align="center">49.45</td>
<td align="center">48.97</td>
</tr>
<tr>
<td align="center">NIMCGCN</td>
<td align="center">70.84</td>
<td align="center">77.81</td>
<td align="center">52.05</td>
</tr>
<tr>
<td align="center">NSAMDA</td>
<td align="center">50.72</td>
<td align="center">67.26</td>
<td align="center">69.11</td>
</tr>
<tr>
<td align="center">VGAMF</td>
<td align="center">70.69</td>
<td align="center">70.49</td>
<td align="center">64.57</td>
</tr>
<tr>
<td align="center">iPiDA_GCN</td>
<td align="center">61.82</td>
<td align="center">61.85</td>
<td align="center">66.51</td>
</tr>
<tr>
<td align="center">IGCNSDA</td>
<td align="center">71.42</td>
<td align="center">87.46</td>
<td align="center">72.90</td>
</tr>
<tr>
<td align="center">Ours</td>
<td align="center">80.33</td>
<td align="center">81.92</td>
<td align="center">75.17</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-3">
<title>Performance evaluation</title>
<p>A 5-fold cross-validation experiment was conducted to further evaluate the model&#x2019;s performance and minimize the impact of random factors. <xref ref-type="table" rid="T3">Table 3</xref> presents the 5-fold cross-validation results of the proposed model on the RNADisease database. The proposed DK-SDA model performs well, achieving an average of 95.84% AUC, 97.23% AUPR, 92.01% ACC, 93.65% PRE, 90.14% SEN, and 91.86% F1. Additionally, the results for each fold are relatively stable, exhibiting small fluctuations. This further demonstrates that the proposed DK-SDA model exhibits strong adaptability and reliability.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Results of 5-fold cross validation of DK-SDA model on RNADisease dataset (%).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Test set</th>
<th align="center">AUC</th>
<th align="center">AUPR</th>
<th align="center">ACC</th>
<th align="center">PRE</th>
<th align="center">SEN</th>
<th align="center">F1</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">1</td>
<td align="center">94.12</td>
<td align="center">96.28</td>
<td align="center">90.41</td>
<td align="center">91.94</td>
<td align="center">88.58</td>
<td align="center">90.23</td>
</tr>
<tr>
<td align="center">2</td>
<td align="center">95.19</td>
<td align="center">96.87</td>
<td align="center">92.01</td>
<td align="center">93.40</td>
<td align="center">90.41</td>
<td align="center">91.88</td>
</tr>
<tr>
<td align="center">3</td>
<td align="center">95.73</td>
<td align="center">97.18</td>
<td align="center">92.24</td>
<td align="center">93.84</td>
<td align="center">90.41</td>
<td align="center">92.09</td>
</tr>
<tr>
<td align="center">4</td>
<td align="center">96.94</td>
<td align="center">97.80</td>
<td align="center">92.69</td>
<td align="center">94.31</td>
<td align="center">90.87</td>
<td align="center">92.56</td>
</tr>
<tr>
<td align="center">5</td>
<td align="center">97.20</td>
<td align="center">98.03</td>
<td align="center">92.69</td>
<td align="center">94.74</td>
<td align="center">90.41</td>
<td align="center">92.52</td>
</tr>
<tr>
<td align="center">Average</td>
<td align="center">95.84 &#xb1; 1.27</td>
<td align="center">97.23 &#xb1; 0.71</td>
<td align="center">92.01 &#xb1; 0.94</td>
<td align="center">93.65 &#xb1; 1.08</td>
<td align="center">90.14 &#xb1; 0.89</td>
<td align="center">91.86 &#xb1; 0.95</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-4">
<title>Parameter experiments</title>
<p>Several sets of parameter experiments were conducted to evaluate the impact of parameter changes on the performance of the proposed model. First, the impact of varying hidden layer dimensions on model performance was tested. The hidden layer refers to the linear mapping layer in <xref ref-type="disp-formula" rid="e1">Equation 1</xref>, where the mapped features are used to generate regularized feature vectors. In the experiment, all other variables were held constant, with only the hidden layer dimension changed. The results are presented in <xref ref-type="fig" rid="F2">Figure 2</xref>. The results indicate that when the hidden layer dimension is within the range [64,256], model performance gradually improves. When the hidden layer dimension exceeds 256, model performance slightly decreases. This may occur because smaller dimensions cannot fully extract features, while larger dimensions may cause information redundancy, limiting model performance. Therefore, a compromise hidden layer dimension can be selected to ensure the model&#x2019;s adaptability to new data.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Model performance at different hidden layer dimensions.</p>
</caption>
<graphic xlink:href="fphar-15-1529128-g002.tif"/>
</fig>
<p>Secondly, we evaluated the impact of various encoders on model performance. The encoder is positioned in the second feature extraction layer of the node degree-based structure. First, the model applies a linear mapping layer to transform input features, followed by feeding them into the second feature extraction layer. We explore the model&#x2019;s performance with different encoders in the second feature extraction layer, including KAN, MLP, GCN, GAT, SAGE, and GIN. In the experiment, all other variables were held constant, with only the encoder being modified. The results are presented in <xref ref-type="fig" rid="F3">Figure 3</xref>. The results indicate that the model performs well when KAN is used as the encoder. Overall, the model&#x2019;s performance decreases when using GNN, suggesting that high-degree nodes may dominate the message propagation process.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Results of the model with different encoders.</p>
</caption>
<graphic xlink:href="fphar-15-1529128-g003.tif"/>
</fig>
</sec>
<sec id="s3-5">
<title>Statistical significance analysis</title>
<p>In this study, we employed the one-way analysis of variance (ANOVA) technique (<xref ref-type="bibr" rid="B30">St and Wold, 1989</xref>) to systematically compare the AUC performance of various snoRNA target prediction models on the snoRNA-disease association dataset, as shown in <xref ref-type="fig" rid="F4">Figure 4</xref>. The results indicated that the proposed model was significantly superior on the dataset, achieving a p-value as low as 1.0e-07 when compared to most competing models, demonstrating its high statistical significance. Overall, the p-values between the model and all reference models were well below the significance threshold of 0.05, confirming its superior performance. This analysis not only underscored the model&#x2019;s superior performance but also highlighted its stability and reliability across diverse datasets. Importantly, the findings provide robust evidence supporting the use of decoupled representation learning and KAN technology to enhance model generalizability.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Statistical significance analysis of multiple methods based on AUC.</p>
</caption>
<graphic xlink:href="fphar-15-1529128-g004.tif"/>
</fig>
</sec>
<sec id="s3-6">
<title>Case analysis</title>
<p>This study focuses on lung cancer, examining snoRNAs SNORD33 and SNORD78 through three case studies to evaluate the model&#x2019;s performance in realistic scenarios. Lung cancer, a prevalent and serious disease, has high morbidity and mortality rates. Factors such as environmental pollution, smoking, and genetic mutations contribute to the onset and progression of lung cancer. Dysregulation of snoRNA <italic>in vivo</italic> can lead to tumorigenesis, manifested by cancer cell proliferation and altered gene expression. For instance, abnormal expression of SNORD78 has been linked to the initiation and progression of lung cancer (<xref ref-type="bibr" rid="B51">Zheng et al., 2015</xref>). Research indicates that inhibiting SNORD78 expression can curb cancer cell proliferation, suggesting its potential as a target for lung cancer therapy. Investigating lung cancer-related snoRNAs is likely to advance the development of novel therapeutic strategies.</p>
<p>SNORD33, typically located in the nucleolus, primarily facilitates the chemical modification of rRNA, such as guiding rRNA methylation (<xref ref-type="bibr" rid="B25">Mannoor et al., 2012</xref>). Dysregulation of SNORD33, whether through overexpression or underexpression, may contribute to cancer or inflammation development. Research has demonstrated that SNORD33 dysregulation disrupts rRNA methylation, resulting in abnormal protein complex synthesis. This disruption aids the growth and survival of lung cancer cells. Furthermore, SNORD33 dysregulation may facilitate cell metastasis and hasten lung cancer progression (<xref ref-type="bibr" rid="B48">Zhang et al., 2023</xref>). Similarly, SNORD78, a snoRNA with functions akin to SNORD33, promotes lung cancer cell growth, thus accelerating lung cancer progression.</p>
<p>In the first case study, we excluded all snoRNAs associated with lung cancer from the training set, and trained the model on the remaining data. Subsequently, the trained model predicted the likelihood of association between lung cancer and all snoRNAs, selecting the top 20 for further analysis. The results are detailed in <xref ref-type="table" rid="T4">Table 4</xref>. 18 of the predicted snoRNAs were validated using the RNADisease v4.0 database (<xref ref-type="bibr" rid="B2">Chen et al., 2023</xref>). Although SNORA21 has not been validated in the database, multiple studies suggest a strong association with lung cancer (<xref ref-type="bibr" rid="B14">Li L. et al., 2024</xref>). In the second and third case studies, we employed a similar approach to assess the association of SNORD33 and SNORD78 with various diseases, selecting the top 10 diseases for each. Results are presented in <xref ref-type="table" rid="T5">Tables 5</xref>, <xref ref-type="table" rid="T6">6</xref>. For SNORD33, eight predicted diseases, excluding Prostate Neoplasm and Gliomas, were validated in the RNADisease v4.0 database (<xref ref-type="bibr" rid="B2">Chen et al., 2023</xref>). Similarly, for SNORD78, eight predicted diseases, excluding Osteosarcoma and Melanoma, were confirmed in the RNADisease v4.0 database. Additionally, in the second and third case studies, both SNORD33 and SNORD78 appear to be associated with lung cancer and related diseases. Furthermore, integrating findings from immune escape mechanism research can aid the development and optimization of lung cancer immunotherapies (<xref ref-type="bibr" rid="B15">Li et al., 2023</xref>). These results demonstrate the suitability of the proposed DK-SDA method for snoRNA therapeutic target prediction, facilitating the exploration of disease pathogenesis and the development of new treatment strategies.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Top 20 predicted snoRNAs with potential associations with lung cancer.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">snoRNA</th>
<th align="center">RNADisease</th>
<th align="center">snoRNA</th>
<th align="center">RNADisease</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">SNORD28</td>
<td align="center">Confirmed</td>
<td align="center">SNORD14D</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">SNORD88B</td>
<td align="center">Confirmed</td>
<td align="center">SNORD18B</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">SNORD11B</td>
<td align="center">Confirmed</td>
<td align="center">SNORD113-8</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">SNORD112</td>
<td align="center">Confirmed</td>
<td align="center">SNORD78</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">SNORD46</td>
<td align="center">Confirmed</td>
<td align="center">SNORD114-12</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">SNORD104</td>
<td align="center">Confirmed</td>
<td align="center">SNORD51</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">SNORD54</td>
<td align="center">Confirmed</td>
<td align="center">SNORD103A</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">SNORD78</td>
<td align="center">Confirmed</td>
<td align="center">SNORD1C</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">SNORD76</td>
<td align="center">Confirmed</td>
<td align="center">SNORA21</td>
<td align="center">Unconfirmed</td>
</tr>
<tr>
<td align="center">SNORD72</td>
<td align="center">Confirmed</td>
<td align="center">SNORD21</td>
<td align="center">Unconfirmed</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Top 10 predicted diseases with potential associations with SNORD33.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Disease</th>
<th align="center">RNADisease</th>
<th align="center">Disease</th>
<th align="center">RNADisease</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Lung Non-Small Cell Carcinoma</td>
<td align="center">Confirmed</td>
<td align="center">Lung Cancer</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">Lung Carcinoma</td>
<td align="center">Confirmed</td>
<td align="center">Colorectal Cancer</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">Gastric Cancer</td>
<td align="center">Confirmed</td>
<td align="center">Rheumatoid Arthritis</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">Traumatic Brain Injury</td>
<td align="center">Confirmed</td>
<td align="center">Prostate Neoplasm</td>
<td align="center">Unconfirmed</td>
</tr>
<tr>
<td align="center">Prostate Cancer</td>
<td align="center">Confirmed</td>
<td align="center">Glioma</td>
<td align="center">Unconfirmed</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Top 10 predicted diseases with potential associations with SNORD78.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Disease</th>
<th align="center">RNADisease</th>
<th align="center">Disease</th>
<th align="center">RNADisease</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Lung Non-Small Cell Carcinoma</td>
<td align="center">Confirmed</td>
<td align="center">Colorectal Cancer</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">Lung Cancer</td>
<td align="center">Confirmed</td>
<td align="center">Prostate Cancer</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">Clear Cell Renal Cell Carcinoma</td>
<td align="center">Confirmed</td>
<td align="center">Gastric Cancer</td>
<td align="center">Confirmed</td>
</tr>
<tr>
<td align="center">Hepatocellular Carcinoma</td>
<td align="center">Confirmed</td>
<td align="center">Osteosarcoma</td>
<td align="center">Unconfirmed</td>
</tr>
<tr>
<td align="center">Head And Neck Squamous Cell Carcinoma</td>
<td align="center">Confirmed</td>
<td align="center">Melanoma</td>
<td align="center">Unconfirmed</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec sec-type="conclusion" id="s4">
<title>Conclusion</title>
<p>SnoRNA dysregulation significantly contributes to disease onset and progression, making accurate identification of snoRNA therapeutic targets essential for redevelopment of traditional medicines. Timely identification of potential snoRNA therapeutic targets facilitates early disease screening and the development of novel treatment strategies. Existing GNN-based deep learning methods reveal network topology and accurately predict potential snoRNA therapeutic targets. However, these methods encounter significant challenges. Datasets are often imbalanced due to limitations in data collection. During message propagation in GNNs, high-degree nodes often dominate, hindering accurate node representation learning. To address these challenges, this study proposes an effective snoRNA therapeutic target prediction model within the VGAE framework, integrating decoupling and KAN techniques. The model learns snoRNA and disease representations within the VGAE framework, reconstructing the snoRNA-disease graph to identify potential therapeutic targets. The decoupling technique separates similarity from node degree, mitigating the dominance of high-degree nodes in information propagation and enhancing the accuracy of snoRNA therapeutic target prediction, particularly for diseases like lung cancer. Additionally, the model integrates KAN technology to enhance flexibility and adaptability to new data. Moreover, the model indicates that snoRNA SNORA21, SNORD78 and SNORD33 may play a critical role in lung cancer. Experimental results demonstrate the model&#x2019;s high efficiency in disease research, particularly in lung cancer studies, providing valuable references for early screening and the redevelopment of traditional medicines.</p>
<p>This study integrates the decoupling of node similarity and node degree with KAN technology within the VGAE architecture, enhancing the accuracy of snoRNA therapeutic target predictions. However, there are several potential limitations in this study. First, KAN technology has high computational complexity. Future work will focus on exploring parallel designs for KAN. Secondly, the proposed model does not account for the sequence, structure, function, and other biological data of snoRNA, nor does it include clinical information on the disease and specific targets, which limits the model&#x2019;s performance. To address this, we plan to integrate large language models in the future to extract and combine multimodal data, enhancing the discovery of potential snoRNA therapeutic targets. Additionally, the model faces challenges when encountering snoRNA targets that deviate from established topological patterns or are previously unobserved. To improve the model&#x2019;s adaptability to out-of-distribution data and capture interaction patterns differing from traditional topological structures, methods such as transfer learning and meta-learning can be employed.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s5">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.</p>
</sec>
<sec sec-type="author-contributions" id="s6">
<title>Author contributions</title>
<p>ZW: Funding acquisition, Investigation, Methodology, Validation, Writing&#x2013;original draft. YC: Formal Analysis, Investigation, Writing&#x2013;review and editing. HM: Data curation, Formal Analysis, Writing&#x2013;review and editing. HG: Formal Analysis, Investigation, Writing&#x2013;review and editing. YZ: Methodology, Supervision, Writing&#x2013;review and editing. HW: Supervision, Validation, Writing&#x2013;review and editing. NZ: Methodology, Supervision, Validation, Writing&#x2013;review and editing.</p>
</sec>
<sec sec-type="funding-information" id="s7">
<title>Funding</title>
<p>The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.</p>
</sec>
<sec sec-type="COI-statement" id="s8">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s9">
<title>Generative AI statement</title>
<p>The author(s) declare that no Generative AI was used in the creation of this manuscript.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ahn</surname>
<given-names>S. J.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>Variational graph normalized autoencoders</article-title>,&#x201d; in <source>Proceedings of the 30th ACM international conference on information and knowledge management</source>.</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ye</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>RNADisease v4. 0: an updated resource of RNA-associated diseases, providing RNA-disease analysis, enrichment and prediction</article-title>. <source>Nucleic Acids Res.</source> <volume>51</volume> (<issue>D1</issue>), <fpage>D1397</fpage>&#x2013;<lpage>D1404</lpage>. <pub-id pub-id-type="doi">10.1093/nar/gkac814</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Cho</surname>
<given-names>Y.-S.</given-names>
</name>
</person-group> (<year>2024</year>). &#x201c;<article-title>Decoupled variational graph autoencoder for link prediction</article-title>,&#x201d; in <source>Proceedings of the ACM on web conference 2024</source>.</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lei</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Liao</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>F.-X.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Predicting miRNA-disease associations based on multi-view variational graph auto-encoder with matrix factorization</article-title>. <source>IEEE J. Biomed. health Inf.</source> <volume>26</volume> (<issue>1</issue>), <fpage>446</fpage>&#x2013;<lpage>457</lpage>. <pub-id pub-id-type="doi">10.1109/JBHI.2021.3088342</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Esteller</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Non-coding RNAs in human disease</article-title>. <source>Nat. Rev. Genet.</source> <volume>12</volume> (<issue>12</issue>), <fpage>861</fpage>&#x2013;<lpage>874</lpage>. <pub-id pub-id-type="doi">10.1038/nrg3074</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hou</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>iPiDA-GCN: identification of piRNA-disease associations based on Graph Convolutional Network</article-title>. <source>PLOS Comput. Biol.</source> <volume>18</volume> (<issue>10</issue>), <fpage>e1010671</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1010671</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <article-title>IGCNSDA: unraveling disease-associated snoRNAs with an interpretable graph convolutional network</article-title>. <source>Briefings Bioinforma.</source> <volume>25</volume> (<issue>3</issue>), <fpage>bbae179</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbae179</pub-id>
</citation>
</ref>
<ref id="B8">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kipf</surname>
<given-names>T. N.</given-names>
</name>
<name>
<surname>Welling</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2016</year>). <source>Variational graph auto-encoders</source>. <comment>arXiv preprint arXiv:1611.07308</comment>.</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kishore</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Stamm</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>The snoRNA HBII-52 regulates alternative splicing of the serotonin receptor 2C</article-title>. <source>science</source> <volume>311</volume> (<issue>5758</issue>), <fpage>230</fpage>&#x2013;<lpage>232</lpage>. <pub-id pub-id-type="doi">10.1126/science.1118265</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kiss-L&#xe1;szl&#xf3;</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Henry</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bachellerie</surname>
<given-names>J.-P.</given-names>
</name>
<name>
<surname>Caizergues-Ferrer</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kiss</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>1996</year>). <article-title>Site-specific ribose methylation of preribosomal RNA: a novel function for small nucleolar RNAs</article-title>. <source>Cell</source> <volume>85</volume> (<issue>7</issue>), <fpage>1077</fpage>&#x2013;<lpage>1088</lpage>. <pub-id pub-id-type="doi">10.1016/s0092-8674(00)81308-2</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kobayashi</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Kuribayashi</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Yokoi</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Inui</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Attention is not only a weight: analyzing transformers with vector norms</article-title>,&#x201d; in <source>2020 conference on empirical methods in natural language processing, EMNLP 2020</source>.</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Krishnan</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Ghosh</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Heyns</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Graham</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Mackey</surname>
<given-names>J. R.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>Profiling of small nucleolar RNAs by next generation sequencing: potential new players for breast cancer prognosis</article-title>. <source>PloS one</source> <volume>11</volume> (<issue>9</issue>), <fpage>e0162622</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0162622</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Ning</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Neural inductive matrix completion with graph convolutional networks for miRNA-disease association prediction</article-title>. <source>Bioinformatics</source> <volume>36</volume> (<issue>8</issue>), <fpage>2538</fpage>&#x2013;<lpage>2546</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btz965</pub-id>
</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2024b</year>). <article-title>The diagnostic value of serum exosomal SNORD116 and SNORA21 for NSCLC patients</article-title>. <source>Clin. Transl. Oncol.</source>, <fpage>1</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1007/s12094-024-03606-1</pub-id>
</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Meng</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>Integration of single sample and population analysis for understanding immune evasion mechanisms of lung cancer</article-title>. <source>NPJ Syst. Biol. Appl.</source> <volume>9</volume> (<issue>1</issue>), <fpage>4</fpage>. <pub-id pub-id-type="doi">10.1038/s41540-023-00267-8</pub-id>
</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liao</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2024a</year>). <article-title>Multi-source data integration for explainable miRNA-driven drug discovery</article-title>. <source>Future Gener. Comput. Syst.</source> <volume>160</volume>, <fpage>109</fpage>&#x2013;<lpage>119</lpage>. <pub-id pub-id-type="doi">10.1016/j.future.2024.05.055</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Mei</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Guarnera</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2010</year>). <article-title>Small nucleolar RNA signatures as biomarkers for non-small-cell lung cancer</article-title>. <source>Mol. cancer</source> <volume>9</volume>, <fpage>198</fpage>&#x2013;<lpage>210</lpage>. <pub-id pub-id-type="doi">10.1186/1476-4598-9-198</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liao</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Ye</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Prediction of miRNA-disease associations in microbes based on graph convolutional networks and autoencoders</article-title>. <source>Front. Microbiol.</source> <volume>14</volume>, <fpage>1170559</fpage>. <pub-id pub-id-type="doi">10.3389/fmicb.2023.1170559</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2023a</year>). <article-title>MPCLCDA: predicting circRNA&#x2013;disease associations by using automatically selected meta-path and contrastive learning</article-title>. <source>Briefings Bioinforma.</source> <volume>24</volume> (<issue>4</issue>), <fpage>bbad227</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbad227</pub-id>
</citation>
</ref>
<ref id="B20">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Wen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Raj</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Sphereface: deep hypersphere embedding for face recognition</article-title>,&#x201d; in <source>Proceedings of the IEEE conference on computer vision and pattern recognition</source>.</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2023b</year>). <article-title>NSRGRN: a network structure refinement method for gene regulatory network inference</article-title>. <source>Briefings Bioinforma.</source> <volume>24</volume> (<issue>3</issue>), <fpage>bbad129</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbad129</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Vaidya</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Ruehle</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Halverson</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Solja&#x10d;i&#x107;</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <source>Kan: Kolmogorov-arnold networks</source>. <comment>arXiv preprint arXiv:2404.19756</comment>.</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ma</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2024a</year>). <article-title>GraphADT: empowering interpretable predictions of acute dermal toxicity with multi-view graph pooling and structure remapping</article-title>. <source>Bioinformatics</source> <volume>40</volume> (<issue>7</issue>), <fpage>btae438</fpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btae438</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ma</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2024b</year>). <article-title>Advancing cancer driver gene detection via Schur complement graph augmentation and independent subspace feature extraction</article-title>. <source>Comput. Biol. Med.</source> <volume>174</volume>, <fpage>108484</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.108484</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mannoor</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Liao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Small nucleolar RNAs in cancer</article-title>. <source>Biochimica Biophysica Acta (BBA)-Reviews Cancer</source> <volume>1826</volume> (<issue>1</issue>), <fpage>121</fpage>&#x2013;<lpage>128</lpage>. <pub-id pub-id-type="doi">10.1016/j.bbcan.2012.03.005</pub-id>
</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ning</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>B.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>MNDR v3. 0: mammal ncRNA&#x2013;disease repository with increased coverage and annotation</article-title>. <source>Nucleic Acids Res.</source> <volume>49</volume> (<issue>D1</issue>), <fpage>D160</fpage>&#x2013;<lpage>D164</lpage>. <pub-id pub-id-type="doi">10.1093/nar/gkaa707</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ning</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>AMHMDA: attention aware multi-view similarity networks and hypergraph learning for miRNA&#x2013;disease associations identification</article-title>. <source>Briefings Bioinforma.</source> <volume>24</volume> (<issue>2</issue>), <fpage>bbad094</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbad094</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qi</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Purtell</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>N. J.</given-names>
</name>
<name>
<surname>Aepler</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>Snord116 is critical in the regulation of food intake and body weight</article-title>. <source>Sci. Rep.</source> <volume>6</volume> (<issue>1</issue>), <fpage>18614</fpage>. <pub-id pub-id-type="doi">10.1038/srep18614</pub-id>
</citation>
</ref>
<ref id="B29">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ranjan</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Castillo</surname>
<given-names>C. D.</given-names>
</name>
<name>
<surname>Chellappa</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2017</year>). <source>L2-constrained softmax loss for discriminative face verification</source>.</citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>St</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wold</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>1989</year>). <article-title>Analysis of variance (ANOVA)</article-title>. <source>Chemom. intelligent laboratory Syst.</source> <volume>6</volume> (<issue>4</issue>), <fpage>259</fpage>&#x2013;<lpage>272</lpage>. <pub-id pub-id-type="doi">10.1016/0169-7439(89)80095-4</pub-id>
</citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sun</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Lv</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>PSnoD: identifying potential snoRNA-disease associations based on bounded nuclear norm regularization</article-title>. <source>Briefings Bioinforma.</source> <volume>23</volume> (<issue>4</issue>), <fpage>bbac240</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbac240</pub-id>
</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
<etal/>
</person-group> (<year>2024a</year>). <article-title>Diff-AMP: tailored designed antimicrobial peptide framework with all-in-one generation, identification, prediction and optimization</article-title>. <source>Briefings Bioinforma.</source> <volume>25</volume> (<issue>2</issue>), <fpage>bbae078</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbae078</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2023b</year>). <article-title>An effective plant small secretory peptide recognition model based on feature correction strategy</article-title>. <source>J. Chem. Inf. Model.</source> <volume>64</volume> (<issue>7</issue>), <fpage>2798</fpage>&#x2013;<lpage>2806</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.3c00868</pub-id>
</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2024c</year>). <article-title>MultiCBlo: enhancing predictions of compound-induced inhibition of cardiac ion channels with advanced multimodal learning</article-title>. <source>Int. J. Biol. Macromol.</source> <volume>276</volume>, <fpage>133825</fpage>. <pub-id pub-id-type="doi">10.1016/j.ijbiomac.2024.133825</pub-id>
</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2024b</year>). <article-title>MS-BACL: enhancing metabolic stability prediction through bond graph augmentation and contrastive learning</article-title>. <source>Briefings Bioinforma.</source> <volume>25</volume> (<issue>3</issue>), <fpage>bbae127</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbae127</pub-id>
</citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2024d</year>). <article-title>ECD-CDGI: an efficient energy-constrained diffusion model for cancer driver gene identification</article-title>. <source>PLOS Comput. Biol.</source> <volume>20</volume> (<issue>8</issue>), <fpage>e1012400</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1012400</pub-id>
</citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Kuang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Identification of eight small nucleolar RNAs as survival biomarkers and their clinical significance in gastric cancer</article-title>. <source>Front. Oncol.</source> <volume>9</volume>, <fpage>788</fpage>. <pub-id pub-id-type="doi">10.3389/fonc.2019.00788</pub-id>
</citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhai</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2023a</year>). <source>SBSM-pro: support bio-sequence machine for proteins</source>.</citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>2024a</year>). <article-title>Enhancing drug&#x2013;food interaction prediction with precision representations through multilevel self-supervised learning</article-title>. <source>Comput. Biol. Med.</source> <volume>171</volume>, <fpage>108104</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.108104</pub-id>
</citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2024d</year>). <article-title>BloodPatrol: revolutionizing blood cancer diagnosis - advanced real-time detection leveraging deep learning and cloud technologies</article-title>. <source>IEEE J. Biomed. health Inf.</source>, <fpage>1</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1109/JBHI.2024.3496294</pub-id>
</citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2024b</year>). <article-title>Efficient deep model ensemble framework for drug-target interaction prediction</article-title>. <source>J. Phys. Chem. Lett.</source> <volume>15</volume> (<issue>30</issue>), <fpage>7681</fpage>&#x2013;<lpage>7693</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpclett.4c01509</pub-id>
</citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
<etal/>
</person-group> (<year>2024c</year>). <article-title>DrugReAlign: a multisource prompt framework for drug repurposing based on large language models</article-title>. <source>BMC Biol.</source> <volume>22</volume> (<issue>1</issue>), <fpage>226</fpage>. <pub-id pub-id-type="doi">10.1186/s12915-024-02028-3</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Lian</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2023a</year>). <article-title>Headtailtransfer: an efficient sampling method to improve the performance of graph neural network method in predicting sparse ncrna&#x2013;protein interactions</article-title>. <source>Comput. Biol. Med.</source> <volume>157</volume>, <fpage>106783</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2023.106783</pub-id>
</citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Lian</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2023b</year>). <article-title>GCFMCL: predicting miRNA-drug sensitivity using graph collaborative filtering and multi-view contrastive learning</article-title>. <source>Briefings Bioinforma.</source> <volume>24</volume> (<issue>4</issue>), <fpage>bbad247</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbad247</pub-id>
</citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2020b</year>). <article-title>Deep metric learning with spherical embedding</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>33</volume>, <fpage>18772</fpage>&#x2013;<lpage>18783</lpage>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="https://proceedings.neurips.cc/paper/2020/hash/d9812f756d0df06c7381945d2e2c7d4b-Abstract.html">https://proceedings.neurips.cc/paper/2020/hash/d9812f756d0df06c7381945d2e2c7d4b-Abstract.html</ext-link>
</comment>
</citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<etal/>
</person-group> (<year>2020a</year>). <article-title>ncRPheno: a comprehensive database platform for identification and validation of disease related noncoding RNAs</article-title>. <source>RNA Biol.</source> <volume>17</volume> (<issue>7</issue>), <fpage>943</fpage>&#x2013;<lpage>955</lpage>. <pub-id pub-id-type="doi">10.1080/15476286.2020.1737441</pub-id>
</citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2024a</year>). <article-title>Fusion of multi-source relationships and topology to infer lncRNA-protein interactions</article-title>. <source>Mol. Therapy-Nucleic Acids</source> <volume>35</volume> (<issue>2</issue>), <fpage>102187</fpage>. <pub-id pub-id-type="doi">10.1016/j.omtn.2024.102187</pub-id>
</citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Xia</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Xiao</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>The emerging role of snoRNAs in human disease</article-title>. <source>Genes and Dis.</source> <volume>10</volume> (<issue>5</issue>), <fpage>2064</fpage>&#x2013;<lpage>2081</lpage>. <pub-id pub-id-type="doi">10.1016/j.gendis.2022.11.018</pub-id>
</citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2024b</year>). <article-title>CardiOT: towards interpretable drug cardiotoxicity prediction using optimal transport and Kolmogorov-arnold networks</article-title>. <source>IEEE J. Biomed. health Inf.</source>, <fpage>1</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1109/jbhi.2024.3510297</pub-id>
</citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>You</surname>
<given-names>Z.-H.</given-names>
</name>
<name>
<surname>Nie</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Zhong</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Predicting Mirna-disease associations based on neighbor selection graph attention networks</article-title>. <source>IEEE/ACM Trans. Comput. Biol. Bioinforma.</source> <volume>20</volume> (<issue>2</issue>), <fpage>1298</fpage>&#x2013;<lpage>1307</lpage>. <pub-id pub-id-type="doi">10.1109/TCBB.2022.3204726</pub-id>
</citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zheng</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ni</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>Small nucleolar RNA 78 promotes the tumorigenesis in non-small cell lung cancer</article-title>. <source>J. Exp. and Clin. Cancer Res.</source> <volume>34</volume>, <fpage>49</fpage>&#x2013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1186/s13046-015-0170-5</pub-id>
</citation>
</ref>
<ref id="B52">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zheng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>D. K.</given-names>
</name>
<name>
<surname>Savvides</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Ring loss: convex feature normalization for face recognition</article-title>,&#x201d; in <source>Proceedings of the IEEE conference on computer vision and pattern recognition</source>.</citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2024d</year>). <article-title>GAM-MDR: probing miRNA-drug resistance using a graph autoencoder based on random path masking</article-title>. <source>Briefings Funct. Genomics</source> <volume>23</volume>, <fpage>475</fpage>&#x2013;<lpage>483</lpage>. <comment>elae005</comment>. <pub-id pub-id-type="doi">10.1093/bfgp/elae005</pub-id>
</citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>MHAM-NPI: predicting ncRNA-protein interactions based on multi-head attention mechanism</article-title>. <source>Comput. Biol. Med.</source> <volume>163</volume>, <fpage>107143</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2023.107143</pub-id>
</citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liao</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2024b</year>). <article-title>Revisiting drug&#x2013;protein interaction prediction: a novel global&#x2013;local perspective</article-title>. <source>Bioinformatics</source> <volume>40</volume> (<issue>5</issue>), <fpage>btae271</fpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btae271</pub-id>
</citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Lv</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2024c</year>). <article-title>Joint masking and self-supervised strategies for inferring small molecule-miRNA associations</article-title>. <source>Mol. Therapy-Nucleic Acids</source> <volume>35</volume> (<issue>1</issue>), <fpage>102103</fpage>. <pub-id pub-id-type="doi">10.1016/j.omtn.2023.102103</pub-id>
</citation>
</ref>
<ref id="B57">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2024a</year>). <article-title>Joint deep autoencoder and subgraph augmentation for inferring microbial responses to drugs</article-title>. <source>Briefings Bioinforma.</source> <volume>25</volume> (<issue>1</issue>), <fpage>bbad483</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbad483</pub-id>
</citation>
</ref>
<ref id="B58">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Predicting ncRNA&#x2013;protein interactions based on dual graph convolutional network and pairwise learning</article-title>. <source>Briefings Bioinforma.</source> <volume>23</volume> (<issue>6</issue>), <fpage>bbac339</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbac339</pub-id>
</citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhuo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>StableDNAm: towards a stable and efficient model for predicting DNA methylation based on adaptive feature correction learning</article-title>. <source>BMC genomics</source> <volume>24</volume> (<issue>1</issue>), <fpage>742</fpage>. <pub-id pub-id-type="doi">10.1186/s12864-023-09802-7</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>