Treffer: DA-MoE: Addressing depth-sensitivity in graph-level analysis through mixture of experts.

Title:
DA-MoE: Addressing depth-sensitivity in graph-level analysis through mixture of experts.
Authors:
Yao Z; School of Computer Science, Wuhan University, China., Chen M; School of Computer Science, Wuhan University, China., Liu C; School of Computer Science, Wuhan University, China., Meng X; School of Computer Science, Wuhan University, China., Zhan Y; JD Explore Academy, China., Wu J; Department of Computing, Macquarie University, Australia., Pan S; School of Information and Communication Technology, Griffith University, Australia., Xu H; Department of abdominal Oncology 1, Hubei Cancer Hospital, China. Electronic address: annexu333@126.com., Hu W; School of Computer Science, Wuhan University, China; Hubei Key Laboratory of Digital Finance Innovation, Hubei University of Economics, China. Electronic address: hwb@whu.edu.cn.
Source:
Neural networks : the official journal of the International Neural Network Society [Neural Netw] 2026 Feb; Vol. 194, pp. 108064. Date of Electronic Publication: 2025 Sep 01.
Publication Type:
Journal Article
Language:
English
Journal Info:
Publisher: Pergamon Press Country of Publication: United States NLM ID: 8805018 Publication Model: Print-Electronic Cited Medium: Internet ISSN: 1879-2782 (Electronic) Linking ISSN: 08936080 NLM ISO Abbreviation: Neural Netw Subsets: MEDLINE
Imprint Name(s):
Original Publication: New York : Pergamon Press, [c1988-
Contributed Indexing:
Keywords: Graph neural networks; Mixture of expert
Entry Date(s):
Date Created: 20250914 Date Completed: 20251216 Latest Revision: 20251216
Update Code:
20251216
DOI:
10.1016/j.neunet.2025.108064
PMID:
40946674
Database:
MEDLINE

Weitere Informationen

Graph neural networks (GNNs) are gaining popularity for processing graph data. In real-world scenarios, graph data within the same dataset can vary significantly in scale. This variability leads to depth-sensitivity, where the optimal depth of GNN layers depends on the scale of the graph data. Empirically, fewer layers are sufficient for message passing in smaller graphs, while larger graphs typically require deeper networks to capture long-range dependencies and global features. However, existing methods generally use a fixed number of GNN layers to generate representations for all graphs, overlooking the depth-sensitivity issue in graph data. To address this challenge, we propose the depth adaptive mixture of expert (DA-MoE) method, which incorporates two main improvements to GNN backbone: 1) DA-MoE employs different GNN layers, each considered an expert with its own parameters. Such a design allows the model to flexibly aggregate information at different scales, effectively addressing the depth-sensitivity issue in graph data. 2) DA-MoE utilizes GNN to capture the structural information instead of the linear projections in the gating network. Thus, the gating network enables the model to capture complex patterns and dependencies within the data. By leveraging these improvements, each expert in DA-MoE specifically learns distinct graph patterns at different scales. Furthermore, comprehensive experiments on the TU dataset and open graph benchmark (OGB) have shown that DA-MoE consistently surpasses existing baselines on various tasks, including graph, node, and link-level analyses. The code are available at https://github.com/Celin-Yao/DA-MoE.
(Copyright © 2025 Elsevier Ltd. All rights reserved.)

Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.