Abstract
Transformer-based graph neural networks have accomplished notable achievements by utilizing the self-attention mechanism for message passing in various domains. However, traditional methods overlook the diverse significance of intra-node representations, focusing solely on internode interactions. To overcome this limitation, we propose a DAG (Dual Attention Graph), a novel approach that integrates both intra-node and internode dynamics for node classification tasks. By considering the information exchange process between nodes from dual branches, DAG provides a holistic understanding of information propagation within graphs, enhancing the interpretability of graph-based machine learning applications. The experimental evaluations demonstrate that DAG excels in node classification tasks, outperforming current benchmark models across ten datasets.
| Original language | English |
|---|---|
| Article number | 3691 |
| Journal | Mathematics |
| Volume | 11 |
| Issue number | 17 |
| DOIs | |
| State | Published - Sep 2023 |
| Externally published | Yes |
Keywords
- graph neural network
- message-passing mechanism
- node classification
Fingerprint
Dive into the research topics of 'DAG: Dual Attention Graph Representation Learning for Node Classification'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver