liu.seSearch for publications in DiVA
Change search
Link to record
Permanent link

Direct link
Khalili Sadaghiani, AbdolvahabORCID iD iconorcid.org/0000-0003-4870-2768
Publications (2 of 2) Show all publications
Khalili Sadaghiani, A. & Nunez-Yanez, J. (2025). Exploring the applicability of Graph Attention Networks in computer vision and their hardware acceleration. In: AccML papers 2025: . Paper presented at 7th Workshop on Accelerated Machine Learning (AccML) on HiPEAC 2025 Conference, 21st January, 2025, Barcelona, Spain. , Article ID 3.
Open this publication in new window or tab >>Exploring the applicability of Graph Attention Networks in computer vision and their hardware acceleration
2025 (English)In: AccML papers 2025, 2025, article id 3Conference paper, Published paper (Other academic)
Abstract [en]

Edge detection is a fundamental task in computer vision, crucial for object recognition, segmentation, and scene understanding. Traditional methods often fail to capture complex edge structures due to their inability to model intricate relationships between pixels. Graph Neural Networks (GNNs), particularly Graph Attention Networks (GATs), have shown promise in addressing these limitations by leveraging graph structures to model pixel relationships. This paper explores the applicability of Graph Attention Networks in edge detection, highlighting their advantages over ordinary Graph convolutional Networks (GCNs) through rigorous mathematical reasoning. We integrate GATs into an edge detection framework based on an encoder-decoder structure with U-Net architecture and provide detailed theoretical and implementation insights. Furthermore, we discuss the hardware acceleration of GCNs and GATs with a reconfigurable dataflow architecture integrated in the Pytorch framework. The experimental results demonstrate the superior performance of GAT-based edge detection and the potential acceleration possible on reconfigurable edge platforms with limited resources. The key advantage of our proposed method is its hardware-friendly design, making it highly suitable for FPGA acceleration while also enabling efficient optimization through pruning of the network. 

Keywords
edge detection, GNN, Graph Attention Networks, encoder-decoder structure, U-Net
National Category
Computer Sciences
Identifiers
urn:nbn:se:liu:diva-220183 (URN)
Conference
7th Workshop on Accelerated Machine Learning (AccML) on HiPEAC 2025 Conference, 21st January, 2025, Barcelona, Spain
Available from: 2025-12-19 Created: 2025-12-19 Last updated: 2025-12-19
Khalili Sadaghiani, A. & Nunez-Yanez, J. L. (2025). n-HDP-GNN: community-aware Bayesian clustering for over-smoothing-resilient, communication-efficient distributed GNNs. Journal of Supercomputing, 81(16), Article ID 1560.
Open this publication in new window or tab >>n-HDP-GNN: community-aware Bayesian clustering for over-smoothing-resilient, communication-efficient distributed GNNs
2025 (English)In: Journal of Supercomputing, ISSN 0920-8542, E-ISSN 1573-0484, Vol. 81, no 16, article id 1560Article in journal (Refereed) Published
Abstract [en]

Deep graph neural networks (GNNs) often suffer from over-smoothing, where node embeddings homogenize as depth grows. We present n-HDP-GNN, a probabilistic, community-aware architecture that couples Louvain coarsening with a nested Hierarchical Dirichlet process to learn soft responsibilities that gate message passing. Multi-level attention (node/community/global) then aggregates features while preserving separability. This selective diffusion delays over-smoothing-quantified once using MADGap-and, in distributed training, reduces cross-partition communication by lowering the cross-partition edge ratio and increasing edge-reduction. We evaluate across seven benchmarks spanning citation networks, a co-purchase network, and three large-scale tasks, under supervised, semi-supervised, and label-scarce regimes, against strong baselines. The proposed model delivers over 5% gain in accuracy, greater robustness, and a superior ability to capture long-range dependencies and subtle patterns. Deployed on a multi-node CPU cluster with PyTorch DDP, n-HDP-GNN attains + 11% higher throughput than the best competitor at matched accuracy, demonstrating that the same community-aware gating curbs over-smoothing and improves communication efficiency on commodity interconnects. Together, these results show that probabilistic, community-aware gating yields depth-robust representations without sacrificing scalability: mid-depth performance is strengthened, deep-depth degradation is reduced, and systems metrics improve in tandem turning a representation-level idea into a practical approach for training deep GNNs on modest clusters.

Place, publisher, year, edition, pages
SPRINGER, 2025
Keywords
Graph neural network (GNN); Nested Hierarchical Dirichlet Process (n-HDP); Over-smoothing; Community detection; Distributed training; High-performance computing (HPC)
National Category
Computer Engineering
Identifiers
urn:nbn:se:liu:diva-219780 (URN)10.1007/s11227-025-08017-9 (DOI)001614198600003 ()2-s2.0-105021525468 (Scopus ID)
Note

Funding Agencies|Linkping University

Available from: 2025-12-04 Created: 2025-12-04 Last updated: 2026-01-31
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-4870-2768

Search in DiVA

Show all publications