liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Deep Quantization of Graph Neural Networks with Run-Time Hardware-Aware Training
Linköping University, Department of Electrical Engineering, Computer Engineering. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0001-5326-4999
Linköping University, Department of Electrical Engineering, Computer Engineering. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-1823-4211
Linköping University, Department of Electrical Engineering, Computer Engineering. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-3470-3911
Linköping University, Department of Electrical Engineering, Computer Engineering. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0002-5153-5481
2024 (English)In: APPLIED RECONFIGURABLE COMPUTING. ARCHITECTURES, TOOLS, AND APPLICATIONS, ARC 2024, SPRINGER INTERNATIONAL PUBLISHING AG , 2024, Vol. 14553, p. 33-47Conference paper, Published paper (Refereed)
Abstract [en]

In this paper, we investigate the benefits of hardware-aware quantization in the gFADES hardware accelerator targeting Graph Convolutional Networks (GCNs). GCNs are a type of Graph Neural Networks (GNNs) that combine sparse and dense data compute requirements that are challenging to meet in resource-constrained embedded hardware. The gFADES architecture is optimized to work with the pruned data representations typically present in graph neural networks for the graph structure and features. It is described in High-Level Synthesis (HLS) which enables efficient design-space exploration of mixed precision hardware configurations. In this work, the mixed-precision design is embedded in the forward pass of the PyTorch back-propagation training loop to enable run-time hardware-aware training. It uses different data types to represent adjacency, feature, weight, internal, and output values which allows for a fine-grained optimization at the tensor level. The resulting hardware configuration after training reduces precision to a 4-bit data type for all inputs. It achieves little to no degradation in the classification accuracy, when training on the Planetoid database dataset, compared to the original 32-bit floating-point. The optimized hardware design running on an AMD/Xilinx Zynq Ultrascale+ FPGA device achieves over 600x speedup compared to the optimized PyTorch software implementation running on the multi-core ARM CPU in the processing system.

Place, publisher, year, edition, pages
SPRINGER INTERNATIONAL PUBLISHING AG , 2024. Vol. 14553, p. 33-47
Series
Lecture Notes in Computer Science, ISSN 0302-9743
Keywords [en]
Neural Networks; FPGA; GNN; quantization; low precision
National Category
Computer Engineering
Identifiers
URN: urn:nbn:se:liu:diva-204379DOI: 10.1007/978-3-031-55673-9_3ISI: 001212357700003ISBN: 9783031556722 (print)ISBN: 9783031556739 (electronic)OAI: oai:DiVA.org:liu-204379DiVA, id: diva2:1868948
Conference
20th International Symposium on Applied Reconfigurable Computing (ARC), Aveiro, PORTUGAL, mar 20-22, 2024
Note

Funding Agencies|Wallenberg AI, Autonomous Systems and Software Program (WASP) - Knut and Alice Wallenberg Foundation

Available from: 2024-06-12 Created: 2024-06-12 Last updated: 2024-06-12

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Search in DiVA

By author/editor
Hansson, OlleGrailootanha, MahdiehGustafsson, OscarNunez-Yanez, Jose Luis
By organisation
Computer EngineeringFaculty of Science & Engineering
Computer Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 245 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf