liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Accelerating Graph Neural Networks in Pytorch With HLS and Deep Dataflows
Linköping University, Department of Electrical Engineering, Computer Engineering. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0002-5153-5481
2023 (English)In: Applied Reconfigurable Computing. Architectures, Tools, and Applications. ARC 2023, SPRINGER INTERNATIONAL PUBLISHING AG , 2023, Vol. 14251, p. 131-145Conference paper, Published paper (Refereed)
Abstract [en]

Graph neural networks (GNNs) combine sparse and densedata compute requirements that are challenging to meet in resourceconstrained embedded hardware. In this paper, we investigate a dataflowof dataflows architecture that optimizes data access and processing element utilization. The architecture is described with high-level synthesisand offers multiple configuration options including varying the number ofindependent hardware threads, the interface data width and the numberof compute units per thread. Each hardware thread uses a fine-graineddataflow to stream words with a bit-width that depends on the network precision while a coarse-grained dataflow links the thread stagesstreaming partially-computed matrix tiles. The accelerator is mappedto the programmable logic of a Zynq Ultrascale device whose processing system runs Pytorch extended with PYNQ overlays. Results basedon the citation networks show a performance gain of up to 140x withmulti-threaded hardware configurations compared with the optimizedsoftware implementation available in Pytorch. The results also show competitive performance of the embedded hardware compared with otherhigh-performance state-of-the-art hardware accelerators.

Place, publisher, year, edition, pages
SPRINGER INTERNATIONAL PUBLISHING AG , 2023. Vol. 14251, p. 131-145
Series
Lecture Notes in Computer Science, ISSN 0302-9743, E-ISSN 1611-3349
Keywords [en]
neural network, FPGA, sparse, HLS, GNN, Pytorch
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:liu:diva-199899DOI: 10.1007/978-3-031-42921-7_9ISI: 001162213700009ISBN: 9783031429200 (print)ISBN: 9783031429217 (electronic)OAI: oai:DiVA.org:liu-199899DiVA, id: diva2:1823736
Conference
19th International Symposium on Applied Reconfigurable Computing (ARC) - Architectures, Tools, and Applications, Cottbus, GERMANY, sep 27-29, 2023
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP), 308397
Note

Funding Agencies|Wallenberg AI autonomous autonomous systems and software (WASP) program - Knut and Alice Wallenberg Foundation

Available from: 2024-01-03 Created: 2024-01-03 Last updated: 2024-09-04

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Authority records

Nunez-Yanez, Jose Luis

Search in DiVA

By author/editor
Nunez-Yanez, Jose Luis
By organisation
Computer EngineeringFaculty of Science & Engineering
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 41 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf