liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Leveraging Dynamic Range Analysis for Efficient Post-Training Quantization in Graph Convolutional Networks
Linköping University, Department of Electrical Engineering, Electronics and Computer Engineering. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-1083-5864
Linköping University, Department of Electrical Engineering, Electronics and Computer Engineering. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-1823-4211
Linköping University, Department of Electrical Engineering, Electronics and Computer Engineering. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0002-5153-5481
2024 (English)In: 2024 IEEE NORDIC CIRCUITS AND SYSTEMS CONFERENCE, NORCAS, IEEE , 2024Conference paper, Published paper (Refereed)
Abstract [en]

This paper presents a novel tensor-wise post-training quantization flow suitable for the sparse and dense tensors present in Graph Convolutional Networks (GCNs), which are a popular type of graph neural networks. The quantization approach employs KL (Kullback-Leibler) divergence and range analysis at the tensor granularity level to address the distinct sources of quantization errors in GCNs and its attractiveness lies in its independence from retraining or access to the full training dataset. The evaluation is performed with the popular citation datasets and shows that our method is competitive with the accuracy of the original floating-point model and also with quantization-aware training (QAT) approaches tailored for INT8 and INT4 precision. This is despite the significant potential for precision optimization of QAT at the cost of retraining. The obtained quantized and sparse tensors are used by a hardware overlay accelerator obtained using High-level synthesis (HLS) and integrated in Pytorch. The results using the Zynq Z7020 device available on the PYNQ-Z2 board show an 11x speedup over the integrated CPU and a 4.2x reduction in memory consumption.

Place, publisher, year, edition, pages
IEEE , 2024.
Keywords [en]
FPGA HLS; graph neural network; hardware accelerator; quantization; streaming PYNQ overlay
National Category
Computer Systems
Identifiers
URN: urn:nbn:se:liu:diva-213072DOI: 10.1109/NorCAS64408.2024.10752486ISI: 001444043400048Scopus ID: 2-s2.0-85211903968ISBN: 9798331517663 (electronic)ISBN: 9798331517670 (print)OAI: oai:DiVA.org:liu-213072DiVA, id: diva2:1952903
Conference
10th Nordic Circuits and Systems Conference, Lund, SWEDEN, oct 29-30, 2024
Note

Funding Agencies|Wallenberg AI autonomous systems and software (WASP) program - Knut and Alice Wallenberg Foundation

Available from: 2025-04-16 Created: 2025-04-16 Last updated: 2025-10-17

Open Access in DiVA

fulltext(8379 kB)20 downloads
File information
File name FULLTEXT01.pdfFile size 8379 kBChecksum SHA-512
6fca730ac256e63b92f5e9f76bc6e543172c2d805bc2bd5ff0324bc348daa127b4d7a30053c02e2709b9e49c9ac9eabccfe6a9fb6064963da6b6df074ce4d138
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Search in DiVA

By author/editor
Mousanejad Jeddi, HadiGrailootanha, MahdiehNunez-Yanez, Jose Luis
By organisation
Electronics and Computer EngineeringFaculty of Science & Engineering
Computer Systems

Search outside of DiVA

GoogleGoogle Scholar
Total: 20 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 72 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf