liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
FT-DeepNets: Fault-Tolerant Convolutional Neural Networks with Kernel-based Duplication
Carnegie Mellon Univ, PA 15213 USA.
Purdue Univ, IN 47907 USA.
Carnegie Mellon Univ, PA 15213 USA.
Linköping University, Department of Computer and Information Science, Software and Systems. Linköping University, Faculty of Science & Engineering.
Show others and affiliations
2022 (English)In: 2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), IEEE COMPUTER SOC , 2022, p. 1878-1887Conference paper, Published paper (Refereed)
Abstract [en]

Deep neural network (deepnet) applications play a crucial role in safety-critical systems such as autonomous vehicles (AVs). An AV must drive safely towards its destination, avoiding obstacles, and respond quickly when the vehicle must stop. Any transient errors in software calculations or hardware memory in these deepnet applications can potentially lead to dramatically incorrect results. Therefore, assessing and mitigating any transient errors and providing robust results are important for safety-critical systems. Previous research on this subject focused on detecting errors and then recovering from the errors by re-running the network. Other approaches were based on the extent of full network duplication such as the ensemble learning-based approach to boost system fault-tolerance by leveraging each models advantages. However, it is hard to detect errors in a deep neural network, and the computational overhead of full redundancy can be substantial. We first study the impact of the error types and locations in deepnets. We next focus on selecting which part should be duplicated using multiple ranking methods to measure the order of importance among neurons. We find that the duplication overhead for computation and memory is a tradeoff between algorithmic performance and robustness. To achieve higher robustness with less system overhead, we present two error protection mechanisms that only duplicate parts of the network from critical neurons. Finally, we substantiate the practical feasibility of our approach and evaluate the improvement in the accuracy of a deepnet in the presence of errors. We demonstrate these results using a case study with real-world applications on an Nvidia GeForce RTX 2070Ti GPU and an Nvidia Xavier embedded platform used by automotive OEMs.

Place, publisher, year, edition, pages
IEEE COMPUTER SOC , 2022. p. 1878-1887
Series
IEEE Winter Conference on Applications of Computer Vision, ISSN 2472-6737
National Category
Embedded Systems
Identifiers
URN: urn:nbn:se:liu:diva-187590DOI: 10.1109/WACV51458.2022.00194ISI: 000800471201093ISBN: 9781665409155 (electronic)ISBN: 9781665409162 (print)OAI: oai:DiVA.org:liu-187590DiVA, id: diva2:1690594
Conference
22nd IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, jan 04-08, 2022
Available from: 2022-08-26 Created: 2022-08-26 Last updated: 2022-08-26

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Search in DiVA

By author/editor
Samii, Soheil
By organisation
Software and SystemsFaculty of Science & Engineering
Embedded Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 116 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf