liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Detection of outliers in classification by using quantified uncertainty in neural networks
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-0695-0720
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0001-6957-2603
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-3270-171X
2022 (English)In: 25th International Conference of Information Fusion, IEEE, 2022Conference paper, Published paper (Refereed)
Abstract [en]

Neural Networks (NNS) can solve very hard classification and estimation tasks but are less well suited to solve complex sensor fusion challenges, such as end-to-end control of autonomous vehicles. Nevertheless, NN can still be a powerful tool for particular sub-problems in sensor fusion. This would require a reliable and quantifiable measure of the stochastic uncertainty in the predictions that can be compared to classical sensor measurements. However, current NN'S output some figure of merit, that is only a relative model fit and not a stochastic uncertainty. We propose to embed the NN'S in a proper stochastic system identification framework. In the training phase, the stochastic uncertainty of the parameters in the (last layers of the) NN is quantified. We show that this can be done recursively with very few extra computations. In the classification phase, Monte-Carlo (MC) samples are used to generate a set of classifier outputs. From this set, a distribution of the classifier output is obtained, which represents a proper description of the stochastic uncertainty of the predictions. We also show how to use the calculated uncertainty for outlier detection by including an artificial outlier class. In this way, the NN fits a sensor fusion framework much better. We evaluate the approach on images of handwritten digits. The proposed method is shown to be on par with MC dropout, while having lower computational complexity, and the outlier detection almost completely eliminates false classifications.

Place, publisher, year, edition, pages
IEEE, 2022.
Keywords [en]
Training, Uncertainty, Current measurement, Stochastic systems, Measurement uncertainty, Stochastic processes, Artificial neural networks
National Category
Control Engineering
Identifiers
URN: urn:nbn:se:liu:diva-188333DOI: 10.23919/FUSION49751.2022.9841376ISI: 000855689000146ISBN: 9781665489416 (print)ISBN: 9781737749721 (electronic)OAI: oai:DiVA.org:liu-188333DiVA, id: diva2:1694594
Conference
25th International Conference of Information Fusion, FUSION 2022, July 4-7, 2022, Linköping, Sweden
Note

Funding: Swedens innovation agency, Vinnova, through project iQDeep [2018-02700]

Available from: 2022-09-09 Created: 2022-09-09 Last updated: 2023-10-17
In thesis
1. Approximative Uncertainty in Neural Network Predictions
Open this publication in new window or tab >>Approximative Uncertainty in Neural Network Predictions
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Suppose data-driven black-box models, e.g., neural networks, should be used as components in safety-critical systems such as autonomous vehicles. In that case, knowing how uncertain they are in their predictions is crucial. However, this needs to be provided for standard formulations of neural networks. Hence, this thesis aims to develop a method that can, out-of-the-box, extend the standard formulations to include uncertainty in the prediction. The proposed method in the thesis is based on a local linear approximation, using a two-step linearization to quantify the uncertainty in the prediction from the neural network. First, the posterior distribution of the neural network parameters is approximated using a Gaussian distribution. The mean of the distribution is at the maximum a posteriori estimate of the parameters, and the covariance is estimated using the shape of the likelihood function in the vicinity of the estimated parameters. The second linearization is used to propagate the uncertainty in the parameters to uncertainty in the model’s output. Hence, to create a linear approximation of the nonlinear model that a neural network is. 

The first part of the thesis considers regression problems with examples of road-friction experiments using simulated and experimentally collected data. For the model-order selection problem, it is shown that the method does not under-estimate the uncertainty in the prediction of overparametrized models. 

The second part of the thesis considers classification problems. The concept of calibration of the uncertainty, i.e., how reliable the uncertainty is and how close it resembles the true uncertainty, is considered. The proposed method is shown to create calibrated estimates of the uncertainty, evaluated on classical image data sets. From a computational perspective, the thesis proposes a recursive update of the parameter covariance, enhancing the method’s viability. Furthermore, it shows how quantified uncertainty can improve the robustness of a decision process by formulating an information fusion scheme that includes both temporal correlational and correlation between classifiers. Moreover, having access to a measure of uncertainty in the prediction is essential when detecting outliers in the data, i.e., examples that the neural network has yet to see during the training. On this task, the proposed method shows promising results. Finally, the thesis proposes an extension that enables a multimodal representation of the uncertainty. 

The third part of the thesis considers the tracking of objects in image sequences, where the object is detected using standard neural network-based object detection algorithms. It formulates the problem as a filtering problem with the prediction of the class and the position of the object viewed as the measurements. The filtering formulation improves robustness towards false classifications when evaluating the method on examples from animal conservation in the Swedish forests. 

Place, publisher, year, edition, pages
Linköping: Linköping University Electronic Press, 2023. p. 59
Series
Linköping Studies in Science and Technology. Dissertations, ISSN 0345-7524 ; 2358
National Category
Control Engineering
Identifiers
urn:nbn:se:liu:diva-198552 (URN)10.3384/9789180754064 (DOI)9789180754057 (ISBN)9789180754064 (ISBN)
Public defence
2023-11-17, Ada Lovelace, B-building and online via Zoom (contact ninna.stensgard@liu.se), Campus Valla, Linköping, 10:15 (English)
Opponent
Supervisors
Note

Funding: The research work in this thesis has been supported by the Sweden's Innovation Agency, Vinnova, through project iQDeep (project number 2018-02700).

Available from: 2023-10-17 Created: 2023-10-17 Last updated: 2024-02-21Bibliographically approved

Open Access in DiVA

fulltext(729 kB)536 downloads
File information
File name FULLTEXT01.pdfFile size 729 kBChecksum SHA-512
65ced686bc70560c1260fc6e4c2207e53ad706963308daf6f874f44d4fd4869ae2d9f9ca31f5282729ef8095cd8ce00b4a89451c53e8a68ce48cf3d6faa26e84
Type fulltextMimetype application/pdf

Other links

Publisher's full text

Authority records

Malmström, MagnusSkog, IsaacAxehill, DanielGustafsson, Fredrik

Search in DiVA

By author/editor
Malmström, MagnusSkog, IsaacAxehill, DanielGustafsson, Fredrik
By organisation
Automatic ControlFaculty of Science & Engineering
Control Engineering

Search outside of DiVA

GoogleGoogle Scholar
Total: 539 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 434 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf