liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Uncertainty quantification in neural network classifiers—A local linear approach
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-0695-0720
Uppsala University, Uppsala, Sweden.
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0001-6957-2603
Swedish Defence Research Agency (FOI), Linköping, Sweden.
2024 (English)In: Automatica, ISSN 0005-1098, E-ISSN 1873-2836, Vol. 163, article id 111563Article in journal (Refereed) Published
Abstract [en]

lassifiers based on neural networks (nn) often lack a measure of uncertainty in the predicted class. We propose a method to estimate the probability mass function (pmf) of the different classes, as well as the covariance of the estimated pmf. First, a local linear approach is used during the training phase to recursively compute the covariance of the parameters in the nn. Secondly, in the classification phase, another local linear approach is used to propagate the covariance of the learned nn parameters to the uncertainty in the output of the last layer of the nn. This allows for an efficient Monte Carlo (mc) approach for; (i) estimating the pmf; (ii) calculating the covariance of the estimated pmf; and (iii) proper risk assessment and fusion of multiple classifiers. Two classical image classification tasks, i.e., mnist, and cfar10, are used to demonstrate the efficiency of the proposed method.

Place, publisher, year, edition, pages
Elsevier, 2024. Vol. 163, article id 111563
Keywords [en]
Neural networksUncertainty descriptionsInformation and sensor fusionIdentification and model reduction
National Category
Probability Theory and Statistics
Identifiers
URN: urn:nbn:se:liu:diva-200886DOI: 10.1016/j.automatica.2024.111563ISI: 001173877200001OAI: oai:DiVA.org:liu-200886DiVA, id: diva2:1837802
Funder
Swedish Research CouncilVinnova
Note

Funding Agencies|Sweden's innovation agency, Vinnova [2018-02700]; Swedish Research Council

Available from: 2024-02-14 Created: 2024-02-14 Last updated: 2024-03-20
In thesis
1. Approximative Uncertainty in Neural Network Predictions
Open this publication in new window or tab >>Approximative Uncertainty in Neural Network Predictions
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Suppose data-driven black-box models, e.g., neural networks, should be used as components in safety-critical systems such as autonomous vehicles. In that case, knowing how uncertain they are in their predictions is crucial. However, this needs to be provided for standard formulations of neural networks. Hence, this thesis aims to develop a method that can, out-of-the-box, extend the standard formulations to include uncertainty in the prediction. The proposed method in the thesis is based on a local linear approximation, using a two-step linearization to quantify the uncertainty in the prediction from the neural network. First, the posterior distribution of the neural network parameters is approximated using a Gaussian distribution. The mean of the distribution is at the maximum a posteriori estimate of the parameters, and the covariance is estimated using the shape of the likelihood function in the vicinity of the estimated parameters. The second linearization is used to propagate the uncertainty in the parameters to uncertainty in the model’s output. Hence, to create a linear approximation of the nonlinear model that a neural network is. 

The first part of the thesis considers regression problems with examples of road-friction experiments using simulated and experimentally collected data. For the model-order selection problem, it is shown that the method does not under-estimate the uncertainty in the prediction of overparametrized models. 

The second part of the thesis considers classification problems. The concept of calibration of the uncertainty, i.e., how reliable the uncertainty is and how close it resembles the true uncertainty, is considered. The proposed method is shown to create calibrated estimates of the uncertainty, evaluated on classical image data sets. From a computational perspective, the thesis proposes a recursive update of the parameter covariance, enhancing the method’s viability. Furthermore, it shows how quantified uncertainty can improve the robustness of a decision process by formulating an information fusion scheme that includes both temporal correlational and correlation between classifiers. Moreover, having access to a measure of uncertainty in the prediction is essential when detecting outliers in the data, i.e., examples that the neural network has yet to see during the training. On this task, the proposed method shows promising results. Finally, the thesis proposes an extension that enables a multimodal representation of the uncertainty. 

The third part of the thesis considers the tracking of objects in image sequences, where the object is detected using standard neural network-based object detection algorithms. It formulates the problem as a filtering problem with the prediction of the class and the position of the object viewed as the measurements. The filtering formulation improves robustness towards false classifications when evaluating the method on examples from animal conservation in the Swedish forests. 

Place, publisher, year, edition, pages
Linköping: Linköping University Electronic Press, 2023. p. 59
Series
Linköping Studies in Science and Technology. Dissertations, ISSN 0345-7524 ; 2358
National Category
Control Engineering
Identifiers
urn:nbn:se:liu:diva-198552 (URN)10.3384/9789180754064 (DOI)9789180754057 (ISBN)9789180754064 (ISBN)
Public defence
2023-11-17, Ada Lovelace, B-building and online via Zoom (contact ninna.stensgard@liu.se), Campus Valla, Linköping, 10:15 (English)
Opponent
Supervisors
Note

Funding: The research work in this thesis has been supported by the Sweden's Innovation Agency, Vinnova, through project iQDeep (project number 2018-02700).

Available from: 2023-10-17 Created: 2023-10-17 Last updated: 2024-02-21Bibliographically approved

Open Access in DiVA

fulltext(854 kB)268 downloads
File information
File name FULLTEXT01.pdfFile size 854 kBChecksum SHA-512
d49fbd4c5efc98bb49cc858ba52ed2a70cd8b51924ef845664ab841e162054d0f3273cc108247ad8f5d770fb0bf2840847c4a559886fc13e6ab4a58be30191ad
Type fulltextMimetype application/pdf

Other links

Publisher's full text

Authority records

Malmström, MagnusAxehill, Daniel

Search in DiVA

By author/editor
Malmström, MagnusAxehill, Daniel
By organisation
Automatic ControlFaculty of Science & Engineering
In the same journal
Automatica
Probability Theory and Statistics

Search outside of DiVA

GoogleGoogle Scholar
Total: 270 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 373 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf