liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Asymptotic Prediction Error Variance for Feedforward Neural Networks
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-0695-0720
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0001-6957-2603
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0003-3270-171X
2020 (English)Conference paper, Published paper (Refereed)
Abstract [en]

The prediction uncertainty of a neural network is considered from a classical system identification point of view. To know this uncertainty is extremely important when using a network in decision and feedback applications. The asymptotic covariance of the internal parameters in the network due to noise in the observed dependent variables (output) and model class mismatch, i.e., the true system cannot be exactly described by the model class, is first surveyed. This is then applied to the prediction step of the network to get a closed form expression for the asymptotic, in training data information, prediction variance. Another interpretation of this expression is as the non-asymptotic Cramér-Rao Lower Bound. To approximate this expression, only the gradients and residuals, already computed in the gradient descent algorithms commonly used to train neural networks, are needed. Using a toy example, it is illustrated how the uncertainty in the output of a neural network can be estimated.

Place, publisher, year, edition, pages
Elsevier, 2020. Vol. 53, no 2, p. 1108-1113
Series
IFAC-PapersOnLine, ISSN 2405-8963
Keywords [en]
Neural Networks, Feedforward Networks, Uncertainty, System Identification, Estimation Theory, Cramér-Rao Bound, Identification for Control, Machine Learning
National Category
Control Engineering
Identifiers
URN: urn:nbn:se:liu:diva-175185DOI: 10.1016/j.ifacol.2020.12.1310ISI: 000652592500179Scopus ID: 2-s2.0-85104184850OAI: oai:DiVA.org:liu-175185DiVA, id: diva2:1546958
Conference
21st IFAC World Congress on Automatic Control - Meeting Societal Challenges, electronic meeting, UL 11-17, 2020
Note

Funding: Swedish Governmental Agency for Innovation SystemsVinnova [2018-02700]

Available from: 2021-04-23 Created: 2021-04-23 Last updated: 2023-10-17Bibliographically approved
In thesis
1. Approximative Uncertainty in Neural Network Predictions
Open this publication in new window or tab >>Approximative Uncertainty in Neural Network Predictions
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Suppose data-driven black-box models, e.g., neural networks, should be used as components in safety-critical systems such as autonomous vehicles. In that case, knowing how uncertain they are in their predictions is crucial. However, this needs to be provided for standard formulations of neural networks. Hence, this thesis aims to develop a method that can, out-of-the-box, extend the standard formulations to include uncertainty in the prediction. The proposed method in the thesis is based on a local linear approximation, using a two-step linearization to quantify the uncertainty in the prediction from the neural network. First, the posterior distribution of the neural network parameters is approximated using a Gaussian distribution. The mean of the distribution is at the maximum a posteriori estimate of the parameters, and the covariance is estimated using the shape of the likelihood function in the vicinity of the estimated parameters. The second linearization is used to propagate the uncertainty in the parameters to uncertainty in the model’s output. Hence, to create a linear approximation of the nonlinear model that a neural network is. 

The first part of the thesis considers regression problems with examples of road-friction experiments using simulated and experimentally collected data. For the model-order selection problem, it is shown that the method does not under-estimate the uncertainty in the prediction of overparametrized models. 

The second part of the thesis considers classification problems. The concept of calibration of the uncertainty, i.e., how reliable the uncertainty is and how close it resembles the true uncertainty, is considered. The proposed method is shown to create calibrated estimates of the uncertainty, evaluated on classical image data sets. From a computational perspective, the thesis proposes a recursive update of the parameter covariance, enhancing the method’s viability. Furthermore, it shows how quantified uncertainty can improve the robustness of a decision process by formulating an information fusion scheme that includes both temporal correlational and correlation between classifiers. Moreover, having access to a measure of uncertainty in the prediction is essential when detecting outliers in the data, i.e., examples that the neural network has yet to see during the training. On this task, the proposed method shows promising results. Finally, the thesis proposes an extension that enables a multimodal representation of the uncertainty. 

The third part of the thesis considers the tracking of objects in image sequences, where the object is detected using standard neural network-based object detection algorithms. It formulates the problem as a filtering problem with the prediction of the class and the position of the object viewed as the measurements. The filtering formulation improves robustness towards false classifications when evaluating the method on examples from animal conservation in the Swedish forests. 

Place, publisher, year, edition, pages
Linköping: Linköping University Electronic Press, 2023. p. 59
Series
Linköping Studies in Science and Technology. Dissertations, ISSN 0345-7524 ; 2358
National Category
Control Engineering
Identifiers
urn:nbn:se:liu:diva-198552 (URN)10.3384/9789180754064 (DOI)9789180754057 (ISBN)9789180754064 (ISBN)
Public defence
2023-11-17, Ada Lovelace, B-building and online via Zoom (contact ninna.stensgard@liu.se), Campus Valla, Linköping, 10:15 (English)
Opponent
Supervisors
Note

Funding: The research work in this thesis has been supported by the Sweden's Innovation Agency, Vinnova, through project iQDeep (project number 2018-02700).

Available from: 2023-10-17 Created: 2023-10-17 Last updated: 2024-02-21Bibliographically approved

Open Access in DiVA

fulltext(925 kB)496 downloads
File information
File name FULLTEXT01.pdfFile size 925 kBChecksum SHA-512
6f7a4f3894619cee206ba3f1ab4e82697ec5ccea473a508971c799d33dd6b268c4259b192247c27a5784e7d07aed2a162adcb593e0952b40911fcc253dcd2cc7
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Malmström, MagnusSkog, IsaacAxehill, DanielGustafsson, Fredrik

Search in DiVA

By author/editor
Malmström, MagnusSkog, IsaacAxehill, DanielGustafsson, Fredrik
By organisation
Automatic ControlFaculty of Science & Engineering
Control Engineering

Search outside of DiVA

GoogleGoogle Scholar
Total: 499 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 484 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf