liu.seSök publikationer i DiVA
Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Calibration tests beyond classification
Department of Information Technology Uppsala University, Sweden.ORCID-id: 0000-0001-9282-053x
Linköpings universitet, Institutionen för datavetenskap, Statistik och maskininlärning. Linköpings universitet, Filosofiska fakulteten.ORCID-id: 0000-0003-3749-5820
Department of Information Technology Uppsala University, Sweden.ORCID-id: 0000-0002-6698-0166
2021 (Engelska)Ingår i: ICLR 2021 - 9th International Conference on Learning Representations Proceedings, International Conference on Learning Representations, ICLR , 2021, s. 1-37Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Most supervised machine learning tasks are subject to irreducible prediction errors. Probabilistic predictive models address this limitation by providing probability distributions that represent a belief over plausible targets, rather than point estimates. Such models can be a valuable tool in decision-making under uncertainty, provided that the model output is meaningful and interpretable. Calibrated models guarantee that the probabilistic predictions are neither over- nor under-confident. In the machine learning literature, different measures and statistical tests have been proposed and studied for evaluating the calibration of classification models. For regression problems, however, research has been focused on a weaker condition of calibration based on predicted quantiles for real-valued targets. In this paper, we propose the first framework that unifies calibration evaluation and tests for probabilistic predictive models. It applies to any such model, including classification and regression models of arbitrary dimension. Furthermore, the framework generalizes existing measures and provides a more intuitive reformulation of a recently proposed framework for calibration in multi-class classification.

Ort, förlag, år, upplaga, sidor
International Conference on Learning Representations, ICLR , 2021. s. 1-37
Nyckelord [en]
calibration, uncertainty quantification, framework, integral probability metric, maximum mean discrepancy
Nationell ämneskategori
Sannolikhetsteori och statistik
Forskningsämne
Matematisk statistik
Identifikatorer
URN: urn:nbn:se:liu:diva-188940Scopus ID: 2-s2.0-85147937089OAI: oai:DiVA.org:liu-188940DiVA, id: diva2:1700801
Konferens
International Conference on Learning Representations, Virtual conference, May 3 - May 7, 2021
Tillgänglig från: 2020-12-23 Skapad: 2022-10-03 Senast uppdaterad: 2024-08-23Bibliografiskt granskad

Open Access i DiVA

Fulltext saknas i DiVA

Övriga länkar

ScopusLink to paperVideo presentationSource code of the experiments

Person

Widmann, DavidLindsten, FredrikZachariah, Dave

Sök vidare i DiVA

Av författaren/redaktören
Widmann, DavidLindsten, FredrikZachariah, Dave
Av organisationen
Statistik och maskininlärningFilosofiska fakulteten
Sannolikhetsteori och statistik

Sök vidare utanför DiVA

GoogleGoogle Scholar

urn-nbn

Altmetricpoäng

urn-nbn
Totalt: 71 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf