liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Designing Explainable and Counterfactual-Based AI Interfaces for Operators in Process Industries
ABB AB Corporate Research, Sweden.ORCID iD: 0000-0001-9645-6990
Umeå University, Sweden.ORCID iD: 0000-0002-9808-2037
ABB AB Corporate Research, Sweden.ORCID iD: 0000-0003-4238-5976
Linköping University, Department of Science and Technology, Media and Information Technology. Linköping University, Faculty of Science & Engineering. (iVis, INV)ORCID iD: 0000-0001-6741-4337
Show others and affiliations
2025 (English)In: Proceedings of the 20th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP '25): Volume 1: GRAPP, HUCAPP and IVAPP, SciTePress, 2025, p. 831-842Conference paper, Published paper (Refereed)
Abstract [en]

Industrial applications of Artificial Intelligence (AI) can be hindered by the issues of explainability and trust from end users. Human-computer interaction and eXplainable AI (XAI) concerns become imperative in such scenarios. However, the prior evidence of applying more general principles and techniques in specialized industrial scenarios is often limited. In this case study, we focus on designing interactive interfaces of XAI solutions for operators in the pulp and paper industry. The explanation techniques supported and compared include counterfactual and feature importance explanations. We applied the user-centered design methodology, including the analysis of requirements elicited from operators during site visits and interactive interface prototype evaluation eventually conducted on site with five operators. Our results indicate that the operators preferred the combination of counterfactual and feature importance explanations. The study also provides lessons learned for researchers and practitioners.

Place, publisher, year, edition, pages
SciTePress, 2025. p. 831-842
Series
VISIGRAPP, ISSN 2184-4321
Keywords [en]
Explainable AI(XAI), Human-Centered AI, Counterfactual Explanations, Feature Importance, Visualization, Process Industry, User-Centered Design
National Category
Human Computer Interaction Computer Sciences
Identifiers
URN: urn:nbn:se:liu:diva-210848DOI: 10.5220/0013107700003912ISBN: 978-989-758-728-3 (electronic)OAI: oai:DiVA.org:liu-210848DiVA, id: diva2:1925648
Conference
International Conference on Information Visualization Theory and Applications (IVAPP), 26-28 February, 2025
Projects
EXPLAIN
Funder
Vinnova, 2021-04336
Note

The present study is funded by VINNOVA Sweden (2021-04336), Bundesministerium für Bildung und Forschung (BMBF; 01IS22030), and Rijksdienst voor Ondernemend Nederland (AI2212001) under the project Explanatory Artificial Interactive Intelligence for Industry (EXPLAIN).

Available from: 2025-01-09 Created: 2025-01-09 Last updated: 2025-03-11

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textPDF

Authority records

Zohrevandi, ElmiraKucher, Kostiantyn

Search in DiVA

By author/editor
Zhang, YanqingMethnani, LeilaBrorsson, EmmanuelZohrevandi, ElmiraKucher, Kostiantyn
By organisation
Media and Information TechnologyFaculty of Science & Engineering
Human Computer InteractionComputer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 219 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf