liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Robot Sound in Interaction: Analyzing and Designing Sound for Human-Robot Coordination
Linköping University, Department of Culture and Society, Division of Language, Culture and Interaction. Linköping University, Faculty of Arts and Sciences.ORCID iD: 0000-0003-0992-5176
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Robots naturally emit sound, but we still know little about how sound can serve as an interface that makes a robot’s behavior explainable to humans. This dissertation draws on insights about human practices for coordinating bodily activities through sound, investigating how they could inform robot design. My work builds on three video corpora, involving i) a Cozmo robot in ten family homes, ii) autonomous public shuttle buses in an urban environment, and iii) a teamwork robot prototype controlled by a researcher and interacting with study participants in an experimental setting. I approached the data from two methodological angles, exploring how they can speak to each other: I first carried out an empirical analysis of the video data from an Ethnomethodology and Conversation Analysis (EMCA) perspective, focusing on how humans make sense of robot sound on a moment-by-moment basis in naturally occurring interaction. Subsequently, taking an Interaction Design perspective, I used my video recordings as a design material for exploring how robot sound could be designed in and for real-time interaction. My work contributes to Human-Robot Interaction through detailed studies of robots in the world (rather than in the lab), focusing on how participants make sense of robot sounds. I present a novel framework for designing sound in and for interaction and a prototyping practice that allows practitioners to embed an EMCA stance into their designs. The dissertation contributes to EMCA by describing how members embed autonomous machines into the social organization of activities and how humans treat robots as participants in the interaction. I make a contribution to the development of EMCA hybrid studies by seeking a synthesis between EMCA and robot interaction design.

Abstract [sv]

Trots att ljud är en naturlig del av en robots närvaro vet vi fortvarande väldigt lite om hur ljud kan användas i gränssnitt för att göra robotars beteende förståeligt för människor. Denna avhandling utgår från nya insikter om hur människor använder sina röster i kroppsliga aktiviteter, för att undersöka hur denna kunskap kan användas vid gestaltning av robotar. Avhandlingen bygger på tre videokorpusar som visar i) leksaksroboten Cozmo i tio olika barnfamiljers hemmiljö, ii) två autonoma bussar i stadsmiljö och iii) en forskarstyrd prototyp av en robot för grupparbete i en experimentell miljö. Korpusarna studerades utifrån ett etnometodologiskt och interaktionsanalytiskt perspektiv (eng. ethnomethodology and conversation analysis, EMCA). Analysen fokuserade på hur människor visar sin förståelse av robotljud i naturligt förekommande interaktioner. De två sistnämnda korpusarna användes dessutom som material för interaktionsdesign i syfte att utforska hur robotljud kan utformas för att stödja realtidsinteraktion. Arbetet bidrar till fältet människa-robotinteraktion genom att erbjuda detaljerade studier av robotar i världen (i motsats till laboratoriemiljö) med fokus på hur deltagare i sampel med en robot förstår dess ljud. Avhandlingen föreslår ett nytt ramverk för att utforma ljud för interaktionella syften och en metod för att implementera ett EMCA-förhållningssätt inom designpraktiker. Arbetet beskriver även hur autonoma maskiner kan ingå i socialt organiserade aktiviteter och hur robotar kan behandlas som deltagare i interaktion med människor. Slutligen bidrar avhandlingen även till utvecklingen av EMCA-hybridstudier genom att utforska möjligheten att utveckla en EMCA-informerad metod för design av robotinteraktion.

Place, publisher, year, edition, pages
Linköping: Linköping University Electronic Press, 2023. , p. 113
Series
Linköping Studies in Arts and Sciences, ISSN 0282-9800 ; 853Studies in Language and Culture, ISSN 1403-2570 ; 36
Keywords [en]
Conversation analysis, Ethnomethodology, Human-robot interaction, Interaction design, Multimodality, Non-lexical sounds, Robot sound, Video analysis
Keywords [sv]
Etnometodologi, Icke-lexikala ljud, Interaktionsanalys, Interaktionsdesign, Multimodalitet, Människa-robotinteraktion, Robotljud, Videoanalys
National Category
Human Computer Interaction
Identifiers
URN: urn:nbn:se:liu:diva-193560DOI: 10.3384/9789180751179ISBN: 9789180751162 (print)ISBN: 9789180751179 (electronic)OAI: oai:DiVA.org:liu-193560DiVA, id: diva2:1754957
Public defence
2023-06-08, KEY1, Building Key, Campus Valla, Linköping, 10:15 (English)
Opponent
Supervisors
Available from: 2023-05-05 Created: 2023-05-05 Last updated: 2023-05-05Bibliographically approved
List of papers
1. "Are you sad, Cozmo?" How humans make sense of a home robot's emotion displays
Open this publication in new window or tab >>"Are you sad, Cozmo?" How humans make sense of a home robot's emotion displays
2020 (English)In: Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction (HRI'20), ACM Press, 2020, p. 461-470Conference paper, Published paper (Refereed)
Abstract [en]

This paper explores how humans interpret displays of emotion pro- duced by a social robot in real world situated interaction. Taking a multimodal conversation analytic approach, we analyze video data of families interacting with a Cozmo robot in their homes. Focusing on one happy and one sad robot animation, we study, on a turn-by-turn basis, how participants respond to audible and visible robot behavior designed to display emotion. We show how emotion animations are consequential for interactional progres- sivity: While displays of happiness typically move the interaction forward, displays of sadness regularly lead to a reconsideration of previous actions by humans. Furthermore, in making sense of the robot animations people may move beyond the designer’s re- ported intentions, actually broadening the opportunities for their subsequent engagement. We discuss how sadness functions as an interactional "rewind button" and how the inherent vagueness of emotion displays can be deployed in design.

Place, publisher, year, edition, pages
ACM Press, 2020
Keywords
emotion; affect; social robots; robots in the home; conversation analysis; non-lexical sounds; long-term interaction
National Category
Humanities and the Arts Computer and Information Sciences Human Computer Interaction
Identifiers
urn:nbn:se:liu:diva-167357 (URN)10.1145/3319502.3374814 (DOI)000570011000049 ()978-1-4503-6746-2 (ISBN)
Conference
2020 ACM/IEEE International Conference on Human-Robot Interaction (HRI'20)
Funder
Swedish Research Council, VR 2016-00827
Note

Funding agencies:  Swedish Research CouncilSwedish Research Council [2016-00827]

Available from: 2020-07-02 Created: 2020-07-02 Last updated: 2023-05-05
2. Managing Delays in Human-Robot Interaction
Open this publication in new window or tab >>Managing Delays in Human-Robot Interaction
2022 (English)In: ACM Transactions on Computer-Human Interaction, ISSN 1073-0516, E-ISSN 1557-7325, Vol. 30, no 4, article id 50Article in journal (Refereed) Published
Abstract [en]

Delays in the completion of joint actions are sometimes unavoidable. How should a robot communicate that it cannot immediately act or respond in a collaborative task? Drawing on video recordings of a face scanning activity in family homes, we investigate how humans make sense of a Cozmo robot’s delays on a moment-by-moment basis. Cozmo’s sounds and embodied actions are recognized as indicators of delay but encourage human participants to act in ways that undermine the scanning process. In comparing the robot’s delay management strategies with human-human vocal and embodied practices, we demonstrate key differences in the sequences that impact how the robot is understood. The study demonstrates how delay events are accomplished as embodied displays that are distributed across co-participants. We present a framework for making delay transparent through situated explanations, particularly in the form of non-lexical sounds and bodily actions.

Place, publisher, year, edition, pages
ACM Digital Library, 2022
Keywords
Conversation analysis, Time delay, Embodiment, Ethnomethodology, System response time, Sound, Engagement
National Category
Human Computer Interaction
Identifiers
urn:nbn:se:liu:diva-193557 (URN)10.1145/3569890 (DOI)001067749200001 ()
Note

Funding: Swedish Research Council [2016-00827]

Available from: 2023-05-05 Created: 2023-05-05 Last updated: 2023-11-03
3. Why Autonomous Driving Is So Hard: The Social Dimension of Traffic
Open this publication in new window or tab >>Why Autonomous Driving Is So Hard: The Social Dimension of Traffic
2021 (English)In: ACM/IEEE International Conference on Human-Robot Interaction (HRI), IEEE Computer Society, 2021, p. 81-85Conference paper, Published paper (Refereed)
Abstract [en]

Smooth traffic presupposes fine coordination between different actors, such as pedestrians, cyclists and car drivers. When autonomous vehicles join regular traffic, they need to coordinate with humans on the road. Prior work has often studied and designed for interaction with autonomous vehicles in structured environments such as traffic intersections. This paper describes aspects of coordination also in less structured situations during mundane maneuvers such as overtaking. Taking an ethnomethodological and conversation analytic approach, the paper analyzes video recordings of self-driving shuttle buses in Sweden. Initial findings suggest that the shuttle buses currently do not comply with cyclists’ expectations of social coordination in traffic. The paper highlights that communication and coordination with human road users is crucial for smooth flow of traffic and successful deployment of autonomous vehicles also in less structured traffic environments.

Place, publisher, year, edition, pages
IEEE Computer Society, 2021
Series
ACM/IEEE International Conference on Human-Robot Interaction, ISSN 2167-2148
Keywords
autonomous vehicle, shuttle bus, cyclist, conversation analysis, ethnomethodology, field study, mobile interaction
National Category
Computer and Information Sciences Languages and Literature
Identifiers
urn:nbn:se:liu:diva-173058 (URN)10.1145/3434074.3447133 (DOI)000767970100013 ()2-s2.0-85102771331 (Scopus ID)9781450382908 (ISBN)
Conference
ACM/IEEE International Conference on Human-Robot Interaction (HRI), ELECTR NETWORK, aug 08-11, 2021
Funder
Swedish Research Council, 2016-00827
Note

Funding: Swedish Research CouncilSwedish Research CouncilEuropean Commission [2016-00827]

Available from: 2021-01-29 Created: 2021-01-29 Last updated: 2023-05-05Bibliographically approved
4. Designing Robot Sound-In-Interaction: The Case of Autonomous Public Transport Shuttle Buses
Open this publication in new window or tab >>Designing Robot Sound-In-Interaction: The Case of Autonomous Public Transport Shuttle Buses
2023 (English)In: HRI '23: Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction, ACM Digital Library, 2023, p. 172-182Conference paper, Published paper (Refereed)
Abstract [en]

Horns and sirens are important tools for communicating on the road, which are still understudied in autonomous vehicles. While HRI has explored different ways in which robots could sound, we focus on the range of actions that a single sound can accomplish in interaction. In a Research through Design study involving autonomous shuttle buses in public transport, we explored sound design with the help of voice-overs to video recordings of the buses on the road and Wizard-of-Oz tests in live traffic. The buses are slowed down by (unnecessary) braking in response to people getting close. We found that prolonged jingles draw attention to the bus and invite interaction, while repeated short beeps and bell sounds can instruct the movement of others away from the bus. We highlight the importance of designing sound in sequential interaction and describe a new method for embedding video interaction analysis in the design process.

Place, publisher, year, edition, pages
ACM Digital Library, 2023
Series
HRI ’23
Keywords
Conversation analysis, Sound design, Ethnomethodology
National Category
Human Computer Interaction
Identifiers
urn:nbn:se:liu:diva-193558 (URN)10.1145/3568162.3576979 (DOI)9781450399647 (ISBN)
Conference
HRI: ACM/IEEE International Conference on Human-Robot Interaction, Stockholm, March 13-16, 2023
Available from: 2023-05-05 Created: 2023-05-05 Last updated: 2023-05-05

Open Access in DiVA

fulltext(4614 kB)622 downloads
File information
File name FULLTEXT01.pdfFile size 4614 kBChecksum SHA-512
8dc8f8757ca3ee5bcfd9b9205362d5848bfdb2015938e55511f2eef9e5a620a36f6d84b7bfecfc7f1efadc17214fac07946194c2043ed1a1cf6ab2f1714c5595
Type fulltextMimetype application/pdf
Order online >>

Other links

Publisher's full text

Authority records

Pelikan, Hannah

Search in DiVA

By author/editor
Pelikan, Hannah
By organisation
Division of Language, Culture and InteractionFaculty of Arts and Sciences
Human Computer Interaction

Search outside of DiVA

GoogleGoogle Scholar
Total: 622 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 3452 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf