liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Predicting EEG Responses to Attended Speech via Deep Neural Networks for Speech
Linköping University, Department of Electrical Engineering, Automatic Control. Linköping University, Faculty of Science & Engineering. Eriksholm Res Ctr, Denmark.ORCID iD: 0000-0002-4655-9112
Eriksholm Res Ctr, Denmark; Univ Stuttgart, Germany.
Oticon AS, Denmark.
Oticon AS, Denmark.
Show others and affiliations
2023 (English)In: 2023 45TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY, EMBC, IEEE , 2023Conference paper, Published paper (Refereed)
Abstract [en]

Attending to the speech stream of interest in multi-talker environments can be a challenging task, particularly for listeners with hearing impairment. Research suggests that neural responses assessed with electroencephalography (EEG) are modulated by listener's auditory attention, revealing selective neural tracking (NT) of the attended speech. NT methods mostly rely on hand-engineered acoustic and linguistic speech features to predict the neural response. Only recently, deep neural network (DNN) models without specific linguistic information have been used to extract speech features for NT, demonstrating that speech features in hierarchical DNN layers can predict neural responses throughout the auditory pathway. In this study, we go one step further to investigate the suitability of similar DNN models for speech to predict neural responses to competing speech observed in EEG. We recorded EEG data using a 64-channel acquisition system from 17 listeners with normal hearing instructed to attend to one of two competing talkers. Our data revealed that EEG responses are significantly better predicted by DNN-extracted speech features than by hand-engineered acoustic features. Furthermore, analysis of hierarchical DNN layers showed that early layers yielded the highest predictions. Moreover, we found a significant increase in auditory attention classification accuracies with the use of DNN-extracted speech features over the use of hand-engineered acoustic features. These findings open a new avenue for development of new NT measures to evaluate and further advance hearing technology.

Place, publisher, year, edition, pages
IEEE , 2023.
Series
IEEE Engineering in Medicine and Biology Society Conference Proceedings, ISSN 1557-170X, E-ISSN 2694-0604
National Category
Otorhinolaryngology
Identifiers
URN: urn:nbn:se:liu:diva-201201DOI: 10.1109/EMBC40787.2023.10340027ISI: 001133788300084PubMedID: 38083171ISBN: 9798350324471 (electronic)ISBN: 9798350324488 (print)OAI: oai:DiVA.org:liu-201201DiVA, id: diva2:1840880
Conference
45th Annual International Conference of the IEEE-Engineering-in-Medicine-and-Biology-Society (EMBC), Sydney, AUSTRALIA, jul 24-27, 2023
Available from: 2024-02-27 Created: 2024-02-27 Last updated: 2024-02-27

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textPubMed

Search in DiVA

By author/editor
Alickovic, EminaSkoglund, Martin
By organisation
Automatic ControlFaculty of Science & Engineering
Otorhinolaryngology

Search outside of DiVA

GoogleGoogle Scholar

doi
pubmed
isbn
urn-nbn

Altmetric score

doi
pubmed
isbn
urn-nbn
Total: 227 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf