liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
SepTr: Separable Transformer for Audio Spectrogram Processing
Univ Politehn Bucuresti, Romania; Mohamed bin Zayed Univ Artificial Intelligence, U Arab Emirates.
Univ Bucharest, Romania.
Linköping University, Department of Electrical Engineering, Computer Vision. Linköping University, Faculty of Science & Engineering. Mohamed bin Zayed Univ Artificial Intelligence, U Arab Emirates.
2022 (English)In: INTERSPEECH 2022, ISCA-INT SPEECH COMMUNICATION ASSOC , 2022, p. 4103-4107Conference paper, Published paper (Refereed)
Abstract [en]

Following the successful application of vision transformers in multiple computer vision tasks, these models have drawn the attention of the signal processing community. This is because signals are often represented as spectrograms (e.g. through Discrete Fourier Transform) which can be directly provided as input to vision transformers. However, naively applying transformers to spectrograms is suboptimal. Since the axes represent distinct dimensions, i.e. frequency and time, we argue that a better approach is to separate the attention dedicated to each axis. To this end, we propose the Separable Transformer (SepTr), an architecture that employs two transformer blocks in a sequential manner, the first attending to tokens within the same time interval, and the second attending to tokens within the same frequency bin. We conduct experiments on three benchmark data sets, showing that our separable architecture outperforms conventional vision transformers and other state-of-the-art methods. Unlike standard transformers, SepTr linearly scales the number of trainable parameters with the input size, thus having a lower memory footprint. Our code is available as open source at https://github.com/ristea/septr.

Place, publisher, year, edition, pages
ISCA-INT SPEECH COMMUNICATION ASSOC , 2022. p. 4103-4107
Series
Interspeech, ISSN 2308-457X
Keywords [en]
separable transformer; multi-head attention; audio spectrogram processing; sound recognition
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:liu:diva-192358DOI: 10.21437/Interspeech.2022-249ISI: 000900724504057OAI: oai:DiVA.org:liu-192358DiVA, id: diva2:1744003
Conference
Interspeech Conference, Incheon, SOUTH KOREA, sep 18-22, 2022
Note

Funding Agencies|Romanian Ministry of Education and Research, CNCS -UEFISCDI within PNCDI III [PN-III-P1-1.1-TE-2019-0235]

Available from: 2023-03-16 Created: 2023-03-16 Last updated: 2023-03-16

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Search in DiVA

By author/editor
Khan, Fahad
By organisation
Computer VisionFaculty of Science & Engineering
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 25 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf