Guest Editorial Introduction to the Special Section on Transformer Models in VisionShow others and affiliations
2023 (English)In: IEEE Transactions on Pattern Analysis and Machine Intelligence, ISSN 0162-8828, E-ISSN 1939-3539, Vol. 45, no 11, p. 12721-12725Article in journal, Editorial material (Other academic) Published
Abstract [en]
Transformer models have achieved outstanding results on a variety of language tasks, such as text classification, ma- chine translation, and question answering. This success in the field of Natural Language Processing (NLP) has sparked interest in the computer vision community to apply these models to vision and multi-modal learning tasks. However, visual data has a unique structure, requiring the need to rethink network designs and training methods. As a result, Transformer models and their variations have been suc- cessfully used for image recognition, object detection, seg- mentation, image super-resolution, video understanding, image generation, text-image synthesis, and visual question answering, among other applications.
Place, publisher, year, edition, pages
IEEE COMPUTER SOC , 2023. Vol. 45, no 11, p. 12721-12725
Keywords [en]
Special issues and sections; Transformers; Text categorization; Machine translation; Natural language processing
National Category
Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:liu:diva-199243DOI: 10.1109/TPAMI.2023.3306164ISI: 001085050900001OAI: oai:DiVA.org:liu-199243DiVA, id: diva2:1813849
2023-11-222023-11-222025-02-07