liu.seSök publikationer i DiVA
Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Discriminative Learning and Target Attention for the 2019 DAVIS Challenge onVideo Object Segmentation
Linköpings universitet, Institutionen för systemteknik, Datorseende. Linköpings universitet, Tekniska fakulteten.
Linköpings universitet, Institutionen för systemteknik, Datorseende. Linköpings universitet, Tekniska fakulteten.
Linköpings universitet, Institutionen för systemteknik, Datorseende. Linköpings universitet, Tekniska fakulteten. ETH Zürich.ORCID-id: 0000-0001-6144-9520
Linköpings universitet, Institutionen för systemteknik, Datorseende. Linköpings universitet, Tekniska fakulteten.ORCID-id: 0000-0002-6096-3648
2019 (Engelska)Ingår i: CVPR 2019 workshops: DAVIS Challenge on Video Object Segmentation, 2019Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

In this work, we address the problem of semi-supervised video object segmentation, where the task is to segment a target object in every image of the video sequence, given a ground truth only in the first frame. To be successful it is crucial to robustly handle unpredictable target appearance changes and distracting objects in the background. In this work we obtain a robust and efficient representation of the target by integrating a fast and light-weight discriminative target model into a deep segmentation network. Trained during inference, the target model learns to discriminate between the local appearances of target and background image regions. Its predictions are enhanced to accurate segmentation masks in a subsequent refinement stage.To further improve the segmentation performance, we add a new module trained to generate global target attention vectors, given the input mask and image feature maps. The attention vectors add semantic information about thetarget from a previous frame to the refinement stage, complementing the predictions provided by the target appearance model. Our method is fast and requires no network fine-tuning. We achieve a combined J and F-score of 70.6 on the DAVIS 2019 test-challenge data

Ort, förlag, år, upplaga, sidor
2019.
Nyckelord [en]
video object segmentation, computer vision, machine learning
Nationell ämneskategori
Datorgrafik och datorseende
Identifikatorer
URN: urn:nbn:se:liu:diva-163334OAI: oai:DiVA.org:liu-163334DiVA, id: diva2:1390580
Konferens
The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
Tillgänglig från: 2020-02-01 Skapad: 2020-02-01 Senast uppdaterad: 2025-02-07

Open Access i DiVA

Fulltext saknas i DiVA

Person

Robinson, AndreasJäremo-Lawin, FelixDanelljan, MartinFelsberg, Michael

Sök vidare i DiVA

Av författaren/redaktören
Robinson, AndreasJäremo-Lawin, FelixDanelljan, MartinFelsberg, Michael
Av organisationen
DatorseendeTekniska fakulteten
Datorgrafik och datorseende

Sök vidare utanför DiVA

GoogleGoogle Scholar

urn-nbn

Altmetricpoäng

urn-nbn
Totalt: 285 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf