Definition
The semi-supervised scenario assumes the user inputs a full mask of the object of interest in the first frame of a video sequence. Methods have to produce the segmentation mask for that object in the subsequent frames.More information in the DAVIS 2017 publication and the Codalab submission site.
Dates and Phases
- Test-Dev: Unlimited number of submissions, open end.- Test-Challenge: Limited to 5 submissions in total from 3rd May 2020 23:59 UTC to 15th May 2020 23:59 UTC.
Prizes
All the participants invited to the workshop will get a subscription to Adobe CC for 1 year.Submission
Submissions to all phases are done through the new Codalab site of the challenge.Please register to the site and read carefully the instructions in the Learn the Details -> Overview and Evaluation sections.
Datasets (Download here, 480p resolution)
- Train + Val: 90 sequences from DAVIS 2017.- Test-Dev 2017: 30 sequences from DAVIS 2017 Semi-supervised. Ground truth not publicly available.
- Test-Challenge 2017: 30 sequences from DAVIS 2017 Semi-supervised. Ground truth not publicly available.
Feel free to train or pre-train your algorithms on any other dataset apart from DAVIS (Youtube-VOS, MS COCO, Pascal, etc.) or use the full resolution DAVIS annotations and images.
Evaluation
- The per-object measures are those described in the original DAVIS CVPR 2016 paper: Region Jaccard (J) and Boundary F measure (F).- The overall ranking measures are computed as the mean between J and F, both averaged over all objects. Precise definitions available in the papers DAVIS 2016 and DAVIS 2017 publications.
Papers
- Right after the Challenge closes (15th May) we will invite participants to submit a short abstract (400 words maximum) of their method (Deadline 19th May 23:59 UTC).- Together with the results obtained, we will decide which teams are accepted at the workshop. Date of notification 20th May.
- Accepted teams will be able to submit a paper describing their approach (Deadline 4th June 23:59 UTC). The template of the paper is the same as CVPR, but length will be limited to 4 pages including references.
- Papers will also be invited to the workshop in form of oral presentation or poster.
- Accepted papers will be self-published in the web of the challenge (not in the official proceedings, although they have the same value).
Other considerations
- Each entry must be associated to a team and provide its affiliation.- The best entry of each team will be public in the leaderboard at all times.
- We will only consider the "semi-supervised" scenario: the mask of the first frame is given, no human interaction/refinement allowed. Even though the class of each object was recently released, it should not be used as additional information in the methods. Although we have no way to check the latter at the challenge stage, we will make our best to detect it a posteriori before the workshop.
- We reserve the right to remove one of the entry methods to the competition when there is a high technical similarity to methods published in previous conferences or workshops. We do so in order to keep the workshop interesting and to push state of the art to move forward.
- The new annotations in this dataset belong to the organizers of the challenge and are licensed under a Creative Commons Attribution 4.0 License.
Citation
The 2017 DAVIS Challenge on Video Object Segmentation
J. Pont-Tuset,
F. Perazzi,
S. Caelles,
P. Arbeláez,
A. Sorkine-Hornung, and
L. Van Gool
arXiv:1704.00675, 2017
[PDF] [BibTex]
@article{Pont-Tuset_arXiv_2017, author = {Jordi Pont-Tuset and Federico Perazzi and Sergi Caelles and Pablo Arbel\'aez and Alexander Sorkine-Hornung and Luc {Van Gool}}, title = {The 2017 DAVIS Challenge on Video Object Segmentation}, journal = {arXiv:1704.00675}, year = {2017} }
A Benchmark Dataset and Evaluation Methodology for Video Object Segmentation
F. Perazzi,
J. Pont-Tuset,
B. McWilliams,
L. Van Gool,
M. Gross, and
A. Sorkine-Hornung
Computer Vision and Pattern Recognition (CVPR) 2016
[PDF] [Supplemental] [BibTex]
@inproceedings{Perazzi2016,
author = {F. Perazzi and J. Pont-Tuset and B. McWilliams and L. {Van Gool} and M. Gross and A. Sorkine-Hornung},
title = {A Benchmark Dataset and Evaluation Methodology for Video Object Segmentation},
booktitle = {Computer Vision and Pattern Recognition},
year = {2016}
}