VBS2020 will be part of the International Conference on MultiMedia Modeling 2020 (MMM2020) in Daejeon, Korea, and organized as a special side event to the Welcome Reception on January 6, 2020. It will be a moderated session where participants solve Known-Item Search (KIS) and Ad-Hoc Video Search (AVS) tasks that are issued as live presentation of scenes of interest, either as a visual 20 seconds clip (randomly selected), or as a textual description. The goal is to find correct segments (for KIS exactly one segment, for AVS many segments) as fast as possible and submit the segment description (video id and frame number) to the VBS server, which evaluates if the found segment is correct.
VBS2020 will use the V3C1 dataset in collaboration with NIST, i.e. TRECVID 2019 (i.e. with the Ad-Hoc Video Search (AVS) Task) , which consists of 7475 video files, amounting for 1000h of video content (1082659 predefined segments) and 1.3 TB in size. In order to download the dataset (which is provided by NIST), please complete this data agreement form and send a scan to firstname.lastname@example.org with CC to email@example.com and firstname.lastname@example.org. You will be provided with a link for downloading the data.
AVS and KIS Tasks
We plan to test at least 20 search tasks:
- 10 AVS tasks, randomly selected in collaboration with TRECVID AVS. Each AVS task has several/many target shots that should be found.
- 10 KIS tasks, which are selected completely random on site. Each KIS task has only one single 20-seconds long target segment
VBS Server and Testing
The VBS uses an own VBS Server to evaluate found segments for correctness. Therefore, all participants need to connect to the server via a dedicated network (typically Ethernet with CAT-5) and need to submit found segments to the server via a simple HTTP-like protocol. The server is connected to a beamer and presents the current score of all teams in a live manner (in addition to presenting task descriptions). The server as well as example tasks from the previous years are provided here.
Anyone with an exploratory video search tool that allows for retrieval, interactive browsing, exploration in a video collection may participate.
There are no restrictions in terms of allowed features, except for presentation screen recording during the competition, which is disallowed. That means in addition to interactive content search you can use any automatic content search as well.
Available Analysis Data
In order to give new teams an easy entry, we provide results of content analysis to all teams. The V3C1 dataset already comes with segmentation information and includes shot boundaries as well as keyframes. Moreover, we provide resulting data from different content analysis steps (e.g., color, faces, text, detected ImageNet classes, etc.). The analysis data is available here and described in this article. Also the ASR data has been released here (many thanks to Luca Rossetto et al.)! Moreover, the SIRET team shared their shot detection network too (many thanks to Jakub Lokoc ans his team)!
To participate please submit an extended demo paper (4-6 pages in Springer LNCS format) until the deadline via the MMM 2020 Submission System (please select „MMM2020_VBS“ track). The submission should include a detailed description of the video search tool (including a screenshot of the tool) and show describe how it supports interactive search in video. Submissions will be peer-reviewed to ensure maximum quality. Accepted papers will be published in the proceedings of the MMM conference and should also be presented as poster during the VBS session.
We plan to write a joint journal paper after the VBS competition, where each participating team should contribute to. The winning team will be honored to be in charge of the journal paper (as a main author).