Deprecated: The each() function is deprecated. This message will be suppressed on further calls in /home/zhenxiangba/zhenxiangba.com/public_html/phproxy-improved-master/index.php on line 456
VoxCeleb Speaker Recognition Challenge
[go: Go Back, main page]

Audio Wave

The VoxCeleb Speaker Recognition Challenge 2019
(VoxSRC-19)

Welcome to the first VoxCeleb speaker recognition challenge! The goal of this challenge is to probe how well current methods can recognize speakers from speech obtained 'in the wild'. The dataset is obtained from YouTube videos of celebrity interviews, consisting of audio from both professionally edited and red carpet interviews as well as more casual conversational audio in which background noise, laughter, and other artefacts are observed in a range of recording environments.


The 2019 challenge is now over. Congratulations to all the winners! Details about the 2020 challenge can be found here.

A report summarising the challenge baselines, results and discussions can be found below. If you would like to reference the challenge mechanics or your performance in the challenge, please cite the report below. If you would like to cite the methods of the winners, please see if the authors have uploaded any papers describing their methods and be sure to give them the appropriate credit. Slides of the winners can be found at the workshop website.

ISCA Challenges, 2019.

Tasks

VoxSRC 2019 consists of the following two tasks:

  1. Audio only speaker verification - Fixed training data: This task requires that participants train only on the VoxCeleb2 dev dataset for which we have already released speaker verification labels. The dev dataset contains 1,092,009 utterances from 5,994 speakers.
  2. Audio only speaker verification - Open training data: For the open training condition, participants can use the VoxCeleb datasets and any other data (including that which is not publicly released) except the challenge's test data

Fixed and open training conditions allow cross-system comparisons and an understanding of the effect of additional and unconstrained amounts of training data on system performance.

Evaluation

We have an evaluation server to allow teams to benchmark their performance on a blind test set over the course of (and even after) the evaluation period. In both training conditions, i.e. fixed and open, the test data can be used strictly for reporting of results alone - it cannot be used in any way to train or tune systems. Teams will only be allowed a limited number of submissions to prevent overfitting to the test set.

Results will be announced at the challenge workshop, where we will invite presentations from the most exciting and novel submissions, as well as from the challenge winners. The challenge workshop will be held on the 14th of September, 2019 in conjunction with Interspeech 2019 in Austria.

Further details about the evaluation can be found at the competition site of each tasks. See the registration section or the use the direct links below:

  1. Evaluation - Audio only speaker verification - Fixed training data
  2. Evaluation - Audio only speaker verification - Open training data

Data

The fixed training condition requires that participants train ONLY on the VoxCeleb2 development set (5994 identities only).

Dates

Early May: Release of the development kit.
Early July: Release of test data
Mid July: Evaluation server open
23th 30th August, 8 am, UTC: Deadline for submission of results
2nd September: Deadline for Technical Description. More details here.
14th September: Challenge workshop at University of Graz, Austria

Challenge Registration

The challenge will be hosted at CodaLab. Please create an account if you do not have one. We kindly ask you to associate your account to an institutional e-mail. We reserve the right to revoke your access to the competition sites otherwise, please read the Terms and Conditions of each competition. If you are part of a Team, at least one person in your team will need a CodaLab account to participate. Make sure to set the name of your Team in the user's profile, or it will not be visible on the leaderboard.

There is one competition site available for each of the tasks. Participants can register in just one competition or in both. If you do participate in both, we kindly ask you to use the same user account to participate in both tasks. The competition site for each task can be accessed at the following links:

  1. Audio only speaker verification - Fixed training data
  2. Audio only speaker verification - Open training data

If you are looking for registration on the Challenge Workshop, see our VoxSRC Workshop page. Representatives from the top 5 teams will be able to attend the VoxSRC Workshop regardless of whether they have registered or not.

Technical Description

All teams are required to submit a brief technical report describing their method. Please submit this using the latest Interspeech paper template. All reports must be a minimum of 1 page (including references) and a maximum of 4 pages (with an additional 1 page for references following the Interspeech format). Reports must be written in English. For the deadline see the dates.

We will e-mail the teams via CodaLab with the information about the submission process.

Organisers

Arsha Nagrani, VGG, University of Oxford,
Joon Son Chung, Naver, South Korea,
Andrew Zisserman, VGG, University of Oxford,
Ernesto Coto, VGG, University of Oxford,
Weidi Xie, VGG, University of Oxford,
Mitchell McLaren, Speech Technology and Research Laboratory, SRI International, CA,
Douglas A Reynolds, Lincoln Laboratory, MIT.

Please contact voxsrc[at]googlegroups.com if you have any queries, or if you would be interested in sponsoring this challenge.

Sponsors

The VoxCeleb Speaker Verification Challenge and Workshop are proudly sponsored by:

Acknowledgements

This work is supported by the EPSRC programme grant Seebibyte EP/M013774/1: Visual Search for the Era of Big Data.