The first SASV challenge has received 23 valid submissions. This page contains the ranks, system description, and the correlation between metrics.
Below table presents the results sorted by SASV-EER(%) on the evaluation set.
Baseline1_v2
is an improved version of the Baseline1 from the evaluation plan with additional softmax applied to the CM output. Details will be addressed in the challenge summary paper.Rank | Team Name | SASV-EER | SV-EER | SPF-EER |
---|---|---|---|---|
1 | IDVoice | 0.13 | 0.11 | 0.17 |
2 | DKU-OPPO | 0.21 | 0.46 | 0.19 |
3 | HYU | 0.28 | 0.28 | 0.28 |
4 | DoubleRoc | 0.37 | 0.45 | 0.26 |
5 | FlySpeech | 0.56 | 0.86 | 0.35 |
5 | IRLAB | 0.56 | 0.73 | 0.43 |
7 | VicomSpeech | 0.84 | 0.97 | 0.58 |
8 | CUHK-NTU | 0.89 | 1.06 | 0.73 |
9 | MARG | 1.15 | 1.42 | 0.76 |
10 | NII_TJU | 1.19 | 1.32 | 1.14 |
11 | UR-AIR | 1.34 | 1.70 | 1.08 |
12 | Clips | 1.36 | 1.75 | 0.76 |
13 | Orange_Lium | 1.47 | 2.44 | 0.56 |
Baseline1-v2 | 1.71 | 1.67 | 1.77 | |
14 | VTCC | 1.86 | 2.42 | 0.97 |
15 | DeepASV | 2.48 | 3.80 | 0.65 |
16 | SHELEZYAKA | 2.77 | 3.54 | 0.54 |
17 | Xmuspeech | 2.89 | 4.28 | 0.89 |
18 | HCCL | 4.30 | 5.62 | 1.21 |
19 | Magnum | 4.48 | 7.10 | 1.08 |
20 | CAU-KU | 4.95 | 9.05 | 0.28 |
21 | Tandem | 6.22 | 11.92 | 0.77 |
Baseline2 | 6.54 | 11.99 | 0.78 | |
22 | DHU | 12.48 | 19.14 | 7.00 |
Baseline1 | 19.31 | 35.31 | 0.67 | |
23 | Souvik | 24.32 | 48.92 | 0.71 |
Below are system descriptions from each team sorted in alphabetical order.