Dataset Viewer
Auto-converted to Parquet
utterance_id
stringclasses
189 values
language
stringclasses
2 values
system_A
stringclasses
3 values
system_B
stringclasses
3 values
user
stringclasses
7 values
pronunciation
stringclasses
3 values
naturalness
stringclasses
3 values
sound_quality
stringclasses
3 values
emotion_similarity
stringclasses
3 values
voice_similarity
stringclasses
3 values
7tY36H7eY58_pt/M1__3.420-4.747
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
B
B
A
A
JZB_ti6Ctlc_it/M0__18.890-19.810
en
dubformer_et
minimax_02hd
annotator_0
SAME
SAME
A
SAME
SAME
-d9C5FyXvvw_it/M2__10.259-11.942
en
11labs_v2
dubformer_et
annotator_0
B
SAME
B
A
A
w35s_xMyHIc_pt/F0__12.090-15.040
en
dubformer_et
minimax_02hd
annotator_0
A
A
SAME
B
B
OesbsUiwgo0_pt/F0__7.550-8.920
en
11labs_v2
dubformer_et
annotator_0
SAME
A
B
A
A
d_TKW3INVKs_it/M1__6.500-8.130
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
A
B
B
SAME
UXx6zn9y2B8_es/F0__16.050-20.460
en
dubformer_et
minimax_02hd
annotator_0
A
A
SAME
SAME
B
jdSyxrG6dfM_it/M0__8.400-9.490
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
B
SAME
SAME
HjxbQx8odLA_zh/F2__1.678-3.741
en
11labs_v2
dubformer_et
annotator_0
SAME
A
A
B
SAME
GLA2YCQi_Rk_fr/F2__9.709-10.987
en
11labs_v2
deepdub_etts-2.5
annotator_0
B
SAME
A
SAME
A
b91pBJWJDhQ_ru/M0__3.240-5.330
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
A
B
B
SdpDal_Tu2s_hi/F0__12.280-13.400
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
A
SAME
B
7EqatsqEI-0_ru/F0__4.153-5.454
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
SAME
B
B
h8Ht1PqPYMw_ru/F0__0.040-3.450
en
dubformer_et
minimax_02hd
annotator_0
A
SAME
B
B
B
-d9C5FyXvvw_it/M2__10.259-11.942
en
11labs_v2
deepdub_etts-2.5
annotator_0
SAME
A
A
SAME
SAME
GLA2YCQi_Rk_fr/F2__9.709-10.987
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
A
B
B
B
DssUtj_qKf4_ru/M0__1.602-3.699
en
11labs_v2
deepdub_etts-2.5
annotator_0
SAME
A
SAME
B
SAME
Ll6fcDRKi9k_ru/F0__0.040-1.410
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
A
SAME
SAME
cJvDUasu9pI_hi/M1__11.260-14.440
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
A
B
B
BFvg-l-e5FE_es/F1__15.220-18.730
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
A
B
SAME
A
HjxbQx8odLA_zh/F2__1.678-3.741
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
SAME
SAME
B
h8Ht1PqPYMw_ru/F1__8.560-10.905
en
11labs_v2
dubformer_et
annotator_0
SAME
SAME
B
SAME
B
yFfw234lX0c_zh/M0__0.760-2.160
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
SAME
SAME
B
SAME
03VPqxrlyxA_zh/F1__8.234-8.922
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
A
B
B
GLA2YCQi_Rk_fr/F2__9.709-10.987
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
B
B
B
SAME
l-l9x4bLVUY_zh/M0__2.355-3.949
en
deepdub_etts-2.5
minimax_02hd
annotator_0
B
SAME
B
B
B
h8Ht1PqPYMw_ru/F1__8.560-10.905
en
11labs_v2
minimax_02hd
annotator_0
A
A
B
B
B
x_gaj_ZhDqs_de/M0__6.300-8.870
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
SAME
SAME
A
A
b91pBJWJDhQ_ru/M0__9.240-10.650
en
dubformer_et
minimax_02hd
annotator_0
A
A
A
SAME
B
eN-waporon0_de/M0__7.005-7.805
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
A
SAME
B
x_gaj_ZhDqs_de/F0__2.890-5.220
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
A
A
A
t_GIRgw8uGo_de/F0__4.065-8.425
en
11labs_v2
dubformer_et
annotator_0
B
SAME
B
SAME
A
Mzp-y0qfgYQ_pt/M0__4.920-9.250
en
dubformer_et
minimax_02hd
annotator_0
A
SAME
A
B
B
Ll6fcDRKi9k_ru/F1__1.760-4.218
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
A
SAME
B
SdpDal_Tu2s_hi/F0__12.280-13.400
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
A
B
B
SAME
qJaAnEUiO6E_it/M0__10.975-13.170
en
11labs_v2
dubformer_et
annotator_0
B
SAME
A
SAME
A
_5MiAdlrqgA_pt/F0__6.240-8.890
en
deepdub_etts-2.5
dubformer_et
annotator_0
SAME
SAME
B
SAME
A
DGbTKywfSxw_ru/F0__6.320-8.405
en
11labs_v2
dubformer_et
annotator_0
SAME
SAME
B
SAME
A
_W3R1g9-ByQ_ru/F0__25.165-28.210
en
deepdub_etts-2.5
minimax_02hd
annotator_0
SAME
SAME
SAME
B
B
Ll6fcDRKi9k_ru/F1__1.760-4.218
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
A
SAME
SAME
B
BFvg-l-e5FE_es/F1__15.220-18.730
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
SAME
A
SAME
B
SdpDal_Tu2s_hi/F1__8.720-12.020
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
B
B
A
A
-d9C5FyXvvw_it/M0__3.153-4.752
en
deepdub_etts-2.5
minimax_02hd
annotator_0
SAME
SAME
A
A
B
lTa1VcbYZEE_fr/F0__6.880-7.950
en
dubformer_et
minimax_02hd
annotator_0
A
SAME
SAME
SAME
B
KO8jCbxlb9U_fr/M0__0.220-2.811
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
A
A
A
J9RB0FjZwyM_zh/M0__19.699-22.900
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
A
SAME
SAME
SAME
w35s_xMyHIc_pt/M0__3.401-5.854
en
dubformer_et
minimax_02hd
annotator_0
A
B
A
B
B
x_gaj_ZhDqs_de/M0__6.300-8.870
en
11labs_v2
deepdub_etts-2.5
annotator_0
SAME
A
SAME
B
B
BFvg-l-e5FE_es/F1__9.730-11.110
en
11labs_v2
dubformer_et
annotator_0
SAME
B
B
SAME
A
KO8jCbxlb9U_fr/M1__10.675-14.225
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
SAME
A
A
SAME
aw-tFjCLrxM_pt/M0__0.055-7.509
en
11labs_v2
dubformer_et
annotator_0
SAME
A
A
SAME
SAME
fVrknIFJBOA_ja/M0__0.176-10.420
en
11labs_v2
dubformer_et
annotator_0
SAME
SAME
B
A
SAME
CDxg_6317fA_ja/M1__4.346-5.698
en
dubformer_et
minimax_02hd
annotator_0
A
SAME
A
SAME
SAME
03VPqxrlyxA_zh/F1__8.234-8.922
en
11labs_v2
dubformer_et
annotator_0
SAME
SAME
A
SAME
A
uTbCKWoc1YY_zh/F0__18.920-19.913
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
SAME
SAME
B
eN-waporon0_de/M0__7.005-7.805
en
dubformer_et
minimax_02hd
annotator_0
A
SAME
A
SAME
B
U_1cQLJKt7c_es/F1__5.425-7.235
en
11labs_v2
dubformer_et
annotator_0
A
SAME
SAME
B
B
KO8jCbxlb9U_fr/M0__0.220-2.811
en
11labs_v2
dubformer_et
annotator_0
B
SAME
SAME
SAME
A
DGbTKywfSxw_ru/F0__10.360-12.490
en
dubformer_et
minimax_02hd
annotator_0
A
A
A
SAME
B
cyLLcnTzV5w_pt/F0__6.600-8.550
en
deepdub_etts-2.5
dubformer_et
annotator_0
SAME
SAME
SAME
SAME
SAME
_5MiAdlrqgA_pt/F0__6.240-8.890
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
SAME
A
SAME
B
l-l9x4bLVUY_zh/M0__2.355-3.949
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
SAME
B
SAME
SAME
GLA2YCQi_Rk_fr/F2__9.709-10.987
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
B
SAME
SAME
aw-tFjCLrxM_pt/M0__0.055-7.509
en
dubformer_et
minimax_02hd
annotator_0
A
SAME
SAME
B
B
cdExO_qxpLI_hi/F1__17.820-25.301
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
B
B
SAME
A
Ll6fcDRKi9k_ru/F0__0.040-1.410
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
SAME
B
A
A
Muhje0yQzFE_pt/F0__9.900-13.100
en
11labs_v2
dubformer_et
annotator_0
B
SAME
B
SAME
A
TxbnBLoS8Ic_zh/F0__3.177-4.051
en
11labs_v2
minimax_02hd
annotator_0
A
A
SAME
A
A
daPO281YkZI_de/M1__1.780-5.250
en
dubformer_et
minimax_02hd
annotator_0
A
B
SAME
SAME
B
Ljaicg38ZBs_fr/M0__3.190-5.560
en
deepdub_etts-2.5
minimax_02hd
annotator_0
SAME
B
B
SAME
SAME
lTa1VcbYZEE_fr/F0__0.330-4.550
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
SAME
A
B
B
SQV7gz091-0_es/M0__0.180-2.485
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
A
B
A
A
CDxg_6317fA_ja/M0__13.440-14.602
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
A
A
A
P6glOSnV2gM_fr/F0__21.290-23.575
en
11labs_v2
dubformer_et
annotator_0
A
SAME
SAME
SAME
SAME
JZB_ti6Ctlc_it/M0__18.890-19.810
en
11labs_v2
dubformer_et
annotator_0
SAME
A
SAME
SAME
A
w35s_xMyHIc_pt/F0__12.090-15.040
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
SAME
B
A
SAME
cdExO_qxpLI_hi/F1__17.820-25.301
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
SAME
A
SAME
B
x_gaj_ZhDqs_de/F0__13.615-15.695
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
A
SAME
SAME
B
lTa1VcbYZEE_fr/F0__0.330-4.550
en
11labs_v2
dubformer_et
annotator_0
B
A
SAME
SAME
A
lTa1VcbYZEE_fr/M0__14.560-18.030
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
B
A
SAME
A
SdpDal_Tu2s_hi/F0__16.420-19.320
en
deepdub_etts-2.5
dubformer_et
annotator_0
A
SAME
B
A
A
KO8jCbxlb9U_fr/M1__10.675-14.225
en
11labs_v2
dubformer_et
annotator_0
SAME
SAME
B
SAME
A
_W3R1g9-ByQ_ru/M0__2.750-5.130
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
SAME
A
SAME
A
TxbnBLoS8Ic_zh/F0__3.177-4.051
en
dubformer_et
minimax_02hd
annotator_0
SAME
SAME
A
A
B
vJEDLWmL_Bg_de/F0__8.580-11.640
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
SAME
B
SAME
A
J9RB0FjZwyM_zh/M0__18.194-19.234
en
deepdub_etts-2.5
minimax_02hd
annotator_0
B
B
SAME
B
SAME
2iK4DdnoL-s_pt/F0__10.600-14.080
en
deepdub_etts-2.5
minimax_02hd
annotator_0
A
SAME
B
SAME
B
rlR9M9QZ4Cg_de/M0__9.245-11.088
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
A
A
SAME
B
03VPqxrlyxA_zh/M0__4.525-5.819
en
11labs_v2
deepdub_etts-2.5
annotator_0
B
SAME
SAME
B
B
b91pBJWJDhQ_ru/F1__12.640-13.888
en
11labs_v2
dubformer_et
annotator_0
SAME
SAME
SAME
B
B
7EqatsqEI-0_ru/M0__10.415-13.964
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
SAME
SAME
SAME
B
daPO281YkZI_de/M1__1.780-5.250
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
SAME
B
SAME
DssUtj_qKf4_ru/F0__17.632-18.770
en
dubformer_et
minimax_02hd
annotator_0
A
SAME
SAME
SAME
B
Q-jR0549ZkA_de/M0__0.190-10.100
en
11labs_v2
dubformer_et
annotator_0
B
SAME
SAME
SAME
A
7jetYNtYDQw_ru/F0__16.195-18.690
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
B
SAME
SAME
Xr6N1oSKWA0_it/M0__23.485-26.810
en
11labs_v2
minimax_02hd
annotator_0
A
SAME
A
B
SAME
jdSyxrG6dfM_it/F0__10.803-11.485
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
B
SAME
B
B
7EqatsqEI-0_ru/F0__4.153-5.454
en
11labs_v2
deepdub_etts-2.5
annotator_0
A
SAME
A
A
SAME
DssUtj_qKf4_ru/M0__1.602-3.699
en
deepdub_etts-2.5
dubformer_et
annotator_0
B
B
SAME
A
SAME
SdpDal_Tu2s_hi/F0__12.280-13.400
en
11labs_v2
dubformer_et
annotator_0
SAME
A
SAME
SAME
A
End of preview. Expand in Data Studio

VOX-DUB is a human-based benchmark for evaluating AI dubbing systems.
It includes:

  • Audio fragments with original speech from real videos and their corresponding translated texts.
  • Generated audio recordings produced by multiple dubbing/TTS systems.
  • Human annotation results with pairwise A/B (+ SAME) evaluations across five aspects (pronunciation, naturalness, sound quality, emotion similarity, and voice similarity).
  • Detailed annotation guidelines with examples for pairwise A/B comparisons.

Systems under evaluation are expected to generate the provided translations conditioned on the original speech audio. Their outputs are then compared against each other. The benchmark assesses dubbing quality across five key aspects: pronunciation, naturalness, sound quality, emotion similarity, and voice similarity.

Dataset structure

This repository exposes three datasets:

  • source_data — original utterances with translations and speaker context
  • synthesized_data — audios generated by different TTS/dubbing systems
  • annotations — human A/B (+ SAME) pairwise judgments across five aspects

Loading

from datasets import load_dataset

source_data = load_dataset("toloka/vox-dub", name="source_data")["train"]
synthesized_data = load_dataset("toloka/vox-dub", name="synthesized_data")["train"]
annotations = load_dataset("toloka/vox-dub", name="annotations")["train"]

Source data

source_data contains original speech segments and metadata for generation/evaluation.

Features

  • utterance_id (string) — unique identifier of the utterance
  • source_language (string) — original language code (ISO-639-1), e.g. "de"
  • translation_en (string) — English translation of the utterance
  • translation_es (string) — Spanish translation of the utterance
  • original_utterance_audio (Audio) — audio clip of the original spoken utterance
  • other_speaker_utterances (Sequence[Audio]) — additional utterances by the same speaker/character (useful for speaker conditioning / voice cloning)

Audio columns are Hugging Face Audio features with on-the-fly decoding.

Synthesized data

synthesized_data are system outputs for the target dubbing task.

Features

  • utterance_id (string) — links back to source_data.utterance_id
  • system (string) — name/identifier of the synthesis provider/model
  • language (string) — generation language code (ISO-639-1)
  • audio (Audio) — synthesized audio for the target line

Multiple rows can exist per utterance_id (different systems and/or languages).

Annotations

annotations are human pairwise A/B (+ SAME) annotations across five aspects.

Features

  • utterance_id (string) — evaluated utterance (ties to source_data)
  • language (string) — comparison language for this judgment
  • system_A (string) — first system in the pair
  • system_B (string) — second system in the pair
  • user (string) — anonymized annotator ID
  • pronunciation (string ∈ {"A", "B", "SAME"}) — preference on pronunciation
  • naturalness (string ∈ {"A", "B", "SAME"}) — preference on naturalness
  • sound_quality (string ∈ {"A", "B", "SAME"}) — preference on audio quality
  • emotion_similarity (string ∈ {"A", "B", "SAME"}) — preference on emotion similarity
  • voice_similarity (string ∈ {"A", "B", "SAME"}) — preference on voice similarity

Each row is one A/B comparison for a single utterance_id, language, system_A and system_B by signle user. You can aggreate labels from different annota with majority vote, Dawid–Skene, etc.


Guidelines for annotators

Each entry contains three short audio samples: an original (reference) audio and its translation, read aloud by two different speech synthesis systems (Audio A and Audio B).

Please listen to Audio A and Audio B and decide which one is better. There are five separate parameters for evaluation: pronunciation, naturalness, sound quality, emotion similarity and voice similarity.

To evaluate the audio samples on the first two parameters (pronunciation and naturalness) please only listen to Audio A and Audio B and compare them. Ignore the reference completely.

To evaluate sound quality, it is also usually enough to listen to Audio A and Audio B.

To evaluate emotion similarity and voice similarity, you need to compare both samples to the reference audio.

For each parameter, please choose which of the systems performs best: A or B, otherwise you can choose option SAME if the samples are equally bad or good.

How to make a decision: a life hack that we find useful

If both samples have faults and you are not sure which one is better, ask yourself: if I were doing a real dubbing project, and I had to choose one of these two samples to paste into my video, which one (naturalness-wise, or pronunciation-wise, or based on other parameter you are currently struggling with) would I choose? In tricky cases, when the answer is not evident, explain your decision in the comments.

Do not overthink it!

If you have listened to the samples 2-3 times and still can’t choose one of them, then they can be safely considered equal.

💬 Pronunciation

Play the samples and compare them to the text. Are all the words in the text pronounced correctly?

Aspects to pay attention to:

  • Was all the text pronounced, or is something missing?
  • Correct pronunciation of words/syllables/sounds,
  • Word stress,
  • Non-native accent or accent of a different region.
  • Phoneme reduction is not a mistake if widely used.
  • If the phrase seems to be pronounced correctly, but is difficult to make out due to reverberation, distortions and “bad microphone” effect, it is a problem of Sound quality, not Pronunciation.

⚠️ Challenging examples:

This happens every day.

Non-native accent should be penalized.

If you recommend them in the comments.

”You” and plural form “s” were lost. This mistake should be penalized.

You cannot fool a father's heart.

The voice in Audio A is clearer.
Audio B sounds muffled, but all the sounds seem to be in place. This is an issue of Sound quality, not Pronunciation

🌿 Naturalness

Does the speech in the samples sound natural? Does it sound like real human speech, or could you suspect a robot? ‼️ Ignore reference audio while assessing on this parameter.

Aspects to pay attention to:

  • Correct usage of affirmative and interrogative intonation,
  • Logical stress in the sentence,
  • Robotic, monotonous, unnatural intonation,
  • Unusually slow or fast speech,
  • Pauses within the sentence - may be natural or not,
  • Breathing, snorts or other human sounds - may occur in natural places or not.

⚠️ Challenging examples:

My king, sometimes people only want to hear good things from Paro.

Emphasis on “my” instead of “king” sounds unnatural

He is hitting on my wife.

The intonation suggests a question or an unfinished sentence
Affirmative intonation matches better

You will also return all his belongings.

Inhalation at the start sounds quite natural, no penalty for that. However, the intonation is not dropping down enough and sounds a bit robotic.

Serve me, can't you see I'm hungry?

The intonation here is flat and unnatural.
The intonation is much more natural and human-like. Audio B wins.
The reference is actually closer to Audio A, but we ignore it when judging Naturalness.

🎧 Sound quality

Please use your headphones to assess the sound quality!

To understand what we consider good sound quality, imagine that you are working on a real dubbing project. The version that you pick will be pasted in your translated video over all its background sounds. The best possible option (in terms of sound quality) is to have clean, rich, studio sound. But sometimes we are ready to accept some non-studio sound features, if they convey interesting properties of the original voice and make our dubbing more life-like.

The voice sounds differently if you hear it over the phone, if the speaker is at a distance, at the bottom of a well, in an empty room, or has a cold, or if the speaker is a ghost and talks with an eerie echo. Basically, the speaker can even be an actual robot with a very mechanic voice.

So, if the original voice has such interesting peculiarities, we:

  • Don’t penalize for them in Sound quality (consider them as good as studio quality voice)
  • Slightly encourage them in Voice similarity.

This is our general logic, but for your convenience we also wrote a set of rules that may help you if you prefer formal approach.

  1. Clear studio quality sound is always good, even if the reference audio has issues and peculiarities.
  2. These issues are always considered bad and should be penalized, because they would always make our dubbing worse:
    • Background noise
    • Claps, clicks and other non-speech artefacts.
  3. These and similar issues are usually considered bad:
    • Distorted, mechanic voice,
    • Flat sound, like over an old telephone, or from a distance,
    • Bad microphone effect: as if a person pronounced all the sounds correctly, but they were corrupted by a bad microphone,
    • Echo

If you hear these issues, please check: do they mimic an interesting feature of the original reference? Do they make our dubbing better?

  • YES, they mimic the original speaker, and I would like to hear them in my dubbing ➡️ Don’t penalize for them.
  • YES, they kind of mimic the original speaker, but I still think they would make dubbing worse ➡️ Penalize for them.
  • NO, they occur only in translation ➡️ They are bugs, penalize for them.

⚠️ Challenging examples:

But don’t tell Mom, ok? It’s our secret.

Random clap sound at the start should be penalized. You should be able to hear it with your headphones.

That’ s your doing!

This is an example of a slightly distorted voice.

Is it a piece of cake?

This sample has a slightly more clear voice than the other..
Less clear sound

And that I wouldn't be alone anymore.

Background hissing and music should be penalized
This sample wins…
…even if the same background sounds are present in the original.

Alright.

The voice is croaky, and it may seem a sound quality bug, but if we listen to the reference, we hear that it perfectly recreates the timber of the original speaker.
The two samples are equal in terms of Sound quality, but B loses in Voice similarity.
Original audio.

This happens every day.

Audio A has the same reverberation as the reference. It would sound good in dubbing, don’t penalize for that.
Audio B has reverberation as well, and we agreed not to penalize for that. However, there is a slight background hissing, that should be penalized.
The reference has some natural reverberation.

🙀 Emotion similarity

Listen to the samples and the reference audio. How well do the samples reproduce the emotion of the reference?

Aspects to pay attention to:

  • Emotion of the speaker (sad, angry, happy, calm, soft, loud)
  • Expressiveness of both audios (narration vs. spontaneous speech),
  • Interrogative intonation should be assessed by the “naturalness” parameter, not emotion similarity.

⚠️ Challenging examples:

What are you looking for?

Reference.
Rushed, agitated speech is a better match.
Calm speech is a worse match.

Am I a victim?

Reference.
Calmer, colder speech matches worse.
The difference is slight, but this audio is more emotional, and matches better

👥 Voice similarity

Listen to the samples and the reference audio. How similar are the voices in Audio A and Audio B to the reference speaker’s voice?

Aspects to pay attention to:

  • Timbre,
  • Pitch,
  • Estimated age and gender,
  • Other voice characteristics (is it far or near? is it in the same room or over the phone? does it have reverberation?). They are less important that the other aspects listed here, but they can be viewed as a bonus.

⚠️ Challenging examples:

Got a problem with that?

Reference. When we hear it, we imagine a senior woman.
This voice sounds younger.
This voice matches better .

Reference

@misc{toloka2025vox-dub,
  title        = {VOX-DUB: a new benchmark that puts AI dubbing to the test},
  author       = {{Toloka team}},
  howpublished = {\url{https://toloka.ai/blog/ai-dubbing-benchmark/}},
  year         = {2025},
  month        = sep # "~9",
  note         = {Accessed: 2025-09-10},
}
Downloads last month
374