Uncertainty Quantification for Evaluating Machine Translation Bias
By: Ieva Raminta Staliūnaitė, Julius Cheng, Andreas Vlachos
Potential Business Impact:
Makes computer translators less biased about gender.
In machine translation (MT), when the source sentence includes a lexeme whose gender is not overtly marked, but whose target-language equivalent requires gender specification, the model must infer the appropriate gender from the context and/or external knowledge. Studies have shown that MT models exhibit biased behaviour, relying on stereotypes even when they clash with contextual information. We posit that apart from confidently translating using the correct gender when it is evident from the input, models should also maintain uncertainty about the gender when it is ambiguous. Using recently proposed metrics of semantic uncertainty, we find that models with high translation and gender accuracy on unambiguous instances do not necessarily exhibit the expected level of uncertainty in ambiguous ones. Similarly, debiasing has independent effects on ambiguous and unambiguous translation instances.
Similar Papers
What Triggers my Model? Contrastive Explanations Inform Gender Choices by Translation Models
Computation and Language
Finds why computer translations get gender wrong.
Voice, Bias, and Coreference: An Interpretability Study of Gender in Speech Translation
Computation and Language
Translates speech, guessing gender from sound, not just pitch.
Addressing speaker gender bias in large scale speech translation systems
Computation and Language
Fixes translation mistakes for female speakers.