Authors
Adithya Renduchintala, Denise Diaz, Kenneth Heafield, Xian Li, Mona Diab
Publication date
2021/6/1
Journal
arXiv preprint arXiv:2106.00169
Description
Is bias amplified when neural machine translation (NMT) models are optimized for speed and evaluated on generic test sets using BLEU? We investigate architectures and techniques commonly used to speed up decoding in Transformer-based models, such as greedy search, quantization, average attention networks (AANs) and shallow decoder models and show their effect on gendered noun translation. We construct a new gender bias test set, SimpleGEN, based on gendered noun phrases in which there is a single, unambiguous, correct answer. While we find minimal overall BLEU degradation as we apply speed optimizations, we observe that gendered noun translation performance degrades at a much faster rate.
Total citations
2021202220232024618108
Scholar articles
A Renduchintala, D Diaz, K Heafield, X Li, M Diab - arXiv preprint arXiv:2106.00169, 2021