Working Around Machine Translation’s Need for Large-Scale Training Data
Unlike some fields in machine learning, machine translation still requires large sets of training data. The solution? Creating more data when none (or not enough) exists.
Unlike some fields in machine learning, machine translation still requires large sets of training data. The solution? Creating more data when none (or not enough) exists.
New paper by Facebook, Amazon, Twitter, and University of Melbourne examines vulnerability of systems using back translation to attacks on monolingual training data.
Association for Computational Linguistics names Best Research Papers of 2020 from a field of more than 3,000 submissions; paper that argues for retiring BLEU is runner-up.
Slator Weekly: Join over 15,800 subscribers and get the latest language industry intelligence every Friday
Tool Box: Join 10,000 subscribers for your monthly linguist technology update.
Your information will not be shared with third parties. No Spam.
This will close in 0 seconds