Lilt Labs | Bridging Translation Research and Practice

Here's Why Neural Machine Translation is a Huge Leap Forward

Written by Chris Healy | August, 2, 2018

 

Though machine translation has been around for decades, the most you’ll read about it is the perceived proximity to the mythical “Babel Fish” --an instantaneous personal translation device-- itself ready to replace each and every human translator. The part that gets left out is machine translation's relationship with human translators. For a long time, this relationship was no more complex than post-editing badly translated text, a process most translators find to be a tiresome chore. With the advent of neural machine translation, however, machine translation is not just something that creates more tedious work for translators. It is now a partner to them, making them faster and their output more accurate.

 

So What’s the Big Deal?

Before we jump into the brave new translating world of tomorrow, let’s put the technology in context. Prior to neural machine translation, there have been two main paradigms in the history of the field. The first was rules-based machine translation (RBMT) and the second, dominant until very recently, was phrase-based statistical machine translation (SMT).

When building rules-based machine translation systems, linguists and computer scientists joined forces to write thousands of rules for translating text from one language to another. This was good enough for monolingual reviewers to be able to get the general idea of important documents in an otherwise unmanageable body of content in a language they couldn’t read. But for the purposes of actually creating good translations, this approach has obvious flaws: it’s time consuming and, naturally, results in low quality translations.

Phrase-based SMT, on the other hand, looks at a large body of bilingual text and creates a statistical model of probable translations. The trouble with SMT is its reliance on systems. For instance, it is unable to associate synonyms or derivatives of a single word, requiring the use of a supplemental system responsible for morphology. It also requires a language model to ensure fluency, but this is limited to a given word's immediate surroundings. SMT is therefore prone to grammatical errors, and relatively inflexible when it encounters phrases that are different from those included in its training data.

Finally, here we are at the advent of neural machine translation. Virtually all NMT systems use what is known as "attentional encoder-decoder" architecture. The system has two main neural networks, one that receives a sentence (the encoder) and transforms it into a series of coordinates, or “vectors”. A decoder neural network then gets to work transforming those vectors back into text in another language, with an attention mechanism sitting in between, helping the decoder network focus on the important parts of the encoder output.

The effect of this encoding is that an NMT system learns the similarity between words and phrases, grouping them together in space, whereas an SMT system just sees a bunch of unrelated words that are more or less likely to be present in a translation.

Interestingly, this architecture is what makes Google's "zero-shot translation" possible. A well-trained multilingual NMT can decode the same encoded vector into different languages it knows, regardless of whether that particular source/target language combination was used in training.

As the decoder makes its way through the translation, it predicts words based on the entire sentence up to that point, which means it produces entire coherent sentences, unlike SMT. Unfortunately, this also means that any flaws appearing early in the sentence tend to snowball, dragging down the quality of the result. Some NMT models also struggle with words it doesn't know, which tend to be rare words or proper nouns.

Despite its flaws, NMT represents a huge improvement in MT quality, and the flaws it does have happen to present opportunities.

 

Translators and Machine Translation: Together at Last

While improvements to MT typically mean increases in its usual applications (i.e. post-editing, automatic translation), the real winner with NMT is translators. This is particularly true when a translator is able to use it in real time as they translate, as opposed to post-editing MT output. When the translator actively works with an NMT engine to create a translation, they are able to build and learn from each other, the engine offering up a translation the human may not have considered, and the human serving as a moderator, and in so doing, a teacher of the engine.

For example, during the translation process, when the translator corrects the beginning of a sentence, it improves the system’s chances getting the rest of the translation right. Often all it takes is a nudge at the beginning of a sentence to fix the rest, and the snowball of mistakes unravels.

Meanwhile, NMT's characteristic improvements in grammar and coherence mean that when it reaches a correct translation, the translator spends less time fixing grammar, beating MT output and skipping post-editing all together. When they have the opportunity to work together, translators and their NMT engines quite literally finish each other's sentences. Besides speeding up the process, and here I’m speaking as a translator, it's honestly a rewarding experience.

 

Where Do We Go Now?

Predicting the future is always a risky business, but provided the quality and accessibility of NMT continues to improve, it will gradually come to be an indispensable part of a translator's toolbox, just as CAT tools and translation memory already have.

A lot of current research has to do with getting better data, and with building systems that need less data. Both of these areas will continue to improve MT quality and accelerate its usefulness to translators. Hopefully this usefulness will also reach more languages, especially ones with less data available for training. Once that happens, translators in those languages could get through more and more text, gradually improving the availability of quality text both for the public and for further MT training, in turn allowing those translators, having already built the groundwork, to move on to bigger challenges.

When done right, NMT has the potential to not just improve translators' jobs, but to move the entire translation industry closer to its goal of being humanity's Babel Fish. Not found in an app, or in an earbud, but in networks of people.