Morphing into the Promised Land

by Kyle Paice
5 Minute Read

Guest post by Jost Zetzsche, originally published in Issue 16–12–268 of The Tool Box Journal.


Some of you know that I’ve been very interested in morphology. No, let me put that differently: I’ve been very frustrated that the translation environment tools we use don’t offer morphology. There are some exceptions — such as SmartCat, Star Transit, Across, and OmegaT — that offer some morphology support. But all of them are limited to a small number of languages, and any effort to expand these would require painful and manual coding.

Other tools, such as memoQ, have decided that they’re better off with fuzzy recognition than specific morphological language rules, but that clearly is not the best possible answer either.

So, what is the problem? And what is morphology in translation environment tools about in the first place?

Well, wouldn’t it be nice to have all inflected forms of any given word in your source text be automatically associated with the uninflected form that is located in your termbase or glossary and have that displayed in your terminology search results? And does it feel a little silly to even have to ask that question at a point when it should be a no-brainer to have any given tool provide that service? In case you wondered: The answer to both questions is “Yes, yes, resoundingly yes!”

On the other hand, there is a reason why we’re stuck where we are. It happens to be cost. If you really have to manually enter morphology rules for all languages, it quickly becomes a Sisyphean exercise (starting with: “What exactly are all languages?”). If you do it just for the “important” (which in the eyes of the technology vendors means “profitable”) languages, you end up with the situation we already have with the tools mentioned above.

A few years ago, a group of folks including myself had the idea to crowdsource the collection of morphology rules for and with each language-specific group of translators. Once the rules were collected, they could then be integrated into the various technologies. It sounded good, but it was hard to get the project started due to a lack of funds to build the necessary infrastructure and/or the time it would have taken to raise funds, among other issues.

Enter translation environment tool Lilt with a very cool proposal that may very well be the solution. Lilt’s latest version introduces a “neural morphology” engine for all presently supported languages minus Chinese (so: EN, DA, NL, FR, DE, IT, NO, PO, PT, RU, ES, SV).

Here is the honest truth, though: When I first read the press release a couple of weeks ago, I fondly rolled my eyes and thought to myself that the folks from Lilt were just thinking it was wise to throw a little “neural” around while it’s hot.

It turns out I was mistaken, however, as I found out when I talked with Lilt’s John DeNero, who is the architect of this part of Lilt’s system. John tried to explain to me what the system does and why it can make a big difference. It was not so hard to understand the second part, but my feeble untechnical mind had a hard time with the first part.

(By the way, we always assume that it’s us, the less-technically-inclined, who are to be pitied when we don’t understand technology. But can you imagine how pitiful life is for the more-technically-inclined who have to speak baby talk when communicating to us?)

This article provides a good summary of the system, which essentially analyzes large monolingual corpora, detects morphological modifications (in theory, they could be any kind of modification; in practice, Lilt focuses on suffixes right now), and classifies them. Since any word is evaluated and also classified within a context, the system is able to distinguish between the adverbial ending -ly in English when it encounters “gladly” vs. “only.” Using the same contextual analysis, the system is also able to make very educated guesses about the morphological transformation of unknown words. (For instance, it might never have encountered “loquacious,” but chances are it would assume — correctly — that the adverbial transformation would be “loquaciously”).

This works with every language (that uses morphology — therefore excluding Chinese, for instance), provided there is enough corpus material to train the system. The time it takes for a new language to be trained is about 2.5 days (on very powerful computers). That’s it.

Now, it’s not perfect (whatever is??). John was very open in his assessment about where the system fails. It tends to fail with irregular morphology (it might not recognize “geese” as the plural of “goose” or “well” as the adverbial form of “good”), and there are about 5% of all cases where John felt that the engine should have made a correct judgment and it did not.

On the other hand, terminology hits have increased by a third for its users since Lilt introduced the system two weeks ago.

I consider this a quantum leap — in particular because it will not only benefit the large European and Asian languages (where applicable) but the long tail-end of other languages as well. Well, you might say, Lilt covers only a handful of languages, so doesn’t that end up being the same thing? The answer to that is (a two-fold) no. First of all, you can expect Lilt to continue to add languages, and — even more importantly — the module used to build these neural morphology engines is open-source and available for every translation technology developer right here.

This is what John said about the available engine and its usability:

“Here’s our open-source release of the morphology system. It’s released as an academic project and does not have any formal support, so it’s not a product. If someone wanted to use it, they’d have to figure it out on their own (though of course I’m happy to answer questions).”

So, get on it, Kilgray and SDL and Atril and Wordfast and, and, and . . . .

It’s also very promising that there are other areas where morphological knowledge can be used by a translation system: How about actively changing the inflection of a term that is automatically inserted based on its usage in the source? Or how about changing that inflection when repairing fuzzy matches? Or when repairing machine translation suggestions?

The sky’s the limit with this. Be creative!


FREE THE TRANSLATORS! How Adaptive MT turns post-editing janitors into cultural consultants

4 Minute Read

Originally posted on LinkedIn by Greg Rosner. I saw the phrase “linguistic janitorial work” in this Deloitte whitepaper on “AI-augmented government, using cognitive technologies to redesign public sector work”, used to describe the drudgery of translation work that so many translators are required to do today through Post-editing of Machine Translation. And then it hit me what’s really going on. The sad reality over the past several years is that many professional linguists, who have decades of particular industry experience, expertise in professional translation and have earned degrees in writing, whose jobs have been reduced to sentence-by-sentence clean-up of translations that flood out of Google Translate or other Machine Translation (MT) systems.

Read More
pexels-photo-373543 (1)

Here's Why Neural Machine Translation is a Huge Leap Forward

4 Minute Read

  Though machine translation has been around for decades, the most you’ll read about it is the perceived proximity to the mythical “Babel Fish” --an instantaneous personal translation device-- itself ready to replace each and every human translator. The part that gets left out is machine translation's relationship with human translators. For a long time, this relationship was no more complex than post-editing badly translated text, a process most translators find to be a tiresome chore. With the advent of neural machine translation, however, machine translation is not just something that creates more tedious work for translators. It is now a partner to them, making them faster and their output more accurate.

Read More