Neural networks advances improve machine translation

by Lindsay Jolivet News 29.04.2015

Breakthroughs in machine translation using neural networks have improved our ability to translate words and sentences between many languages. Researchers say their techniques will soon surpass tools such as Google Translate.

Traditional machine translation trains computers with paired examples of text in both languages, keeping track of how a word or a short phrase in one language gets translated into another. These shorter pieces can then be combined into overall sentences. This approach works reasonably well given enough training pairs, but it ignores any underlying sense of meaning.

Researchers have been improving upon this traditional technique for about two decades, but recently they’ve taken on a new approach using recurrent neural networks.

The technique uses an encoder and a decoder. The encoder reads a word, phrase or sentence and computes it to a set of numbers that represents the meaning. Two sentences that are different but mean the same thing would be represented in almost the same way, and sentences in two different languages with the same meaning would also be represented by a very similar set of numbers.

“You could imagine some kind of intermediate lingua franca, but it’s not in words, it’s in vectors,” says CIFAR Senior Fellow Yoshua Bengio (Université de Montréal).

Given a meaning representation, the decoder takes these numbers and produces a sequence of words in another language. If you are translating from English to French, the system uses the English encoder to convert to the numerical common language, then the French decoder to produce the translation.

In parallel with colleagues at Google also working on recurrent networks for machine translation, Bengio and his colleagues have recently produced a series of papers on improvements to this technique, including a recent breakthrough that overcame a problem with translating long sentences by incorporating an ability to pay attention to one element at a time.

“Let’s say I ask you to translate a whole page from one language to another one, you wouldn’t try to understand the whole page in your head and then try to spit out the translation,” he says. “After having read the original French text, as you produce translated words in English, you would go over appropriate segments of the French text to guide your translation as you do it.”

Bengio and his team have also collaborated with their CIFAR Senior Fellow Richard Zemel (University of Toronto) to incorporate attention into a model for generating image captions from scratch.

Improving translation allows scientists to study one of the big challenges of artificial intelligence — natural language understanding. “It’s a task that requires semantic understanding,” Bengio says. “It needs to really understand what the person’s intent was.”

It is also a way to advance the applications of neural networks, which have already vastly improved image and speech recognition. Bengio says the successes of their models contradict a long-held view in the AI community that they could never work to handle linguistic meaning.

“A few decades ago, people in AI thought that it would be ridiculous to think that neural networks would be able to do a good job at something like machine translation,” he says. In fact, their models are at par with the best existing methods and improving rapidly.

“It’s almost sure that within a year we’ll be substantially above the state of the art.”

Related Ideas

Research Brief | Quantum Materials

Weyl magnons predicted in quantum magnets

Focus of Study In this study, CIFAR Fellows Yong Baek Kim and Leon Balents of CIFAR’s program in Quantum Materials...

News | Institutions, Organizations & Growth

Review: Daron Acemoglu on Robotics, AI and the Future of Work

As society looks to the future and the potential of AI and robotics, CIFAR Senior Fellow and MIT economist Daron...

Feature | News | Azrieli Program in Brain, Mind & Consciousness | CIFAR Azrieli Global Scholars

EEG in the Wild

What if we could track our brain activity the same way a smart watch tracks our heart rate? But instead...

Reach Magazine | Quantum Information Science

Quantum Insecurity

The ability to send gigabytes of data around the planet in the blink of an eye has transformed our world....

News | Quantum Information Science

Physicists discover a way to increase the resolution of microscopes and telescopes

Physicists have found a way to increase the resolution of microscopes and telescopes beyond long-accepted limitations by tapping into previously...