Neural networks are employed almost everywhere. They make predictions, perform complex calculations, generate articles, stories, and even create paintings. But what about coherent translations — is that possible, and how close are the processed texts to the original? And is it true that AI will soon replace human translators? Let’s find out.
Of course, as we all rely on Google.Translate already. Nowadays, algorithms can recognize speech; they perform the transliteration of image captions and can even reproduce in audio format, how a phrase sounds in French, Chinese, or even the fictional Elvish language.
It all started at the end of the 1940s, with the development of the concept of machine information processing. The Georgetown-IBM experiment in 1954 marked a milestone when a machine correctly translated 60 sentences from Russian to English and even corrected spelling errors in words.
Although the experiment’s results were later criticized — the sentences were simple and clear in meaning, so the machine had no chance to err — this led to the technology known as Rule-Based Machine Translation (RBMT), based on grammatical rules.
The second model, SMT, which is widely used now (both independently and in conjunction with other technologies), was developed in the 2000s. It compares arrays of uploaded data — texts in the source language and those translated by humans. Algorithms analyze the statistical semantic combinations of words in sentences in both data blocks and choose the correct option based on these data.
In November 2016, Google introduced Google’s Neural Machine Translation, which utilized solely neural networks, announcing the program’s shift to deep learning. This new model does not break sentences into words but focuses on the phrase, comparing it with other phrases, allowing for a more accurate conveyance of meaning. As for learning, Google’s Neural Machine Translation can “understand” language pairs it was not taught.
Indeed, there are many applications. There are paid and free ones, specialized (for programmers, doctors, or lawyers, for instance), and for the general public. Let’s talk about the most well-known:
All programs recognize speech and can automatically identify captions in photos. They can edit completed documents, and Smodin even evaluates authors’ works and checks for plagiarism.
Moreover, services have emerged that allow working with video. HeyGen is a neural network for nearly professional dubbing into French, Spanish, Portuguese, and Indian, while preserving the character’s voice on screen. And that’s not all. The neural network adapts the person’s facial expressions and lip movements, making the video sequence look realistic. The capabilities of HeyGen can be assessed by numerous dubbed meme clips.
Translation programs help with studying, work, and entertainment. Of the main advantages:
However, there are also downsides. And the main one is inaccuracy. This is even acknowledged by Chat GPT, confirming that the result of processing information depends on software settings and the type of task being solved.
Artificial intelligence does not understand the wordplay in literary works and cannot convey emotional coloring. When processing specialized documents, AI can distort the meaning or give an incorrect interpretation to specific definitions. Therefore, it’s better not to trust legal and technical documentation to AI.
Technologies are evolving and becoming smarter. Algorithms can now accurately convey the overall meaning of the source and no longer commit gross errors, as was the case 5 years ago. But it is still too early to say that neural networks will soon replace humans; rather, they can be considered assistants.