ATA SLD

Slavic Languages Division (American Translators Association)

American Translators Association: The Voice of Interpreters and Translators

  • Home
  • About Us
  • Blog
    • Comments Policy and Disclaimer
  • SlavFile
  • Resources
    • Slavic Languages Presentations Archive
  • Contact Us
  • SLD Podcast

ATA65 Review: I Can’t Place the Accent

March 5, 2025

A Review of: I Can’t Place the Accent: Identifying the Characteristics Traits of Computer Translation, presented by Eugenia Tietz-Sokolskaya, CT and John Riedl, CT on Friday, November 1 at ATA65

Review by Christine Pawlowski; slides from the presentation available in the Slavic Languages Presentation Archive

I do not feel threatened by AI,  and this is not because I think my language skills are superior to the machine. Perhaps it is mostly that my monthly social security check allows me to lighten my workload to a manageable volume. And maybe it is also that I am technologically challenged (being provided with a modem to do my work for FBIS in the 90’s was a huge step).

In my very limited experience with AI projects, I have enjoyed “beating” the machine, as when the AI translation of the word “wygodny” in Polish, which may be translated variously as “convenient” or “comfortable,” resulted in an English version of an apartment advertisement that sported a comfortable bedroom armoire—perhaps a magician’s prop?

Well before the October/November ATA65 Conference, we received a survey from Eugenia and John in which we were asked to select the best translation of several Russian texts, but we were not told who (or what) did the translation. When we arrived at the session, we learned that the translations had been done by Deepl, an NMT (Neural Machine Translation) service launched in 2017; ChatGPT 3.5, an LLM (Large Language Model) service launched in 2022; and a few different humans.

Given its timeliness, it is not surprising that generative AI figured prominently in many of the conference sessions. John and Eugenia’s session dovetailed beautifully with the subject of Holly Mikkelson’s Wednesday training for ATA graders: Prescriptive and Descriptive Language. In a nutshell, we investigated how we really speak.

In both Holly’s presentation and that of our Slavists, we looked for the “tell”—a clue or indication that reveals information or suggests a hidden truth. All translators—human or generative AI–have these tells. To find them, we looked to cohesion, fluency, syntax and terminology.

From the survey results and our on-site bantering about some of the linguistic conundrums, we learned that:

  • ChatGPT’s renditions will be grammatical and flow deceptively well but may not be accurate.
  • DeepL is easier to peg as a computer translation.
  • Human translators take liberties, which can be a blessing or a curse (hence the dangers of prescriptivism and the difficulties encountered in evaluating translation).

The good news disseminated by Eugenia and John is that humans can achieve higher quality by:

  • recasting or rewriting clauses
  • splitting or combining sentences
  • choosing subject-appropriate terms
  • substituting phrases for words and vice versa

This list suggests that skills in manipulating syntax are critical. The bad news is that humans also misspell words, misuse collocations and struggle with job fatigue.

Comparatively speaking, for the three passages we studied in the session, in every case ChatGPT came out on top of DeepL, which sticks very close to the original syntax, even to the point of unreadability. In two of the three examples, the human translation won.

Discussions of AI are ubiquitous—on Linkedin, in journals and magazines. An opinion piece in the latest edition of the journal First Things offers the suggestion that society can resist the techno-tyranny trend, which is making us miserable, by demanding human-to-human businesses because “People…will pay for happiness.”  There is some nostalgia for the way things were before the modem and the ease of searching the web for the contextually right word—but not much, in my opinion.

Christine Pawlowski is a freelance Polish and Russian translator with an M.A. in Slavic Languages and Literatures from Indiana University, Bloomington. She is retired from teaching elementary-school music and delights in being Busia to her 17 grandchildren and in directing and accompanying her church choir. She is ATA-certified for Polish-into-English and an ATA Certification grader for that language pair. She may be contacted at pawlow@verizon.net.

Filed Under: ATA65, Tools Tagged With: AI, ata65, Russian, session review

Generative AI and What It Means for Translators and Interpreters. Part I.

July 29, 2024

By Viktoryia Baum

Over the past year and a half, the world has been thrown into the hype (or pain?) of artificial intelligence, with the advancement of many products, and the race between the companies to create their own, unique and individual intelligent chatbots capable of many ordinary human tasks. I have been hearing from many colleagues of different ages how this is becoming an existential threat to the profession as a whole, for both translators and interpreters alike. But the real question remains: is it?

I’m a firm believer that if there’s a will, there’s a way. Meaning that I don’t think any of these “robots” will eliminate the need for the profession and the professionals, or somehow bring about our total and ultimate demise. If anything, they can be trained to help us in any imaginative way or method possible. Such opinions of mine have been met with frustration and disbelief by many of my colleagues, yet welcomed and shared by another many.

Recently, Eugenia Tietz-Sokolskaya invited and encouraged me to consider writing for the SLD blog on any subject and anything at all (thank you, Eugenia!), and although I don’t do much writing in any capacity, I figured I might as well give it a try. I usually have a lot to say about things, so why not do it here? Hopefully, the readers will not have a plate of rotten tomatoes handy to smash against the monitor as they read these lines. Maybe, just by sharing some of my experiences I may help alleviate some technology fears, or help someone learn something new. I will probably anger some colleagues and make them reach for their plate of rotten tomatoes, but that’s where technology is in my favor—unlike 200 years ago, I don’t have to deliver this speech in front of you (thank you, industrial civilization!), and I can hide behind the screen.


Let’s start with some basics and try to break down the concept of generative AI, its purpose, and what it can do. Basically, generative AI refers to artificial intelligence systems capable of generating new data and producing content. It can create not just text, but also images or even audio and video files. Here are some examples of generative AI:

  • Language models (GPT-4) that can generate human-quality text on almost any subject based on a prompt
  • Image generators used to create novel images from text descriptions (DALL-E 2 and Stable Diffusion)
  • Music generators that can compose new songs, musical compositions, or other audio using training data (MuseNet)
  • Video generators that can make video clips by processing and synthesizing existing data (Sora)

The capability behind generative AI is machine learning models, especially large neural networks. They can analyze thousands of datasets and learn the patterns and representations within that data. Having learned and captured those patterns, the models are then capable of generating new content that is similar to the training data statistically, yet differs from it in a new way. One type of machine learning model is large language models (a phrase everyone in the industry has been hearing more and more); those are built on neural architecture, and they are referred to as large because they use huge datasets. There has been a boost and a rapid increase in research and development of large language models in big tech, with any major player buckling up and racing to develop their own models and tools. The now very infamous ChatGPT is a large language model, or LLM. In simple terms, this is a computer program that has received and analyzed enough data that it is now capable of generating its own responses. The quality of the responses always depends on the quality of the datasets the LLM has been fed. The more they know, the better they are at producing content.


I’ve been experimenting with AI chatbots for quite some time now. When I said they can be trained to help us in any imaginative way or method possible, that’s because I’ve tried and largely succeeded on some level, achieving the objective I set out for myself. A few months ago, for example, I chatted with one of the publicly available chatbots (not ChatGPT) about building up a glossary specifically for use in interpretation, and I’d like to share the outcome.

The background for this experiment was fairly simple. I wear both hats of being a translator and an interpreter. In my state, I am a certified per diem court interpreter, but I live in an area with not a lot of need for court interpreting in my language combination. If anything, it’s approximately 4-5 calls a year plus a few depositions. This means that enough time passes between court appearances that my skills and language databanks typically grow rusty and need a solid refreshment before assignments. Preparing for any court appearances can be daunting due to lack of information and/or materials. If all I know is that tomorrow I am heading over to family court, how should I prepare? What should I look for?

I started with a relatively easy prompt, asking my new AI friend (let’s call him Sam) to help me put together a Russian to English legal glossary to use for interpreting purposes. In a few seconds, the system produced a glossary with some common legal terms, roughly 20 to 25, including terms like lawsuit/claim, indictment, defendant, and court order. The chatbot then asked me if I wanted to know any other specifics to be added “for my translation purposes,” to which I happily said yes, also requesting a set of legal terms in Latin, giving an example prompt of “amicus curiae.” The result was a compilation of another 25-30 terms including bona fide, nolo contendere, ex parte, pro se, inter alia, etc.

For the third part of the conversation, I congratulated Sam on providing me with the definitions of Latin legal terms in the English language, then asking if he could take these same Latin terms and give me their Russian equivalents. He obliged. It is important to note here that the glossary was provided to me in the LATIN to RUSSIAN variant, although for some reason I expected Sam to use the English translations he pulled for the Latin and only then give me the Russian equivalents. I was wrong. Sam was smart, even if my prompt was quite poorly written.

Next, I asked for more terminology, the more advanced, the better. The resulting table included terms like defamation, pre-trial agreement, acquittal, and statute of limitations. I felt that it could have been a bit more advanced, but ok. The next query asked for terms used in family and traffic courts. Then I made another query, and another, and another. But I think I made my point already. Perhaps the best glossary (in my humble opinion) was produced when I asked for terminology specifically used in an arraignment hearing. This turned up terms like request for leniency, motion to suppress evidence, own recognizance release, and recusal motion.

The entire encounter and the prompts along with their results took me approximately 5-7 minutes. The end of the conversation was quite comical, since I asked for all of the terms to be exported to Excel and sent to me via email. Sam happily agreed, took down my email address, and nothing happened. When told that no email had come, he apologized profusely for letting me down and leading me to believe he had the capability of exporting and sending emails. He said that by design he actually did not have that capability, but in his inherent desire to please me, he misrepresented his abilities. The apologetics went on for a few more rounds, much to my amusement and comic relief. It was clear that I wasn’t speaking with a human, but I got amused as if I really were.

Overall, I found the quality of the glossaries produced by AI very high. I did not find any incorrect translations, nor did I find any inconsistencies (some terms were repeated with each of the prompts, and the output was the same). If you ask me personally, I believe that for a real assignment, this approach would have saved me hours of googling terminology and trying to think about all the possible terms I could encounter. Of course, if I were unfamiliar with any of the terms or translations produced, I would have double-checked everything and cross-referenced items using regular old-fashioned dictionaries. For example, if I were to go and interpret at a mining conference, or any other completely foreign subject, like rare and valuable gemstones, I would follow up on what the AI told me.

I forgot to mention that I went into this experiment with very low expectations. I can’t say that I wanted the AI to give me erroneous translations of the terms, to prove that humans and only humans are capable of creating a glossary, but I did think there would be errors. AI proved me wrong, at least for that exercise. I intend to keep practicing and conversing with it regarding my needs in the profession. If you are given the tool, why not use it? And if you are skeptical, I would invite you to try for yourself. Or you can reach for that plate of rotten tomatoes……just kidding. In short, you don’t have to love it or hate it, endorse it or promote it. It’s just a tool. If it can help you save time and do a better job, why not. If anything, you will definitely get some comic relief from experimenting.

If you have any comments or feedback regarding this blog post, I invite you to reach out directly to me via email at vbaum00@gmail.com. In the meantime, I’ll work on Part II of this series over the next weeks. Thanks for reading, and stay tuned!


Viktoryia Baum is an avid technology buff and skilled researcher who spends at least part of her spare time studying new and existing language tools. Having started as an aerospace interpreter many years ago, she discovered she was equally good at translation, so she does both, working with Eastern European languages and specializing in technical, legal, and medical translation and interpreting. She lives in upstate New York and can be contacted at vbaum00@gmail.com.

Filed Under: Interpreting, Legal, Tools Tagged With: artificial intelligence, interpreting, Russian

Achieving High-Quality Translation: The Final Step

June 24, 2024

High-quality translation requires in-depth domain knowledge, meticulous attention, and a series of well-defined steps. In this article, I’ll focus on the last stage of a standard translation project: automated Quality Assurance (QA).

QA ensures that hard-to-spot and easy-to-miss errors are eliminated. Common issues include source and target inconsistencies, capitalization errors, incorrect spacing around punctuation marks, numerical mistakes, missing or extra tags, incorrect quotation marks, terminology mistakes, measurement unit discrepancies, etc.

Below are the steps I use when doing QA.

  1. Built-in CAT tool QA:

Most computer-assisted translation (CAT) tools (e.g., Trados Studio or memoQ) offer built-in QA functionality. Enable the relevant options in the CAT tool menus and run QA.

  1. Export to Word and use Word’s proofing features:

If your CAT tool allows it, export the translation to Microsoft Word. Press F7 to check for additional mistakes that the CAT tool may have missed. Ensure that Word’s proofing options (grammar, repeated words, uppercase words, etc.) are activated.

  1. Standalone QA tools:

Use standalone QA tools for comprehensive checks. Xbench, Verifika, and QA Distiller are some of the oldest and most popular ones.

QA Distiller is completely free, Xbench has a free limited-functionality version, and Verifika (my personal favorite) has a fully functional free web version, provides language-dependent checks, and covers numerous mistake categories. Take the time to configure Verifika with the options you need—it’s worth the effort.

  1. Double-check and run QA again:

Correct any mistakes the tool found and run another round of QA to catch any overlooked errors or newly introduced mistakes.

  1. Use multiple QA tools:

If you’re doing a test translation or working on a particularly important project, consider running QA using multiple tools. It’s better to spend time reviewing false positives than to miss an embarrassing error.

  1. Impress the client:

Go the extra mile by exporting a final QA report containing only false positives. Demonstrating your commitment to quality will leave a positive impression.

Let no mistakes slip through into your translations!


This is the third and final post in a series of posts on translation quality. The first post can be found here, and the second here.

headshot of Mikhail YashchukMikhail Yashchuk is an industry veteran. In 2002, he received his university degree in English, and six years later he founded a boutique agency where he gained experience in linguist recruitment, project management, translation, editing, and quality assurance. He has recently been admitted as a sworn translator to the Belarusian Notary Chamber.
In 2018, Mikhail joined the American Translators Association and is now working as an English-to-Russian translator, actively sharing knowledge with younger colleagues. He is the moderator for the
SLD LinkedIn group. He may be contacted at mikhail@lexicon.biz.

Filed Under: Tools, Translation Tagged With: CAT tools, series, translation

A [Better] CAT Breed for the Slavic Soul

July 12, 2017

A review by Jennifer Guernsey

Aha! I said to myself upon spying this presentation among the 2013 ATA Conference’s offerings. At last, I will find out which elusive CAT tool actually does a good job with Slavic languages! I had tried several tools, but hadn’t yet run across one that was able to accommodate the peculiarities of my language, Russian, particularly when it came to all of the inflected forms.

Alas, it took no more than two slides for me to be sorely disappointed – not in Konstantin Lakshin’s presentation, but in the sad news that there is, in fact, no such thing as a good CAT tool for Slavic languages. Or, at least, there isn’t yet.

Despite my initial dismay at the news, I fortunately stayed to hear the entire presentation. It can be briefly summarized as follows: A combination of technical, linguistic, and particularly market forces have conspired to make CAT tools what they are today: decidedly Slavic-unfriendly. The good news is that many of the pieces needed to improve them already exist, and it’s up to us to put pressure on developers and companies to make use of those pieces.

The reason it took the better part of an hour to provide this information is that the presentation included a lot of very interesting history, examples, and details. It really was quite educational, at least for me.

Kostya started by outlining the history of computer use in translation, and the development of CATs in particular. He began with a discussion of a 1966 government-funded report by the Automatic Language Processing Advisory Committee on the use of computer technology in translation. The gist of this report as it applies to our CAT tool discussion is that machine translation doesn’t work well, but that something vaguely resembling what we now consider a CAT tool, with a similar workflow, might be useful. This pseudo-CAT workflow used the punch card operator – i.e., a human being – as a morphology analyzer. This is interesting, because one of our principal complaints about today’s CAT tools is that they do not have morphology analysis capability. The report also compared use of this early form of CAT with a standard translation process, and found that while it might save some time, its primary advantage was that it “relieve[d] the translator of the unproductive and tiresome search for the correct technical terms.” The report emphasized that compiling the proper termbase was really the key to an effective translation tool.

In the decade or so following the report, the emphasis in computer-assisted translation was thus on building termbanks. In other words, the focus was on words and phrases – small subsegments, if you will – and these termbanks were generally compiled for specific large organizations operating in specific contexts and were not readily transferrable to other entities.

The philosophy that drives current CAT tools – the “recycling” of previously translated texts – emerged fully only in 1979, though large corporations had begun exploring this starting in the late 1960s. This philosophy was in great part a result of the requirements and technologies in place at the time. In the 1960s, for instance, the world was a less integrated place, and there was limited control over the input side – the source text content, editing, and so on. The example Kostya provided was scientific texts coming out of the USSR that were being translated. Fast-forward to the 1980s and 1990s: large corporations have end-to-end control of processes and utilize translation (and translation technology) for their own documents. In this latter context, being able to retrieve and reuse entire sentences made a lot of sense. Note also that in the prevailing markets in which the early CAT tools developed, the primary languages were not highly inflected.

In the late 1980s and early 1990s, the first commercially available CAT tools appeared: IBM Translation Manager II, XL8, Eurolang, and two still-familiar tools, Trados and Star Transit. Trados, in particular, started life as a language services provider trying to get an IBM contract.

The mid- to late 1990s saw the emergence of tools being created ostensibly for translators: Déjà Vu, Memo Q, and WordFast. However, rather than being fundamentally different from their larger predecessors, these often turned out to be essentially smaller, less functional versions of Trados. This era also witnessed the development of smaller commercial players, such as WordFisher (a set of Word macros) and in-house tools such as LionBridge, Foreign Desk, and Rainbow (specifically for software localization), as well as Omega T, the first open-source CAT tool.

That brings us to the present day, the 2000s, when there are too many CAT tools to list, and there have been many mergers and acquisitions among them. However, NONE of the existing tools can be considered very useful for Slavic or other highly inflected languages. In addition to the reasons noted above, there were other issues that contributed to this situation as the software was being developed. First, there were no obvious ways to incorporate Cyrillic into early software. Second, there were additional market forces, such as software piracy, the cross-border digital divide, and the lack of major clients, that provided little incentive to software developers to make CAT tools that would be particularly useful in Slavic-language markets.

Today, we have a much wider playing field in terms of the market for translation. Translation work is “messier” now, and involves things like corporate rebranding and renaming, a variety of dialects and non-native speech, outsourcing, rewrites for search engine optimization, and bidirectional editing in which both source and target documents are being modified. In this environment, the old “termbase plus recycled text” CAT model is not sufficient.

From this historical background, Kostya next proceeded to illustrate just what the difficulties are that Slavic languages present for today’s CAT tools. These can be boiled down to their relatively free word order, their rich morphology, and their highly inflected nature. The CAT tool’s “fuzzy match” capabilities are insufficient for Slavic languages.

Kostya then provided a number of illustrative examples. Consider the following pairs of segments:

To open the font menu, press CTRL+1.

Press CTRL+1 to open the font menu.

Analyzing and characterizing behaviors

Analysing and characterising behaviours

He ran these and other examples through about a half-dozen CAT tools using a 50% match cutoff, and found that the first example was considered only a 60-80% match, and the second was 0% (in other words, below the 50% threshold). The CAT tools on the market generally do not recognize partial segments in a different order, nor can they tell that “analyzing” and “analysing” are essentially the same word. In other words, they lack language-specific subsegment handling, and morphology-aware matching, searching, and term management. They are also missing form agreement awareness (e.g., noun/adjective case agreement). This diminishes their utility for those translating out of Slavic languages, to be sure, but it also complicates matters for those translating into Slavic languages, as word endings in retrieved fuzzy matches must constantly be checked and corrected.

The obvious question that Kostya next asked is, can this situation be fixed? In theory, yes. Kostya believes that many software tools already in use by search engines, machine translation, and the like could be integrated into CAT tools. These include Levenshtein distance analyzers that can handle differences within words; computational linguistics tools such as taggers, parsers, chunkers, tokenizers, stemmers, and lemmatizers, which analyze such things as syntax and word construction; morphology modules; and even Hunspell, the engine already in use by numerous CAT tools for spellchecking but not for analyzing matches.

Developers continue to cite obstacles to integrating these tools: it’s complicated, they are too language-specific, we don’t know how to set up the interface, there are licensing issues, we have limited resources. While all of these are legitimate factors, Kostya believes that they do not present insurmountable obstacles. He is hopeful that developers will start seeing these tools as data abstraction tools that enable the software to break down the data into something that is no longer language-specific.

So what can we do about this lack of suitable CAT tools? Kostya’s recommendation is principally that we talk to software developers and vendors and explain what we want. We need to create our own market pressure to move things along. In addition, we need to educate developers and vendors about the existing tools that are available; for instance, we might point them to non-English search engines that utilize morphology analyzers.

Alas, there is neither a good CAT tool for the Slavic soul nor a quick fix to this situation. But after listening to Kostya’s presentation, I have a much better understanding of how this situation developed and how we might take action to prompt vendors and developers to move in a new direction.

Filed Under: Annual Conferences, Tools, Translation Tagged With: CAT tools

Russian Language Style Guide Resources

July 12, 2017

 

Article by Natalie Shahova – published in 2015

At the ATA 55th Annual Conference in Chicago a question was raised whether there is a Russian Guide similar to The Chicago Manual of Style for English language. I tried then to answer this question orally while below are some formal links to the sources I cited. One must keep in mind that Russian rules are much stricter than English. Though they do leave some freedom to the users, in most cases the absence (or presence) of a comma or of any other punctuation sign is an obvious mistake.

Please also note that numbers 2 & 3 of my list exist in various versions (titles, authors and dates of publication vary) but they are generally referred as Розенталь and Мильчин accordingly.

  1. Правила русской орфографии и пунктуации

https://www.rusyaz.ru/pr/
Утверждены в 1956 году Академией наук СССР, Министерством высшего образования СССР и Министерством просвещения РСФСР. На сегодня эти Правила, установившиеся почти полвека назад, – по-прежнему базовый источник для составителей словарей и справочников по русскому языку. На них основаны все многочисленные учебники и пособия для школьников и абитуриентов.

  1. Справочник по правописанию, произношению, литературному редактированию

Розенталь Д.Э., Джанджакова Е.В., Кабанова Н.П.

https://evartist.narod.ru/text1/20.htm
Дитмар Эльяшевич Розенталь (1899-1994) — советский и российский лингвист, автор многочисленных трудов по русскому языку.

  1. Справочник издателя и автора

А.Э. Мильчин и Л.К. Чельцова

https://www.redaktoram.ru/izdat_books_download_1_2.php – первые 12 разделов в виде pdf

https://diamondsteel.ru/useful/handbook/ – первые 7 разделов книги online

https://www.artlebedev.ru/everything/izdal/spravochnik-izdatelya-i-avtora/ – описание книги, покупка бумажной версии

  1. Запятание трудных слов и выражений – правила постановки запятых

https://www.konorama.ru/igry/zapatan/

  1. Корпус русского языка

https://www.ruscorpora.ru/search-main.html
На этом сайте помещен корпус современного русского языка общим объемом более 500 млн слов. Корпус русского языка — это информационно-справочная система, основанная на собрании русских текстов в электронной форме.

Корпус предназначен для всех, кто интересуется самыми разными вопросами, связанными с русским языком: профессиональных лингвистов, преподавателей языка, школьников и студентов, иностранцев, изучающих русский язык.

  1. Переводим служебные знаки

Наталья Шахова

Статья о различиях между правилами русской и английской пунктуации

https://atasld.org/sites/atasld.org/files/slavfile/fall-2008.pdf
SlavFile, Fall 2008, Vol. 17, No. 4, p.5

  1. Ководство

Артемий Лебедев

Подборка статей о дизайне и веб-дизайне, а также о российском интернете и событиях в нем.

Многие статьи касаются пунктуации и оформления текстов.

https://www.artlebedev.ru/everything/izdal/kovodstvo4/
Некоторые главы книги online: https://www.artlebedev.ru/kovodstvo/sections/

Filed Under: Tools, Translation Tagged With: Russian

Recent Posts

  • Turkic Languages SIG: Seeking Moderator
  • ATA65 Review: On Interpreting for Russian-Speaking LGBTQ+ Individuals
  • ATA65 Review: I Can’t Place the Accent
  • SLD Announcements: Networking Zoom and ATA66 Deadline Extended
  • Speak at ATA66 – Proposals Due March 3

SLD on Twitter

My Tweets

SLD on Social Media

Facebook: ATA Slavic Languages Division LinkedIn: Slavic Languages Division of the American Translators Association

Tags

Administrative AI annual dinner ATA ATA58 ATA59 ATA60 ATA61 ATA63 ATA64 ata65 ATA66 audiovisual AVT business CAT tools certification ceu watch conference editing events feedback interpreting interview legal literary localization marketing medical member profile networking podcast Polish professional development project management Russian series session review SlavFile SLD specializations survey translation webinar workshop

SLD Blog Categories

Search This Website

Copyright © 2025 · American Translators Association

 

Loading Comments...