Google translation AI botches legal terms 'enjoin,' 'garnish' – research | Inquirer Technology

Google translation AI botches legal terms ‘enjoin,’ ‘garnish’ – research

/ 09:14 PM April 19, 2021

Translation tools from Alphabet Inc’s Google and other companies could be contributing to significant misunderstanding of legal terms with conflicting meanings such as “enjoin,” according to research due to be presented at an academic workshop on Monday.

Google’s translation software turns an English sentence about a court enjoining violence or banning it, into one in the Indian language of Kannada that implies the court ordered violence, according to the new study

“Enjoin” can refer to either promoting or restraining an action. Mistranslations also arise with other contronyms, or words with contradictory meanings depending on the context, including “all over,” “eventual,” and “garnish,” the paper said.

ADVERTISEMENT

Google said machine translation is “is still just a complement to specialized professional translation” and that it is “continually researching improvements, from better handling ambiguous language, to mitigating bias, to making large-quality gains for under-resourced languages.”

FEATURED STORIES

The study’s findings add to the scrutiny of automated translations generated by artificial intelligence software. Researchers previously have found programs that learn translations by studying non-diverse text perpetuate historical gender biases, such as associating “doctor” with “he.”

The new paper raises concerns about a popular method that companies use to broaden the vocabulary of their translation software. They translate foreign text into English and then back into the foreign language, aiming to teach the software to associate similar ways of saying the same phrase.

Known as back translation, this process struggles with contronyms, said Vinay Prabhu, chief scientist at authentication startup UnifyID and one of the paper’s authors.

When they translated a sentence about a court enjoining violence into 109 languages supported by Google’s software, most results erred. When spun back to English, 88 back translations said the court called for violence and only 10 properly said the court prohibited it. The remainder generated other issues.

Another researcher, Abubakar Abid, tweeted in December that he found possible bias in back translation through Turkish. Using Google, short phrases with “enjoin” translated to “people” and “Muslims” ordering violence but the “government” and “CIA” outlawing it.

The new paper said translation issues could lead to severe consequences as more businesses use AI to generate or translate legal text. One example in the paper is a news headline about nonlethal domestic violence turning “hit” into “killed” during translation, a potentially true but problematic association.

ADVERTISEMENT

Authors also expressed concern about the lack of warnings and confidence scores in tools from Google and others. Google in support materials warns it may not have the best solution “for specialized translation in your own fields.”

Your subscription could not be saved. Please try again.
Your subscription has been successful.

Subscribe to our daily newsletter

By providing an email address. I agree to the Terms of Use and acknowledge that I have read the Privacy Policy.

TOPICS: AI, Alphabet Inc., Google, Software, Translation
TAGS: AI, Alphabet Inc., Google, Software, Translation

© Copyright 1997-2024 INQUIRER.net | All Rights Reserved

We use cookies to ensure you get the best experience on our website. By continuing, you are agreeing to our use of cookies. To find out more, please click this link.