Google rules out using artificial intelligence for weapons | Inquirer Technology

Google rules out using artificial intelligence for weapons

/ 10:26 AM June 08, 2018

Sundar Pichai. AFP File Photo

SAN FRANCISCO – Google announced on Thursday it would not use artificial intelligence for weapons or to “cause or directly facilitate injury to people,” as it unveiled a set of principles for the technologies.

Chief executive Sundar Pichai, in a blog post outlining the company’s artificial intelligence policies, noted that even though Google would not use AI for weapons, “we will continue our work with governments and the military in many other areas” such as cybersecurity, training, or search and rescue.

Article continues after this advertisement

The news comes with Google facing an uproar from employees and others over a contract with the United States military, which the California tech giant said last week would not be renewed.

FEATURED STORIES

Pichai set out seven principles for Google’s application of artificial intelligence, or advanced computing that can simulate intelligent human behavior.

He said Google is using AI “to help people tackle urgent problems” such as prediction of wildfires, helping farmers, diagnosing disease or preventing blindness.

Article continues after this advertisement

“We recognize that such powerful technology raises equally powerful questions about its use,” Pichai said in the blog.

Article continues after this advertisement

“How AI is developed and used will have a significant impact on society for many years to come. As a leader in AI, we feel a deep responsibility to get this right,” Pichai also wrote.

Article continues after this advertisement

The chief executive said Google’s AI programs would be designed for applications that are “socially beneficial” and “avoid creating or reinforcing unfair bias.”

He said the principles also called for AI applications to be “built and tested for safety,” to be “accountable to people” and to “incorporate privacy design principles.”

Article continues after this advertisement

Google will avoid the use of any technologies “that cause or are likely to cause overall harm,” Pichai wrote.

That means steering clear of “weapons or other technologies whose principal purpose or implementation is to cause or directly facilitate injury to people” and systems “that gather or use information for surveillance violating internationally accepted norms.”

Google also will ban the use of any technologies “whose purpose contravenes widely accepted principles of international law and human rights,” Pichai said.

‘A good start’

Some initial reaction to the announcement was positive.

The Electronic Frontier Foundation (EFF), which had led opposition to Google’s Project Maven contract with the Pentagon, called the news “a big win for ethical AI principles.”

“Congratulations to the Googlers and others who have worked hard to persuade the company to cancel its work on Project Maven,” EFF said on Twitter.

Ryan Calo, a University of Washington law professor and fellow at the Stanford Center for Internet & Society, tweeted: “Google’s AI ethics principles owe more to (English philosopher Jeremy) Bentham and the positivists than (German philosopher) Kant. Nevertheless, a good start.”

Calo added, “The clear statement that they won’t facilitate violence or totalitarian surveillance is meaningful.”

The move comes amid growing concerns that automated or robotic systems could be misused and spin out of control, leading to chaos. At the same time, Google has faced criticism that it has drifted away from its original founders’ motto of “don’t be evil.”

Several technology firms have already agreed to the general principles of using artificial intelligence for good but Google appeared to offer a more precise set of standards.

The company, which is already a member of the Partnership on Artificial Intelligence including dozens of tech firms committed to AI principles, had faced criticism for the contract with the Pentagon on Project Maven, which uses machine learning and engineering talent to distinguish people and objects in drone videos.

Faced with a petition signed by thousands of employees and criticism outside the company, Google indicated the $10 million contract would not be renewed, according to media reports.

Your subscription could not be saved. Please try again.
Your subscription has been successful.

Subscribe to our daily newsletter

By providing an email address. I agree to the Terms of Use and acknowledge that I have read the Privacy Policy.

But Google is believed to be competing against other tech giants such as Amazon and Microsoft for lucrative “cloud computing” contracts with the US government, including for military and intelligence agencies.                   /kga

TOPICS: AI, Computers, Information, IT, Military, technology, US, Weaponry
TAGS: AI, Computers, Information, IT, Military, technology, US, Weaponry

Your subscription could not be saved. Please try again.
Your subscription has been successful.

Subscribe to our newsletter!

By providing an email address. I agree to the Terms of Use and acknowledge that I have read the Privacy Policy.

© Copyright 1997-2024 INQUIRER.net | All Rights Reserved

This is an information message

We use cookies to enhance your experience. By continuing, you agree to our use of cookies. Learn more here.