Google renounces AI weapons; Will still work with military

Register now

(Bloomberg) -- Google pledged not to use its powerful artificial intelligence for weapons, illegal surveillance and technologies that cause "overall harm." But the company said it will keep working with the military in other areas, giving its cloud business the chance to pursue future lucrative government deals.

Sundar Pichai, chief executive officer for Alphabet Inc.’s Google, released a set of principles on Thursday after a revolt by thousands of employees of the internet giant. The charter sets "concrete standards" for how Google will design its AI research, implement its software tools and steer clear of certain work, Pichai said in a blog post.

"How AI is developed and used will have a significant impact on society for many years to come," Pichai wrote. "As a leader in AI, we feel a special responsibility to get this right."

Some Google employees and outside critics cautiously welcomed the principles, although they voiced reservations, particularly about language that gives the company ample wiggle room in future decisions.

The seven principles were drawn up to quell concern over Google’s work on Project Maven, a Defense Department initiative to apply AI tools to drone footage. Staff protests forced Google to retreat from the contract last week. The company said on Thursday that if the principles had existed earlier, Google would not have bid for Project Maven.

Yet Google’s cloud-computing unit, where the company is investing heavily, wants to work with the government and the Department of Defense because they are spending billions of dollars on cloud services. The charter shows Google’s pursuit of these contracts will continue.

"While we are not developing AI for use in weapons, we will continue our work with governments and the military in many other areas," Pichai wrote. "These collaborations are important and we’ll actively look for more ways to augment the critical work of these organizations and keep service members and civilians safe."

Google’s charter is a watershed moment for the company and AI as a field. Technology giants, like Google, have stretched far ahead in developing software and services that give machines more control over decisions. However, these capabilities are now spreading to more industries, such as automotive, health care and government sectors. A driving force behind the spread is the easier access to AI building blocks that Google, Inc. and Microsoft Corp. have provided through their cloud services.

AI advances are helping medical research and provide other benefits. But the use of the technology in other areas has sparked concern among lawmakers and advocacy groups. Civil liberties organizations recently called out Amazon for offering facial recognition tech to local police departments.

Microsoft CEO Satya Nadella proposed similar principles in 2016, without mentioning the military. “Microsoft may decide to forgo the pursuit of business proposals for numerous reasons, including the company’s commitment to upholding human rights," the company said in an emailed statement on Thursday. Amazon didn’t respond to a request for comment.

In Google’s new principles, the company pledges not to pursue AI applications for weapons and technologies that "gather or use information for surveillance," in violation of accepted human rights laws. The principles also state that the company will work to avoid "unjust impacts" in its AI algorithms by injecting racial, sexual or political bias into automated decision-making.

In addition to outside criticism, Google has faced a rare spate of objections from its own staff. More than 4,000 employees signed a petition calling for the cancellation of the Project Maven contract, citing Google’s history of avoiding military work and worries about autonomous weapons. Last week, cloud chief Diane Greene said Google would not renew the deal when it expires next year -- an unusual withdrawal from a business deal.

One Google employee who signed the petition said the principles don’t go far enough, but it’s good that the company has finally addressed the issue. The proposed limit on the use of AI for surveillance is positive, but the language was too cautious, the person said. Other employees described the internal reception as lukewarm. They asked not to be identified criticizing their employer.

The principles about surveillance were not specific enough, according to Peter Asaro, an associate professor at The New School who organized a letter from academics against Project Maven.

"The international norms surrounding espionage, cyberoperations, mass information surveillance, and even drone surveillance are all contested and debated in the international sphere," he said. "Ultimately, how the company enacts these principles is what will matter more than statements such as this."

Google’s principles state that the company won’t design or deploy AI for "weapons or other technologies whose principal purpose or implementation is to cause or directly facilitate injury to people."

Miles Brundage, an AI policy research fellow at the University of Oxford, said the statement’s focus on "injury to people" suggests Google AI could still be used for cyberattacks and autonomous weapons aimed at buildings and other non-human targets. "Bit vague in places," he wrote on Twitter. "But it’s a start."

The staff who opposed the Project Maven deal also noted, in an internal email on Friday, that they would look closely at the charter and weigh in. Google will integrate the principles into existing product-review processes and plans to set up an internal review board to enforce the guidelines this year, according to a person familiar with the company.

"While this is our chosen approach to AI development, we also understand that there is room for many voices in this conversation," Pichai wrote in the blog post. "And we will continue to share what we have learned about ways to improve AI technologies and practices."

--With assistance from Dina Bass and Spencer Soper

Bloomberg News