英语资讯
News

VOA慢速英语:谷歌禁止其人工智能技术用于武器

Source: 恒星英语学习网    2018-06-14   English BBS   Favorite  

Google says it will no longer permit its artificial intelligence, or AI technology to be used in any activities involving weapons.

The company's chief executive officer, Sundar Pichai, announced the decision in an internet post. He wrote that the new policy was one of several newly-launched "principles" aimed at guiding the company's AI work in the future.

The principles are a set of ethical guidelines covering the company's development and sale of AI technology and tools.

Google says it will no longer design or launch AI for weapons or other technologies whose main purpose is to cause harm to people. It will also not permit its AI technology to be used for surveillance activities that violate "internationally accepted norms."

"We believe these principles are the right foundation for our company and the future development of AI," Pichai wrote.

The principles were announced after more than 4,000 Google employees signed a document calling for the company to cancel an AI agreement with the U.S. Department of Defense. That agreement, known as Project Maven, involves the use of Google's AI technology to examine drone images for the U.S. military.

A Google official recently told employees Project Maven would not be extended after it ends next year. Google is expected to discuss with military officials how to complete the project without violating its new principles.

Kirk Hanson is director of the Markkula Center for Applied Ethics at Santa Clara University in California. The center examines how ethics can be used to guide technology development.

He told VOA the opposition by Google employees to the U.S. military agreement was based on fears that AI technology could lead to the creation of "autonomous weapons."

"If you have artificial intelligence which identifies targets and automatically launches weapons, you have what is known as an autonomous weapon -- there is no human decision to launch the weapon."

Hanson said other companies could also face pressure from employees or the public if their AI technology is used to develolp autonomous weapons. Just as with driverless vehicles, autonomous weapon systems may not be as safe as their supporters promise.

"We should be more concerned about how an autonomous weapon might make a mistake. Is that artificial intelligence targeting system as good as we think it is? And until we have trust that those systems will not make mistakes, we're going to have a lot of doubts about the use of artificial intelligence."

Hanson says even though Project Maven does not directly use Google AI to power autonomous weapons, AI systems do help with military targeting.

"If you have better targeting, presumably that's a good thing. But the critics say if you have better targeting, it raises your level of confidence in the targeting, which may lead you to then apply independent autonomous decision making by machine, which will launch the weapons."

A top Department of Defense official was asked about the use of autonomous weapons during an event last year at the Center for Strategic and International Studies in Washington. Air Force Gen. Paul J. Selva, vice chairman of the Joint Chiefs of Staff, said such systems should never be used to replace human commanders.

Google chief Pichai said the company does not plan to stop providing AI technology for all military uses. He said Google will still seek government projects in areas such as military training, internet security and search and rescue.

I'm Bryan Lynn.

Bryan Lynn wrote this story for VOA Learning English, based on information from Google, and reports from the Associated Press and VOA News reporter Michelle Quinn. Kelly Jean Kelly was the editor.

We want to hear from you. Write to us in the Comments section, and visit www.hxen.com .


将本页收藏到:
上一篇:VOA慢速英语:北美国家将举办2026世界杯
下一篇:VOA慢速英语:近70%的美国人“新闻疲劳”

最新更新
论坛精彩内容
网站地图 - 学习交流 - 恒星英语论坛 - 关于我们 - 广告服务 - 帮助中心 - 联系我们
Copyright ©2006-2007 www.Hxen.com All Rights Reserved