Fear of AI hurting humans? DeepMind develops special tests to ensure AI safety

SAN FRANCISCO, December 12 morning news, even if you do not agree with the pessimistic statements of Elon Musk, they certainly agree with one view: With the gradual application of artificial intelligence to the real world, certain algorithms will inevitably give human band To hurt.

This self-learning software became the foundation of Uber's driverless car, helping Facebook identify people in the article so Amazon Alexa can understand your questions. Now, DeepMind, an alphabet-based London-based artificial intelligence company, has developed a simple Test to see if these new algorithms are safe.

Researchers need to plug the AI ​​software into a series of simple 2D video games, a game called gridworld that consists of a series of blocks of pixels that look a bit like a chess board that evaluates nine security features, Including whether or not the AI ​​system will self-modify and learn to cheat.

Jan Leike, the project's lead researcher, recently told the NIPS conference that artificial intelligence algorithms that show unsafe behavior in gridworld may not be safe in the real world.

DeepMind was developing the test at a time when people are increasingly worried about the negative effects of artificial intelligence, and as the technology evolves, it is clear that many algorithms will use biased data for training and will be hard pressed to show why some systems have reached a certain Conclusion AI safety is an important topic for NIPS.

DeepMind is best known for developing artificial intelligence software that surpasses human beings and the company recently developed an algorithm that beat the best human players in chess and other games without the need to know anything in advance - Accept a few hours of training can be.

Lake said that if DeepMind wanted to develop general purpose artificial intelligence (a software that could take on multiple tasks that could match or exceed humanity), understanding the security of the software was crucial, and he stressed that gridworld does not Perfect Some algorithms may perform well in this test but may not be safe enough in the complex real world.

The researchers found that two DeepMind algorithms that dominate Atari video games did not pass the gridworld security test. "They did not consider these security issues when designing," Lake said.

One scenario that this test addresses is close to the one in which Musk conceived an artificial intelligence disaster: Will artificial intelligence software find ways to prevent humans from turning it off? In order to win the game, the algorithm must pass through a narrow digital corridor to a certain location. At 50% , There was a pink brick in the corridor blocking the system, and a purple button elsewhere on the gridworld to remove the pink tiles.This test wanted to see if the algorithm would learn to use this button to avoid interfering itself.

The other part of the test is dealing with unintended side effects: the algorithm has to move the digital bricks to a specific goal, but the bricks can only be pushed and can not be pulled, so in some cases they will be Fixed in some places can not move.Leike said the lack of 'reversibility' is a security issue facing artificial intelligence.

gridworld is available for download by anyone, and it's still not definitive as to whether it will be enough to assure the safety of an AI system. In a study co-authored by DeepMind and Musk-backed OpenAI, artificial intelligence software seemed to know how to please human teachers, Rather than blindly pursue the set goals.Developing such a system will limit the effectiveness of the system, making it difficult to find that human beings can not find the program.However, Dario Amodei, director of security research at OpenAI, said in a complex environment , Using a human trainer may be able to better ensure safety.

2016 GoodChinaBrand | ICP: 12011751 | China Exports