- News Home
5 December 2013 11:26 am ,
Vol. 342 ,
An animal rights group known as the Nonhuman Rights Project filed lawsuits in three New York courts this week in an...
Researchers have been hot on the trail of the elusive Denisovans, a type of ancient human known only by their DNA and...
Thousands of scientists in the Russian Academy of Sciences (RAS) are about to lose their jobs as a result of the...
Dyslexia, a learning disability that hinders reading, hasn't been associated with deficits in vision, hearing, or...
Exotic, elusive, and dangerous, snakes have fascinated humankind for millennia. They can be hard to find, yet their...
Researchers have sequenced and analyzed the first two snake genomes, which represent two evolutionary extremes. The...
Snake venoms are remarkably complex mixtures that can stun or kill prey within minutes. But more and more researchers...
At age 30, Dutch biologist Freek Vonk has built up a respectable career as a snake scientist. But in his home country,...
- 5 December 2013 11:26 am , Vol. 342 , #6163
- About Us
How to Train Your Robot (to Lie)
10 September 2010 6:10 pm
A military base has just fallen to enemy fighters. A robot containing top-secret information has to escape detection by the invading army. The robot is facing three corridors: right, center, and left. It could randomly pick a corridor and hope the enemy soldiers pick a different one. Or it could leave a false trail—assuming robots can be trained to lie. A new study using this scenario suggests that they can be.
This is not the first time machines have displayed deceptive tendencies. In 2007, Swiss researchers discovered that robots programmed to learn from experience spontaneously evolved the habit of lying to one another under some conditions. But the idea of deliberately implanting deceit is new.
The first step wasn't teaching the robots how to lie, but when to lie. Computer scientists Alan Wagner and Ronald Arkin of the Georgia Institute of Technology in Atlanta came up with an algorithm that set two conditions: First, a robot had to be in conflict with someone or something else. And second, it had to be able to influence its adversary's actions. If both conditions checked out, the robot was cleared to lie.
Then they tested the idea in two-wheeled, camera-equipped robots that were playing a simple game of hide-and-seek.
The "hider" robot had to choose one of three compartments in which to conceal itself. The pathway to each was blocked by a green, red, or blue marker. Whichever way the robot went, it knocked down one marker. After some training, the hider bot and seeker bot figured out that a fallen marker indicated which direction the hider had gone. But there was one catch: The seeker didn't know the hider bot had been programmed to lie.
When the game began, the hider randomly chose one of the compartments. Then its deceptive programming kicked in and told it to go another direction first—knocking over a different marker—and then turn back toward its choice. The seeker, seeing the fallen marker and suspecting no trickery, would follow the fake trail and leave the hider undiscovered.
After 20 trials, they found the deception worked 75% of the time. When it failed, Wagner says, the robots were clumsy. The hider might try to go right but end up going toward the center, for example, or its slow perception of the marker might make it hit the wrong one. "It thinks it's going to knock over one, but lo and behold, it plows through all of them." Barring such blunders, however, the seeker robots fell for the ruse every time, the researchers reported last week in the International Journal of Social Robotics.
One of the hardest parts of deception is knowing how an adversary will interpret a false message, notes game theorist Philippe Jehiel of the Paris School of Economics, who was not involved in the work. He said he wasn't certain that the authors' way of modeling what the seeker would do was clear. "As far as I could see in the present study, this is one of the most challenging issues."
Wagner and Arkin also recognized the importance of knowing the enemy. In this study, they made the deception work by not revealing to the seeker the hider's capacity to lie. How the hider would fare against smarter robots remains an open question.
Wagner says he's aware people might be leery of creating deceitful robots. But he thinks robots that know how to lie could benefit society in the long run. "There are a lot of important situations in which humans deceive for the better of the other person," he says. For example, "If I'm trying to get a person with Alzheimer's to take medicine, we may be in a temporary state of conflict, but overall it's better for them to take that medicine," Wagner says. "Deception is not necessarily nefarious."
If you enjoyed this article, you might also like It's a Bot-Eat-Bot World.