People are more likely to program an autonomous vehicle to behave cooperatively with other cars than if they were driving themselves, a US study suggests.
It surprised researchers who, based on prior work, expected the involvement of AI in cars to make people “more selfish”.
Timeline for Automation
- February 19, 2019
- February 15, 2019
- February 13, 2019
The study involved 1,225 volunteers taking part in computerised experiments involving a social dilemma – a situation where individuals benefit from a selfish decision unless everyone in the group makes a selfish decision – with autonomous vehicles.
Based on the results, the researchers concluded people are less selfish programming autonomous vehicles because the same short-term rewards are less evident.
In other words, a human driver choosing not to let a car into a queue ahead of them in a traffic jam will be instantly rewarded by a higher position in the queue.
However, those programming a vehicle in an environment that is detached from the same situation, such as an office, are less likely to be swayed by the chance of an instant reward.
Beyond programming autonomous vehicles
The joint study between the US Combat Capabilities Development Command’s Army Research Laboratory (ARL), the Army’s Institute for Creative Technologies and Northeastern University published their findings in the Proceedings of the National Academy of Sciences.
The researchers said the results can be generalised beyond the domain of autonomous vehicles.
“Autonomous machines that act on people’s behalf — such as robots, drones and autonomous vehicles — are quickly becoming a reality and are expected to play an increasingly important role in the battlefield of the future,” said ARL’s Dr Celso de Melo, who led the research.
“People are more likely to make unselfish decisions to favour collective interest when asked to program autonomous machines ahead of time versus making the decision in real-time on a moment-to-moment basis.”
However, it is unclear whether this mentality will apply to an individual’s self-interest against a collective interest, said De Melo.
“For instance, should a recognition drone prioritize intelligence gathering that is relevant to the squad’s immediate needs or the platoon’s overall mission?” asked de Melo. “Should a search-and-rescue robot prioritize local civilians or focus on mission-critical assets?”
3 Things That Will Change the World Today
The UK government has committed to having driverless cars on the road by 2021.