Last updated October 25, 2019 at 3:20 pm
When military robots become valuable members of the team, there’s a tendency to treat them like colleagues rather than machines.
Why This Matters: Making human-like robots changes how we interact with them. But that empathy can also put humans in danger trying to protect them.
It is increasingly common to use robots in war zones to examine and disarm hazards or recover objects with the understanding that the loss of a robot is a far more acceptable outcome than the death of a solider.
Mark Billinghurst from the University of South Australia, along with members from the United States Air Force Academy (USAFA) wanted to explore troops connections with robots, as well as the robot’s impact on team efficiency and productivity on the front line.
Their research shows that for military robots to be fully integrated within a human-machine team (HMT), they first must be accepted as teammates. To facilitate this, a lot of work has been done over the years to make robots more ‘human-like’ by altering their physical characteristics and capabilities.
Deeper: Can we control killer robots?
While humanising robots strengthens the working relationships between soldiers and their robots, it also inflates the value of the robot team members in the minds of military personnel, leading to an increased emotional response when the robot is put under stress.
But, researchers have found that thinking of robots are a part of the team, could lead to life-changing consequences.
Personified robots are less likely to be put in jeopardy
Designing a simulation-based application, researchers tracked the emotional responses of two teams of participants who undertook a range of simulated tasks with either a personified or non-personified robot.
The study showed that teams working with a personified robot were 12 per cent less likely to put their robot at jeopardy of destruction compared to teams working with a non-personified robot.
The research also found that troops were more sensitive to the robot’s health and the possibility of seeing the robot ‘killed’ in action.
This is the first time that research has measured how actions can be altered by empathy when potential harm is induced in a simulation.
Billinghurst says the results show first-hand how emotional connections can impact decision making in the field.
“We have evidence to show that teams working with a personified robot are significantly more mindful about limiting damage and harm towards it – but this can have significant consequences.”
“Participants who limited their use of robots or chose not the use the robots had a similar overall achievement to the teams that did, the result of an increased level of self-sacrifice in the form of working harder to gain the same result.”
An empathic response can be dangerous
For most of us, an emotional attachment to a robot is considered harmless. Creating a bond with your Roomba vacuum or Google Home speaker can be fun and comforting, but empathy shown by a solider towards a military robot has the potential to interfere with performance on the front-line.
“Rather than sacrificing the robot, participants who were working with a personified robot had to increase their workloads and were willing to take more personal risks and would stop before putting the robot at risk – impacting their decision making under pressure,” Billinghurst says.
“Such hesitation and having an empathic response in these circumstances could have dangerous consequences for military personnel.”
Where split second decisions can determine the difference between life and death, it will become increasingly important to monitor soldiers working in collaboration with robots.
It is expected that military robots will be increasingly used in the future necessitating further research, training and evaluation on the topic.
The research also has implications for a wide range of other human/robot collaborative tasks in non-military settings, such as on the factory floor, in hospitals, or even in the home.