Countries are rapidly developing “killer robots” – machines with artificial intelligence (AI) that independently kill – but are moving at a snail’s pace on agreeing global rules over their use in future wars, warn technology and human rights experts.
From drones and missiles to tanks and submarines, semi-autonomous weapons systems have been used for decades to eliminate targets in modern day warfare – but they all have human supervision.
Nations such as the United States, Russia and Israel are now investing in developing lethal autonomous weapons systems (LAWS) which can identify, target, and kill a person all on their own – but to date there are no international laws governing their use.
“Some kind of human control is necessary … Only humans can make context-specific judgements of distinction, proportionality and precautions in combat,” said Peter Maurer, President of the International Committee of the Red Cross (ICRC).
“(Building consensus) is the big issue we are dealing with and unsurprisingly, those who have today invested a lot of capacities and do have certain skill which promise advantages to them, are more reluctant than those who don’t.”
The ICRC oversaw the adoption of the 1949 Geneva Conventions that define the laws of war and the rights of civilians to protection and assistance during conflicts and it engages with governments to adapt these rules to modern warfare.
AI researchers, defence analysts and roboticists say LAWS such as military robots are no longer confined to the realm of science fiction or video games, but are fast progressing from graphic design boards to defence engineering laboratories.
Within a few years, they could be deployed by state militaries to the battlefield, they add, painting dystopian scenarios of swarms of drones moving through a town or city, scanning and selectively killing their targets within seconds.
This has raised ethical concerns from human rights groups and some tech experts who say giving machines the power of life and death violates the principles of human dignity.
Not only are LAWS vulnerable to interference and hacking which would result in increased civilian deaths, they add, but their deployment would raise questions over who would be held accountable in the event of misuse.
“Don’t be mistaken by the nonsense of how intelligent these weapons will be,” said Noel Sharkey, chairman of the International Committee for Robot Arms Control.
“You simply can’t trust an algorithm – no matter how smart – to seek out, identify and kill the correct target, especially in the complexity of war,” said Sharkey, who is also an AI and robotics expert at Britain’s University of Sheffield.
Experts in defence-based AI systems argue such weapons, if developed well, can make war more humane.
They will be more precise and efficient, not fall prey to human emotions such as fear or vengeance and minimise deaths of civilians and soldiers, they add.
“From a military’s perspective, the primary concern is to protect the security of the country with the least amount of lives lost – and that means its soldiers,” said Anuj Sharma, chairman of India Research Centre, which works on AI warfare.
“So if you can remove the human out of the equation as much as possible, it’s a win because it means less body bags going back home – and that’s what everyone wants.”
A 2019 survey by Human Rights Watch (HRW) and the Campaign to Stop Killer Robots, a global coalition, found 61% of people across 26 countries, including the United States and Israel, opposed the development of fully autonomous lethal weapons.
Mary Wareham, HRW’s arms division advocacy director, said countries have held eight meetings under the United Nations Convention on Certain Conventional Weapons since 2014 to discuss the issue, but there has been no progress.
Thirty nations including Brazil, Austria and Canada are in favour of a total ban, while dozens of others want a treaty to establish some form of control over the use of LAWS, said Wareham, also coordinator of the Campaign to Stop Killer Robots.
This is largely due to a few, powerful nations like the United States and Russia who say it is premature to move towards regulation, without first defining such weapons, she said.
The US State Department said it supported the discussions at the U.N., but that “dictating a particular format for an outcome before working through the substance” would not result in the best outcome.
“The United States has opposed calls to develop a ban and does not support opening negotiations, whether on a legally binding instrument or a political declaration, at this time,” said a U.S. state department spokesperson in a statement.
“We must not be anti-technology and must be cautious not to make hasty judgments about emerging or future technologies especially given how ‘smart’ precision-guided weapons have allowed responsible militaries to reduce risks to civilians in military operations,” it added.
Officials from Russia’s foreign ministry did not immediately respond to requests for comment.
“It’s not good enough that we moving forward into the 21st century without set rules,” said Wareham, adding that without oversight, more nations would develop lethal autonomous weapons.
“We need to have a new international treaty as we have for landmines and cluster munitions. We have to prevent the avoidable tragedy that is coming if we do not regulate our killer robots.”