Who’s calling the shots on lethal autonomous weapons systems?
Killer, robotic weapons with artificial intelligence are already being made, but there are no global rules over their use, warn tech and rights experts
Nairobi — Countries are rapidly developing “killer robots” — machines with artificial intelligence (AI) that independently kill — but are moving at a snail’s pace on agreeing global rules over their use in future wars, warn technology and human rights experts.
From drones and missiles to tanks and submarines, semi-autonomous weapons systems have been used for decades to eliminate targets in modern day warfare — but they all have human supervision.
Nations such as the US, Russia and Israel are now investing in developing lethal autonomous weapons systems (LAWS) which can identify, target, and kill a person on their own — but to date there are no international laws governing their use.
“Some kind of human control is necessary ... Only humans can make context-specific judgments of distinction, proportionality and precautions in combat,” said Peter Maurer, president of the International Committee of the Red Cross (ICRC).
“[Building consensus] is the big issue we are dealing with and, unsurprisingly, those who have today invested a lot of capacities and do have certain skills that promise advantages to them, are more reluctant than those who don’t.”
The ICRC oversaw the adoption of the 1949 Geneva Conventions that define the laws of war and the rights of civilians to protection and assistance during conflicts and it engages with governments to adapt these rules to modern warfare.
AI researchers, defence analysts and roboticists say LAWS, such as military robots, are no longer confined to the realm of science fiction or video games, but are fast progressing from graphic design boards to defence engineering laboratories.
Within a few years, they could be deployed by state militaries to the battlefield, they add, painting dystopian scenarios of swarms of drones moving through a town or city, scanning and selectively killing their targets within seconds.
Death by algorithm
This has raised ethical concerns from human rights groups and some tech experts who say giving machines the power of life and death violates the principles of human dignity.
Not only are LAWS vulnerable to interference and hacking, which would result in increased civilian deaths, they add, but their deployment would raise questions over who would be held accountable in the event of misuse.
“Don’t be mistaken by the nonsense of how intelligent these weapons will be,” said Noel Sharkey, chair of the International Committee for Robot Arms Control and an AI and robotics expert at Britain’s University of Sheffield. “You simply can’t trust an algorithm — no matter how smart — to seek out, identify and kill the correct target, especially in the complexity of war.”
Experts in defence-based AI systems argue that such weapons, if developed well, can make war more humane. They will be more precise and efficient, not fall prey to human emotions, such as fear or vengeance, and minimise deaths of civilians and soldiers.
“From a military’s perspective, the primary concern is to protect the security of the country with the least amount of lives lost — and that means its soldiers,” said Anuj Sharma, chair of India Research Centre, which works on AI warfare. “So if you can remove the human out of the equation as much as possible, it’s a win because it means less body bags going back home — and that’s what everyone wants.”
A 2019 survey by Human Rights Watch (HRW) and the Campaign to Stop Killer Robots, a global coalition, found 61% of people across 26 countries, including the US and Israel, opposed the development of fully autonomous lethal weapons.
Mary Wareham, HRW’s arms division advocacy director, said countries have held eight meetings under the UN convention on certain conventional weapons since 2014 to discuss the issue, but there has been no progress.
Thirty nations, including Brazil, Austria and Canada, are in favour of a total ban, while dozens of others want a treaty to establish some form of control over the use of LAWS, said Wareham, who is also co-ordinator of the Campaign to Stop Killer Robots. This is largely due to a few, powerful nations, such as the US and Russia, who say it is premature to move towards regulation, without first defining such weapons, she said.
The US state department said it supports the discussions at the UN, but that “dictating a particular format for an outcome before working through the substance” would not result in the best outcome.
“The US has opposed calls to develop a ban and does not support opening negotiations, whether on a legally binding instrument or a political declaration, at this time,” said a US state department spokesperson in a statement. “We must not be anti-technology and must be cautious not to make hasty judgments about emerging or future technologies, especially given how ‘smart’, precision-guided weapons have allowed responsible militaries to reduce risks to civilians in military operations.”
Officials from Russia’s foreign ministry did not immediately respond to requests for comment.
“It’s not good enough that we moving forward into the 21st century without set rules,” said Wareham, adding that without oversight, more nations would develop lethal autonomous weapons. “We need to have a new international treaty as we have for landmines and cluster munitions. We have to prevent the avoidable tragedy that is coming if we do not regulate our killer robots.”
Thomson Reuters Foundation
Would you like to comment on this article or view other readers' comments?
Register (it’s quick and free) or sign in now.
Please read our Comment Policy before commenting.