Given the perceived military success of unmanned drones and other semi-autonomous weapons, many proponents of robotic warfare are pushing for the next phase of development: fully autonomous weapons. Once developed and deployed, these weapons — killer robots, as they have become known — would be able to select their own targets and fire on them, without human intervention.
Monday, in opposition to such developments, Canadians are launching a campaign to stop killer robots.
Joining an international movement of more than 50 non-governmental organizations in 24 countries, Canadian advocates will hold a public meeting Monday and convene with parliamentarians Tuesday, calling for a pre-emptive prohibition on the development, production and use of fully autonomous weapons.
As a concept, the killer robot represents a stark shift in automation policy — a wilful, intentional and unprecedented removal of humans from the kill decision loop.
Proponents of this stark shift in policy believe that the substitution of machines for humans is justified because robot soldiers will outperform human soldiers physically, emotionally and ethically. Robots, they say, are not vulnerable to the perils that plague humans in battlespace: bias, exhaustion, elevated emotions or the need to seek retribution for the death of a comrade. Consequently, proponents believe that robots will better comport with international standards and the ethical rules of just war, since those rules and standards can be programmed into the machines’ operations.
Is this a reasonable argument or just wishful thinking?
The underlying utopic vision for killer robots originated with Aristotle, who imagined in his Politics that we could delegate to automatons all undesirable roles in society, ultimately eliminating the need for slaves, soldiers and the like.
But it was the science fiction of Asimov, many centuries later, that gave hygiene to Aristotle’s idea.
Asimov’s brilliance lay in his recognition that robots might be universally programmed to obey humans. Taking Aristotle to the next level, the elite could not only use robots to their great advantage but the general public could also be made to trust robots as their companions and co-workers. Programming obedience into all robots, coding a kind of slave-morality that would allay people’s fears, would ensure that robots are “more faithful, more useful and absolutely devoted.” It would also avoid what Asimov called humankind’s Frankenstein Complex — “its gut fears that any artificial man they created would turn upon its creator.”
As compelling and reassuring as the Asimovian narrative may be — plain and simple — the deployment of fully autonomous weapons entails that we delegate crucial moral decisions of life and death away from robust human decision-makers in favour of relatively limited software algorithms.
Here, I suggest, the Frankenstein complex is the least of our concerns.
The concern is not a robot uprising. It is the voluntary relinquishment of human control to machines and the dependencies created thereby.
Relinquishment is an emerging topic in the field of roboethics.
However, the relinquishment question is quite different for autonomous weapons than, say, for autonomous vehicles. Whether to relinquish control of the steering wheel may simply turn out to be a matter of safety and efficacy, as determined by evidence over time.
By contrast, the rules of armed conflict require a much more challenging set of threshold determinations. First, we would have to determine that killer robots could successfully discriminate between combatants and non-combatants in the moment of conflict. Second, we would have to determine that killer robots have the ability to morally assess every possible conflict in order to justify whether a particular use of force is proportional. Third, we would have to determine that killer robots are able to assess and comprehend military operations sufficiently well to be able to decide whether the use of force on a particular occasion is of military necessity.
When we consider even these basic threshold determinations, it is not difficult to see that killer robots operate only in the realm of science fiction. The capacities required to comply with international humanitarian law vastly exceed what robots are capable of today and in the foreseeable future.
But even if robots could comply with IHL, killer robots are still a bad idea.
Allowing life-or-death decisions to be made by robots crosses a fundamental moral line. And, although it may seem attractive to send robots rather than people into the line of fire, the ability to do so would make the decision to go to war easier, generating more rather than less armed conflict. Finally, even if killer robots are compliant with IHL, their available use will operate as a force multiplier of perceived military needs, thus skewing the proportionality metric and amplifying new forms of destructive, lethal force.
As we launch the Campaign to Stop Killer Robots, the government of Canada must come to see it has a role to play as well. At a minimum, it should:
- create a national moratorium on the use of killer robots;
declare a commitment to abide by and, if necessary, strengthen IHL and international human rights law in all activities surrounding robot weapons (and implement rigorous processes to ensure compliance);
commit to being as transparent as possible about internal weapons review processes, including metrics used to test robot systems; and
participate in international debate and trans-governmental dialogue, and be prepared to exchange best practices with other states.
A military technology that would seek to achieve humanitarian ends by taking humans out of the kill decision loop is highly suspect. Let’s not build them just because we can. Let’s keep killer robots fiction.
Ian Kerr holds the Canada Research Chair in Ethics, Law and Technology at the University of Ottawa, and is co-author of the forthcoming book Robot Law.
Source: Ottawa Citizen