Can replacing human soldiers with robot warriors save lives and make war more humane? We try to find out in this episode. But as we learn, the laws of war are not written in computer code. Modern warfare is not ready for killer robots that "decide" without human input.
"When a robot gets blown up, that's another life saved." - Mark Belanger, iRobot.
In this episode, we hear from the people making the robots as they show off their lethal products. We meet a former fighter pilot who touts the values of automation and likes lawyers sitting side by side with soldiers. Several experts tell us about the terrifying moral risks of letting machines think too far ahead of people in battle. We learn there could be lives to be saved, war could be made less atrocious if -- and it is a huge if -- the technology can advance side by side with the antiquated laws. In the end, we hear from the activists who want autonomous lethal weapons banned before they march on the enemy. A U.N. body has just begun to consider it.
Quotes heard in this episode:
- "Maybe we can make war -- as horrible as it sounds -- less devastating to the non-combatants than it currently is." -Ronald Arkin director of the Mobile Robot Lab at Georgia Tech
- When to unleash the machines: "They must do better than human beings before they should be deployed in the battlefield." -Ronald Arkin
- On why Las Vegas could be considered a target: "With Napoleonic-era combat, you knew where the battlefield was, right? With modern warfare, modern conflict, you really don't know, where the battlefield is." -Brad Allenby, Arizona State University
- "Robotics has been trying to do visual recognition for... a bit more than 50 years and we can just about tell the difference between a lion and a car. So the idea of putting one of these things onto a battlefield... and thinking it should discriminate between [innocent people] and insurgents is just insane." -Noel Sharkey Professor of Artificial Intelligence and Robotics at the University of Sheffield in the U.K.
- "In today's warfare, a drone pilot is looking on a screen, talking to potentially five to ten other people looking at that same screen, one of which is a lawyer." -Missy Cummings Duke professor and former fighter pilot
- About autonomous lethal weapons: "These machines for the foreseeable future would fail to meet the requirements of international law." -Peter Asaro, International Committee for Robot Arms Control
- "The preemptive ban is the only thing that makes sense." -Stephen Goose, of Human Rights Watch
If you like this episode why not share it with that friend of yours who always posts about military issues?