Killer Robots Aren't Science Fiction. A Push to Ban Them is Growing.
It may have seemed like an obscure United Nations conclave, but a meeting this week in Geneva was followed intently by experts in artificial intelligence, military strategy, disarmament and humanitarian law. The reason for the interest? Killer robots — drones, guns and bombs that decide on their own, with artificial brains, whether to attack and kill — and what should be done, if anything, to regulate or ban them. Once the domain of science fiction films like the “Terminator” series and “RoboCop,” killer robots, more technically known as Lethal Autonomous Weapons Systems, have been invented and tested at an accelerated pace with little oversight. Some prototypes have even been used in actual conflicts. Sign up for The Morning newsletter from the New York Times The evolution of these machines is considered a potentially seismic event in warfare, akin to the invention of gunpowder and nuclear bombs. This year, for the first time, a majority of the 125 nations that belong to an agreemen