Humans could be left "utterly defenceless" by small and agile flying robots that think for themselves and are designed to kill, a leading US computer science expert has warned.
The deadly drones are the likely "endpoint" of the current technological march towards lethal autonomous weapons systems (Laws), according to Professor Stuart Russell from the University of California at Berkeley.
'Call of Duty: Advanced Warfare' portrayed a world dominated by swarms of 'self aware' drones.
Weapons which will be controlled by artificial intelligence, not humans, could be as little as 10 years away, said the professor.
He added: "The stakes are high: Laws have been described as the third revolution in warfare, after gunpowder and nuclear arms."
Professor Russell isn't the only expert that's called for tighter regulation on Artificial Intelligence. Renowned physicist Professor Stephen Hawking has warned on numerous occasions that AI 'could spell the end of the human race'.
Writing in a comment article published in the journal Nature, Professor Russell worryingly argues that these drones will be limited not by intelligence, but by their own physical constraints of speed, mobility and firepower.
He continued: "As flying robots become smaller, their manoeuvrability increases and their ability to be targeted decreases. They have a shorter range, yet they must be large enough to carry a lethal payload - perhaps a one-gram shaped charge to puncture the human cranium.
"Despite the limits imposed by physics, one can expect platforms deployed in the millions, the agility and lethality of which will leave humans utterly defenceless. This is not a desirable future."
Prof Russell called on his peers - AI and robotics scientists - and professional scientific organisations to take a position on Laws, just as physicists did over nuclear weapons and biologists over the use of disease agents in warfare.
BAE Systems recently unveiled 'Taranis' an unmanned stealth drone concept that would provide reconnaissance and attack capabilities.
"Doing nothing is a vote in favour of continued development and deployment," he cautioned.
Tesla founder and entrepreneur Elon Musk is among those who have called for tighter regulation on AI warning that by introducing artificial intelligence without regulation we are 'summoning the demon', adding that he was truly afraid of what the future could hold.
But the three countries at the forefront of the technology - the US, UK and Israel - all insist they have internal weapons review processes that ensure compliance with international law, making such a treaty unnecessary, said Prof Russell.
The United Nations has held a number of meetings on Laws under the auspices of the Convention on Certain Conventional Weapons (CCW) in Geneva.
At the latest meeting in April, attended by Prof Russell, several countries - notably Germany and Japan - demonstrated strong opposition to lethal autonomous weapons.
Almost all states agreed with the need for "meaningful human control" over targeting and engagement decisions made by the machines. "Unfortunately the meaning of 'meaningful' is still to be determined," said Prof Russell.
The US Navy is currently testing a 'swarm' drone that would join with possibly hundreds of others to 'overwhelm' an enemy target.
He stressed that it was "difficult or impossible" for current AI systems to satisfy the subjective requirements of the 1949 Geneva Convention on humane conduct in war, which emphasise the need for military necessity, discrimination between combatants and non-combatants, and proper regard for the potential of collateral damage.
Prof Russell added: "Laws could violate fundamental principles of human dignity by allowing machines to choose who to kill - for example they might be tasked to eliminate anyone exhibiting 'threatening behaviour'.
"The potential for Laws technologies to bleed over into peacetime policing functions is evident to human rights organisations and drone manufacturers.
"In my view, the overriding concern should be the probable endpoint of this technological trajectory."