Sunday, October 6, 2024
Contact    |    RSS icon Twitter icon Facebook icon  
Unexplained Mysteries
You are viewing: Home > News > Science & Technology > News story
Welcome Guest ( Login or Register )  
All ▾
Search Submit

Science & Technology

What will autonomous killer robots mean for the future of warfare ?

January 10, 2023 · Comment icon 4 comments

Will we ever see something like The Terminator ? Image Credit: CC BY-SA 2.0 Daniel Jurena
Could the T-800 become a reality and how will countries respond to the threat posed by autonomous killing machines ?
You might have heard of killer robots, slaughterbots or terminators - officially called lethal autonomous weapons (LAWs) - from films and books. And the idea of super-intelligent weapons running rampant is still science fiction. But as AI weapons become increasingly sophisticated, public concern is growing over fears about lack of accountability and the risk of technical failure.

Already we have seen how so-called neutral AI have made sexist algorithms and inept content moderation systems, largely because their creators did not understand the technology. But in war, these kinds of misunderstandings could kill civilians or wreck negotiations.

For example, a target recognition algorithm could be trained to identify tanks from satellite imagery. But what if all of the images used to train the system featured soldiers in formation around the tank? It might mistake a civilian vehicle passing through a military blockade for a target.

Why do we need autonomous weapons?

Civilians in many countries (such as Vietnam, Afghanistan and Yemen) have suffered because of the way global superpowers build and use increasingly advanced weapons. Many people would argue they have done more harm than good, most recently pointing to the Russian invasion of Ukraine early in 2022.

In the other camp are people who say a country must be able to defend itself, which means keeping up with other nations' military technology. AI can already outsmart humans at chess and poker. It outperforms humans in the real world too. For example Microsoft claims its speech recognition software has an error rate of 1% compared to the human error rate of around 6%. So it is hardly surprising that armies are slowly handing algorithms the reins.

But how do we avoid adding killer robots to the long list of things we wish we had never invented? First of all: know thy enemy.

What are Lethal Autonomous Weapons (LAWs)?

The US Department of Defence defines an autonomous weapon system as: "A weapon system that, once activated, can select and engage targets without further intervention by a human operator."

Many combat systems already fit this criteria. The computers on drones and modern missiles have algorithms that can detect targets and fire at them with far more precision than a human operator. Israel's Iron Dome is one of several active defence systems that can engage targets without human supervision.

While designed for missile defence, the Iron Dome could kill people by accident. But the risk is seen as acceptable in international politics because the Iron Dome generally has a reliable history of protecting civilian lives.

There are AI enabled weapons designed to attack people too, from robot sentries to loitering kamikaze drones used in the Ukraine war. LAWs are already here. So, if we want to influence the use of LAWs, we need to understand the history of modern weapons.

The rules of war
International agreements, such as the Geneva conventions establish conduct for the treatment of prisoners of war and civilians during conflict. They are one of the few tools we have to control how wars are fought. Unfortunately, the use of chemical weapons by the US in Vietnam, and by Russia in Afghanistan, are proof these measures aren't always successful.

Worse is when key players refuse to sign up. The International Campaign to Ban Landmines (ICBL) has been lobbying politicians since 1992 to ban mines and cluster munitions (which randomly scatter small bombs over a wide area). In 1997 the Ottawa treaty included a ban of these weapons, which 122 countries signed. But the US, China and Russia didn't buy in.

Landmines have injured and killed at least 5,000 soldiers and civilians per year since 2015 and as many as 9,440 people in 2017. The Landmine and Cluster Munition Monitor 2022 report said:

"Casualties...have been disturbingly high for the past seven years, following more than a decade of historic reductions. The year 2021 was no exception. This trend is largely the result of increased conflict and contamination by improvised mines observed since 2015. Civilians represented most of the victims recorded, half of whom were children."

Despite the best efforts of the ICBL, there is evidence both Russia and Ukraine (a member of the Ottawa treaty) are using landmines during the Russian invasion of Ukraine. Ukraine has also relied on drones to guide artillery strikes, or more recently for "kamikaze attacks" on Russian infrastructure.

Our future

But what about more advanced AI enabled weapons? The Campaign to Stop Killer Robots lists nine key problems with LAWs, focusing on the lack of accountability, and the inherent dehumanisation of killing that comes with it.

While this criticism is valid, a full ban of LAWs is unrealistic for two reasons. First, much like mines, pandora's box has already been opened. Also the lines between autonomous weapons, LAWs and killer robots are so blurred it's difficult to distinguish between them. Military leaders would always be able to find a loophole in the wording of a ban and sneak killer robots into service as defensive autonomous weapons. They might even do so unknowingly.

We will almost certainly see more AI enabled weapons in the future. But this doesn't mean we have to look the other way. More specific and nuanced prohibitions would help keep our politicians, data scientists and engineers accountable.

For example, by banning:
  • black box AI: systems where the user has no information about the algorithm beyond inputs and outputs
  • unreliable AI: systems that have been poorly tested (such as in the military blockade example mentioned previously).
And you don't have to be an expert in AI to have a view on LAWs. Stay aware of new military AI developments. When you read or hear about AI being used in combat, ask yourself: is it justified? Is it preserving civilian life? If not, engage with the communities that are working to control these systems. Together, we stand a chance at preventing AI from doing more harm than good.

Jonathan Erskine, PhD Student, Interactive AI, University of Bristol and Miranda Mowbray, Lecturer in Interactive AI, University of Bristol

This article is republished from The Conversation under a Creative Commons license.

Read the original article. The Conversation

Source: The Conversation | Comments (4)




Other news and articles
Recent comments on this story
Comment icon #1 Posted by mdbuilder 2 years ago
It means no one has to get their hands dirty or wear those silly uniforms?
Comment icon #2 Posted by Desertrat56 2 years ago
I want to know what it means for peace, obviously nothing but fear.  It will be the demise of the human race.
Comment icon #3 Posted by DreadLordAvatar 2 years ago
Until a new power source is invented, all this terminator type AI is fantasy.
Comment icon #4 Posted by DanL 2 years ago
Battle robots should not be allowed!! Any nation that puts them in the field should be outlawed and boycotted totally and restricted from doing any international commerce. If that doesn't work maybe a nuke in each of their major cities would get the point across. Wars end when there are enough loses that one side loses heart and surrenders. Without the human element wars would become endless and sooner or later they are going to invade the homes and cities with those soulless things. It would be like the atomic tombs. The day someone launches one and the bubble is busted there will be radioact... [More]


Please Login or Register to post a comment.


Our new book is out now!
Book cover

The Unexplained Mysteries
Book of Weird News

 AVAILABLE NOW 

Take a walk on the weird side with this compilation of some of the weirdest stories ever to grace the pages of a newspaper.

Click here to learn more

We need your help!
Patreon logo

Support us on Patreon

 BONUS CONTENT 

For less than the cost of a cup of coffee, you can gain access to a wide range of exclusive perks including our popular 'Lost Ghost Stories' series.

Click here to learn more

Top 10 trending mysteries
Recent news and articles