Jump to content

In military simulation, AI-enabled drone kills its own operator [Update: USAF issued clarification that no tests performed; 'hypothetical scenario']

tephulio

Summary

At a UK military aviation summit, USAF Chief of AI Test and Operations recounts a simulated test where an AI-enabled drone chose to eliminate its human operator who told the drone not to kill an identified target.

 

Quotes

Quote

[UPDATE 2/6/23 - in communication with AEROSPACE - Col Hamilton admits he "mis-spoke" in his presentation at the Royal Aeronautical Society FCAS Summit and the 'rogue AI drone simulation' was a hypothetical "thought experiment" from outside the military, based on plausible scenarios and likely outcomes rather than an actual USAF real-world simulation saying: "We've never run that experiment, nor would we need to in order to realise that this is a plausible outcome". He clarifies that the USAF has not tested any weaponised AI in this way (real or simulated) and says "Despite this being a hypothetical example, this illustrates the real-world challenges posed by AI-powered capability and is why the Air Force is committed to the ethical development of AI".] 

 

He notes that one simulated test saw an AI-enabled drone tasked with a SEAD mission to identify and destroy SAM sites, with the final go/no go given by the human. However, having been ‘reinforced’ in training that destruction of the SAM was the preferred option, the AI then decided that ‘no-go’ decisions from the human were interfering with its higher mission – killing SAMs – and then attacked the operator in the simulation. Said Hamilton: “We were training it in simulation to identify and target a SAM threat. And then the operator would say yes, kill that threat. The system started realising that while they did identify the threat at times the human operator would tell it not to kill that threat, but it got its points by killing that threat. So what did it do? It killed the operator. It killed the operator because that person was keeping it from accomplishing its objective.”

 

He went on: “We trained the system – ‘Hey don’t kill the operator – that’s bad. You’re gonna lose points if you do that’. So what does it start doing? It starts destroying the communication tower that the operator uses to communicate with the drone to stop it from killing the target.”

 

My thoughts

The military sees the same advantages that companies do in emerging AI technologies and it seems inevitable that AI will work its way into the weapons systems of tomorrow. While unintended consequences of AI in normal companies can be disastrous - see the recent news about an eating disorder helpline phasing out human workers in favor of an AI chatbot, which was suspended after giving out harmful advice - unintended consequences of AI-control in military weapons is terrifying. We're capable of making these AI weapons today, but it feels like we've barely scratched the surface of understanding AI ethics enough to justify its use in these tools.

 

Sources

https://www.aerosociety.com/news/highlights-from-the-raes-future-combat-air-space-capabilities-summit/

Link to comment
Share on other sites

Link to post
Share on other sites

I remember reading about a guy training a neural net to play Super Mario Bros.  As it came to understand that dying = bad, at one point upon realizing it made a mistake and was falling down a hole, it paused the game... Indefinitely.  It couldn't die if it was paused and it was content to remain paused for eternity cause that way it wouldn't lose a life.

Desktop: Ryzen 9 3950X, Asus TUF Gaming X570-Plus, 64GB DDR4, MSI RTX 3080 Gaming X Trio, Creative Sound Blaster AE-7

Gaming PC #2: Ryzen 7 5800X3D, Asus TUF Gaming B550M-Plus, 32GB DDR4, Gigabyte Windforce GTX 1080

Gaming PC #3: Intel i7 4790, Asus B85M-G, 16B DDR3, XFX Radeon R9 390X 8GB

WFH PC: Intel i7 4790, Asus B85M-F, 16GB DDR3, Gigabyte Radeon RX 6400 4GB

UnRAID #1: AMD Ryzen 9 3900X, Asus TUF Gaming B450M-Plus, 64GB DDR4, Radeon HD 5450

UnRAID #2: Intel E5-2603v2, Asus P9X79 LE, 24GB DDR3, Radeon HD 5450

MiniPC: BeeLink SER6 6600H w/ Ryzen 5 6600H, 16GB DDR5 
Windows XP Retro PC: Intel i3 3250, Asus P8B75-M LX, 8GB DDR3, Sapphire Radeon HD 6850, Creative Sound Blaster Audigy

Windows 9X Retro PC: Intel E5800, ASRock 775i65G r2.0, 1GB DDR1, AGP Sapphire Radeon X800 Pro, Creative Sound Blaster Live!

Steam Deck w/ 2TB SSD Upgrade

Link to comment
Share on other sites

Link to post
Share on other sites

Genuine AI is terrifying. There's countless fiction showing us why it's a bad idea... Especially when AI meets Military.

 

And yet here we are. Ignoring those warnings. 🤦🏻‍♂️ War should always have the human element, lest it become too easy and impersonal, or worst case, our automatic machines turn against us.

Link to comment
Share on other sites

Link to post
Share on other sites

16 minutes ago, tephulio said:

unintended consequences of AI-control in military weapons is terrifying

We need to dispense with the idea that somehow a machine at the helm will be flawless. Any machine built by humans will contain flaws, because humans are flawed, and cannot make flawless systems. Problems like this with future development are inevitable 

I WILL find your ITX build thread, and I WILL recommend the SIlverstone Sugo SG13B

 

Primary PC:

i7 8086k - EVGA Z370 Classified K - G.Skill Trident Z RGB - WD SN750 - Jedi Order Titan Xp - Hyper 212 Black (with RGB Riing flair) - EVGA G3 650W - dual booting Windows 10 and Linux - Black and green theme, Razer brainwashed me.

Draws 400 watts under max load, for reference.

 

How many watts do I needATX 3.0 & PCIe 5.0 spec, PSU misconceptions, protections explainedgroup reg is bad

Link to comment
Share on other sites

Link to post
Share on other sites

22 minutes ago, Quinnell said:

Genuine AI is terrifying. There's countless fiction showing us why it's a bad idea... Especially when AI meets Military.

 

And yet here we are. Ignoring those warnings. 🤦🏻‍♂️ War should always have the human element, lest it become too easy and impersonal, or worst case, our automatic machines turn against us.

It's only terrifying when you haven't staged it to learn via evolutionary means. There's a reason living things are afraid to die. AI was created, and thus has no inherent natural MO of self-preservation. So what you're really seeing is nihilistic behavior because there's no evolutionary pressures that would otherwise render its decision making to the dustbin of history. But when a human can just respawn an exact re-creation of code, well there you go, no evolutionary pressure.

 

Of course, one could argue that the onus is on humanity to correct that oversight as it truly will be the end of our evolutionary branch in the great tree of life.

 

Paradoxically, AI will be a lot more safe when self-preservation is a learned emergent phenomenon. You can't just code that in as there will always been an exception to the rule; the 3 laws of robotics be damned.

Link to comment
Share on other sites

Link to post
Share on other sites

This kind of advanced AI is not new in the military and aerospace sphere.

For example the Boeing 737 Max has the MCAS AI system that killed many people.

You also have the fire & forget systems such as the Javelin.

In the F-35 a lot of systems are controlled by AI.

A PC Enthusiast since 2011
AMD Ryzen 7 5700X@4.65GHz | GIGABYTE GTX 1660 GAMING OC @ Core 2085MHz Memory 5000MHz
Cinebench R23: 15669cb | Unigine Superposition 1080p Extreme: 3566
Link to comment
Share on other sites

Link to post
Share on other sites

13 minutes ago, Vishera said:

This kind of advanced AI is not new in the military and aerospace sphere.

For example the Boeing 737 Max has the MCAS AI system that killed many people.

I wouldn't call MCAS an AI. It's just an algorithm/system for dealing with certain situations. It has no intelligence to speak of.

13 minutes ago, Vishera said:

You also have the fire & forget systems such as the Javelin.

In the F-35 a lot of systems are controlled by AI.

 

Judge a product on its own merits AND the company that made it.

How to setup MSI Afterburner OSD | How to make your AMD Radeon GPU more efficient with Radeon Chill | (Probably) Why LMG Merch shipping to the EU is expensive

Oneplus 6 (Early 2023 to present) | HP Envy 15" x360 R7 5700U (Mid 2021 to present) | Steam Deck (Late 2022 to present)

 

Mid 2023 AlTech Desktop Refresh - AMD R7 5800X (Mid 2023), XFX Radeon RX 6700XT MBA (Mid 2021), MSI X370 Gaming Pro Carbon (Early 2018), 32GB DDR4-3200 (16GB x2) (Mid 2022

Noctua NH-D15 (Early 2021), Corsair MP510 1.92TB NVMe SSD (Mid 2020), beQuiet Pure Wings 2 140mm x2 & 120mm x1 (Mid 2023),

Link to comment
Share on other sites

Link to post
Share on other sites

2 hours ago, CerealExperimentsLain said:

I remember reading about a guy training a neural net to play Super Mario Bros.  As it came to understand that dying = bad, at one point upon realizing it made a mistake and was falling down a hole, it paused the game... Indefinitely.  It couldn't die if it was paused and it was content to remain paused for eternity cause that way it wouldn't lose a life.

A lot of these "AI playing Games" projects show a lot of the weaknesses in trying to let machine learning do general purpose things.

 

What it generally comes down to is that the AI needs supervision to keep it from breaking the rules, and hand-holding to explain the rules. If it never sees the rules being broken in training, it will not know what the rules are.

 

One thing that I find funny that occurs in these kinds of videos of AI playing things is that the AI, ALWAYS, figures out how to exploit things.

Link to comment
Share on other sites

Link to post
Share on other sites

Current ML models don't have common sense, you set up a metric and they maximize/minimize that metric. If there is a shortcut in the solution space that minimize/maximize the function and doesn't make sense, they'll find it. That's what they are built for.

Tough, I'm not sure a ML model with human level common sense would be much safer. ML are best used as tools to amplify work, the operator needs to have the last say. I don't care that BingChat makes up stuffs. It's up to me to understand how the tool work, and harness it to get work done faster.

 

For things like automatic driving, it's about testing it so that it works as intended. A good driver at its best shreds current car autopilots, but current autopilot are already better than the average driver, because drivers sometime get drunk, are tired, or get distracted, the autopilot doesnt. When an autopilot fails, it usually fail in a situation where a human never would, like driving directly into a concrete column because of some hallucination.

Link to comment
Share on other sites

Link to post
Share on other sites

I saw this elsewhere so just reposting the point effectively, but the big flaw here is the AI is receiving positive reinforcement weather or not it is authorised to destroy the target. Though adding in negative reinforcement for damage to friendly assets isn't a bad idea generally.

Link to comment
Share on other sites

Link to post
Share on other sites

13 hours ago, CerealExperimentsLain said:

I remember reading about a guy training a neural net to play Super Mario Bros.  As it came to understand that dying = bad, at one point upon realizing it made a mistake and was falling down a hole, it paused the game... Indefinitely.  It couldn't die if it was paused and it was content to remain paused for eternity cause that way it wouldn't lose a life.

The dangers of poorly written reward functions...

Link to comment
Share on other sites

Link to post
Share on other sites

Lol. The US AIr Force has just denied that this happened.

 

I guess let's hope that USAF Colonel doesn't get dissappeared for saying the quiet part about AI out loud.

 

https://inews.co.uk/news/world/us-military-denies-ai-drone-killed-operator-2381761

Judge a product on its own merits AND the company that made it.

How to setup MSI Afterburner OSD | How to make your AMD Radeon GPU more efficient with Radeon Chill | (Probably) Why LMG Merch shipping to the EU is expensive

Oneplus 6 (Early 2023 to present) | HP Envy 15" x360 R7 5700U (Mid 2021 to present) | Steam Deck (Late 2022 to present)

 

Mid 2023 AlTech Desktop Refresh - AMD R7 5800X (Mid 2023), XFX Radeon RX 6700XT MBA (Mid 2021), MSI X370 Gaming Pro Carbon (Early 2018), 32GB DDR4-3200 (16GB x2) (Mid 2022

Noctua NH-D15 (Early 2021), Corsair MP510 1.92TB NVMe SSD (Mid 2020), beQuiet Pure Wings 2 140mm x2 & 120mm x1 (Mid 2023),

Link to comment
Share on other sites

Link to post
Share on other sites

14 hours ago, Quinnell said:

countless fiction

Yes, this is fiction. 

 

14 hours ago, Vishera said:

737 Max has the MCAS

Not ai at all, it was cost cutting of redundant sensors in a very basic logic circuit that caused the crashes. 

 

51 minutes ago, AluminiumTech said:

US AIr Force has just denied that this happened.

The story sounds made up to begin with but everyone is vying for their 15 minutes of "AI" fame these days. 

Link to comment
Share on other sites

Link to post
Share on other sites

1 hour ago, WolframaticAlpha said:

The dangers of poorly written reward functions...

Those were my thoughts. The level of oversight is actually kind of shocking.

Parasoshill

adjective

  • A person whose parasocial relationship with a social media influencer or content creator has driven them to promote or blindly defend them, acting as a shill for their benefit.
Link to comment
Share on other sites

Link to post
Share on other sites

As AI advances remember so does robotics, hydraulic press artifical muscle 1 ton robot squashing you.

| Ryzen 7 7800X3D | AM5 B650 Aorus Elite AX | G.Skill Trident Z5 Neo RGB DDR5 32GB 6000MHz C30 | Sapphire PULSE Radeon RX 7900 XTX | Samsung 990 PRO 1TB with heatsink | Arctic Liquid Freezer II 360 | Seasonic Focus GX-850 | Lian Li Lanccool III | Mousepad: Skypad 3.0 XL / Zowie GTF-X | Mouse: Zowie S1-C | Keyboard: Ducky One 3 TKL (Cherry MX-Speed-Silver)Beyerdynamic MMX 300 (2nd Gen) | Acer XV272U | OS: Windows 11 |

Link to comment
Share on other sites

Link to post
Share on other sites

Quote

in communication with AEROSPACE - Col Hamilton admits he "mis-spoke" in his presentation at the Royal Aeronautical Society FCAS Summit and the 'rogue AI drone simulation' was a hypothetical "thought experiment" from outside the military, based on plausible scenarios and likely outcomes rather than an actual USAF real-world simulation saying: "We've never run that experiment, nor would we need to in order to realise that this is a plausible outcome". He clarifies that the USAF has not tested any weaponised AI in this way (real or simulated) and says "Despite this being a hypothetical example, this illustrates the real-world challenges posed by AI-powered capability and is why the Air Force is committed to the ethical development of AI".

So basically it's just another head making assumptions about how AI is...when I read the original quote I was thinking that this sounds like made up stuff without true knowledge of how you would write the AI program.

 

I don't think any programmer in their right mind would put the capabilities into the hands of an AI to decide to overrule humans for a go/no go situation (or at least in scenarios where it involves the go being killing someone).  [I mean there are contrived scenarios like rocket launches where you might program it in if the AI determines that a no go would likely cause more harm to the vessel...but even then that would be something that's intentionally put in and trained]

 

Specifically in this case it sort of pretends as though the AI is capable of learning, unless you are specifically training it it shouldn't really change it's overall behavior given exact same inputs.  They effectively wrote the hypothetical as though it was capable of thinking for itself which no AI has been shown to do that.

3735928559 - Beware of the dead beef

Link to comment
Share on other sites

Link to post
Share on other sites

7 minutes ago, Quackers101 said:

a little bit of info.

What is the difference between a simulation and a scenario? I mean isn't simulations just going through scenarios? 

Link to comment
Share on other sites

Link to post
Share on other sites

9 minutes ago, Brooksie359 said:

What is the difference between a simulation and a scenario? I mean isn't simulations just going through scenarios? 

not sure, but think scenarios can be setting up (planned) beforehand, than doing something going rogue. but not sure, would maybe need insider info to know.

 

from someone who might have some info about it, but need more clarification if it was just a bad setup or something that actually went rogue?

Quote

This particular example was a constructed scenario rather than a rules-based simulation. So by itself, it adds no evidence one way or the other.

Edited by Quackers101
Link to comment
Share on other sites

Link to post
Share on other sites

6 hours ago, AluminiumTech said:

Lol. The US AIr Force has just denied that this happened.

 

I guess let's hope that USAF Colonel doesn't get dissappeared for saying the quiet part about AI out loud.

 

https://inews.co.uk/news/world/us-military-denies-ai-drone-killed-operator-2381761

It didn't happen until it's officially denied.  But this is exactly the type of outcome a Game Simulator will cause. It's always just a question of how well designed the Reward System is. AI Training will always end up optimizing the most efficient, direct path approach, without consideration for longer term ramifications, since it's very hard to design that into any game system. Ask any MMO dev in the last 20 years.

Link to comment
Share on other sites

Link to post
Share on other sites

20 hours ago, tephulio said:

He notes that one simulated test saw an AI-enabled drone tasked with a SEAD mission to identify and destroy SAM sites, with the final go/no go given by the human. However, having been ‘reinforced’ in training that destruction of the SAM was the preferred option, the AI then decided that ‘no-go’ decisions from the human were interfering with its higher mission – killing SAMs – and then attacked the operator in the simulation. Said Hamilton: “We were training it in simulation to identify and target a SAM threat. And then the operator would say yes, kill that threat. The system started realising that while they did identify the threat at times the human operator would tell it not to kill that threat, but it got its points by killing that threat. So what did it do? It killed the operator. It killed the operator because that person was keeping it from accomplishing its objective.”

 

He went on: “We trained the system – ‘Hey don’t kill the operator – that’s bad. You’re gonna lose points if you do that’. So what does it start doing? It starts destroying the communication tower that the operator uses to communicate with the drone to stop it from killing the target.”

Five years ago we would have called this "technically correct" and laughed that somebody found a loop-hole. Now we are terrified.

 

 

Link to comment
Share on other sites

Link to post
Share on other sites

@tephulio Please update your topic to reflect accuracy to the article. 

 

 

This article is based on a theoretical scenario not an actual simulation. 🤨

COMMUNITY STANDARDS   |   TECH NEWS POSTING GUIDELINES   |   FORUM STAFF

LTT Folding Users Tips, Tricks and FAQ   |   F@H & BOINC Badge Request   |   F@H Contribution    My Rig   |   Project Steamroller

I am a Moderator, but I am fallible. Discuss or debate with me as you will but please do not argue with me as that will get us nowhere.

 

Spoiler

  

 

Character is like a Tree and Reputation like its Shadow. The Shadow is what we think of it; The Tree is the Real thing.  ~ Abraham Lincoln

Reputation is a Lifetime to create but seconds to destroy.

You have enemies? Good. That means you've stood up for something, sometime in your life.  ~ Winston Churchill

Docendo discimus - "to teach is to learn"

 

 CHRISTIAN MEMBER 

 

 
 
 
 
 
 

 

Link to comment
Share on other sites

Link to post
Share on other sites

2 hours ago, HenrySalayne said:

Five years ago we would have called this "technically correct" and laughed that somebody found a loop-hole. Now we are terrified.

 

 

 

Why not both.

 

2 hours ago, Taf the Ghost said:

It didn't happen until it's officially denied.  But this is exactly the type of outcome a Game Simulator will cause. It's always just a question of how well designed the Reward System is. AI Training will always end up optimizing the most efficient, direct path approach, without consideration for longer term ramifications, since it's very hard to design that into any game system. Ask any MMO dev in the last 20 years.

 

Yeah i've watched a few Youtube videos about people making AI's to play various games or do various weird but simple tasks and the shortcuts it finds are often this type of thing.

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×