Warning about race dynamics.

Rob Miles is an AI safety researcher. In early 2023, days after GPT-4’s release, he expressed his fears about arms race dynamics in AI development, in a computerphile video. His worst fears seem to have turned into reality.

Robert Miles

Rob Miles is a science communicator focused on AI Safety and Alignment.
He is probably my biggest influence. He was the first person to bring the complicated AI Safety concepts to the general public, many years ago, way before the LLM revolution started, in a time where today’s reality seemed sci-fi and the whole Alignment Study was an academic curiosity, a cool way for intellectuals to tickle the most rational and analytical parts of their brain.
His Youtube Channel Robert Miles AI Safety and his interviews at Computerphile popularized the frontier ideas coming from LessWrong, Machine Intelligence Research Institute, Centre for the Study of Existential Risk and leading AI Labs like OpenAI and DeepMind.

Robert Miles is a true hero of the AI Safety saga and an inspiring force for lethalintelligence.ai.

AI "Stop Button" Problem - Computerphile

March 3, 2017 4:32 pm

How do you implement an on/off switch on a General Artificial Intelligence? Rob Miles explains the perils.

Part 1:
...

There's No Rule That Says We'll Make It

January 16, 2022 8:36 pm

We're not doomed! But doom is definitely an actual possibility, and we need to act like it.
If you're thinking about working on AI Safety, check
...

Why Do I Avoid Sci-fi?

August 16, 2022 8:30 am

How come I tend to avoid discussing specific scenarios when talking about AI risks?

Intro to AI Safety, Remastered

June 24, 2021 5:25 pm

An introduction to AI Safety, remastered from a talk I gave at "AI and Politics" in London

The second channel:
...

Ep. 136: The case for taking AI Safety seriously | Rob Miles

July 4, 2023 10:00 am

Through a series of popular explainer videos, Rob has become one of the most prominent voices in the AI safety community, exploring topics like ...

ROBERT MILES - "There is a good chance this kills everyone"

May 21, 2023 6:16 pm

Please check out Numerai - our sponsor @
http://numer.ai/mlst

Numerai is a groundbreaking platform which is taking the data science
...

Robert Miles–Youtube, Doom

August 19, 2022 6:00 pm

Robert Miles has been making videos for Computerphile, then decided to create his own Youtube channel about AI Safety. Lately, he's been working on ...

Would You Help an AI to Escape?

July 8, 2022 5:01 pm

Could an AGI get a human to help them, just through a text conversation? Recently we got some new evidence about that question.

Rob Miles - Why should I care about AI safety?

December 2, 2020 5:01 pm

In the latest episode of the TDS podcast, host Jeremie Harris talks with guest Rob Miles about AI safety, AI and the course of human evolution, and ...

Ep 12 - Education & advocacy for AI safety w/ Rob Miles (YouTube host)

March 9, 2024 1:05 am

We speak with Rob Miles. Rob is the host of the “Robert Miles AI Safety” channel on YouTube, the single most popular AI alignment video series out ...

9 Examples of Specification Gaming

April 29, 2020 6:41 pm

AI systems do what you say, and it's hard to say exactly what you mean.
Let's look at a list of real life examples of specification
...

Why Would AI Want to do Bad Things? Instrumental Convergence

March 24, 2018 9:51 pm

How can we predict that AGI with unknown goals would behave badly by default?

The Orthogonality Thesis video:
...

Is AI Safety a Pascal's Mugging?

May 16, 2019 4:11 pm

An event that's very unlikely is still worth thinking about, if the consequences are big enough. What's the limit though?

Do we have to
...

Intelligence and Stupidity: The Orthogonality Thesis

January 11, 2018 9:53 pm

Can highly intelligent agents have stupid goals?
A look at The Orthogonality Thesis and the nature of stupidity.

The 'Stamp Collector'
...

A Response to Steven Pinker on AI

March 31, 2019 3:39 pm

Steven Pinker wrote an article on AI for Popular Science Magazine, which I have some issues with.

The article:
...

What can AGI do? I/O and Speed

October 17, 2017 12:20 pm

Suppose we make an algorithm that implements general intelligence as well as the brain. What could that system do?
It might have better input and
...

Why Not Just: Think of AGI Like a Corporation?

December 23, 2018 10:01 pm

Corporations are kind of like AIs, if you squint. How hard do you have to squint though, and is it worth it?
In this video we ask: Are
...

AI learns to Create ̵K̵Z̵F̵ ̵V̵i̵d̵e̵o̵s̵ Cat Pictures: Papers in Two Minutes #1

October 29, 2017 1:49 pm

Some beautiful new GAN results have been published, so let's have a quick look at the pretty pictures.
More AI Safety coming soon of
...

The other "Killer Robot Arms Race" Elon Musk should worry about

August 22, 2017 1:19 pm

Elon Musk is in the news, talking to the UN about autonomous weapons. This seems like a good time to explain one area where we don't quite agree ...

Why Not Just: Raise AI Like Kids?

July 22, 2017 3:58 pm

Newly made Artificial General Intelligences are basically like children, right? So we already know we can teach them how to behave, right? ...

Are AI Risks like Nuclear Risks?

June 10, 2017 5:22 pm

Concerns about AI cover a really wide range of possible problems. Can we make progress on several of these problems at once?

With thanks to
...

Deadly Truth of General AI? - Computerphile

June 17, 2015 3:09 pm

The danger of assuming general artificial intelligence will be the same as human intelligence. Rob Miles explains with a simple example: The deadly ...

Holy Grail of AI (Artificial Intelligence) - Computerphile

May 1, 2015 5:19 pm

Audible free book: http://www.audible.com/computerphile
Why can't artificial intelligence do what humans can? Rob Miles talks about generality
...

Why Asimov's Laws of Robotics Don't Work - Computerphile

November 6, 2015 3:01 pm

Audible Free Book: http://www.audible.com/computerphile
Three or four laws to make robots and AI safe - should be simple right? Rob Miles on why
...

AI Safety - Computerphile

February 3, 2016 8:22 pm

Safety in AI is important, but more important is to work it out before working out the AI itself. Rob Miles on AI safety.

Brain Scanner:
...

AI's Game Playing Challenge - Computerphile

March 24, 2016 10:18 pm

AlphaGo is beating humans at Go - What's the big deal? Rob Miles explains what AI has to do to play a game.

What on Earth is Recursion?:
...

General AI Won't Want You To Fix its Code - Computerphile

February 28, 2017 9:02 pm

Part 1 of a Series on AI Safety Research with Rob Miles. Rob heads away from his 'Killer Stamp Collector' example to find a more concrete example of ...

AI? Just Sandbox it... - Computerphile

June 23, 2017 4:44 pm

Why can't we just disconnect a malevolent AI? Rob Miles on some of the simplistic solutions to AI safety.

Out of focus shots caused by
...

Using Dangerous AI, But Safely?

November 15, 2024 10:35 pm

Can we keep powerful AI under control, using AI?

The Paper: https://arxiv.org/abs/2312.06942

AI Lab Watch:
...

Using Dangerous AI, But Safely?

November 15, 2024 10:35 pm

Can we keep powerful AI under control, using AI?

The Paper: https://arxiv.org/abs/2312.06942

AI Lab Watch:
...

Learn AI Safety at MATS #shorts

September 28, 2024 12:12 am

Apply for MATS at matsprogram.org by Oct 6

#shorts

Apply to Study AI Safety Now! #shorts

April 28, 2023 6:37 pm

Apply to SERI MATS at https://www.serimats.org/ by May 7th, and check out http://aisafety.training to stay up to date with events and programs!

Why Does AI Lie, and What Can We Do About It?

December 9, 2022 10:10 pm

How do we make sure language models tell the truth?

The new channel!: https://www.youtube.com/@aisafetytalks
Evan Hubinger's Talk:
...

Apply Now for a Paid Residency on Interpretability #short

November 11, 2022 8:07 pm

https://www.redwoodresearch.org/remix
Real main channel video coming soon!

#short #shorts

$100,000 for Tasks Where Bigger AIs Do Worse Than Smaller Ones #short

October 14, 2022 1:05 pm

Check out http://inversescaling.com for more details and to apply

#shorts #short

Free ML Bootcamp for Alignment #shorts

May 24, 2022 7:30 pm

Apply for the second MLAB (Machine Learning for Alignment ...

Win $50k for Solving a Single AI Problem? #Shorts

February 8, 2022 9:17 pm

The Alignment Research Center is offering up to $50k for proposals for their Eliciting Latent Knowledge problem.

The report:
...

Apply to AI Safety Camp! #shorts

November 19, 2021 8:03 pm

Trying out #shorts
Applications are open for next year's AI Safety Camp!
http://aisafety.camp

We Were Right! Real Inner Misalignment

October 10, 2021 10:50 pm

Researchers ran real versions of the thought experiments in the 'Mesa-Optimisers' videos!
What they found won't shock you (if you've been paying
...

Intro to AI Safety, Remastered

June 24, 2021 5:25 pm

An introduction to AI Safety, remastered from a talk I gave at "AI and Politics" in London

The second channel:
...

Deceptive Misaligned Mesa-Optimisers? It's More Likely Than You Think...

May 23, 2021 10:15 pm

The previous video explained why it's *possible* for trained models to end up with the wrong goals, even when we specify the goals perfectly. This ...

The OTHER AI Alignment Problem: Mesa-Optimizers and Inner Alignment

February 16, 2021 8:37 pm

This "Alignment" thing turns out to be even harder than we thought.

# Links
The Paper: https://arxiv.org/pdf/1906.01820.pdf
Discord
...

Quantilizers: AI That Doesn't Try Too Hard

December 13, 2020 10:46 pm

How do you get an AI system that does better than a human could, without doing anything a human wouldn't?

A follow-up to "Maximizers and
...

Sharing the Benefits of AI: The Windfall Clause

July 6, 2020 6:53 pm

AI might create enormous amounts of wealth, but how is it going to be distributed?

The Paper:
...

10 Reasons to Ignore AI Safety

June 4, 2020 5:28 pm

Why do some ignore AI Safety? Let's look at 10 reasons people give (adapted from Stuart Russell's list).

Related Videos from Me:
Why
...

9 Examples of Specification Gaming

April 29, 2020 6:41 pm

AI systems do what you say, and it's hard to say exactly what you mean.
Let's look at a list of real life examples of specification
...

Training AI Without Writing A Reward Function, with Reward Modelling

December 13, 2019 6:39 pm

How do you get a reinforcement learning agent to do what you want, when you can't actually write a reward function that specifies what that ...

AI That Doesn't Try Too Hard - Maximizers and Satisficers

August 23, 2019 5:05 pm

Powerful AI systems can be dangerous in part because they pursue their goals as strongly as they can. Perhaps it would be safer to have systems that ...

Is AI Safety a Pascal's Mugging?

May 16, 2019 4:11 pm

An event that's very unlikely is still worth thinking about, if the consequences are big enough. What's the limit though?

Do we have to
...

A Response to Steven Pinker on AI

March 31, 2019 3:39 pm

Steven Pinker wrote an article on AI for Popular Science Magazine, which I have some issues with.

The article:
...

How to Keep Improving When You're Better Than Any Teacher - Iterated Distillation and Amplification

March 11, 2019 2:14 pm

[2nd upload] AI systems can be trained using demonstrations from experts, but how do you train them to out-perform those experts? Can this still be ...

Why Not Just: Think of AGI Like a Corporation?

December 23, 2018 10:01 pm

Corporations are kind of like AIs, if you squint. How hard do you have to squint though, and is it worth it?
In this video we ask: Are
...

Safe Exploration: Concrete Problems in AI Safety Part 6

September 21, 2018 1:20 pm

To learn, you need to try new things, but that can be risky. How do we make AI systems that can explore safely?

Playlist of the series so
...

Friend or Foe? AI Safety Gridworlds extra bit

June 25, 2018 1:31 am

The last video about the AI Safety Gridworlds paper. How does an agent detect and adapt to friendly and adversarial intentions in the ...

AI Safety Gridworlds

May 25, 2018 6:20 pm

Got an AI safety idea? Now you can test it out! A recent paper from DeepMind sets out some environments for evaluating the safety of AI systems, and ...

Experts' Predictions about the Future of AI

March 31, 2018 2:12 pm

When will AI systems surpass human performance? I don't know, do you? No you don't. Let's see what 352 top AI researchers ...

Why Would AI Want to do Bad Things? Instrumental Convergence

March 24, 2018 9:51 pm

How can we predict that AGI with unknown goals would behave badly by default?

The Orthogonality Thesis video:
...

Superintelligence Mod for Civilization V

February 13, 2018 7:17 pm

Let's play this new mod for Civ 5 that makes AGI an available technology!
Can we guide humanity to a utopian AI future, or will we destroy
...

Intelligence and Stupidity: The Orthogonality Thesis

January 11, 2018 9:53 pm

Can highly intelligent agents have stupid goals?
A look at The Orthogonality Thesis and the nature of stupidity.

The 'Stamp Collector'
...

Scalable Supervision: Concrete Problems in AI Safety Part 5

November 29, 2017 11:47 pm

Why can't we just have humans overseeing our AI systems?

The Concrete Problems in AI Safety Playlist:
...

AI Safety at EAGlobal2017 Conference

November 16, 2017 9:21 pm

I attended a charity conference to learn about AI Safety!

Correction: Alan Dafoe is funded by a grant from the Open Philanthropy Project,
...

AI learns to Create ̵K̵Z̵F̵ ̵V̵i̵d̵e̵o̵s̵ Cat Pictures: Papers in Two Minutes #1

October 29, 2017 1:49 pm

Some beautiful new GAN results have been published, so let's have a quick look at the pretty pictures.
More AI Safety coming soon of
...

What can AGI do? I/O and Speed

October 17, 2017 12:20 pm

Suppose we make an algorithm that implements general intelligence as well as the brain. What could that system do?
It might have better input and
...

What Can We Do About Reward Hacking?: Concrete Problems in AI Safety Part 4

September 24, 2017 2:09 pm

Three different approaches that might help to prevent reward hacking.

New Side Channel with no content yet!:
...

Reward Hacking Reloaded: Concrete Problems in AI Safety Part 3.5

August 29, 2017 12:08 pm

Goodhart's Law, Partially Observed Goals, and Wireheading: some more reasons for AI systems to find ways to 'cheat' and get more reward than we ...

The other "Killer Robot Arms Race" Elon Musk should worry about

August 22, 2017 1:19 pm

Elon Musk is in the news, talking to the UN about autonomous weapons. This seems like a good time to explain one area where we don't quite agree ...

Reward Hacking: Concrete Problems in AI Safety Part 3

August 12, 2017 9:24 pm

Sometimes AI can find ways to 'cheat' and get more reward than we intended by doing something unexpected.

The Concrete Problems in AI
...

Why Not Just: Raise AI Like Kids?

July 22, 2017 3:58 pm

Newly made Artificial General Intelligences are basically like children, right? So we already know we can teach them how to behave, right? ...

Interviews and Talks

Industry Leaders and Notable Public Figures

Lethal Intelligence Microblog

Blow your mind with the latest stories

Favorite Microbloggers

×