Rob Miles is an AI safety researcher. In early 2023, days after GPT-4’s release, he expressed his fears about arms race dynamics in AI development, in a computerphile video. His worst fears seem to have turned into reality.

Rob Miles is an AI safety researcher. In early 2023, days after GPT-4’s release, he expressed his fears about arms race dynamics in AI development, in a computerphile video. His worst fears seem to have turned into reality.
Rob Miles is a science communicator focused on AI Safety and Alignment.
He is probably my biggest influence. He was the first person to bring the complicated AI Safety concepts to the general public, many years ago, way before the LLM revolution started, in a time where today’s reality seemed sci-fi and the whole Alignment Study was an academic curiosity, a cool way for intellectuals to tickle the most rational and analytical parts of their brain.
His Youtube Channel Robert Miles AI Safety and his interviews at Computerphile popularized the frontier ideas coming from LessWrong, Machine Intelligence Research Institute, Centre for the Study of Existential Risk and leading AI Labs like OpenAI and DeepMind.
Robert Miles is a true hero of the AI Safety saga and an inspiring force for lethalintelligence.ai.
AI Ruined My Year
June 1, 2024 8:17 am
How to Help: https://aisafety.info/questions/8TJV/How-can-I-help
https://www.aisafety.com/
AI Safety Talks: ...
AI "Stop Button" Problem - Computerphile
March 3, 2017 4:32 pm
How do you implement an on/off switch on a General Artificial Intelligence? Rob Miles explains the perils.
Part 1: ...
There's No Rule That Says We'll Make It
January 16, 2022 8:36 pm
We're not doomed! But doom is definitely an actual possibility, and we need to act like it.
If you're thinking about working on AI Safety, check ...
Why Do I Avoid Sci-fi?
August 16, 2022 8:30 am
How come I tend to avoid discussing specific scenarios when talking about AI risks?
Intro to AI Safety, Remastered
June 24, 2021 5:25 pm
An introduction to AI Safety, remastered from a talk I gave at "AI and Politics" in London
The second channel: ...
Ep. 136: The case for taking AI Safety seriously | Rob Miles
July 4, 2023 10:00 am
Through a series of popular explainer videos, Rob has become one of the most prominent voices in the AI safety community, exploring topics like ...
ROBERT MILES - "There is a good chance this kills everyone"
May 21, 2023 6:16 pm
Please check out Numerai - our sponsor @
http://numer.ai/mlst
Numerai is a groundbreaking platform which is taking the data science ...
Robert Miles–Youtube, Doom
August 19, 2022 6:00 pm
Robert Miles has been making videos for Computerphile, then decided to create his own Youtube channel about AI Safety. Lately, he's been working on ...
Would You Help an AI to Escape?
July 8, 2022 5:01 pm
Could an AGI get a human to help them, just through a text conversation? Recently we got some new evidence about that question.
Rob Miles - Why should I care about AI safety?
December 2, 2020 5:01 pm
In the latest episode of the TDS podcast, host Jeremie Harris talks with guest Rob Miles about AI safety, AI and the course of human evolution, and ...
Ep 12 - Education & advocacy for AI safety w/ Rob Miles (YouTube host)
March 9, 2024 1:05 am
We speak with Rob Miles. Rob is the host of the “Robert Miles AI Safety” channel on YouTube, the single most popular AI alignment video series out ...
9 Examples of Specification Gaming
April 29, 2020 6:41 pm
AI systems do what you say, and it's hard to say exactly what you mean.
Let's look at a list of real life examples of specification ...
Why Would AI Want to do Bad Things? Instrumental Convergence
March 24, 2018 9:51 pm
How can we predict that AGI with unknown goals would behave badly by default?
The Orthogonality Thesis video: ...
Is AI Safety a Pascal's Mugging?
May 16, 2019 4:11 pm
An event that's very unlikely is still worth thinking about, if the consequences are big enough. What's the limit though?
Do we have to ...
Intelligence and Stupidity: The Orthogonality Thesis
January 11, 2018 9:53 pm
Can highly intelligent agents have stupid goals?
A look at The Orthogonality Thesis and the nature of stupidity.
The 'Stamp Collector' ...
A Response to Steven Pinker on AI
March 31, 2019 3:39 pm
Steven Pinker wrote an article on AI for Popular Science Magazine, which I have some issues with.
The article: ...
What can AGI do? I/O and Speed
October 17, 2017 12:20 pm
Suppose we make an algorithm that implements general intelligence as well as the brain. What could that system do?
It might have better input and ...
Why Not Just: Think of AGI Like a Corporation?
December 23, 2018 10:01 pm
Corporations are kind of like AIs, if you squint. How hard do you have to squint though, and is it worth it?
In this video we ask: Are ...
AI learns to Create ̵K̵Z̵F̵ ̵V̵i̵d̵e̵o̵s̵ Cat Pictures: Papers in Two Minutes #1
October 29, 2017 1:49 pm
Some beautiful new GAN results have been published, so let's have a quick look at the pretty pictures.
More AI Safety coming soon of ...
The other "Killer Robot Arms Race" Elon Musk should worry about
August 22, 2017 1:19 pm
Elon Musk is in the news, talking to the UN about autonomous weapons. This seems like a good time to explain one area where we don't quite agree ...
Why Not Just: Raise AI Like Kids?
July 22, 2017 3:58 pm
Newly made Artificial General Intelligences are basically like children, right? So we already know we can teach them how to behave, right? ...
Are AI Risks like Nuclear Risks?
June 10, 2017 5:22 pm
Concerns about AI cover a really wide range of possible problems. Can we make progress on several of these problems at once?
With thanks to ...
Deadly Truth of General AI? - Computerphile
June 17, 2015 3:09 pm
The danger of assuming general artificial intelligence will be the same as human intelligence. Rob Miles explains with a simple example: The deadly ...
Holy Grail of AI (Artificial Intelligence) - Computerphile
May 1, 2015 5:19 pm
Audible free book: http://www.audible.com/computerphile
Why can't artificial intelligence do what humans can? Rob Miles talks about generality ...
Why Asimov's Laws of Robotics Don't Work - Computerphile
November 6, 2015 3:01 pm
Audible Free Book: http://www.audible.com/computerphile
Three or four laws to make robots and AI safe - should be simple right? Rob Miles on why ...
AI Safety - Computerphile
February 3, 2016 8:22 pm
Safety in AI is important, but more important is to work it out before working out the AI itself. Rob Miles on AI safety.
Brain Scanner: ...
AI's Game Playing Challenge - Computerphile
March 24, 2016 10:18 pm
AlphaGo is beating humans at Go - What's the big deal? Rob Miles explains what AI has to do to play a game.
What on Earth is Recursion?: ...
General AI Won't Want You To Fix its Code - Computerphile
February 28, 2017 9:02 pm
Part 1 of a Series on AI Safety Research with Rob Miles. Rob heads away from his 'Killer Stamp Collector' example to find a more concrete example of ...
AI? Just Sandbox it... - Computerphile
June 23, 2017 4:44 pm
Why can't we just disconnect a malevolent AI? Rob Miles on some of the simplistic solutions to AI safety.
Out of focus shots caused by ...
Using Dangerous AI, But Safely?
November 15, 2024 10:35 pm
Can we keep powerful AI under control, using AI?
The Paper: https://arxiv.org/abs/2312.06942
AI Lab Watch: ...
Using Dangerous AI, But Safely?
November 15, 2024 10:35 pm
Can we keep powerful AI under control, using AI?
The Paper: https://arxiv.org/abs/2312.06942
AI Lab Watch: ...
Learn AI Safety at MATS #shorts
September 28, 2024 12:12 am
Apply for MATS at matsprogram.org by Oct 6
#shorts
AI Ruined My Year
June 1, 2024 8:17 am
How to Help: https://aisafety.info/questions/8TJV/How-can-I-help
https://www.aisafety.com/
AI Safety Talks: ...
Apply to Study AI Safety Now! #shorts
April 28, 2023 6:37 pm
Apply to SERI MATS at https://www.serimats.org/ by May 7th, and check out http://aisafety.training to stay up to date with events and programs!
Why Does AI Lie, and What Can We Do About It?
December 9, 2022 10:10 pm
How do we make sure language models tell the truth?
The new channel!: https://www.youtube.com/@aisafetytalks
Evan Hubinger's Talk: ...
Apply Now for a Paid Residency on Interpretability #short
November 11, 2022 8:07 pm
https://www.redwoodresearch.org/remix
Real main channel video coming soon!
#short #shorts
$100,000 for Tasks Where Bigger AIs Do Worse Than Smaller Ones #short
October 14, 2022 1:05 pm
Check out http://inversescaling.com for more details and to apply
#shorts #short
Free ML Bootcamp for Alignment #shorts
May 24, 2022 7:30 pm
Apply for the second MLAB (Machine Learning for Alignment ...
Win $50k for Solving a Single AI Problem? #Shorts
February 8, 2022 9:17 pm
The Alignment Research Center is offering up to $50k for proposals for their Eliciting Latent Knowledge problem.
The report: ...
Apply to AI Safety Camp! #shorts
November 19, 2021 8:03 pm
Trying out #shorts
Applications are open for next year's AI Safety Camp!
http://aisafety.camp
We Were Right! Real Inner Misalignment
October 10, 2021 10:50 pm
Researchers ran real versions of the thought experiments in the 'Mesa-Optimisers' videos!
What they found won't shock you (if you've been paying ...
Intro to AI Safety, Remastered
June 24, 2021 5:25 pm
An introduction to AI Safety, remastered from a talk I gave at "AI and Politics" in London
The second channel: ...
Deceptive Misaligned Mesa-Optimisers? It's More Likely Than You Think...
May 23, 2021 10:15 pm
The previous video explained why it's *possible* for trained models to end up with the wrong goals, even when we specify the goals perfectly. This ...
The OTHER AI Alignment Problem: Mesa-Optimizers and Inner Alignment
February 16, 2021 8:37 pm
This "Alignment" thing turns out to be even harder than we thought.
# Links
The Paper: https://arxiv.org/pdf/1906.01820.pdf
Discord ...
Quantilizers: AI That Doesn't Try Too Hard
December 13, 2020 10:46 pm
How do you get an AI system that does better than a human could, without doing anything a human wouldn't?
A follow-up to "Maximizers and ...
Sharing the Benefits of AI: The Windfall Clause
July 6, 2020 6:53 pm
AI might create enormous amounts of wealth, but how is it going to be distributed?
The Paper: ...
10 Reasons to Ignore AI Safety
June 4, 2020 5:28 pm
Why do some ignore AI Safety? Let's look at 10 reasons people give (adapted from Stuart Russell's list).
Related Videos from Me:
Why ...
9 Examples of Specification Gaming
April 29, 2020 6:41 pm
AI systems do what you say, and it's hard to say exactly what you mean.
Let's look at a list of real life examples of specification ...
Training AI Without Writing A Reward Function, with Reward Modelling
December 13, 2019 6:39 pm
How do you get a reinforcement learning agent to do what you want, when you can't actually write a reward function that specifies what that ...
AI That Doesn't Try Too Hard - Maximizers and Satisficers
August 23, 2019 5:05 pm
Powerful AI systems can be dangerous in part because they pursue their goals as strongly as they can. Perhaps it would be safer to have systems that ...
Is AI Safety a Pascal's Mugging?
May 16, 2019 4:11 pm
An event that's very unlikely is still worth thinking about, if the consequences are big enough. What's the limit though?
Do we have to ...
A Response to Steven Pinker on AI
March 31, 2019 3:39 pm
Steven Pinker wrote an article on AI for Popular Science Magazine, which I have some issues with.
The article: ...
How to Keep Improving When You're Better Than Any Teacher - Iterated Distillation and Amplification
March 11, 2019 2:14 pm
[2nd upload] AI systems can be trained using demonstrations from experts, but how do you train them to out-perform those experts? Can this still be ...
Why Not Just: Think of AGI Like a Corporation?
December 23, 2018 10:01 pm
Corporations are kind of like AIs, if you squint. How hard do you have to squint though, and is it worth it?
In this video we ask: Are ...
Safe Exploration: Concrete Problems in AI Safety Part 6
September 21, 2018 1:20 pm
To learn, you need to try new things, but that can be risky. How do we make AI systems that can explore safely?
Playlist of the series so ...
Friend or Foe? AI Safety Gridworlds extra bit
June 25, 2018 1:31 am
The last video about the AI Safety Gridworlds paper. How does an agent detect and adapt to friendly and adversarial intentions in the ...
AI Safety Gridworlds
May 25, 2018 6:20 pm
Got an AI safety idea? Now you can test it out! A recent paper from DeepMind sets out some environments for evaluating the safety of AI systems, and ...
Experts' Predictions about the Future of AI
March 31, 2018 2:12 pm
When will AI systems surpass human performance? I don't know, do you? No you don't. Let's see what 352 top AI researchers ...
Why Would AI Want to do Bad Things? Instrumental Convergence
March 24, 2018 9:51 pm
How can we predict that AGI with unknown goals would behave badly by default?
The Orthogonality Thesis video: ...
Superintelligence Mod for Civilization V
February 13, 2018 7:17 pm
Let's play this new mod for Civ 5 that makes AGI an available technology!
Can we guide humanity to a utopian AI future, or will we destroy ...
Intelligence and Stupidity: The Orthogonality Thesis
January 11, 2018 9:53 pm
Can highly intelligent agents have stupid goals?
A look at The Orthogonality Thesis and the nature of stupidity.
The 'Stamp Collector' ...
Scalable Supervision: Concrete Problems in AI Safety Part 5
November 29, 2017 11:47 pm
Why can't we just have humans overseeing our AI systems?
The Concrete Problems in AI Safety Playlist: ...
AI Safety at EAGlobal2017 Conference
November 16, 2017 9:21 pm
I attended a charity conference to learn about AI Safety!
Correction: Alan Dafoe is funded by a grant from the Open Philanthropy Project, ...
AI learns to Create ̵K̵Z̵F̵ ̵V̵i̵d̵e̵o̵s̵ Cat Pictures: Papers in Two Minutes #1
October 29, 2017 1:49 pm
Some beautiful new GAN results have been published, so let's have a quick look at the pretty pictures.
More AI Safety coming soon of ...
What can AGI do? I/O and Speed
October 17, 2017 12:20 pm
Suppose we make an algorithm that implements general intelligence as well as the brain. What could that system do?
It might have better input and ...
What Can We Do About Reward Hacking?: Concrete Problems in AI Safety Part 4
September 24, 2017 2:09 pm
Three different approaches that might help to prevent reward hacking.
New Side Channel with no content yet!: ...
Reward Hacking Reloaded: Concrete Problems in AI Safety Part 3.5
August 29, 2017 12:08 pm
Goodhart's Law, Partially Observed Goals, and Wireheading: some more reasons for AI systems to find ways to 'cheat' and get more reward than we ...
The other "Killer Robot Arms Race" Elon Musk should worry about
August 22, 2017 1:19 pm
Elon Musk is in the news, talking to the UN about autonomous weapons. This seems like a good time to explain one area where we don't quite agree ...
Reward Hacking: Concrete Problems in AI Safety Part 3
August 12, 2017 9:24 pm
Sometimes AI can find ways to 'cheat' and get more reward than we intended by doing something unexpected.
The Concrete Problems in AI ...
Why Not Just: Raise AI Like Kids?
July 22, 2017 3:58 pm
Newly made Artificial General Intelligences are basically like children, right? So we already know we can teach them how to behave, right? ...
Learn about the issue by some of the best explainers out there
Subscribe for important updates !!!
© 2024 Lethal Intelligence – Ai. All rights reserved.
![]() | ![]() ![]() | ![]() | ||
![]() | ![]() | ![]() | ||
![]() ![]() | ![]() ![]() | |||
![]() |
| ![]() |