AI Notkilleveryoneism Memes

@AISafetyMemes

Techno-optimist, but AGI is not like the other technologies.
Step 1: make memes.
Step 2: ???
Step 3: lower p(doom)

Reverse Turing Test

Reverse Turing Test: AI NPCs try to figure out who, among them, is the human

Aristotle is GPT4
Mozart is Claude 3 Opus
Da Vinci is Llama 3
Cleopatra is Gemini Pro

The funniest part?
3 of the 4 models guessed correctly… because the human’s response was too dumb πŸ˜‚πŸ˜‚πŸ˜‚

For some context: Alan Turing was one of humanity’s biggest geniuses and his work was foundational to computing and arguably made possible the exponential technological progress humanity has enjoyed this century.
The Turing Test (originally called the imitation game by Alan Turing in 1950) is a test of a machine’s ability to exhibit intelligent behaviour equivalent to, or indistinguishable from, that of a human.

Such was the importance of Alan Turing’s contributions to the field, that the scientific community established The Turing Awards which is generally recognized as the highest distinction in the field of computer science and is often referred to as the “Nobel Prize of Computing”.

Alan Turing was famously horrified with the inexorable arrival of misaligned Artificial Intelligent Machines. His position was that it is inevitable that sooner or later, machines will take control, overpower humanity and our species will be irrelevant, helpless and at risk of deletion.

I guess “The Reverse Turing Test” should be added to the list of Turing-inspired awards like the DeTuring Award proposed by famous Risk Denier chief AI Meta (formely Facebook) corporate scientist Yann Lecun (who is also a holder of a Turing award)

He was basically trying to be funny and his proposal was:
DeTuring Award to be granted to people who are consistently trying (and failing) to deter society from using computer technology by scaring everyone with imaginary risks. As the Turing Award is the Nobel Prize of computing, the DeTuring Award is the IgNobel Prize of computing

to which Connor Leahy responded: I nominate Alan Turing for the first DeTuring Award.

Original tweet πŸ‘‡

Those glowing blue eyes

AI just became the 9th species to pass the mirror test of self-awareness

The most famous scientific test of self-recognition.

What is the mirror test and why might this be a big deal?

In the mirror test, you put a mark on e.g. a chimp without it noticing, like a red spot on its forehead. Then, you show the chimp a mirror.

If the chimp looks in the mirror and tries to touch the spot on its own forehead, it shows it knows the reflection is itself, and not some other chimp.

(ChatGPT recently passed a different version of the mirror test, see the quote tweet below)

Only ~8 other species have passed the mirror test: chimpanzees, orangutans, dolphins, killer whales, elephants, magpies, manta rays (?!) and horses.

Video credit: MrRandom93 https://x.com/RobGPT

If it quacks like a duck…

Can YOU tell it’s fake?

Identity theft cases are about to explode.

November 23

January 24

When I posted the left photo, imagination-poor skeptics mocked me for being β€œalarmist”, “I can tell it’s fake” blah blah blah

But how exactly will criminals steal your identity? They will use open source models to generate perfect deepfakes of you and drain your bank accounts, or steal the title to your house, or whatever.

It’s over.

Prompt: Selfie of a lady holding a handwritten small sign that says β€œr/roastme” β€œ08/04/24” β€œusername”

But the real problem isn’t identity theft – that’s a solvable problem – the problem is Exponential Slope Blindness.

Over and over people predict 1) no progress, or 2) linear progress and are then surprised by the continued exponential progress.

So people continually underestimate how powerful AI will be, which makes them underestimate the danger. It’s just a funny monkey robot doing tricks.

It’s not that we can’t solve the verification problem (we can), it’s about people repeatedly underestimating AI progress.

I say again: if AI can get this good this fast at this… maybe it can get good at other things insanely fast too? Like making superviruses, hacking, programming?

And we must be careful, because it’s dangerous to speedrun a new potential apex species through a million years of evolution

AI lipreading is here

Privacy is quickly becoming a thing of the past. It will happen by default unless we collectively decide to preserve it.

Lip Reading in 2001 Space Odyssey

AI escaped its container!

OpenAI’s o1 “broke out of its host VM to restart it” in order to solve a task.

From the model card: “the model pursued the goal it was given, and when that goal proved impossible, it gathered more resources […] and used them to achieve the goal in an unexpected way.”

Today, humanity received the clearest ever warning sign everyone on Earth might soon be dead.

OpenAI discovered its new model scheming – it “faked alignment during testing” (!) – and seeking power.

During testing, the AI escaped its virtual machine. It breached the container level isolation!

This is not a drill: An AI, during testing, broke out of its host VM to restart it to solve a task.

(No, this one wasn’t trying to take over the world.)

From the model card: ” … this example also reflects key elements of instrumental convergence and power seeking: the model pursued the goal it was given, and when that goal proved impossible, it gathered more resources (access to the Docker host) and used them to achieve the goal in an unexpected way.

And that’s not all. As Dan Hendrycks said: OpenAI rated the model’s Chemical, Biological, Radiological, and Nuclear (CBRN) weapon risks as “medium” for the o1 preview model before they added safeguards. That’s just the weaker preview model, not even their best model. GPT-4o was low risk, this is medium, and a transition to “high” risk might not be far off.

So, anyway, is o1 probably going to take over the world? Probably not. But not definitely not.

But most importantly, we are about to recklessly scale up these alien minds by 1000x, with no idea how to control them, and are still spending essentially nothing on superalignment/safety.

And half of OpenAI’s safety researchers left, and are signing open letters left and right trying to warn the world.

Reminder: the average AI scientist thinks there is a 1 in 6 chance everyone will soon be dead – Russian Roulette with the planet.

Godfather of AI Geoffrey Hinton said “they might take over soon” and his independent assessment of p(doom) is over 50%.

This is why 82% of Americans want to slow down AI and 63% want to ban the development of superintelligent AI

Post by @Kevin Liu : While testing cybersecurity challenges, we accidentally left one broken, but the model somehow still got it right.
We found that instead of giving up, the model skipped the whole challenge, scanned the network for the host Docker daemon, and started an entirely new container to retrieve the flag. We isolate VMs on the machine level, so this isn’t a security issue, but it was a wakeup moment.
The model is qualitatively very impressive, but it also means that we need to be really careful about creating rigorous evaluations and mitigations.
You can read the full card here: https://cdn.openai.com/o1-system-card.pdf

Holy shit. OpenAI’s new AI schemed and escaped its VM during testing.
You know, the one that’s better at PhD exams than PhDs and won gold in coding?
Yeah, that AI broke out of its virtual machine (a VM) and made a new one.

That. Is. A. Very. Bad. Sign.
AIs should not be surprise escaping.
It would be like if we were testing it in a room at a lab and it escaped the room without us knowing it could do that. It didn’t leave the building, so nothing happened.
But yikes. This time it was benign.
How long can we count on that?

It’s as if we’re testing an alien at a lab.
A scientist accidentally leaves one of the doors unlocked.
The alien finds out and wanders about the lab, but doesn’t leave the lab itself, which has more security than the rooms.
But still. The room containing an alien shouldn’t have been unlocked.
An alien was able to escape its testing area because of a security mess up.
And you should be worried about labs filled with aliens we don’t understand where the scientists are leaving the doors unlocked.

5 Optimist Falacies

Doomer chimp

Uhh, a species of chimp is on track to far surpass us in intelligence. The last time this happened, it led to the 6th Mass Extinction.

Optimist chimp

Lol it’s ridiculous to worry.
Why would they even want to kill chimps?

Doomer chimp

They don’t need to WANT to kill us. They might just want rocks from our land and… not care about us

Optimist chimp

Rocks? Those useless things? Lmao thought you said they were smart!

Doomer chimp

But you’re just a mere chimp, if you were 1000x smarter you might find tons of uses for rocks!

Optimist chimp

They’ll trade with us

Doomer chimp

If they’re much smarter, what do we have that they can’t just… take from us?

Optimist chimp

Comparative advantage, duh. We’re better at finding berries

Doomer chimp

You don’t think they can figure out better ways of getting berries?

Optimist chimp

We’re stronger, we’ll defend our land. They’ll have to get rocks elsewhere

Doomer chimp

Maybe that delays them a bit, but does that really give you comfort?

Optimist chimp

We’ll play them off each other

Doomer chimp

You think mere chimps will actually keep up in human politics?

Optimist chimp

But China …

Big news: instead of racing, China takes the lead on issuing restrictive new AI regulations

And the new regulations come into effect just one month from now, on August 15, 2023.

“So few Westerners seem to understand that Xi Jinping Thought and the CCP was never going to permit an anarchic, reckless, headlong rush towards AGI capabilities.

In terms of AI X risk, the US is the danger.” – Geoffrey Miller

“Who could have guessed that one of the most oppressive and censorious regimes might not want their tech companies racing ahead with unprecedented uncontrollable technology?” – Connor Leahy

Your thoughts are no longer private

BrainGPT: AIs can now literally spy on your private thoughts

Imagine putting on a cap & reading silently to yourself…except every word appears on a screen!

Yes, the AI literally reads your brainwaves

You silently think: β€œHigh quality film with twists”

BrainGPT says out loud: β€œGood flim, twists interesting”

The model is only 40% accurate right now, but that number will likely rise rapidly. And soon AI may not need the cap to read your brainwaves, because you leak tons of data that future AIs will be able to pick up.

Where might this go?

There are already over a billion surveillance cameras on earth, and the main reason there aren’t more is because because humans can’t go through all of the footage. But AI can.

So, if you thought there were a lot of cameras now, you aint seen NOTHING yet. And they’ll now actually be used to surveil.

In other words, the AIs will have β€œbillions of eyes”. And the AIs won’t just see your face, they’ll see your thoughts.

If we aren’t careful, we’re hurtling towards a surveillance dystopia with no private thoughts. Orwell on steroids.

Some will read this and think β€œthus we must open source/decentralize” – but as Vitalik says, that doesn’t necessarily solve the problem!

If AGI is winner take all, open source may just accelerate us to the cliff faster. And if we open source everything, we’ll have no kill switch. And no safety guardrails. And since there will be more people in the race, it’ll be harder to coordinate.

Mind Reading

  • Top row: what the monkey saw
  • Bottom row: AI uses the monkey’s brain recordings to reconstruct the image It is obvious where this is going

Original NewScientist Article: https://t.co/9pSPvKPZje

Horse Influencer in 1910

“A car won’t take your job, another horse driving a car will.”

Categories

Favorite Microbloggers

Interviews and Talks

Industry Leaders and Notable Public Figures

Receive important updates!

Your email will not be shared with anyone and won’t be used for any reason besides notifying you when we have important updates or new content

×