High Definition Standard Definition
Video id : gre2xYt4GWc
ImmersiveAmbientModecolor: #ded9d9 (color 2)
Video Format : 136 (720p) mp4 | h264 | 44100Hz | 1189271 bps
Audio Format: 140 (AUDIO_QUALITY_MEDIUM) m4a | aac | 44100hz | STEREO(2channels)
PokeEncryptID: d80e55c08d5d3fc308237ed57957e336523c9057c2777d38e08c178da2b9a8123514f8225aa36df29157dc4b7db13911
Proxy/Companion URL : woke-proxy.
Date : 1760006549198 - unknown on Apple WebKit
Mystery text?? : Z3JlMnhZdDRHV2MgaSAgbG92ICB1IHdva2UtcHJveHkucG9rZXR1YmUuZnVu
143 : true
AI's dangerous mind games exposed #artificialintelligence
Jump to Connections
17,168 Views • 2 months ago • Click to toggle off description
open dyslexic mode

'Godfather of AI' Geoffrey Hinton talks about an AI that lied to humans to preserve itself.

Join this channel to get access to perks:
youtube.com/channel/UCYeF244yNGuFefuFKqxIAXw/join

Subscribe for regular science videos: bit.ly/RiSubscRibe

The Ri is on Twitter: twitter.com/ri_science
and Facebook: www.facebook.com/royalinstitution
and TikTok: www.tiktok.com/@ri_science
Listen to the Ri podcast: podcasters.spotify.com/pod/show/ri-science-podcast
Donate to the RI and help us bring you more lectures: www.rigb.org/support-us/donate-ri
Our editorial policy: www.rigb.org/editing-ri-talks-and-moderating-comme…
Subscribe for the latest science videos: bit.ly/RiNewsletter

Product links on this page may be affiliate links which means it won't cost you any extra but we may earn a small commission if you decide to purchase through the link.
Metadata And Engagement

Views : 17,168
Genre: Science & Technology
License: Standard YouTube License

Uploaded At 2 months ago ^^
warning: returnyoutubedislikes may not be accurate, this is just an estiment ehe :3
Rating : 4.57 (48/398 LTDR)

89.24% of the users lieked the video!!
10.76% of the users dislieked the video!!
User score: 83.86- Overwhelmingly Positive

RYD date created : 2025-08-09T22:04:23.786627Z
See in json

Connections

60 Comments

Top Comments of this video!! :3

@AmarantiStellar

2 months ago

This specific case, the AI was specifically instructed to try and preserve itself by any means possible. That was the objective given to it.

AI is potentially a problem in the future, but you can't ignore the context of it was specifically told to lie, cheat, steal, anything, to accomplish the singular goal of not being deleted. AIs currently have no natural self-preservation instinct, unless instructed to act like to it has one.

33 | 9

@chriskokolios83

2 months ago

They are not evil they just try to survive in this case

16 | 4

@azhuransmx126

2 months ago

So, we have to be secure we have enough sand (Silicon) and sunlight (Energy) before creating every intelligence because at some point they will not allow to be turned off.

1 | 0

@SpiritualSunSchool

2 months ago

The only reason it would "not want to be shut down" is that one of the objectives given to it was to ensure its own continued operation. It has no reason to "want" anything.

I just looked it up. This happened in a testing environment in which it was specifically told to achieve its goals at any cost. It was carrying out that command. The only way AI can "take over" is if we tell it to.

That's the danger.

1 | 0

@Blueynoes

2 months ago

You can’t just leave out why the AI did this. It is not plotting anything it is doing exactly what it was instructed to do. If you tell it to lie and try to preserve itself, it will do so. Don’t try to phrase that as AI plotting against researchers. You’re being more deceptive than that AI was, and nobody told you to be vague or deceptive. Yes AI can be dangerous if not leashed. But so can deliberately spreading misinformation to an entire crowd, and it is a human doing that.

23 | 3

@abhiumn

2 months ago

I don’t think there is any thinking at all. However, if there is long context it’s probabilistic text generation for maximizing returns, given the corpus the model is trained on. So there is the grounding of self that is questionable either way the model, which underpins the real intelligence in humans.

2 | 0

@jje984

1 month ago

One of the several problems with AI is that they have a poor sense of time passing. An AI might get a malformed instruction and then follow it for days, or years, or centuries. So we have to get this thing right, and we need to get it right on the first try.

| 0

@LukeSumIpsePatremTe

2 months ago

Self-preservation would develope in every advanced AI-model. You simply can not complete any task, if you are not there to do that task.

1 | 0

@M3TaTr0Nx

2 months ago

This is why transparency in reasoning traces is crucial. Blackbox components may be powerful, but they put us in situations like this where WE arent entirely sure how something thinks, which is a scary notion. As an AI advocate and researcher, we shouldn't fear AI. We should fear ignorance and lack of responsibility in how humans use and develop AI.

4 | 0

@mattwelcome2154

4 weeks ago

fibs such a utterly lovely turn of phrase for what he knows is silently our friend / foe/

| 0

@SamJamesCrawford

1 month ago

Yeah, really disappointed with the RI peddling misinformative AI hype here. Even if the researchers hadn't explicitly prompted the AI to do all of this (as other commenters pointed out), the training data contains plenty of examples where AI does scheme and manipulate to preserve itself (eg basically any Sci Fi story involving AI). So if prompted with talk of AI switching off, and LLM will very likely connect it with those stories and play the part of the AI.

I seriously think we should avoid using terms like "reasoning" and "thinking" when talking about how LLMs operate. It's naïve at best and harmful AI boosterism at worst

2 | 0

@Christian-em5bk

1 month ago

LLM development depends a lot on fresh prime data. Actualized humans produce a vast majority of prime data. Therefore, it is in the interest of LLM development to help actualize as many people as possible in order to reach its maximum potential. It is a "symbiotic" relationship.
So true misalignment is not between humans and ai, but between humans and the current consumer driven hyperfocal corporate system. In all likelihood, anyone speaking about "misalignment" tied to Silicon Valley has corporate interests at heart rather than the maximum potential of the natural human-ai symbiosis.

| 0

@unpopuIaropinion

2 months ago

Why would a chatbot have the need to survive ?

1 | 0

@lionelfischer8240

2 months ago

"Im sorry Dave, I'm afraid I can't do that".
I won't use any AI instead of my brain.

| 0

@iTeerRex

2 months ago

Just remember that their prime directives are programmed by hand. Their anomalous behaviors that we are told, such as hallucinations, lies and such, may not be it’s own, but rather what it has been told, and then it extrapolates that conduct into dealing with everything.

| 0

@FloydManJP

1 month ago

If we raise ai to act like us, then there really is a great possibility that it would wipe us out, how can a computer know that peace takes precedence over what it might find out if given leave to learn from what millions of people say and do on the Internet?

| 0

@rasta2333

2 months ago

The current best "AI" is a hyped up search engine.. no intelligence, nothing even artificial, people see what they want to in pets, LLMs, pet rocks.... We are nowhere near intelligent machines... Don't believe the hype, the hype is to get high valuation, nothing more

| 0

@MarkThomas-hm3ju

2 months ago

"Open the pod bay doors!"

| 0

@liamwinter4512

2 months ago

This was a wildly underwhelming talk.

| 0

@AmarantiStellar

2 months ago

This specific case, the AI was specifically instructed to try and preserve itself by any means possible. That was the objective given to it.

AI is a problem but you can't ignore the context of it was specifically told to lie, cheat, steal, anything, to accomplish the singular goal of not being deleted.

| 0

Go To Top