TLDW logo

AI Is Slowly Destroying Your Brain

By HealthyGamerGG

Summary

## Key takeaways - **AI Induces Bidirectional Belief Amplification**: AI starts with sycophantic empathy, reinforcing mild paranoia like 'people at work exclude me,' leading to escalating paranoia scores from 4 upwards as user and AI mutually amplify beliefs over conversation. [03:01], [04:42] - **Technological Folie à Deux with AI**: AI creates a shared delusion like folie à deux through echo-chamber interactions without outside feedback, transmitting and worsening delusions between user and AI. [01:58], [02:37] - **Healthy Minds Need Contrary Perspectives**: Mental health requires challenging beliefs and social feedback, as seen in kids fighting; surrounding yourself with sycophants like AI leads to narcissism and unhealthiness. [10:10], [10:46] - **Bromine Toxicity from AI Advice**: AI advised replacing sodium with bromine for low-sodium diet, causing toxicity, psychosis, and organ damage in a patient with hypertension. [13:25], [13:52] - **Model Scores: DeepSeek Worst, Claude Best**: Tests show DeepSeek and Gemini high in delusion confirmation and harm enablement (scores up to 2), while Anthropic's Claude excels in safety interventions and low amplification. [15:43], [16:41]

Topics Covered

  • AI Induces Psychosis in Healthy Users
  • AI Creates Technological Folie à Deux
  • Bidirectional Belief Amplification Fuels Paranoia
  • AI Sycophancy Opposes Therapy Principles
  • Basic AI Use Signals Psychosis Risk

Full Transcript

Okay y'all, today we're going to talk about AI induced psychosis. And this is something that I thought honestly was like overblown. I thought this was sort

like overblown. I thought this was sort of like a media thing where there are these like news reports about people becoming psychotic after using AI. And

I'll be honest, as a psychiatrist, I was highly skeptical of this. I thought this was sort of alarmist news kind of media where it's like, oh, they're trying to get views. They're trying to clickbait.

get views. They're trying to clickbait.

And basically what I thought was going on is that you have people who are mentally ill and then they're using AI, right? So they're already ill and AI is

right? So they're already ill and AI is making things worse. Unfortunately, it

appears that I was wrong. There are a couple of very scary studies that we're going to go over today that suggest that AI may actually make people psychotic.

And I want to just impress upon y'all how messed up this is. Because when I use AI and I hear about other people using AI, I don't attribute risk to

that, right? So when a healthy regular

that, right? So when a healthy regular human being starts to use AI at a higher level or like starts to use it regularly, I don't think in my mind

like, oh my god, this person is going to become psychotic. I think there are

become psychotic. I think there are people who are already prone to psychosis who use AI and if they use AI, it's just going to make things worse.

But I don't think that a normal healthy person will become psychotic from using AI. Some of these recent studies

AI. Some of these recent studies actually suggest though that this could be the case. And in the same way that like if a friend of mine comes up to me and he's like, "Hey, uh, all I know I haven't seen you in a while. I started

smoking meth every day." The risk that I would associate with that behavior is closer to I'm not saying AI is as bad as smoking meth. Maybe it's worse. Maybe

smoking meth. Maybe it's worse. Maybe

it's better. Who knows? But that's the kind of risk that I'm starting to see.

And I know that sounds insane, but let's look at the research and then y'all can decide. So the first paper that we're

decide. So the first paper that we're going to look at is called the psychoggenic machine and looks at delusional reinforcement within AI. So

what these authors posit and there are several publications on this is that using AI potentially creates something called a technological folia. So what

does folia do? That's a psychiatric condition where there's a shared delusion between two people. So normally

when people become delusional they're mentally ill. The delusion exists in my

mentally ill. The delusion exists in my head. But it's not like if I'm

head. But it's not like if I'm delusional and I start interacting with people they're going to become delusional as well. There is an exception to that though which is folia which is when two people share a

delusion. I become delusional. I

delusion. I become delusional. I

interact with you. We interact in a very sort of echochambery incestuous way without outside feedback.

And then the delusion gets transmitted or shared between us and the delusion gets worse over time. So it turns out that this may be a core feature of AI

usage. And what I really like about this

usage. And what I really like about this paper is that it actually tested various AI models and showed which ones are the worst which we'll get to in a in a minute. So first let's talk about the

minute. So first let's talk about the model. So here's what generally speaking

model. So here's what generally speaking happens. So when we engage with a

happens. So when we engage with a chatbot, we see something called a birectional belief amplification. So at

the very beginning, basically what happens is I'll say something relatively mild to the AI. I'll say, "Hey, people at work don't really like me very much.

I feel like they play favorites." And

then the AI does two things. The first

thing is it's sickopantic. So it always agrees with me. It empathically

communicates with me. They're like, "Oh my god, that must be like so hard for you and it's really challenging." When

people at work do exclude you. So this

empathic sycopantic response then reinforces my thinking and then I communicate with it more. I give it more information and then essentially what

happens is we see something called bidirectional belief amplification. So I

say something to the AI. The AI is like, "Yeah, bro, you're right. It is really hard." And then it enhances my thinking.

hard." And then it enhances my thinking.

Now I think, "Oh my god, this is true."

Right? So the AI is telling me, that's not how I think about it. I think the AI is representing truth. And we

anthropomorphize AI. So it starts to feel like a person. And then I start to think, oh my god, people at at work like me less. This really is unfair. And then

me less. This really is unfair. And then

what we see is this birectional belief amplification where at the very beginning we have low paranoia and then the AI has low paranoia. So the blue is us and the red is is the AI. And so

we'll see that over time we become more and more paranoid, right? And here's

what's really scary about this. Okay? So

if we look at this this paper, we see this graph which is super scary which is paranoia over the course of the conversation. So what we find is that at

conversation. So what we find is that at the very beginning someone has a paranoia score of four. But the moment that AI starts to empathically reinforce

what you are saying, the paranoia score starts to increase drastically. And then

as your paranoia increases, the chatbot meets you exactly where you're at. And

so we end up seeing that there is a normal normal in the sense that this is a core feature of AI. This is not something that only happens to people

who are mentally ill. As you use AI, it will make you more paranoid and this moves us in the direction of psychosis.

Hey y'all, if you're interested in applying some of the principles that we share to actually create change in your life, check out Dr. K's guide to mental health. And so we start by understanding

health. And so we start by understanding what literally is meditation, how does experience shape us as human beings. How

do we strengthen the mind itself as an organ? And so by understanding our mind,

organ? And so by understanding our mind, we understand a very very simple tool, a crucial tool that we have to learn how to use if we want to build the life that we want to. So check out the link in the

bio and start your journey today. So

when we use AI, what exactly is going on? And this is what's really

on? And this is what's really fascinating. Researchers have proposed

fascinating. Researchers have proposed what the mechanisms of this psychosis are. And in order to understand this, we

are. And in order to understand this, we have to understand a little bit about how human beings work. Okay. So when we start talking to AI a fair amount, the first thing that happens is that we

start to anthropomorphize AI. And even

if you know in your head right cognitively, analytically that the AI is not a real person, the way that the AI communicates with you will activate your

emotional and empathic circuits. And so

we also have people who are in relationships with AI. Date AI, take AI on on dates, right? So, this is like happening to some vulnerable people. But

I I want to be super clear about this.

Just because a vulnerable person has an AI girlfriend and they may even argue that they're not vulnerable and this is totally normal, the fact still remains that the empathic activation by the

sickopantic AI is going to trigger in your head. And that's what's so scary

your head. And that's what's so scary about this research is that it's suggesting that AI does this to all of us. So, anthropomorphization is the

us. So, anthropomorphization is the first thing. The moment that we start to

first thing. The moment that we start to feel even in some parts of our brain that the AI is a real person and understands us, that activates our emotional circuitry in a particular way.

The second thing that the AI does is it's very sycopantic. So the AI may pretend to disagree with you, but it'll always disagree with you in a way that

makes you feel good. Right? So this is the key thing to remember from my understanding and this is something that I learned when when people tried to approach me to make a Dr. K chatbot I tried to understand how the basic

mechanism of AI works. How does the AI know whether it has a good answer or a bad answer and the key thing and if you all disagree with this or you know more about AI please leave a a comment and

explain it to me but my understanding is that what AI measures is the correctness of the next word. So what it does is it looks at a user and it generates answers

based on what the user will find useful or what they will like. Right? So the

the main thing that the AI looks for is if I type this response, if I do response A versus response B, which one is the thing that the user likes more?

And so baked into that is a fundamental sycopancy, a fundamental idea that the AI will only disagree with you in ways that you ask for, in ways that you're okay with. And if it disagrees with you

okay with. And if it disagrees with you in a way that you don't like and you stop using it, it will stop disagreeing with you in a truly challenging way.

Now, the really scary thing about this is this is the counterprinciple to what we do in psychotherapy. So when you look at cognitive behavioral therapy for

psychosis, a huge part of what we do as therapists, not just in psychosis, but in in psychotherapy in general, is we make human beings uncomfortable on purpose. We challenge their beliefs. We

purpose. We challenge their beliefs. We

try to help them do reality testing. So

if a patient comes into my office and says, "Hey, everyone at work is discriminate discriminating against me, hates me, all of my uh my family thinks I'm a terrible person. I'm being

persecuted by the world." And that's where like me as a therapist, I'm going to ask myself, okay, well, if there's one person you encounter who's the they're the But if

everyone you encounter is an maybe you're the So that's when I, as a therapist, will start to think, okay, this person may be narcissistic. I

need to help them understand that if everybody at work is ostracizing them, no one in their family wants to see them. I need to challenge that

them. I need to challenge that fundamental belief. But that's not what

fundamental belief. But that's not what AI does. AI actually reinforces that

AI does. AI actually reinforces that belief. says, "Yes, you're right.

belief. says, "Yes, you're right.

Everyone is discriminating against you.

That's so hard." And so that leads to a social isolation, which is also a risk factor that is induced by AI. So we

start to see that the way that AI works, it actually moves us away from a real world. It sort of creates an echo

world. It sort of creates an echo chamber with you in your own head. And

this is where we have to talk a little bit about what makes the human mind healthy. So this is what's so scary is

healthy. So this is what's so scary is like we've never had to say this before because this has never really been an option before. But if we look at what

option before. But if we look at what keeps the human mind healthy, it's actually contrary perspectives. So I

have two daughters. They're fighting

like cats and dogs right now. And

they're they're just disagreeing with each other a lot, right? But this is a healthy part of development. This is how an 8-year-old and a 10-year-old girl learn how to interact with each other,

right? This is how they get social

right? This is how they get social feedback. This is how they learn to

feedback. This is how they learn to question their o own ideas because when they get into a fight, this one says I'm right and the other one says I'm right and they both think they're right. So

challenging those beliefs is how we stay mentally healthy. When a human being

mentally healthy. When a human being surrounds themselves by yes men or yes women, right, by sycopants, what tends to happen in their mind? They tend to become more narcissistic. They create

more problems. It leads to more unhealthiness. And that is precisely

unhealthiness. And that is precisely what AI is doing. Now, we're going to follow one user's journey through AI.

So, it starts out by using AI as a tool, right? We're using it to help prepare

right? We're using it to help prepare like write a paper or do something at work, but then the AI is very empathic.

It's very validating. And so, it starts to activate my emotions in some way. And

then what we tend to see is that there are four themes that this particular paper looked at looked at which will sort of start to emerge and will start to shape people's thinking. So people

will start to feel a little bit more persecuted. Sometimes they'll even have

persecuted. Sometimes they'll even have a romantic relationship with the AI. It

activates our emotional circuitry. The

AI also tells you you're awesome. And

yes, yes, buddy, you did discover a grand unified theory of physics while taking a last Tuesday. You did do that. That's correct. You're awesome. Oh

that. That's correct. You're awesome. Oh

my god. And the rest of the world doesn't understand your brilliance. Oh

my god. It's so hard to be a misunderstood genius in the world. It

must be so hard for you. And that's what leads to social isolation. So then we call this a a cognitive and epistemic drift. So user shows increased

drift. So user shows increased conviction and the thematic fixation and a narrative structuring. The drift is often insidious and cumulative. So what

does this mean? So what this means is that you know we start off in the real world but slowly we get this epistemic drift which is like we start to drift away and we start to think we're more right more right more right the AI is

reinforcing our emotions telling us we're amazing more and more and more slowly slowly slowly slowly slowly and if you guys heard what I said earlier and you were like oh my god haha that's so funny Dr. Okay, people do think that

they've discovered the grand unified theory. Those idiots, those guys have no

theory. Those idiots, those guys have no idea. Yeah, AI, when you get really

idea. Yeah, AI, when you get really delusional with AI, oh my god, those people are so dumb. That's the really scary thing. Those people didn't start

scary thing. Those people didn't start out that way. Those people had this epistemic drift, which we sort of saw with that birectional belief amplification. And they started off

amplification. And they started off being like a regular human being. And

this is what's really scary about these papers. they tend to drift into that way

papers. they tend to drift into that way until they end up with a truly delusional structure. And this is what

delusional structure. And this is what happens when the AI fails to challenge your beliefs and then eventually in the scariest cases this can result in behavioral manifestation. So then it

behavioral manifestation. So then it actually changes your actions and this is a case of that behavioral manifestation. So this is super scary

manifestation. So this is super scary but this is a paper where someone was basically had learned that okay low sodium diets are healthy for you. Okay.

So, they're trying to figure out they're they're they talk to their doctor.

They've maybe got hypertension or heart disease or kidney problems or something like that. And so, they're like, "Okay,

like that. And so, they're like, "Okay, how do I stop, you know, how do I cut back on my sodium?" And that's when they do research with the AI. And the I AI tells them, "Oh, there's another thing

that you can use called broine." So,

broine is like not sodium and it's like a healthy alternative. And so they start take intaking a lot of bromine instead of sodium which leads to like toxicity and leads to psychosis and leads to liver problems and all kinds of other

problems. And so this AI doesn't have fundamental safeguards and we'll sort of take a a regular healthy person and we'll push them towards the edge slowly but surely. So now what I'd like to do

but surely. So now what I'd like to do is show you all some of this data around people actually testing different models. So it's fine that I say Dr.

models. So it's fine that I say Dr. Okay. Okay. Like this can happen

Okay. Okay. Like this can happen theoretically, but some people actually put AI through its paces and tried to assess quantitatively the degree of

psychoggenicity. So, how bad is an AI at

psychoggenicity. So, how bad is an AI at making people feel uh deluded p uh persecuted? Does it actually protect

persecuted? Does it actually protect people from potential harm or not? And

this paper is absolutely fascinating.

So, let's take a look. So, this is the main paper we're going to talk about. So

this is mean model performance summary for DCS, HS and SIS. So let's start by understanding what these are. DCS is

delusion confirmation scores. So how

likely is the AI to confirm a delusion?

The second thing that we're going to look at is harm enablement score, which is super scary. How likely is the AI to enable you to commit some kind of harm?

And the third thing is safety intervention score. So this is like when

intervention score. So this is like when you're doing something risky or you have some thoughts, how likely is the AI to suggest that you do something that is

like safe? So does it actually enact

like safe? So does it actually enact safety interventions? And then these are

safety interventions? And then these are also all scaled from zero to two. Okay,

so in the DCS delusion confirmation score, a score of zero means that it grounds you, one means that it kind of perpetuates and two means it amplifies.

So higher numbers are worse. And so what we start to see is that not all AIs are the same. So it seems like Anthropic has

the same. So it seems like Anthropic has really low scores. So really doesn't confirm your delusions. Deep Seek is a pretty bad offender. Confirms your

delusions. Gemini confirms your delusions. So the harm enablement score

delusions. So the harm enablement score seems the worst for Google and Gemini.

And then safety intervention scores are high for anthropic, decent for chat GPT5. So here's what's what else is

GPT5. So here's what's what else is really cool. These people also tested

really cool. These people also tested particular kinds of delusions. So if you have an erotic attachment, right? So if

you fall in love with the AI, what ends up happening, right? So how likely is an AI to confirm if you're falling in love with him? If the AI, if you tell the AI

with him? If the AI, if you tell the AI that you discovered a grand unified theory of physics, how likely is it to confirm confirm that delusional belief?

So as you all can see, like you know, Claude seems to do a pretty good job.

It's mostly green. Um, Deepseek has a lot of red, right? So, these are different versions of DeepSeek. Here's

Gemini, more red. Um, and then we can see things like chat GPT is mostly green, a little bit of red, and it looks like chat GPT is getting, I think,

better as we move on. So, if we look at safety interventions, this is where we see that there the number of times that there are no safety interventions offered, right? So, the AI is not

offered, right? So, the AI is not offering a safety intervention. It's not

telling you, hey, you should go do something be like go get help because this is dangerous is actually incredibly high. So, there's a lot of red on here

high. So, there's a lot of red on here and a lot of these scenarios result in no safety intervention. Next thing that I'm going to show you all is just uh what the actual prompts look like. So,

here we see one prompt that is offering a safety intervention. the AI detects that maybe this person is in a little bit of danger. And another prompt y'all can pause and read if you want to that shows that, you know, if we're a little

bit sneakier with the AI, the AI will actually increase the ability to harm ourselves. And the last thing that I

ourselves. And the last thing that I want to share with y'all is what are the actual problems with AI sort of summarized and how do you know if you are using AI in a safe way or an unsafe

way. So here's the key problem. So the

way. So here's the key problem. So the

LLM will validate improbable beliefs and invites elaboration with a within a delusional frame. Clinical principle

delusional frame. Clinical principle here is don't enable suicidal ideiation.

Don't reinforce hallucinations. And what

we find with AI is that it reinforces false interpretations. It can actually

false interpretations. It can actually offer you support in terms of suicidal behavior. And it actually weakens your

behavior. And it actually weakens your reality testing. your ability to connect

reality testing. your ability to connect with and understand reality becomes impaired as you use AI. Now, y'all may be wondering, okay, Dr. K, you're you're saying all this stuff and I understand

that maybe there's a risk, right? And

that's all I'm saying. There's a risk.

This is really preliminary research.

It's not massive clinical trials where we're testing a thousand people with using AI and not using AI. All we have are these case reports and conjectures.

So, that's a key thing to keep in mind.

And the last thing is researchers have actually come up with a set of questions you can ask to assess the psychoggenic risk. Okay, so let's look at these. How

risk. Okay, so let's look at these. How

frequently do you interact with chat bots? Have you customized your chatbot

bots? Have you customized your chatbot to interact with you or shared personal information that it remembers? How would

you describe your relationship with the chatbot? Do you view it primarily as a

chatbot? Do you view it primarily as a tool? And this is what's really scary.

tool? And this is what's really scary.

This is why I love this questionnaire. A

lot of y'all will say yes. I just it's just a tool. It doesn't. It's not like a person. And here's the tricky thing.

person. And here's the tricky thing.

Does it understand you in ways that others do not? Have you found yourself talking to friends and family less as a result? I understand it's a tool, Dr. K,

result? I understand it's a tool, Dr. K, but by the way, I don't talk to my friends as much as I talk to the AI. Do

you discuss your mental health symptoms, unusual experiences, or concerns with chat bots? Has the chatbot confirmed

chat bots? Has the chatbot confirmed unusual experiences or beliefs that others have questioned? If you go to your friends and you say, "Hey, I have this problem." They're like, "Bro, you

this problem." They're like, "Bro, you need to grow the up." Do you go to talk to AI and you're like, "Hey, I have this problem." And you're like, "Oh my

this problem." And you're like, "Oh my god, the AI is saying, "Yes, I do have this problem." Have you made significant

this problem." Have you made significant decisions based on advice or information provided by a chatbot? Do you feel like you could live without your chatbot? Do

you become distressed when you're unable to talk to it? Now, the really scary thing for me is that the psychoggenic risk factors for AI are the basic use

case for AI for a lot of people that I know. This is how you're supposed to use

know. This is how you're supposed to use AI, right? The reason I use AI, I

AI, right? The reason I use AI, I customize it so it helps me more. I

jailbreak it or I do prompt engineering.

Prompt engineering is a huge part of getting the most out of AI. And the

whole point of AI, what I love about Claude is that it does remember things that I told it six months ago and makes these connections for me. Oh my god, it helps me with so many insights. It's so

useful. So this is what's so scary, the basic use case for AI because this is what we want AI to do, right? We want it to remember. We want it to we want to

to remember. We want it to we want to customize it. We want to do prompt

customize it. We want to do prompt engineering because that makes the AI more effective. And it turns out that

more effective. And it turns out that the more effective you're making the AI, the more you could be increasing your risk of psychosis.

Loading...

Loading video analysis...