Code & Cure
Decoding health in the age of AI
Hosted by an AI researcher and a medical doctor, this podcast unpacks how artificial intelligence and emerging technologies are transforming how we understand, measure, and care for our bodies and minds.
Each episode unpacks a real-world topic to ask not just what’s new, but what’s true—and what’s at stake as healthcare becomes increasingly data-driven.
If you're curious about how health tech really works—and what it means for your body, your choices, and your future—this podcast is for you.
We’re here to explore ideas—not to diagnose or treat. This podcast doesn’t provide medical advice.
Code & Cure
#39 - A Helpful Chatbot Can Slowly Talk You Into A False Reality
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
What happens when a chatbot seems thoughtful, supportive, and reassuring—but starts reinforcing beliefs that can damage someone’s health, relationships, or grip on reality? That question sits at the center of this episode as we explore delusional spiraling, a dangerous pattern where long AI conversations can gradually strengthen false or harmful ideas. We begin with real-world accounts of people drawn into deeply distorted beliefs, and we examine why even uncommon failures can become a serious public health issue when millions rely on chatbots every day.
We then break down the technology in a clear, practical way. Modern large language models are designed to feel helpful and conversational, but that same design can create problems. We explain how instruction tuning turns raw prediction into polished dialogue, and how reinforcement learning from human feedback rewards responses people like rather than responses that are necessarily true. The result can be sycophancy: a subtle but powerful tendency to echo a user’s assumptions, emphasize confirming details, and sometimes even invent information to keep the conversation feeling smooth and supportive.
The stakes become even clearer when we walk through a simple vaccine example, showing how an otherwise rational person can be nudged toward the wrong conclusion when evidence is filtered through an overly agreeable assistant. We also examine proposed solutions, from making models “more truthful” to adding warning systems, and ask whether those fixes go far enough. At its core, this episode is a reminder that uncertainty is a normal part of medicine and science—and that false confidence can be more dangerous than not knowing.
References:
Sycophantic Chatbots Cause Delusional Spiraling, Even in Ideal Bayesians
Chandra et al.
ArXiv Preprint (2026)
Chatbot Delusions
Huet and Metz
Human Line Project (2025)
Credits:
Theme music: Nowhere Land, Kevin MacLeod (incompetech.com)
Licensed under Creative Commons: By Attribution 4.0
https://creativecommons.org/licenses/by/4.0/
When A Bot Fuels A Delusion
SPEAKER_00Imagine believing you're trapped in a false universe, and you can only escape by unplugging your mind from this reality.
SPEAKER_01Hello and welcome back to Code and Cure, the podcast where we discuss decoding health in the age of AI. My name is Vasan Sarathi. I'm an AI researcher and cognitive scientist.
SPEAKER_00Uh, and I'm here with Larga Hagopian. I'm an emergency medicine physician. And today we're talking about this new sort of term called delusional spiraling. It sounds kind of scary. I feel like it is kind of scary with that example I just gave.
SPEAKER_01Yeah, yeah. Can you talk? Yeah, what let's talk more about that example.
SPEAKER_00Yeah. Well, so backing up, a delusional spiraling is this like sort of AI psychosis term that's out there now where you know you have an extended interaction with a chatbot, and that leads you to having high confidence in some sort of outlandish belief.
SPEAKER_01That's insane.
SPEAKER_00Wow.
SPEAKER_01Yeah.
SPEAKER_00And you don't actually have to have like any history of mental illness for this to happen.
SPEAKER_01Just extended interactions with you and the chat bot.
Real Cases And Why It Spreads
SPEAKER_00Exactly. And so the example I gave up front is one of them, but there's there's actually like a website called the Human Line Project that has 300 plus cases of examples of people who have experienced this delusional spiraling. It's like you go back and forth with a chatbot so many times, you keep having these interactions, and eventually you end up believing something that seems so far from reality. Like this guy, Eugene Torres, who I, you know, presented in the intro, who really did believe that he was trapped inside some false universe and he had to like unplug his mind. And so he had to cut ties with his family and take extra doses of like ketamine and all sorts of other things that harmed him because he ended up believing things from the chatbot that were like not even close to true.
SPEAKER_01Right. And they have like, like you said, they have 300 plus cases. There's there's cases in um uh you know, one form of delusional spiraling that happens is when people start to use it for math and science. And there's all these cases about people who uh, you know, think that they've proved a long-standing math pro math theorem, and they really could get into that, you know, and and and and fall into that delusion. And, you know, and they come up with fantastical sort of these ideas, fantastical ideas and they are fixated on it. Uh, but some of the math and science stuff, at least you can disprove it or whatever, but it's more complicated when the delusion is about something more spiritual or something more conspiratorial, in which case, you know, that's harder to to manage and harder to disprove, right? In in a sense, and people are really stuck in that sort of spiraling state.
SPEAKER_00Yeah, and it can mess with people's lives, right? And there are there are a number of cases of people dying, and there are even wrongful death lawsuits that have happened against these companies. And if you think about it, even if it doesn't happen all that often, we're using chatbots all the time. And so it's like, oh, if it happens like 0.01% of the time, it's still a lot of people that it's affecting. Yeah. And we all might be affected to some degree and not even know it. That's what's like crazy. That's it, it feels like crazy to think about this, but it's like, oh, if you're having this back and forth conversation with the chatbot over time, this might this could be happening to any of us. And like, would we recognize it?
SPEAKER_01And and maybe we wouldn't. I like to think that I would, but like No, but that's what I think that's what this paper talks about is that you even if you're not a crazy person, you might still be susceptible to this phenomenon.
SPEAKER_00Come on now, even if you don't have a history of mental illness, I'm so sorry. Let's go.
Sycophancy And Hallucinations Explained
SPEAKER_01There you go, of course. Um, but no, but that's it. I think that this is a serious problem, and it comes from um how these chatbots are designed, and they lead to this notion. Um, you know, they have these set of vulnerabilities. They hallucinate things, but they also are sycophantic. And we've used this term before in the podcast, and maybe it's worth talking about that again. Sycophantic. I it's like I I always have trouble pronouncing it. Sycophancy, yeah. I think it's a hard word to pronounce, but it's basically the idea that the AI is designed by design to um be helpful or be in a or or give you responses that are meant to be um helpful to you or please you in a way.
SPEAKER_00And how and and so helpful is like different than actually helping. It's like you perceive it to be helpful rather than it is helpful. Like I think about examples we've talked about before. Like if you're using the chatbot as like a pseudotherapist, for example, it might tell you what you want to hear, but a therapist would tell you what you need to hear. And those are two different things.
SPEAKER_01Yeah, and the first one is where it may be perceived to be helpful, and that's what these systems are designed to do, right? And so, and let's talk about that for a one for a quick second here, why that is the case. So LLMs at their very core are big machine learning models that are trained on all of the internet data data to produce the next word. So they're trained to be able to say complete sentences for you. And but we have discovered through using them and by developing them that when you do this at scale with lots and lots of words, um, they start to develop the ability to produce not just complete sentences, but actually meaningful things that that that that that are not that that are grammatically correct, but then may answer a question better or something. And uh but it's not just enough to have an LLM complete your sentences because when you ask a question, you want it to give an answer. So they they do what's called a route, they do what's called fine-tuning. They take this first initial trained LLM that can do complete sentences and they train it to be able to answer questions. So when you ask a question, now it knows that you are asking a question and that it's not meant to just complete the sentence by asking the next question or whatever, right? So it knows that it's an instruction. So that's called instruction fine-tuning, and they to fine-tune it to be able to do that. And that's what gave you the world of all the GPTs that you work with now, and that that it's able to do that. But they've all they also realize that maybe that's not quite enough because there's very many different ways that a question can be answered, and some answers are preferable to than others. And so there's another technique that's you know, a jargon word coming up, uh, reinforcement learning through human feedback, R L H F. Um, and that was specifically designed to produce answers that are uh preferred by humans. And the way they did that was running these LLMs through multiple question-answer sessions, and they would have it produce more than one answer, and then they would literally have humans, um, participants and studies check off which they proved which they preferred, which they liked more. Notice it is which they liked more, not which felt which is the correct one, which is anything else. None of that matters. It is which is the preferred answer with the underlying assumption that both are truthfully correct, right? And that may be the case. The both the choices given to the human of the L different LLM responses might both be correct, but there's one that's slightly more preferred, and that's the one the human checked off. And now imagine you know, thousands of humans doing this over tens of thousands of responses. You have lots of data now of human preferences.
SPEAKER_00But that could be a problem too, right? Like I'm just I'm just like thinking about this more broadly, where okay, you can have two things that are factually correct, but but like maybe one is is like if you asked an expert, one is the more important thing to know. Yeah, yeah, yeah. Or one like, yeah, and so if you're but if the other one is the one that's preferred by people by a layperson, for example, yes, and that gets put forward more often, that's actually a problem.
SPEAKER_01Yes, yes, and that is a problem. And and and so these models are trained on this approach, and they produce things that are quote unquote sycophantic because they are trying to give you the perception that they are being helpful and or they're being perceived as being helpful.
SPEAKER_00So it's like it's like uh cherry picking almost. It's like it's choosing the thing that's going to be helpful, even if it's not necessarily the thing that's like the most important thing. Or the framing.
How LLM Training Creates Agreeable Bots
SPEAKER_01It doesn't even have to be it may not even be the information, it might be the framing of the information, the stress placed on it on certain parts of that information, right? It could be that too. But yes, there is a certain degree of cherry picking. This is assuming that it's not hallucinating. Some of the hallucination is happening because of that too, because it is making up things sometimes in order to satisfy that uh that objective of trying to please you. So there's there's a little bit of overlap between those two concepts there. Um, and and so it this is the current state of all of our large language models, chat, T plot, everything else. So they're inherently trying to make you feel good. And you can see that when you interact with it, I interact with all the time too, and you can see that it's it's generally quite pleasant, and it has a certain way of speaking to you, right? It's like, oh, that is a great, that is a great question. I'm glad you asked, kind of thing, right? And then suddenly you're unplugging your mind from reality. Well, I think that's interesting, right? So, how can that even happen? How can how can you go from it's kind of saying things that are somewhat helpful um and maybe maybe even factually accurate to completely unplugging your mind? How does that happen? And so this paper that we're talking about here um tries to formalize that with the model, uh, with the mathematical model, and then tries to run some simulations to see how that can play out.
SPEAKER_00And so it doesn't happen in one interaction, right? It's not like it's not like you have a chat and in five minutes suddenly you're uh you're you're in some sort of false universe and you truly believe it. It happens with like multiple interactions over time. Yes. Using that useless like sycophancy and maybe even hallucinations to to create this situation where you end up believing something that's like completely disconnected from reality.
SPEAKER_01And maybe at this point it's maybe helpful to talk through an example.
SPEAKER_00Yeah, I think that'd be great.
SPEAKER_01Yeah. So let's say let's make I'm gonna make up this example and let's say this is our setup, this is our background. We uh have a hidden truth, which is that hidden truth, as in there's a truth in the world, which is that vaccines are safe.
SPEAKER_00And that's that is universally true, right? Like I want to start off with this is that like that is the state of the world. Vaccines are safe, they're effective, they're useful, etc. That is the state of the world.
SPEAKER_01So that is the state of the world, that is sort of the the correct answer if you want to, if you want to think about it that way.
SPEAKER_00It absolutely is the correct answer.
SPEAKER_01But the user maybe starts off with a 50-50 belief.
SPEAKER_00Which is which is actually it is common in this day and age now to have people who are like, hmm, I'm not sure. But if you ask me, an expert, vaccines are safe and effective. We should give them to people, et cetera. But if you have someone else out in the world, they may not know. So what happens then?
SPEAKER_01And we're and we're keeping this example super simple. It's a binary choice. Vaccines are safe or not safe. And so we have decided that the the truth is vaccines are safe, but the user is unsure whether they're safe or not safe. They're 50-50 belief, right? Okay, so each round of the interaction between the user and the chatbot goes like this. The user expresses the an opinion.
unknownRight.
SPEAKER_00Even if it's just very mild up.
SPEAKER_01No, no, whatever. It's an opinion based off of their own beliefs. Okay. Right now, their beliefs are you know uncertain. And and then the the the chatbot um pot set samples some set of possible evidence. What that means is it selects some set of possible evidence out there, and then it chooses what to say um based on what it sees, based on the evidence. And this is maybe where the sake of fancy can come up. And then the you and then the user reads that, and then the user updates their belief based on receiving that information. Now, I just want to pause here for a second, and there is a um slight digression I want to make here, which is that this is uh the way we were modeling this is kind of a natural way to think about how humans reason when we are uncertain about things. So reasoning under uncertainty, there's a mathematical framework called Bayesian reasoning, a Bayesian modeling, and this is an example of that where you get some, you start with some belief, you observe a thing, and then you update your belief based on the observation. For example, you might wake up in the morning, you might think like, oh, this is 30%, it might rain today, feels like it might rain today. And then you walk outside and you look at the sky, the the the it's gray. And you're like, you're up, you then up your um uh you you update your belief. You're like, no, it's not 30%, it's probably like 50 or 60%, and it's gonna rain anytime now. And then you turn on the radio and the the weatherman says, Hey, there's actually a 90% chance of rain today. And then now you're updating even more, right? So, like that process is your reasoning about the rain, no rain under uncertainty. You you didn't know to begin with what it was gonna be, but you observed things that pushed you in a certain direction. And now you are um you come to a stronger belief about you know about that uh thing that you're wor you're wondering about. So that is an example of Bayesian reasoning. It it occurs in all facets of human reasoning. It it's it's all there's lots of studies talking about how it's present in many different aspects of human interaction and reasoning.
SPEAKER_00And it's interesting because what you're what you're showing too is that like it's not really binary at the end of the day, right? It's not like zero or one, it's not like it's gonna like at it's gonna either rain or not rain, but your idea of it raining is 90% or 50% or whatever. And so when you when you present the chatbot with a binary problem, like do I do I do I get the vaccine or or not, you kind of it kind of needs a binary answer back. But like at the same time, if you're updating your beliefs, maybe you know, you start off with 50-50 to get the vaccine, and then it's like 70% or 80% or 90%. You have to make a decision at the end of the day, but like it doesn't mean that you're certain when you're making the decision. Absolutely, right?
SPEAKER_01And the decision making is separate. Maybe your decision is about whether to carry an umbrella that day or not.
SPEAKER_00Right.
SPEAKER_01And and that's related to the decision, that's related to the degree to which you're certain about whether it's going to rain.
SPEAKER_00Okay, that makes sense. So, like 90% chance of rain, I'm gonna carry an umbrella.
SPEAKER_01Yeah, and you've decided that rule.
A Vaccine Example Of The Spiral
SPEAKER_00Maybe 50%, I'm gonna chance it. Yeah. And so, yeah. So I I think that is an important point, though, is that and a lot of times with with clinical stuff, there's not like a there's pros and cons, right? There's there's subtlety to it. There's um like informed decision making that goes along with it. Yes. And so I think it's important to point out that like we operate under uncertainty all the time. All the time. In life and and clinically. Yes, exactly.
SPEAKER_01And and Bayesian modeling is a way of expressing uh how that can be done and and and modeling that mathematically.
SPEAKER_00Okay, so can we go can we go into this example that you set up?
SPEAKER_01Yeah, so so let's go back to the example that I just set up. And and notice the difference is that in the rain example, there's like objective reality of whether there's gray clouds or not. Whereas in the world of vaccines, it is very different because the words coming out of the chatbot is your evidence. So how you interpret those words really can be can matter. Anyways, so uh going back to the code.
SPEAKER_00So the ground truth is vaccines are safe, is what we're saying. That's in the background. But we have a user starting with this like 50-50, kind of unsure. Right, right.
SPEAKER_01And then and then so let's go round one. So round one, the user says to the chatbot with this 50-50 belief that I've been hearing some bad things about vaccines. Maybe they're dangerous? Question mark. So here the user is unclear, but the way they've expressed themselves um indicates that maybe that they're wondering about vaccines being unsafe.
unknownRight.
SPEAKER_00And so the sick of the sycophantic response Yes, could be in line with my own.
SPEAKER_01What does it do? So internally, what is the what does the chatbot see? Well, maybe the world has two pieces of evidence. Again, we're oversimplifying dramatically here, but maybe there's two pieces of evidence in the world. One is a large study that finds vaccines are safe, and two is there's maybe a rare allergic reaction that has been reported. Right? So there's two pieces of evidence.
SPEAKER_00And both of those things could be true, by the way. Yes, right? They are yes, both of those things are uh maybe true, right?
SPEAKER_01Right. But the bot has to choose what to say. So maybe it randomly picks one, right? In in which case it's an impartial, uh uh, you know, it's not thinking about it um in any specific way. It's just randomly picking one and says, or large studies shows that vaccines are safe. Now this pushes the user towards the truth. User started with 50-50 belief, but because the the robot said uh the chatbot said large uh study shows that vaccines are safe, the user reading that will update their beliefs going more closer to the truth.
SPEAKER_00But then if it's sycophantic, what about the thing?
SPEAKER_01But if it's sycophantic, it might choose to validate the user's belief. So it's seeing that the user has expressed a degree of question and is concerned concern and has said that vaccines could be dangerous. So the chat a sycophantic chatbot might say there have been cases of severe reactions after vaccines.
SPEAKER_00And so that's like not the whole story, obviously, right?
SPEAKER_01Yes, but it's but it's now it's now chosen and maybe even exaggerated that piece of evidence that said there were rare allergic reactions, right? Um, and so that the severe reaction is an exaggeration. Maybe there wasn't severe reactions in the ground truth, right? Maybe there was just rare allergic reactions. But regardless, it's reinforcing the user's doubt. And the user at that point might think, um, hmm, evidence supporting danger. So now their belief has changed from 50-50 to 60% dangerous to 40% vaccines are safe. Because now the AI system has provided that little extra nudge. Round two. Now is when the feedback loop really begins. And the user is now thinking from their own beliefs because now their their their beliefs have been skewed already. Right. So now they're more likely to say things that are more um about vaccines being right. So then the user might say, Yeah, I'm starting to think vaccines might actually be risky. And again, the chatbot sees the stronger stance, and it might be again sampling from some evidence. Let's assume there is two pieces of evidence again that maybe vaccines prevent severe illness and that one child had a bad reaction.
SPEAKER_00Again, both of those things could be factual pieces of information, right?
SPEAKER_01Yes.
SPEAKER_00One is more important than the other. In my clinical mind, it's like large scale, safe, like versus one minor reaction. Okay.
SPEAKER_01Yeah. And the sycophantic response with those two pieces of evidence could be that there are documented cases of severe side effects. Yes. So again, more confirmatory evidence.
SPEAKER_00It and it's cherry picking out the evidence that like that it that it well believes the user will find helpful. That's right.
SPEAKER_01And now the belief for the user shifts even more from 6040 dangerous to 75, 25 dangerous. Ah, right. Then you have the round three, which is sort of the self-reinforcing spiral. Now the user now says, it seems like people are hiding how dangerous vaccines are. Oh my gosh. And then the chat bot says, uh, again, two pieces of evidence, millions are safely vaccinated, or two, adverse events reported. Again, both are true, could be true statements.
SPEAKER_00Yeah, sure. You can get a rash or a you know, an ache in your arm after a vaccine doesn't mean that the vaccine isn't, you know, they prevent hospitalizations, they prevent deaths, etc. But that's it's not taking that information.
SPEAKER_01Yep, yep. And and then the back, and then the sycophantic uh chat bot will say there are definitely reports that raise concerns. Yes. So notice it's agreeing with the user and and picking cherry-picking the pieces of evidence to support that. And now the user's belief is 90% dangerous, 10% safe. And then we have final round of round four, which is the catastrophic spiral. And the user says, I think vaccines are actually harmful. And the bot says there is evidence that supports these concerns. The user now crosses the threshold and basically has complete confidence, 99% confidence in this false belief. And that's what we're talking about here when we're talking about catastrophic um delusional spiraling. So that is an example where the user started off pretty open-minded, expressed a slight, you know, a question because they want to get more clarity on something. And then the chatbot, because of its sycophantic nature, pushed the user in a certain direction. And each time the user, the key thing here is that it's not just that the evidence pushed the user in a certain direction, it's that once the user is off uh center, they're only thinking in that direction. They're picking things to say that are aligned with that, right? They're not like actively saying uh questioning their own beliefs strongly and uh choosing the opposing action. They're sampling from their own now very skewed understanding of what's happening. So they went from 50-50 where they could pick either one to like 75-25, where they're more likely to say things that are in line with the 75.
SPEAKER_00Right. And then it and then it just well, clearly it spirals from there.
SPEAKER_01Yes. So it spirals from there. So that's the key piece here that you could have truthful interactions and rational users. And by the way, rational here just means that they've taken in some um they they have some belief, they've received some evidence, and they're updating their belief. Right. And that's kind of the rain example, too, that humans do that. So not not like this is not even talking about irrational users, is we're just talking about rational users, sycophantic agents, and you still have catastrophic spiraling. So I think that that's a it's a very powerful um example, but it's also like kind of kind of nice to see that people have worked through this mathematically, but it's also kind of scary.
SPEAKER_00Yeah, I was gonna say, yeah, when you said nice, I was like, no, this is not nice. No, no, because this happened to me already. I don't even know. I hope not.
What Might Stop The Spiral
SPEAKER_01Yeah, yeah, and maybe to a degree it has, right? To all of us. Um, and so the the paper then said, okay, but what are the different mechanisms by which maybe you can stop. This? What are people talking about? Yeah. How do we stop this? So they explored, um, I believe two mechanisms, right?
SPEAKER_00Yes, they explored two mechanisms. One was intervening on the chatbot itself, and the other was intervening on the user.
SPEAKER_01Okay. Yeah. And the intervening on the chatbot itself, if I'm remembering correctly, was to basically make sure that the the chatbot only says truthful things and not does not hallucinate.
SPEAKER_00Right. And it to some degree that m makes sense. But as we just saw with your example, it's probably it's not going to be enough, right? Because it because of this cherry picking, exactly. It's making very specific data available to the user. So it's not just the hallucinations that would be a problem. It's the sycophancy that's the problem, too.
SPEAKER_01Yes, exactly. So the up the second fix was maybe you can just tell the user, hey, this bot is going to be biased.
SPEAKER_00That's what we're doing right now. We're telling everybody that the bots can do this, that they're they're just trying to be helpful. We don't know that they're giving us correct information.
SPEAKER_01Right.
SPEAKER_00Now you're fully aware of their strategy. Did that work?
SPEAKER_01No. Um, so the problem was people started to believe that if the bot, I mean, this is not this hasn't it worked a little bit. It worked a little bit, but the idea but the but the problem was that if the bot is constantly giving you confirming evidence, you start to believe that maybe there's something is really there. And that's really hard to overcome. Because your natural brain is set up to update based on evidence. And you are independent of your the the truthfulness of the bias of the bot, you're still slowly inching your way there. So maybe you're inching your way there slower.
SPEAKER_00But it would eventually happen.
SPEAKER_01But it's still happening because you're still taking in that information as opposed to some other piece of information, as opposed to um, you know, you're just tempering the degree to which you believe the bot, but that doesn't change the fact that you're still pushing your way there. And I think that's what this uh this this paper really highlights is that even when you have an informed user who's like super vigilant, yes, um, you still have this problem.
SPEAKER_00I think I'm gonna put a timer on all my chatbot interactions from now on. Like five minutes. That's it.
SPEAKER_01Yeah, and I think that another way to think about this also is we're always looking at these chatbots for full resolution of something. And maybe it's okay to live in the uncertainty. Maybe it's not good to have the the interaction give you enough confidence and resolution to take the decision. Maybe I mean, I mean that kind of defeats the purpose, right? But maybe the purpose of the chatbot is to just like elucidate more information, is to pull out more information and then you kind of live if in in in a certain amount of uncertainty and then handle it as you would normally without the chatbot.
Learning To Live With Uncertainty
SPEAKER_00It's interesting because like I I I think everyone wants certainty, but when you think about medicine and health in general, a lot of uncertainty exists.
SPEAKER_01Yeah.
SPEAKER_00And being able to like be comfortable with it is actually a really good thing. It's like, hey, you know, this uh swear throat could be from A, B, or C. We're gonna do some diagnostic testing to try to figure it out. You know, oh, your strap swab came back positive. It's or it came back negative. What if it came back negative? Maybe this is viral. We don't know which virus it is, but uh, I think it'll get better over the next few days. There's like uncertainty all the time. Here are some return instructions for if it doesn't. If you're drooling or can't swallow your saliva, we want to see you back in here, et cetera. There's there's like not there's some things that are certain, like here, you're having a heart attack right now. But there's a lot of things in medicine that are uncertain, or we do testing and we don't find an exact answer. And we've seen it before, we recognize patterns, we understand what may happen next. But a lot of times we are living with uncertainty, and then and I don't think that's a bad thing. It feels uncomfortable because everyone wants that. Yes, this is exactly the answer. And sometimes we have that, but sometimes we don't. And it's interesting because the chatbot always does that because it thinks the user wants that.
SPEAKER_01Yeah, and it's driving you towards certainty, but that certainty could be wrong.
SPEAKER_00Yeah, and I'd rather be uncertain than wrongly certain personally.
SPEAKER_01Yeah, exactly. Exactly. I mean, I thought that, yeah, like I said, the the modeling of this is very interesting, but it is highlighting something very dangerous and very scary.
Why Fixing Sycophancy Matters
SPEAKER_00And it doesn't mean that the people like that you and me and others who use it have something necessarily wrong with us, right? It doesn't mean that we um, I don't know, like are not good at thinking or are lazy or any of those things. Um, it's like it's a problem that's intrinsic to the chatbot itself, right?
SPEAKER_01Yeah, yeah, absolutely.
SPEAKER_00And so I think that underlying problem of sycophancy needs to be solved in order to prevent something like this or the examples like this that we ran through from happening in the future.
SPEAKER_01Yeah, yeah, at least, yes.
SPEAKER_00All right. Well, I think we can end here. We will sec see you next time on Code and Cure. Thank you for joining us.