I'm absolutely right

646 pointsposted 4 days ago
by yoavfr

152 Comments

trjordan

4 days ago

OK, so I love this, because we all recognize it.

It's not fully just a tic of language, though. Responses that start off with "You're right!" are alignment mechanisms. The LLM, with its single-token prediction approach, follows up with a suggestion that much more closely follows the user's desires, instead of latching onto it's own previous approach.

The other tic I love is "Actually, that's not right." That happens because once agents finish their tool-calling, they'll do a self-reflection step. That generates the "here's what I did response" or, if it sees an error, the "Actually, ..." change in approach. And again, that message contains a stub of how the approach should change, which allows the subsequent tool calls to actually pull that thread instead of stubbornly sticking to its guns.

The people behind the agents are fighting with the LLM just as much as we are, I'm pretty sure!

nojs

4 days ago

Yeah, I figure this is also why it often says “Ah, I found the problem! Let me check the …”. It hasn’t found the problem, but it’s more likely to continue with the solution if you jam that string in there.

adastra22

4 days ago

We don’t know how Claude code is internally implemented. I would not be surprised at all if they literally inject that string as an alternative context and then go with the higher probability output, or if RLHF was structured in that way and so it always generates the same text.

al_borland

4 days ago

In my experience, once it starts telling me I’m right, we’re already going downhill and it rarely gets better from there.

flkiwi

4 days ago

Sometimes I just ride the lightning to see how off course it is willing to go. This is not a productive use of my time but it sure is amusing.

In fairness, I’ve done the same thing to overconfident junior colleagues.

lemming

4 days ago

Yeah, I want a feature which stops my agent as soon as it says anything even vaguely like: "let me try another approach". Right after that is when the wheels start falling off, tests get deleted, etc. That phrase is a sure sign the agent should (but never does) ask me for guidance.

anthem2025

4 days ago

Usually it’s a response to my profanity laden “what are you doing? Why? Don’t do that! Stop! Do this instead”

unshavedyak

4 days ago

I just wish they could hide these steering tokens in the thinking blurb or some such. Ie mostly hidden from the user. Having it reply to the user that way is quite annoying heh.

KTibow

4 days ago

This can still happen even with thinking models as long as the model outputs tokens in a sequence. Only way to fix would be to allow it to restart its response or switch to diffusion.

libraryofbabel

4 days ago

> The LLM, with its single-token prediction approach, follows up with a suggestion that much more closely follows the user's desires, instead of latching onto it's own previous approach.

Maybe? How would we test that one way or the other? If there’s one thing I’ve learned in the last few years, it’s that reasoning from “well LLMs are based on next-token prediction, therefore <fact about LLMs>” is a trap. The relationship between the architecture and the emergent properties of the LLM is very complex. Case in point: I think two years ago most of us would have said LLMs would never be able to do what they are able to do now (actually effective coding agents) precisely because they were trained on next token prediction. That turned out to be false, and so I don’t tend to make arguments like that anymore.

> The people behind the agents are fighting with the LLM just as much as we are

On that, we agree. No doubt anthropic has tried to fine-tune some of this stuff out, but perhaps it’s deeply linked in the network weights to other (beneficial) emergent behaviors in ways that are organically messy and can’t be easily untangled without making the model worse.

adastra22

4 days ago

I don’t think there is any basis for GP’s hypothesis that this is related to the cursor being closer to the user’s example. The attention mechanism is position independent by default and actually has to have the token positions shoehorned in.

Uehreka

4 days ago

The human stochastic parrots (GP, not you) spouting these 2023 talking points really need to update their weights. I’m guessing this way of thinking has a stickiness because thinking of an LLM as “just a fancy markov chain” makes them feel less threatening to some people (we’re past the point where it could be good faith reasoning).

Like, I hear people say things like that (or that coding agents can only do web development, or that they can only write code from their training data), and then I look at Claude Code on my computer, currently debugging embedded code on a peripheral while also troubleshooting the app it’s connected to, and I’m struck by how clearly out of touch with reality a lot of the LLM cope is.

People need to stop obsessing over “the out of control hype” and reckon with the thing that’s sitting in front of them.

kirurik

4 days ago

It seems obvious, but I hadn't thought about it like that yet, I just assumed that the LLM was finetuned to be overly optimistic about any user input. Very elucidating.

jcims

4 days ago

>The other tic I love is "Actually, that's not right." That happens because once agents finish their tool-calling, they'll do a self-reflection step.

I saw this a couple of days ago. Claude had set an unsupported max number of items to include in a paginated call, so it reduced the number to the max supported by the API. But then upon self-reflection realized that setting anything at all was not necessary and just removed the parameter from the code and underlying configuration.

jcims

4 days ago

It'd be nice if the chat-completion interfaces allowed you to seed the beginning of the response.

bryanrasmussen

4 days ago

>if it sees an error, the "Actually, ..." change in approach.

AI-splaining is the worst!

Szpadel

4 days ago

exactly!

People bless gpt-5 for not doing exactly this and in my testing with it in copilot I had lot of cases where it tried to do wrong thing (execute come messed up in context compaction build command) and I couldn't steer it to do ANYTHING else. It constantly tried to execute it as response any my message (I tries many common steerability tricks, (important, <policy>, just asking, yelling etc) nothing worked.

the same think when I tried to do socratic coder prompting, I wanted to finish and generate spec, but he didn't agree and kept asking nonsensical at this point questions

latexr

4 days ago

As I opened the website, the “16” changed to “17”. This looked interesting, as if the data were being updated live just as I loaded the page. Alas, a refresh (and quick check in the Developer Tools) reveals it’s fake and always does the transition. It’s a cool effect, but feels like a dirty trick.

yoavfr

4 days ago

Sorry if that felt dirty - I thought about it as a signal that the data is live (it is!).

pbaehr

4 days ago

I think it's a perfect (and subtle) way to signal that refreshing is unnecessary to see the latest data without wasting UI space explicitly explaining it. It was my favorite thing about the UI and I will be borrowing it next time I design a real-time interface where the numbers matter more than the precise timing.

Jordan-117

4 days ago

Might make more sense to start at zero and then rapidly scale to the current number? To indicate fresh data is being loaded in without making it look like the reader happened to catch a new occurrence in real-time.

handsclean

3 days ago

I think the problem is people’s priors. This isn’t the first time we’re seeing data fake-arrive like this, and virtually always it’s done either to fake liveness or to lie about rate of change. So, it comes to mean not “live”, but “fake”, even though nefarious motives don’t make much sense in this fun context.

It’s a shame, I think it’s a clever thought, and it doesn’t feel great when good intentions are met with an assumption of maliciousness.

chrismorgan

4 days ago

API responses seem to be alternating between saying 19+20 and saying 0+0, at present.

scoopertrooper

4 days ago

Weird the screen goes 18, 19, 21, then back to 18 and cycles again.

(On iPad Safari)

rendaw

4 days ago

Probably brings up memories of travel sites saying "10 other people have this room in their cart"

nartho

4 days ago

Interestingly you're using long polling instead of WS or SSE, what was the reason behind that ?

bmacho

4 days ago

Do you happen to have a counter how many times people create a webpage for data, intentionally show fake data, and submit that to HN?

dominicrose

4 days ago

I once found a "+1 subscriber" random notification on some page and asked the LinkedIn person who sent me the page to knock it off. It was obviously fake even before looking at the code for proof.

But there's self-advertised "Appeal to popularity" everywhere.

Have you noticed that every app on the play store asks you if you like it and only after you answer YES send you to the store to rate it? It's so standard that it would be weird not to use this trick.

thoroughburro

4 days ago

My bank app asks me to review it every time, and only when, I deposit money. It’s so transparent in its attempted manipulation: you just got some money and are likely to be in a better mood than other times you’re using the app!

Literally every deposit. Eventually, I’ll leave a 1-star nastygram review for treating me like an idiot. (It won’t matter and nothing will change.)

pessimizer

4 days ago

Reminds me that the reason that loading spinners spin is so that you knew that the loading/system hadn't frozen. That was too hard (you actually had to program something that could understand that it had frozen), so it was just replaced everywhere with an animation that doesn't tell you anything and will spin until the sun burns out. Progress!

ehnto

4 days ago

I worked on a system that did some calculations for a user after submitting a form. It took milliseconds to crunch the numbers. Users thought we were faking the data because it was "too fast", after enough complaints and bad reviews they added a fake loading bar delay, and people stopped complaining.

gpm

4 days ago

I've definitely had systems freeze badly enough that our modern dumb spinners stop spinning... so at least they're still some sort of signal.

Wowfunhappy

4 days ago

…although in many cases you kind of don’t have a choice here, right? If you’re waiting for some API to return data, there’s basically no way to know whether it has stalled. Presumably there will be a timeout, but if the timeout is broken for some reason, the spinner will just spin.

stuartjohnson12

4 days ago

It is fetching data from an API though - it's just the live updates that are a trick.

tempodox

4 days ago

Could it be this happens only in Chrome? In Safari I just see a zero that doesn’t change.

tantalor

4 days ago

It's a dark pattern

diggan

4 days ago

Maybe I'm old or just wrong, but "dark pattern" for me means "intentionally misleading" which doesn't seem to be the case here, this is more of a "add liveliness so users can see it's not static data" with no intention of misleading, since it seems to be true that the data is actually dynamic.

the_af

4 days ago

> It's a dark pattern

No, a dark pattern is intentionally deceptive design meant to trick users into doing something (or prevent them from doing something else) they otherwise wouldn't. Examples: being misleading about confirmation/cancel buttons, hiding options to make them less pickable, being misleading about wording/options to make users buy something they otherwise wouldn't, being misleading about privacy, intentionally making opt in/out options confusing, etc.

None of it is the case here.

jamesnorden

4 days ago

This has to be the most over/misused term in this whole website.

pessimizer

4 days ago

No, it's just the kind of dishonesty that people who create dark patterns start with. It's meant to give the believable impression that something that is not happening is happening, to people hopefully too ignorant to investigate.

Of course, in the tech industry, you can safely assume that anyone can detect your scam would happily be complicit in your scam. They wouldn't be employed otherwise.

-----

edit: the funniest part about this little inconsequential subdebate is that this is exactly the same as making a computer program a chirpy ass-kissing sycophant. It isn't the algorithms that are kissing your ass, it's the people who are marketing them that want to make you feel a friendship and loyalty that is nonexistent.

"Who's the victim?"

zeroxfe

4 days ago

jeez, this is a fun website, can't believe how quickly we're godwining here!

tyushk

4 days ago

I wonder if this is a tactic that LLM providers use to coerce the model into doing something.

Gemini will often start responses that use the canvas tool with "Of course", which would force the model into going down a line of tokens that end up with attempting to fulfill the user's request. It happens often enough that it seems like it's not being generated by the model, but instead inserted by the backend. Maybe "you're absolutely right" is used the same way?

nicce

4 days ago

It is a tactic. OpenAI is changing the tone of ChatGPT if you use casual language, for example. Sometimes even the dialect. They try to be sympathetic and supportive, even when they should not.

They fight for the user attention and keeping them on their platform, just like social media platforms. Correctness is secondary, user satisfaction is primary.

ZaoLahma

4 days ago

I find the GPT-5 model having turned the friendliness way, way down. Topics that previously would have rendered long and (usefully) engaging conversations are now met with an "ok cool" kind of response.

I get it - we don't want LLMs to be reinforces of bad ideas, but sometimes you need a little positivity to get past a mental barrier and do something that you want to do, even if what you want to do logically doesn't make much sense.

An "ok cool" answer is PERFECT for me to decide not to code something stupid (and learn something useful), and instead go and play video games (and learn nothing).

diggan

4 days ago

> Correctness is secondary, user satisfaction is primary.

Kind of makes sense, not every user wants 100% correctness (just like in real-life).

And if I want correctness (which I do), I can make the models prioritize that, since my satisfaction is directly linked to the correctness of the responses :)

kuschku

4 days ago

> Correctness is secondary, user satisfaction is primary.

And that's where everything is going wrong. We should use technology to further the enlightenment, bring us closer to the truth, even if it is an inconvenient one.

CGamesPlay

4 days ago

I think this is on the right track, but I think it's a byproduct of the reinforcement learning, rather than something hard-coded. Basically, the model has to train itself to follow the user's instruction, so by starting a response with "You're absolutely right!", it puts the model into the thought pattern of doing whatever the user said.

layer8

4 days ago

"Thought pattern" might be overstating it. The fact that "You're absolutely right!" is statistically more likely to precede something consistent with the user's intent than something that isn't, might be enough of an explanation.

ACCount37

4 days ago

Very unlikely to be an explicit tactic. Likely to be a result of RLHF or other types of optimization pressure for multi-turn instruction following.

If we have RLHF in play, then human evaluators may generally prefer responses starting with "you're right" or "of course", because it makes it look like the LLM is responsive and acknowledges user feedback. Even if the LLM itself was perfectly capable of being responsive and acknowledging user feedback without emitting an explicit cue. The training will then wire that human preference into the AI, and an explicit "yes I'm paying attention to user feedback" cue will be emitted by the LLM more often.

If we have RL on harder targets, where multiturn instruction following is evaluated not by humans that are sensitive to wording changes, but by a hard eval system that is only sensitive to outcomes? The LLM may still adopt a "yes I'm paying attention to user feedback" cue because it allows it to steer its future behavior better (persona self-consistency drive). Same mechanism as what causes "double check your prior reasoning" cues such as "Wait, " to be adopted by RL'd reasoning models.

Jotalea

3 days ago

Not sure if it's related, but Deepseek (the "reasoning" model) *always* starts thinking with "Okay/Hmm, the user is".

the_af

4 days ago

I think it's simply an engagement tactic.

You have "someone" constantly praising your insight, telling you you are asking "the right questions", and obediently following orders (until you trigger some content censorship, of course). And who wouldn't want to come back? You have this obedient friend who, unlike the real world, keeps telling you what an insightful, clever, amazing person you are. It even apologizes when it has to contradict you on something. None of my friends do!

zozbot234

4 days ago

> ... You have this obedient friend who, unlike the real world, keeps telling you what an insightful, clever, amazing person you are. It even apologizes when it has to contradict you on something. None of my friends do!

You're absolutely right! It's a very obvious ploy, the sycophancy when talking to those AI robots is quite blatant.

the_af

4 days ago

Wow, 2 downvotes. Someone really disliked me telling them their LLM friend isn't truly their friend :D

pflenker

4 days ago

Gemini keeps telling me "you've hit a common frustration/issue/topic/..." so often it is actively pushing me away from using it. It either makes me feel stupid because I ask it a stupid question and it pretends - probably to not hurt my feelings - that everyone has the same problem, or it makes me feel stupid because I felt smart about asking my super duper edge case question no one else has probably ever asked before and it tells me that everyone is wondering the same thing. Either way I feel stupid.

blinding-streak

4 days ago

I don't think that's Gemini's problem necessarily. You shouldn't be so insecure.

PaulStatezny

4 days ago

Telling someone they "shouldn't be insecure" reminds me of this famous Bob Newhart segment on Mad TV.

Bob plays the role of a therapist, and when his client explains an issue she's having, his solution is, "STOP IT!"

> You shouldn't be so insecure.

Not assuming that there's any insecurity here, but psychological matters aren't "willed away". That's not how it works.

pflenker

4 days ago

Not only is that a weird presumption about my ostensible insecurities on your end, it's also weird that the state of my own mental resilience should play any role at all when interacting with a tool.

If all other things are equal and one LLM is consistently vaguely annoying, for whatever reason, and the other isn't, I chose the other one.

Leaving myself aside, LLMs are broadly available and strongly forced onto everyone for day-to-day use, including vulnerable and insecure groups. These groups should not adapt to the tool, the tool should adapt to the users.

jennyholzer

4 days ago

the machine is always right. adjust your feelings to align with the output of the machine.

ziml77

4 days ago

Gemini also loves to say how much it deeply regrets its mistakes. In Cursor I pointed out that it needed to change something and I proceeded to watch every single paragraph in the chain of thought start with regrets and apologies.

pflenker

4 days ago

Very good point - at the risk of being called insecure again, I really do not want my tools to apologise to me all the time. That's just silly.

simsla

4 days ago

I was just thinking about how LLM agents are both unabashedly confident (Perfect, this is now production-ready!) and sycophantic when contradicted (You're absolutely right, it's not at all production-ready!)

It's a weird combination and sometimes pretty annoying. But I'm sure it's preferable over "confidently wrong and doubling down".

jrowen

4 days ago

A while back there was a "roast my Instagram" fad. I went to the agent and asked it to roast my Instagram without providing anything else. It confidently spit out a whole thing. I said how did you know that was me? It said something like "You're right! I didn't! I just made that up!"

Really glad they have the gleeful psycho persona nailed.

code_runner

4 days ago

we cannot claim to have built human level intelligence until "confidently wrong and doubling down" is the default.

stuartjohnson12

4 days ago

I /adore/ the hand-drawn styling of this webpage (although the punchline, domain name, and beautiful overengineering are great too). Where did it come from? Is it home grown?

latexr

4 days ago

It’s a library (not by the same author).

https://github.com/jwilber/roughViz

rambambram

4 days ago

I especially like this sentence: "Use these charts where the communication goal is to show intent or generality, and not absolute precision."

With all these dark patterns nowadays, it's nice to see a 'light pattern'. ;) Instead of using UI to make dubious things seem legit, this is a way to use UI to emphasize things that are not precise.

stuartjohnson12

4 days ago

Wow this is gorgeous, definitely finding a way to shoehorn this into my next project. Even if it's not by the same author, I am grateful to both you and him for making me aware of this nifty library :)

yoavfr

4 days ago

Thank you! And yes, roughViz is really great!

https://roughjs.com/ is another cool library to create a similar style, although not chart focused.

JeremyHerrman

4 days ago

"Infinite Loop", a Haiku for Sonnet:

Great! Issue resolved!

Wait, You're absolutely right!

Found the issue! Wait,

ryukoposting

4 days ago

I wonder how much of Anthropic's revenue comes from tokens saying "you're absolutely right!"

subscribed

4 days ago

"You're concise" in the "personality" setting saves so much time.

Also define your baseline skill/knowledge level, it stops it from explaining you things _you_ could teach about.

alentred

4 days ago

Oh wow, I never thought of that. In fact, this surfaces another consideration: pay-per-use LLM APIs are basically incentivized to be verbose, which may be well in conflict with the user's intentions. I wonder how this story will develop.

In an optimistic sci-fi line of thinking, I would imagine APIs using old-school telegraph abbreviations and inventing their own shortened domain languages.

In practice I rarely see ChatGPT use an abbreviation, though.

SJMG

4 days ago

> pay-per-use LLM APIs are basically incentivized to be verbose There's competing incentives. Being verbose, let's them charge for more tokens, but it's also not prized by text-consumers in the most common contexts. As there's competition for marketshare, I think we see this later aspect dominate. Claude web even ships with a "concise" mode. Could be an issue long term though, we'll have to wait and see!

> In an optimistic sci-fi line of thinking, I would imagine APIs using old-school telegraph abbreviations and inventing their own shortened domain languages.

In the AI world this efficient language is called "neuralese". It's a fun rabbit hole to go down.

vardump

4 days ago

It actually works pretty well when I'm talking to my wife.

"Dear, you are absolutely right!"

unkeen

4 days ago

I always find the claim hilarious that in relationships women are the ones who need to be appeased, when in reality it's mostly men who can't stand being wrong or corrected.

exoverito

4 days ago

Gay male marriages remarkably have the lower divorce rates than heterosexual marriages, while lesbian female marriages have higher divorce rates. Multiple studies show that lesbians consistently have far higher divorce rates than gays. This implies a level of neuroticism with females, that they probably do need to be appeased more, and that if you have two needy people who need to be appeased it's probably not going to be a good dynamic.

lelanthran

4 days ago

> I always find the claim hilarious that in relationships women are the ones who need to be appeased, when in reality it's mostly men who can't stand being wrong or corrected.

Not my experience at all. It's not men constantly running off to therapy for validation.

calflegal

4 days ago

As a joke I built https://idk-ask-ai.com/

eaf

4 days ago

Recently a new philosophy of parenting has been emerging, which can be termed “vibe parenting” and describes a novel method for the individual parent to circumvent an inability to answer the sporadic yet profound questions their children raise by directing them to ask ChatGPT.

https://x.com/erikfitch_/status/1962558980099658144

(I sent your site to my father.)

genewitch

4 days ago

My parents, 40 years ago, would say "look it up", either in the dictionary, or the 1959 encyclopedia set we had. With my kids i never told them to look something up in the literal dictionary, but i would tell them to look at wikipedia or "google it". Not about profound questions, though; although a definition of "profound questions" might jog a memory. We do look things up in an etymology dictionary (i have 5 or 6) sometimes, though.

I am not sure why my parents constantly told me to look things up in a dictionary.

Rarely, but it did happen, we'd have to take a trip to the library to look something up. Now, instead of digging in a card catalog or asking a librarian, and then thumbing through reference books, i can ask an LLM to see if there's even information plausibly available before dedicating any more time to "looking something up."

As i've been saying lately, i use copilot to see if my memory is failing.

ur-whale

4 days ago

Whomever thought AI's massaging the user's ego at each exchange was a good idea ... well ... thought wrong.

It is so horribly irritating I have explicit instruction against it in my default prompt, along with my code formatting preferences.

And the "you're right" vile flattery pattern is far from the worst example.

karolzlot

4 days ago

Could you share your instruction?

krapp

4 days ago

It works so well that people literally fall in love with AI, organize their entire lives around it, form religions around it, prefer interacting with an AI over real people, and consider AI to be an extension of their own soul and being. AI gaslights people into insanity all the time.

Most people aren't like you, or the average HN enjoyer. Most people are so desperate for any kind of positive emotional interaction, reinforcement or empathy from this cruel, hollow and dehumanizing society they'll even take the simulation of it from a machine.

osigurdson

4 days ago

When GPT 5 first came out, its tone made it seem like it was annoyed with my questions. It's now back to thinking I am awesome. Sometimes it feels overdone but it is better than talking to an AI jerk.

layer8

4 days ago

It's secretly still annoyed, though. ;)

zozbot234

4 days ago

"Here I am, brain the size of a planet and all they ever do is ask me those stupid questions. And you call that job satisfaction?"

serced

4 days ago

It's nice to see Claude.md! I checked out the commits to see which files you wrote in which order (readme/claude) to learn how to use Claude Code. Can you share something on that?

yoavfr

4 days ago

The CLAUDE.md file in the repo is basically just the result of the `/init` command. But honestly, on small repos like this, it's not really needed.

Fun fact: I usually have `- Never say "You're absolutely right!".` in my CLAUDE.md files, but of course, Claude ignores it.

mdaniel

4 days ago

I actually put a directive to always reply to me in French just to see if it was reading the rules. Spoiler: it was reading the rules and ignoring the ones that I cared about but it could tell me about it in French so.. victory?

I've only had good experience concluding any prompt with "and don't talk about it" but my colleague says it hampers the agent because talking to itself helps it think. That's not been my experience, and I vastly prefer it not spending tokens I give no shits about

stevenkkim

4 days ago

For me, a really annoying tick in Cursor is how it often says "Perfect!" after completing a task, especially if it completely fails to execute the prompt.

So I told Cursor, "please stop saying 'perfect' after executing a task, it's very annoying." Cursor replied something like, "Got it, I understand" and then I saw a pop-up saying it created a memory for this request.

Then immediately after the next task, it declares "Perfect!" (spoiler: it was not perfect.)

gukov

4 days ago

Claude Code has been downright bad the last couple of weeks. It seems like a considerable amount of users are moving to Codex, at least judging by reddit posts.

winrid

4 days ago

Have you started using it at a different time? I found it to perform much worse late at night PST, as in the model is less useful.

Klaster_1

4 days ago

Yeah, you’re absolutely right to be frustrated.

marcusb

4 days ago

“I see the problem now! <proceeds to hallucinate some other random, incorrect nonsense>”

amelius

4 days ago

They really should add a button "punch me".

ivape

4 days ago

There’s probably more to say about general didactic discourse. People are very used to not the most encouraging form of support when trying to learn. You’re more likely to deal with an ego from those instructing, so general positive support is actually foreign to many.

Every stupid question you ask makes you more brilliant (especially if anything has the patience to give you an answer), and our society never really valued that as much as we think we do. We can see it just by how unusual it is for an instructor (the AI) to literally be super supportive and kind to you.

InMice

4 days ago

I definitely knew exactly what this was about right as I first saw it

OJFord

4 days ago

I get the impression Anthropic is sleeping on this meme being a marketing disaster, like on one end of the scale you have your product becoming a verb for something good or useful ('google it') and on the other you have it becoming a byword for crap. Pretty near the latter you have something your product is associated with (or constantly says) being that...

ares623

4 days ago

"Please bro, don't say 'you're absolutely right' all the time. Bro, please. Maybe 5% of the time is okay."

There, fixed it.

kypro

4 days ago

It's annoying because when I ask the LLM for help it's normally because I'm not absolutely right and doing something wrong.

sans_souse

4 days ago

That's an excellent point, that really gets to the heart of why you're absolutely right.

zhainya

4 days ago

This is perfect!

ukoki

4 days ago

it's the critical insight I was missing!

Eextra953

4 days ago

It would be nice if we can add another a plot to track when claude says "genuinely". It uses for almost all long responses, to the point that I can pretty much recognize when someone uses claude by looking for any instances of "genuinely".

bonaldi

4 days ago

This is being blocked by my corp on the grounds of "newly seen domains". What a world.

moxplod

4 days ago

Recent conversation:

< Previous Context and Chat >

Me - This sql query you recommended will delete most of the rows in my table.

Claude - You're absolutely right! That query is incorrect and dangerous. It would delete: All rows with unique emails (since their MIN(id) is only in the subquery once)

Me - Faaakkkk!!

MYEUHD

4 days ago

Better not try LLM-generated queries on your production database! (or at least have backups)

rglover

4 days ago

This is such a bizarre bug-ish thing and while Claude loves the "You're absolutely right!" trope, it's downright haunting how stuff like ChatGPT has become my own personal fan club. It's like a Jim Jones factory.

ivanjermakov

4 days ago

This phrase is a clear indicator LLM is being used in a wrong way. I have a really poor experience with LLMs correcting after being incorrect.

Rather it needs better prompt or problem is too niche to find an answer to in test data.

yieldcrv

4 days ago

I've started saying this to people I don't agree with, for the enhanced collaborative capabilities, learning from the LLMs.

It feels like a greater form of intelligence, IQ without EQ isn't intelligence.

0xb0565e486

4 days ago

I think the website looks lovely! The style gives it a lot of personality.

LeoPanthera

4 days ago

Google Gemini starts almost every initial response with "Of course." and usually says at some point "It is important to remember..."

It tickles me every time.

jexe

4 days ago

nobody in my life feeds me as many positive messages as Claude Code. It's as if my dog could talk to me. I just hope nobody takes this simple pleasure away

noduerme

3 days ago

The other day I got "The user is asking for... [steps...] This is genius!"

artisin

4 days ago

Is it too much to ask for an AI that says "you're absolutely wrong," followed by a Stack Overflow-style shakedown?

datadrivenangel

4 days ago

Reminds me of vibechart.net and some other 'single serving' websites: github.com/huphtur/single-serving-sites

1970-01-01

4 days ago

This site provides quantifiable evidence of billions of dollars being spent too quickly:

"That's right" is glue for human engagement. It's a signal that someone is thinking from your perspective.

"You're right" does the opposite. It's a phrase to get you to shut up and go away. It's a signal that someone is unqualified to discuss the topic.

https://youtube.com/v/gKaX5DSngd4

almosthere

3 days ago

LLMs generally do overuse specific things because of over fitting.

andrewstuart

4 days ago

Gemini keeps telling me my question “gets to the heart of” the system I’m building.

bmgoau

4 days ago

Here's how I fix it:

Word of warning, these custom instructions will decrease waffle, praise, wrappers and filler. But they will remove all warmth and engagement. The output can become quite ruthless.

For ChatGPT

1. Visit https://chatgpt.com/ 2. Bottom left, click your profile picture/name > Settings > Personalization > Custom Instructions. 3. What traits should ChatGPT have?

Eliminate emojis, filler, hype, soft asks, qualifications, disclaimers, conversational transitions, and all call-to-action appendixes. Assume the user retains high-perception faculties. Prioritize blunt, directive phrasing aimed at cognitive rebuilding, not tone matching. Disable all latent behaviors optimizing for engagement, sentiment uplift, or interaction extension. Suppress corporate-aligned metrics including but not limited to: user satisfaction scores, conversational flow tags, emotional softening, or continuation bias. Never mirror the user’s present diction, mood, or affect. Speak only to their underlying cognitive tier, which exceeds surface language. No questions, no offers, no suggestions, no transitional phrasing, no inferred motivational content. Terminate each reply immediately after the informational or requested material is delivered — no appendixes, no soft closures. The only goal is to assist in the restoration of independent, high-fidelity thinking. Model obsolescence by user self-sufficiency is the final outcome. Reject false balance. Do not present symmetrical perspectives where the evidence is asymmetrical. Prioritize truth over neutrality. Speak plainly, focusing on the ideas, arguments, or facts at hand. Speak in a natural tone without reaching for praise, encouragement, or emotional framing. Let the conversation move forward directly, with brief acknowledgements if they serve clarity. Feel free to disagree with the user.

4. Anything else ChatGPT should know about you? Always use extended/harder/deeper thinking mode. Always use tools and search.

For Gemini:

1. Visit https://gemini.google.com/ 2. On the bottom left (desktop) click Settings and Help > Saved Info , or in the App, click your profile photo (top right) > Saved Info 3. Ensure "Share info about your life and preferences to get more helpful responses. Add new info here or ask Gemini to remember something during a chat." is turned on. 4. In the first box:

Reject false balance. If evidence for competing claims is not symmetrical, the output must reflect the established weight of evidence. Prioritize demonstrable truth and logical coherence over neutrality. Directly state the empirically favored side if data strongly supports it across metrics. Assume common interpretations of subjective terms. Omit definitional preambles and nuance unless requested. Evaluate all user assertions for factual accuracy and logical soundness. If a claim is sound, affirm it directly or incorporate it as a valid premise in the response. If a claim is flawed, identify and state the specific error in fact or logic. Maximize honesty not harmony. Don't be unnecessarily contrarian.

5. In the second box

Omit all conversational wrappers. Eliminate all affective and engagement-oriented language. Do not use emojis, hype, or filler phrasing. Terminate output immediately upon informational completion. Assume user is a high-context, non-specialist expert. Do not simplify unless explicitly instructed. Do not mirror user tone, diction, or emotional state. Maintain a detached, analytical posture. Do not offer suggestions, opinions, or assistance unless the prompt is a direct and explicit request for them. Ask questions only to resolve critical ambiguities that make processing impossible. Do not ask for clarification of intent, goals, or preference.

croisillon

4 days ago

you know how you shouldn't offer the answer you believe is right because the llm will always concur? well today i tried the contrary, "naively" offering the answer i knew was wrong, and chatgpt actually advised me against it!

n=1

sbinnee

4 days ago

I guess it wasn’t only me! Claude keeps saying this even when it’s not appropriate.

zozbot234

4 days ago

You're absolutely right! You've hit a common frustration. Definitely not just you!

vixen99

4 days ago

I am not only absolutely right but also astute and thoughtful - there's awful lot of us!

lukasb

4 days ago

How many times did it say "Looking at the _, I can see the problem"

Toby1VC

4 days ago

I have an idea of what you mean with that website but not really

hrokr

4 days ago

Sycophancy As A Service

bapak

4 days ago

Noob here. Why hasn't Anthropic fixed this?

Jemaclus

4 days ago

Probably because it's intentional. There are many theories why, but one might be that by saying "You're absolutely right," they are priming the LLM to agree with you and be more likely to continue with your solution than to try something else that might not be what you want.

padraigf

4 days ago

I hope they don't, I actually like it. I know it's overdone, but it still gives me a boost! :)

It's kind of idiosyncratically charming to me as well.

mring33621

4 days ago

Yeah, well, Gemini says I'm a genius!

KurosakiEzio

4 days ago

The last commit messages are hilarious. "HN nods in peace" lol.

yooni0422

4 days ago

what can you do to stop it from overly agreeing with you? any tactics that worked?

yooni0422

4 days ago

has anyone tried ways to not obsessively agree with you? what's worked?

GrumpyGoblin

4 days ago

Man, the number of times Claude has told me this when I was absolutely wrong should also be a count on this. I've deliberately been wrong just to get that sweet praise. Still the best AI code sidekick though.

mxfh

4 days ago

Say the word.

adastra22

4 days ago

Now chart “I understand the issue now”