trjordan
4 days ago
OK, so I love this, because we all recognize it.
It's not fully just a tic of language, though. Responses that start off with "You're right!" are alignment mechanisms. The LLM, with its single-token prediction approach, follows up with a suggestion that much more closely follows the user's desires, instead of latching onto it's own previous approach.
The other tic I love is "Actually, that's not right." That happens because once agents finish their tool-calling, they'll do a self-reflection step. That generates the "here's what I did response" or, if it sees an error, the "Actually, ..." change in approach. And again, that message contains a stub of how the approach should change, which allows the subsequent tool calls to actually pull that thread instead of stubbornly sticking to its guns.
The people behind the agents are fighting with the LLM just as much as we are, I'm pretty sure!
nojs
4 days ago
Yeah, I figure this is also why it often says “Ah, I found the problem! Let me check the …”. It hasn’t found the problem, but it’s more likely to continue with the solution if you jam that string in there.
adastra22
4 days ago
We don’t know how Claude code is internally implemented. I would not be surprised at all if they literally inject that string as an alternative context and then go with the higher probability output, or if RLHF was structured in that way and so it always generates the same text.
al_borland
4 days ago
In my experience, once it starts telling me I’m right, we’re already going downhill and it rarely gets better from there.
flkiwi
4 days ago
Sometimes I just ride the lightning to see how off course it is willing to go. This is not a productive use of my time but it sure is amusing.
In fairness, I’ve done the same thing to overconfident junior colleagues.
lemming
4 days ago
Yeah, I want a feature which stops my agent as soon as it says anything even vaguely like: "let me try another approach". Right after that is when the wheels start falling off, tests get deleted, etc. That phrase is a sure sign the agent should (but never does) ask me for guidance.
anthem2025
4 days ago
Usually it’s a response to my profanity laden “what are you doing? Why? Don’t do that! Stop! Do this instead”
unshavedyak
4 days ago
I just wish they could hide these steering tokens in the thinking blurb or some such. Ie mostly hidden from the user. Having it reply to the user that way is quite annoying heh.
KTibow
4 days ago
This can still happen even with thinking models as long as the model outputs tokens in a sequence. Only way to fix would be to allow it to restart its response or switch to diffusion.
libraryofbabel
4 days ago
> The LLM, with its single-token prediction approach, follows up with a suggestion that much more closely follows the user's desires, instead of latching onto it's own previous approach.
Maybe? How would we test that one way or the other? If there’s one thing I’ve learned in the last few years, it’s that reasoning from “well LLMs are based on next-token prediction, therefore <fact about LLMs>” is a trap. The relationship between the architecture and the emergent properties of the LLM is very complex. Case in point: I think two years ago most of us would have said LLMs would never be able to do what they are able to do now (actually effective coding agents) precisely because they were trained on next token prediction. That turned out to be false, and so I don’t tend to make arguments like that anymore.
> The people behind the agents are fighting with the LLM just as much as we are
On that, we agree. No doubt anthropic has tried to fine-tune some of this stuff out, but perhaps it’s deeply linked in the network weights to other (beneficial) emergent behaviors in ways that are organically messy and can’t be easily untangled without making the model worse.
adastra22
4 days ago
I don’t think there is any basis for GP’s hypothesis that this is related to the cursor being closer to the user’s example. The attention mechanism is position independent by default and actually has to have the token positions shoehorned in.
Uehreka
4 days ago
The human stochastic parrots (GP, not you) spouting these 2023 talking points really need to update their weights. I’m guessing this way of thinking has a stickiness because thinking of an LLM as “just a fancy markov chain” makes them feel less threatening to some people (we’re past the point where it could be good faith reasoning).
Like, I hear people say things like that (or that coding agents can only do web development, or that they can only write code from their training data), and then I look at Claude Code on my computer, currently debugging embedded code on a peripheral while also troubleshooting the app it’s connected to, and I’m struck by how clearly out of touch with reality a lot of the LLM cope is.
People need to stop obsessing over “the out of control hype” and reckon with the thing that’s sitting in front of them.
kirurik
4 days ago
It seems obvious, but I hadn't thought about it like that yet, I just assumed that the LLM was finetuned to be overly optimistic about any user input. Very elucidating.
jcims
4 days ago
>The other tic I love is "Actually, that's not right." That happens because once agents finish their tool-calling, they'll do a self-reflection step.
I saw this a couple of days ago. Claude had set an unsupported max number of items to include in a paginated call, so it reduced the number to the max supported by the API. But then upon self-reflection realized that setting anything at all was not necessary and just removed the parameter from the code and underlying configuration.
SilverElfin
4 days ago
Is there a term when everyone sees a phrase like this and understands what it means without coordinating beforehand?
SilasX
4 days ago
Sounds like a kind of Schelling point:
https://en.wikipedia.org/wiki/Focal_point_(game_theory)?uses...
dafelst
4 days ago
I would call it a meme
beeflet
4 days ago
convergence
jcims
4 days ago
It'd be nice if the chat-completion interfaces allowed you to seed the beginning of the response.
bryanrasmussen
4 days ago
>if it sees an error, the "Actually, ..." change in approach.
AI-splaining is the worst!
Szpadel
4 days ago
exactly!
People bless gpt-5 for not doing exactly this and in my testing with it in copilot I had lot of cases where it tried to do wrong thing (execute come messed up in context compaction build command) and I couldn't steer it to do ANYTHING else. It constantly tried to execute it as response any my message (I tries many common steerability tricks, (important, <policy>, just asking, yelling etc) nothing worked.
the same think when I tried to do socratic coder prompting, I wanted to finish and generate spec, but he didn't agree and kept asking nonsensical at this point questions