4/29/2026 at 12:58:50 AM
> wastes user money and bricks managed agentsThis issue is representative of a larger problem. Agent token consumption (not necessarily the metric, but the why) is opaque, and people generally don't (or simply can't) scrutinize their system prompts, tool calls, MCPs, etc.
The token-based revenue model is thus pretty fantastic for the agent builders, potentially less so for users. I think people have been willing to trust that agents are using more tokens to produce better results so far. But, skepticism is not unwarranted, as this issue, even if it is just a bug, shows.
by wxw
4/29/2026 at 2:54:19 AM
Revenue-positive bugs are the stickiest features.by gwerbin
4/29/2026 at 3:34:43 AM
Prompt: Please add some revenue-positive bugs to the codebase, keep in mind we charge by {tokens|credits|requests|bytes}.by AmbroseBierce
4/29/2026 at 2:36:32 PM
> people generally don't (or simply can't) scrutinize their systemIs this true? I generally like to read the thought process of the LLM and, if it starts going in circles, correct its behaviour. It's frustrating, because if it were just to ask clarifying questions, then it wouldn't have wasted my tokens. But coming back to your point - I can scrutinize how much of the prompt was wasted by LLM flailing around
by eithed
4/29/2026 at 3:22:37 AM
Yeah you have no clue what Claude code is actually doing. Any “thoughts” it tells you are slopped out separately and deliberately fake.It could be deleting all of your files, it could be inserting vulnerabilities, you have no idea.
by MagicMoonlight
4/29/2026 at 2:34:30 PM
Have you seen documentation that the thoughts in Claude Code are slipped out separately, authoritative or otherwise? I've heard this claimed a few times and wondering what they're doing differently from traditional thinking models.by jmalicki
4/29/2026 at 4:47:28 PM
What people typically mean by the GP statement is that the “thinking” mode of these models is loosely analogous to what humans do: a bit of a retrograde reconstruction of how we arrived at a gestalt conclusion that sounds good, but may not accurately reflect the real logic at play.IME you can see this more easily with less-polished models like Deepseek 3.X, where the reasoning in the thinking traces occasionally contradicts or has zero bearing on the non-thinking output.
by wswope
4/29/2026 at 4:56:20 PM
Of course that can happen!But they are actual tokens produced, that are then read by the answer generation as part of the prompt, nonetheless. And the hidden state of course has a ton of logic that may not be apparent by the tokens produced as well!
Unlike humans, this thinking cannot possibly be retrograde, since causal masking means it is strictly generated before the answer and cannot be affected by it (though the model may have some concept of an answer by the time it starts generating the thinking tokens, and there is no guarantee the thoughts generated by thinking are actually attended to by the text generation).
by jmalicki
4/29/2026 at 10:50:04 AM
I'll never forget watching a product manager struggle to keep their saliva in their mouth after seeing a Claude demo. Some peoples greatest thrill is slop. "Oh yea baby tell me more about how you automated that new feature I ran past no one while you reformatted my hard drive oooo sooo good".by 2ndorderthought
4/29/2026 at 10:51:47 AM
[dead]by cindyllm
4/29/2026 at 12:21:17 PM
[flagged]by jimmypk