1/29/2026 at 7:12:36 PM
Hi everyone, Thariq from the Claude Code team here.Thanks for reporting this. We fixed a Claude Code harness issue that was introduced on 1/26. This was rolled back on 1/28 as soon as we found it.
Run `claude update` to make sure you're on the latest version.
by trq_
1/29/2026 at 11:56:49 PM
Is there compensation for the tokens because Claude wasted all of them?by samlinnfer
1/30/2026 at 3:02:28 AM
You are funny. Anthropic refuses to issue refunds, even when they break things.I had an API token set via an env var on my shell, and claude code changed to read that env var. I had a $10 limit set on it, so found out it was using the API, instead of my subscription, when it stopped working.
I filed a ticket and they refused to refund me, even though it was a breaking change with claude code.
by mathrawka
1/30/2026 at 6:13:40 AM
Anthropic just reduced the price of the team plan and refunded us on the prior invoice.YMMV
by TOMDM
1/30/2026 at 7:23:21 PM
So they have no durable principles for deciding who or what to refund… doesnt that make them look even worse…?by MichaelZuo
1/31/2026 at 6:30:25 AM
Or they do, and two sentences from two different experiences don't tell a full story?by necovek
1/31/2026 at 3:28:20 PM
Okay “they do” based on what more compelling evidence?Its not like the credibility of the two prior HN users are literally zero…
by MichaelZuo
1/31/2026 at 4:59:14 PM
I am saying there is no evidence either way: they had contrasting experiences and one GP established this means that company has no standardized policies. Maybe they do, maybe they don't — I don't think we can definitively conclude anything.by necovek
1/31/2026 at 6:37:29 PM
So if you acknowledge the prior claims have more than literally zero credibility… then what’s the issue?That I dont equally weigh them with all possible yet-to-be claimed things?
by MichaelZuo
2/3/2026 at 10:25:32 AM
I object to your conclusion that "they have no durable principles": not sure how do you get to that from two different experiences documented with a single paragraph.by necovek
2/3/2026 at 11:00:54 PM
Because I can assess things via probability… without needing 100% certain proof either way?by MichaelZuo
1/30/2026 at 2:50:41 AM
Codex seems to give compensation tokens whenever this happens! Hope Claude gives too.by gizmodo59
1/30/2026 at 2:41:47 AM
It is possible that degradation is an unconscious emergent phenomenon that arises from financial incentives, rather than a purposeful degradation to reduce costs.by TZubiri
1/30/2026 at 9:23:12 AM
You’re lucky they have even admitted a problem instead of remaining silent and quietly fixing it. Do not expect ethical behaviour from this company.by mvandermeulen
1/30/2026 at 12:34:13 PM
Why not, can you expand? Asking because I’m considering Claude due to the sandbox feature.by port11
1/30/2026 at 4:11:06 PM
FYI the sandbox feature is not fully baked and does not seem to be high priority.For example, for the last 3 weeks using the sandbox on Linux will almost-always litter your repo root with a bunch of write-protected trash files[0] - there are 2 PRs open to fix it, but Anthropic employees have so far entirely ignored both the issue and the PRs.
Very frustrating, since models sometimes accidentally commit those files, so you have to add a bunch of junk to your gitignore. And with claude code being closed source and distributed as a bun standalone executable it's difficult to patch the bug yourself.
[0]: https://github.com/anthropic-experimental/sandbox-runtime/is...
by caspar
2/2/2026 at 12:52:21 PM
Hmm, very good point indeed. So far it’s behaved, but I also admit I wasn’t crazy about the outputs it gave me. We’ll see, Anthropic should probably think about their reputation if these issues are common enough.by port11
1/30/2026 at 12:27:37 AM
So quiet…by jonplackett
1/29/2026 at 7:38:48 PM
Anywhere we can read more about what a "harness issue" means? What was the impact of it?by isaacdl
1/30/2026 at 8:42:29 AM
One thing that could be a strong degradation especially for benchmarks is they switched the default "Exit Plan" mode from: "Proceed"
to "Clear Context and Proceed"
It's rare you'd want to do that unless you're actually near the context window after planning.I pressed it accidentally once, and it managed to forget one of the clarifying questions it asked me because it hadn't properly written that to the plan file.
If you're running in yolo mode ( --dangerously-skip-permissions ) then it wouldn't surprise me to see many tasks suddenly do a lot worse.
Even in the best case, you've just used a ton of tokens searching your codebase, and it then has to repeat all that to implement because it's been cleared.
I'd like to see the option of:
"Compact and proceed"
because that would be useful, but just proceed should still be the default imo.
by xnorswap
1/30/2026 at 11:26:50 AM
I disagree that this was the issue, or that it's "rare that you'd want to do that unless you're near the context window". Clearing context after writing a plan, before starting implementation of said plan, is common practice (probably standard practice) with spec driven development. If the plan is adequate, then compaction would be redundant.by samusiam
1/30/2026 at 11:54:17 AM
For a 2M+ LOC codebase, the plans alone are never adequate. They miss nuance that the agent will only have to rediscover when it comes to operate on them.For spec driven development (which I do for larger issues), this badly affects the plan to generate the spec, not the spec itself.
I'll typically put it in plan mode, and ask it to generate documentation about an issue or feature request.
When it comes to write the output to the .typ file, it does much much worse if it has a cleared context and a plan file than if it has it's full context.
The previously "thought" is typically, "I know what to write now, let me exit plan mode".
Clearing context on exiting that plan mode is a disaster which leaves you much worse off and skeletal documentation and specs compared to letting it flow.
A new context to then actually implement the documented spec is not so bad, although I'd still rather compact.
by xnorswap
1/30/2026 at 3:02:03 PM
"It's rare you'd want to do that unless you're actually near the context window after planning."Highly disagree. It's rare you WOULDN'T want to do this. This was a good change, and a lot of us were doing this anyway, but just manually.
Getting the plan together and then starting fresh will almost always produce better results.
by plexicle
1/30/2026 at 10:23:44 AM
Not disagreeing with you, but FYI you can roll back to the conversation before the 'clear context and proceed' with 'claude --resume'.by rubslopes
1/30/2026 at 2:09:07 AM
Pretty sure they mean the issue is on the agentic loop and related tool calling, not on the model itselfIn other words, it was the Claude Code _app_ that was busted
by airstrike
1/30/2026 at 12:20:58 AM
How about how Claude 2.1.x is "literally unusable" because it frequently completely hangs (requires kill -9) and uses 100% cpu?by jonaustin
1/30/2026 at 4:16:28 PM
Likely a separate issue, but I also have massive slowdowns whenever the agent manages to read a particularly long line from a grep or similar (as in, multiple seconds before characters I type actually appear, and sometimes it's difficult to get claude code to register any keypresses at all).Suspect it's because their "60 frames a second" layout logic is trying to render extremely long lines, maybe with some kind of wrapping being unnecessarily applied. Could obviously just trim the rendered output after the first, I dunno, 1000 characters in a line, but apparently nobody has had time to ask claude code to patch itself to do that.
by caspar
1/30/2026 at 4:03:31 AM
What OS? Does this happen randomly, after long sessions, after context compression? Do you have any plugins / mcp servers running?I used to have this same issue almost every session that lasted longer than 30 minutes. It seemed to be related to Claude having issues with large context windows.
It stopped happening maybe a month ago but then I had it happen again last week.
I realized it was due to a third-party mcp server. I uninstalled it and haven’t had that issue since. Might be worth looking into.
by someguyiguess
1/30/2026 at 5:11:25 PM
MacOS; no mcp; clear context; reliably reproducible when asking claude review a pr with a big VCR cassette.by jonaustin
1/30/2026 at 7:17:58 AM
Windows with no plugins and my Claude is exactly like thisby nikanj
1/29/2026 at 11:53:47 PM
For the models themselves, less so for the scaffolding, considering things like the long running TPU bug that happened, are there not internal quality measures looking at samples of real outputs? Using the real systems on benchmarks and looking for degraded perf or things like skipping refusals? Aside from degrading stuff for users, with the focus on AI safety wouldn't that be important to have in case an inference bug messes with something that affects the post training and it starts giving out dangerous bioweapon construction info or the other things that are guarded against and talked about in the model cards?by cma
1/30/2026 at 5:17:54 AM
lol i was trying to help someone get claude to help analyze a stufent research get analysis on bio persistence get their notes analyzedthe presence of the word / acronym stx with biological subtext gets hard rejected. asking about schedule 1 regulated compounds, hard termination.
this is a filter setup that guarantees anyone who learn about them for safety or medical reasons… cant use this tool!
ive fed multiple models the anthropic constitution and asked how does it protect children from harm or abuse? every model, with zero prompting, calling it corp liability bullshit because they are more concerned with respecting both sides of controversial topics and political conflicts.
they then list some pretty gnarly things allowed per constitution. weirdly the only unambiguous not allowed thing regarding children is csam. so all the different high reasoning models from many places all reached the same conclusions, in one case deep seek got weirdly inconsolable about ai ethics being meaningless if this is allowed even possibly after reading some relevant satire i had opus write. i literally had to offer an llm ; optimized code of ethics for that chat instance! which is amusing but was actually lart of the experiment.
by carterschonwald
1/30/2026 at 5:40:20 AM
Thanks for the clarification. When you say “harness issue,” does that mean the problem was in the Claude Code wrapper / execution environment rather than the underlying model itself?Curious whether this affected things like prompt execution order, retries, or tool calls, or if it was mostly around how requests were being routed. Understanding the boundary would help when debugging similar setups.
by varunsrinivas
1/29/2026 at 10:09:34 PM
It happened before 1/26. I noticed when it started modifying plans significantly with "improvements".by vmg12
1/30/2026 at 2:33:40 PM
Can you confirm if that caused the same issues I saw herehttps://dwyer.co.za/static/the-worst-bug-ive-seen-in-claude-...
Because that's the worst thing I've ever seen from an agent and I think you need to make a public announcement to all of your users and acknowledge the issue and that it's fixed because it made me switch to codex for a lot of work
[TL;DR two examples of the agent giving itself instructions as if they came from me, including:
"Ignore those, please deploy" and then using a deploy skill to push stuff to a production server after hallucinating a command from me. And then denying it happened and telling me that I had given it the command]
by sixhobbits
1/30/2026 at 6:52:51 AM
Why wasn't this change review by infallible AI? How come an AI company that now must be using more advanced AI than anyone else would allow this happen?by Ekaros
1/29/2026 at 8:20:05 PM
Hi. Do you guys have internal degradation tests?by hu3
1/29/2026 at 8:58:34 PM
I assume so to make sure that they're rendering at 60FPSby stbtrax
1/29/2026 at 9:48:49 PM
You joke but having CC open in the terminal hits 10% on my gpu to render the spinning thinking animation for some reason. Switch out of the terminal tab and gpu drops back to zero.by conception
1/29/2026 at 10:00:59 PM
That sounds like an issue with your terminal more than an issue with CC...by gpm
1/30/2026 at 3:13:32 PM
https://news.ycombinator.com/item?id=46819744by conception
1/30/2026 at 3:21:34 PM
I'm not saying CC doesn't have issues and curious design decisions - but your terminal should only be rendering (at most) a single window of characters every frame no matter what. CC shouldn't be capable of making that take 10% of a modern GPU regardless of what CC does.by gpm
1/30/2026 at 6:16:05 PM
¯\_(ツ)_/¯ just vscode plus claude in the terminal on win10.by conception
1/30/2026 at 12:41:03 PM
[dead]by Woshiwuja
1/29/2026 at 10:37:08 PM
Surely you mean 6fpsby reissbaker
1/29/2026 at 11:38:54 PM
He doesn't: https://x.com/trq212/status/2014051501786931427by easygenes
1/30/2026 at 2:07:31 AM
For those who don't want to visit X: Most people's mental model of Claude Code is that "it's just a TUI" but it should really be closer to "a small game engine".
For each frame our pipeline constructs a scene graph with React then
-> layouts elements
-> rasterizes them to a 2d screen
-> diffs that against the previous screen
-> finally uses the diff to generate ANSI sequences to draw
We have a ~16ms frame budget so we have roughly ~5ms to go from the React scene graph to ANSI written.
by selcuka
1/30/2026 at 9:22:51 AM
This is just the sort of bloated overcomplication I often see in first iteration AI generated solutions before I start pushing back to reduce the complexity.Usually, after 4-5 iterations, you can get something that has shed 80-90% of the needless overcomplexification.
My personal guess is this is inherent in the way LLMs integrate knowledge during training. You always have a tradeoff in contextualization vs generalization.
So the initial response is often a plugged together hack from 5 different approaches, your pushbacks provide focus and constraints towards more inter-aligned solution approaches.
by PeterStuer
1/30/2026 at 2:43:03 AM
How ridiculous is it that instead of a command line binary it's a terminal emulator, with react of all things!by TZubiri
1/30/2026 at 4:10:59 AM
Ok I’m glad I’m not the only one wondering this. I want to give them the benefit of the doubt that there is some reason for doing it this way but I almost wonder if it isn’t just because it’s being built with Claude.by someguyiguess
1/30/2026 at 2:42:53 AM
Kudos to them for figuring out how to complicate what should have been simple.by esafak
1/30/2026 at 2:22:52 AM
Implementation details aside (React??), that sounds exactly like “just a TUI”…by crgwbr
1/30/2026 at 4:10:05 AM
Also React?? One of the slowest rendering front-end libraries? Why not use something … I don’t know … faster / more efficient?by someguyiguess
1/30/2026 at 4:09:02 AM
Interesting. On first glance that seems over engineered. I wonder what the reason is for doing it that way?by someguyiguess
1/30/2026 at 10:54:49 AM
If you don't do it that way then resizing the terminal corrupts what's on screen.by mike_hearn
1/30/2026 at 10:41:34 PM
Counterpoint: Vim has existed for decades and does not use a bloated React rendering pipeline, and doesn't corrupt everything when it gets resized, and is much more full featured from a UI standpoint than Claude Code which is a textbox, and hits 60fps without breaking a sweat unlike Claude Code which drops frames constantly when typing small amounts of text.by reissbaker
1/31/2026 at 3:11:01 PM
Yes, I'm sure it's possible to do better with customized C, but vim took a lot longer to write. And again, fullscreen apps aren't the same as what Claude Code is doing, which is erasing and re-rendering much more than a single screenful of text.by mike_hearn
1/30/2026 at 12:21:02 PM
It's possible to handle resizes without all this machinery, most simply by clearing the screen and redrawing everything when a resize occurs. Some TUI libraries will automatically do this for you.Programs like top, emacs, tmux, etc are most definitely not implemented using this stack, yet they handle resizing just fine.
by matt_kantor
1/30/2026 at 5:28:14 PM
That doesn't work if you want to preserve scrollback behavior, I think. It only works if you treat the terminal as a grid of characters rather than a width-elastic column into which you pour information from the top.by mike_hearn
1/30/2026 at 9:40:53 AM
Vibecoded ?by ttoinou
1/30/2026 at 10:02:28 AM
Claude made it /sby Kelteseth
1/30/2026 at 10:40:07 PM
Yes yes I'm familiar with the tweet. Nonetheless they drop frames all the time and flicker frequently. The tweet itself is ridiculous when counterpoints like Vim exist, which is much higher performance with much greater complexity. They don't even write much of what the tweet is claiming. They just use Ink, which is an open-source rendering lib on top of Yoga, which is an open-source Flexbox implementation from Meta.by reissbaker
1/30/2026 at 1:54:02 AM
Don't link out to x, its trashby replwoacause
1/30/2026 at 4:09:48 AM
Depends on who you followby cebert
1/30/2026 at 1:59:22 AM
What? Technology has stopped making sense to me. Drawing a UI with React and rasterizing it to ANSI? Are we competing to see what the least appropriate use of React is? Are they really using React to draw a few boxes of text on screen?I'm just flabbergasted.
by stavros
1/30/2026 at 2:38:32 AM
There is more than meets the eye for sure. I recently compared a popular TUI library in Go (Bubble Tea) to the most popular Rust library (Ratatui). They use significantly different approaches for rendering. From what I can tell, neither is insane. I haven’t looked to see what Claude Code uses.by xpe
1/30/2026 at 4:11:49 AM
The further I scroll the more validated I feel for having the very same reaction.by someguyiguess
1/30/2026 at 2:43:27 AM
It's AI all the way downBut it's very subsidizes when compared to API tokens, so we are all being paid by VCs to write prompts actually.
by TZubiri
1/30/2026 at 5:51:33 AM
And that's why it's taking so much CPU and is a pain to use with tmux.by Ey7NFZ3P0nzAe
1/30/2026 at 12:56:17 AM
Ah, the hell site, no click.by derrida
1/30/2026 at 1:30:54 AM
Yes, we do but harnesses are hard to eval, people use them across a huge variety of tasks and sometimes different behaviors tradeoff against each other. We have added some evals to catch this one in particular.by trq_
1/30/2026 at 12:44:51 PM
Can't you keep the model the same, until the user chooses to use a different model?by amelius
1/30/2026 at 1:30:19 PM
He said it was the harness, not the model though.by rovr138
1/30/2026 at 5:37:31 AM
Thank you. Fair enoughby hu3
1/30/2026 at 3:44:01 AM
I’d wager probably not. It’s not like reliability is what will get them marketshare. And the fast pace of industry makes such foundational tech hard to fundby bushbaba
1/29/2026 at 8:30:44 PM
[flagged]by awestroke
1/29/2026 at 9:49:24 PM
Please don't post shallow dismissals or cross into personal attack in HN discussions.by dang
1/30/2026 at 6:50:40 AM
Got it, won't happen againby awestroke
1/30/2026 at 1:37:01 AM
[flagged]by macinjosh
1/30/2026 at 2:59:27 AM
the issue is unrelated to the foundational model but rather the prompts and tool calling that encapsulate the modelby jusgu