2/27/2026 at 10:48:50 AM
There isn't any attempt to falsify the "clean room" claim in the article - a rational approach would be to not provide any documents about the Z80 and the Spectrum, and just ask it to one-shot an emulator and compare the outputs...If the one-shot output resembles anything working (and I am betting it will), then obviously this isn't clean room at all.
by jaen
2/27/2026 at 9:10:33 PM
Author just trusts the agent to not use the internet b/c he wrote it so in the instructions should tell you all you need to know. It's great he managed to prompt it w/ the right specification for writing yet another emulator but I don't think he understands how LLMs actually work so most of the commentary on what's going on with the "psychology" of the LLM should be ignored.by measurablefunc
2/27/2026 at 9:24:12 PM
Even without internet access, probably everything there is to say about Z80/Speccy emulators was already in its training set.by the_af
2/27/2026 at 11:04:39 AM
You didn't read the full article. The past paragraph talks about this specifically.by antirez
2/27/2026 at 9:31:05 PM
In the last paragraph you handwave that all the Z80 and ZX Spectrum documentations is likely already in the model anyway... Choosing to not provide the documents/websites might then requiring more prompting to finish the emulator, but the knowledge is there. You can't clean room with a large LLM. That's delusion!by tredre3
2/28/2026 at 8:52:04 PM
Counterpoint: in December, a Polish MP [0] has vibe-coded an interpreter [1] of a 1959 Polish programming language, feeding it the available documentation. _That,_ at least, is unlikely to have appeared in the model’s training data.[0]: https://en.wikipedia.org/wiki/Adrian_Zandberg [1]: https://sako-zam41.netlify.app/
by nathell
3/1/2026 at 12:01:10 PM
Not exactly a counterpoint, since nobody argued that LLMs can not produce "original" code from specs at all - just that this particular exercise was not clean room.(although for SAKO [1], it's an average 1960 programming language, just with keywords in Polish, so it's certainly almost trivial for an LLM to produce an interpreter, since construction via analogy is the bread and butter of LLMs. Also, such interpreters tend to have an order of magnitude less complexity than emulators.)
[1]: https://en.wikipedia.org/wiki/SAKO_(programming_language)
by jaen
2/28/2026 at 8:30:24 AM
I mean, for an article that's titled "clean room", that would be the first thing to do, not as a "maybe follow up in the future"...(I do think the article could have stood on its own without mentioning anything about "clean room", which is a very high standard.)
For the handwavy point about the x86 assembler, I am quite sure that the LLM will remember the entirety of the x86 instruction set without any reference, it's more of a problem of having a very well-tuned agentic loop with no context pollution to extract it. (which you won't get by YOLOing Claude, because LLMs aren't that meta-RLed yet to be able to correct their own context/prompt-engineering problems)
Or alternatively, to exploit context pollution, take half of an open-source project and let the LLM fill in the rest (try to imagine the synthetic "prompt" it was given when training on this repo) and see how far it is from the actual version.
by jaen