4/16/2026 at 5:38:31 PM
I've been running this on my laptop with the Unsloth 20.9GB GGUF in LM Studio: https://huggingface.co/unsloth/Qwen3.6-35B-A3B-GGUF/blob/mai...It drew a better pelican riding a bicycle than Opus 4.7 did! https://simonwillison.net/2026/Apr/16/qwen-beats-opus/
by simonw
4/16/2026 at 10:24:37 PM
Thanks for pointing to the GGUF.I just tried this GGUF with llama.cpp in its UD Q4_K_XL version on my custom agentic oritened task consisiting of wiki exploration and automatic database building ( https://github.com/GistNoesis/Shoggoth.db/ )
I noted a nice improvement over QWen3.5 in its ability to discover new creatures in the open ended searching task, but I've not quantified it yet with numbers. It also seems faster, at around 140 token/s compared to 100 token/s , but that's maybe due to some different configuration options.
Some little difference with QWen3.5 : to avoid crashes due to lack of memory in multimodal I had to pass --no-mmproj-offload to disable the gpu offload to convert the images to tokens otherwise it would crash for high resolutions images. I also used quantized kv store by passing -ctk q8_0 -ctv q8_0 and with a ctx-size 150000 it only need 23099 MiB of device memory which means no partial RAM offloading when I use a RTX 4090.
by GistNoesis
4/16/2026 at 8:23:47 PM
I'm not sure how you can give the flamingo win to Qwen:* It's sitting on the tire, not the seat.
* Is that weird white and black thing supposed to be a beak? If so, it's sticking out of the side of its face rather than the center.
* The wheel spokes are bizarre.
* One of the flamingo's legs doesn't extend to the pedal.
* If you look closely at the sunglasses, they're semi-transparent, and the flamingo only has one eye! Or the other eye is just on a different part of its face, which means the sunglasses aren't positioned correctly. Or the other eye isn't.
* (subjective) The sunglasses and bowtie are cute, but you didn't ask for them, so I'd actually dock points for that.
* (subjective) I guess flamingos have multiple tail feathers, but it looks kinda odd as drawn.
In contrast, Opus's flamingo isn't as detailed or fancy, but more or less all of it looks correct.
by kelnos
4/16/2026 at 9:50:27 PM
He literally said it came down to the comment in the SVG. Points for taste, not correctness. Basically.by withinboredom
4/16/2026 at 10:55:41 PM
Here's a reproduction attempt (LM Studio, same Qwen3.6-35B-A3B-GGUF model as linked in parent, M1 Max 64GB, <90 seconds):https://files.catbox.moe/r3oru2.png
- My Qwen 3.6 result had sun and cloud in sky, similar to the second Opus 4.7 result in Simon's post.
- My Qwen 3.6 result had no grass (except as a green line), but all three results in Simon's post had grass (thick).
- My Qwen 3.6 result had visible "tailing air motion" like Simon's Qwen 3.6 result.
- My Qwen 3.6 result had a "sun with halo" effect that none of Simon's results had.
But, I know, it's more about the pelican and the bicycle.
by realityfactchex
4/17/2026 at 5:58:37 AM
The bicycle frame is ok. Simon's was better but at least it's not broken like Opus 4.7.I can't comment that flamingo.
by _ache_
4/16/2026 at 6:16:42 PM
I wonder when pelican riding a bicycle will be useless as an evaluation task. The point was that it was something weird nobody had ever really thought about before, not in the benchmarks or even something a team would run internally. But now I'd bet internally this is one of the new Shirley Cards.by jubilanti
4/16/2026 at 6:41:07 PM
Simon has an article on thishttps://simonwillison.net/2025/Nov/13/training-for-pelicans-...
by abustamam
4/16/2026 at 10:03:22 PM
Pelicanmaxxingby SwellJoe
4/16/2026 at 7:59:01 PM
Yeah try it with something else, or e.g. add a tiger to the back seat.by amelius
4/17/2026 at 10:36:52 AM
I use this metric now, and I suggest you change it per your imagination:"Make a single-page HTML file using threejs from a CDN. Render a scene of a flying dinosaur orbiting a planet. There are clouds with thunder and lightning, and the background is a beautiful starscape with twinkling stars and a colorful nebula"
This allows me to evaluate several factors across models. It is novel and creative. I generally run it multiple times, though now that I have shared it here, I will come up with new scenes personally to evaluate.
I also consider how well it one shots, errors generated, response to errors being corrected, and velocity of iteration to improvement.
Generally speaking, Claude Sonnet has done the best, Qwen3.5 122B does second, and I have nice results from Qwen3.5 35B.
ChatGPT does not do well. It can complete the task without errors but the creativity is atrocious.
by survirtual
4/16/2026 at 7:25:05 PM
They’ll hardcode it in 4.8, just like they do when they need to “fix” other issuesby MagicMoonlight
4/16/2026 at 6:41:34 PM
I mean look at the result where he asked about a unicycle - the model couldn't even keep the spokes inside the wheels - would be rudimentary if it "learned" what it means to draw a bicycle wheel and could transfer that to unicycle.by rafaelmn
4/16/2026 at 7:27:19 PM
it's the frame that's surprisingly - and consistentnly - wrong. You'd think two triangles would be pretty easy to repro; once you get that the rest is easy. It's not like he's asking "draw a pelican on a four-bar linkage suspension mountainbike..."by duzer65657
4/16/2026 at 7:59:06 PM
This is older, but even humans don't have a great concept of how a bicycle works... https://twistedsifter.com/2016/04/artist-asks-people-to-draw...by Reddit_MLP2
4/16/2026 at 8:31:23 PM
Wouldn't this be more about being capable of mentally remembering how a bicycle looks versus how it works?This reminds me of Pictionary. [0] Some people are good and some are really bad.
I am really bad a remembering how items look in my head and fail at drawing in Pictionary. My drawing skills are tied to being able to copy what I see.
by yndoendo
4/17/2026 at 6:21:32 AM
I think it’s difficult to draw a bike exactly because you remember how it works rather than how it looks, so you worry about placing all the functional parts and get the overall composition wrong. Similar to drawing faces, without training, people will consistently dedicate too much area to the lower part of the face and draw some kind of neanderthal with no forehead.by johanvts
4/16/2026 at 9:38:46 PM
is it possible to have greater success with the specificity? I don't think i ever drew a bike frame properly as a kid despite riding them and understanding the concept of spokes and wheels...by quinnjh
4/17/2026 at 11:43:35 AM
Valid points, but you"d think "superintelligence" would "know" how to draw a pelican on a bike?by hansmayer
4/16/2026 at 5:53:58 PM
It's fascinating that a $999 Mac Mini (M4 32GB) with almost similar wattage as a human brain gets us this far.by bertili
4/17/2026 at 6:33:30 AM
Interesting thought, I looked it up out of curiosity and fund 155w max (but realistically more like 80w sustained) for the mac under load, and just around 20watts for the brain, surprisingly almost constant whether “under load” or not.by johanvts
4/17/2026 at 4:34:31 PM
I clocked my M4 at 108 Watts while running inference using Qwen3.6-35b-a3b via Al dente.by fragmede
4/17/2026 at 10:27:22 AM
> 155w max (but realistically more like 80w sustained)155W PSU seems to be unified with M4 Pro model, plus there's reserve for peripherals (~55W for 5 USB/Thunderbolt ports).
Apple lists 65W for base M4 Mac itself: https://support.apple.com/en-am/103253
Notebookcheck found same number: https://www.notebookcheck.net/Apple-Mac-Mini-M4-review-Small...
by petu
4/16/2026 at 7:06:52 PM
the more I look at these images the more convinced I become that world models are the major missing piece and that these really are ultimately just stochastic sentence machines. Maybe Chomsky was rightby culi
4/17/2026 at 6:10:51 AM
> that these really are ultimately just stochastic sentence machinesI thought that's exactly what they are?
by bmitc
4/17/2026 at 7:44:04 PM
No, they have "attention". There is unique logic going on in the deep layers of the neural network.Even the standard introductory exercise artificial neural networks, handwritten digit recognition, already shows deeper understanding. These simple networks take in raw pixels and somewhere in the many layers recognize "curves" and "edges" and then "circles" and "boxes" and whatnot and eventually "digits".
I think there's a genuine debate about whether or not this is a form of intelligence. I think the oversimplified argument of them just being stochastic sentence machines mostly comes from people who don't understand how they work. But I also think there's a much more nuanced version of this argument offered by people like Chomsky that should be taken seriously
by culi
4/18/2026 at 5:49:15 AM
> No, they have "attention". There is unique logic going on in the deep layers of the neural network.Any specifics? That doesn't say anything about them not being sentence generators. And it's pretty well known that the LLMs constantly spew out fantastically grammatically correct sentences that have no logic to them whatsoever.
> These simple networks take in raw pixels and somewhere in the many layers recognize "curves" and "edges" and then "circles" and "boxes" and whatnot and eventually "digits".
That sounds like a version of anthropomorphizing. It is my understanding that it is a completely open problem as to what neural networks are actually doing in their internal, deep layers.
> I think the oversimplified argument of them just being stochastic sentence machines mostly comes from people who don't understand how they work.
I mean, that's effectively a logical fallacy, so it's not a strong argument.
by bmitc
4/17/2026 at 6:28:24 AM
I am so perplexed what exactly where people thinking they were. Its nothing else than highly sofisticated statistics.by mastermage
4/17/2026 at 8:53:52 AM
From that perspective, which is totally correct, it makes you wonder what other domains of knowledge look like when pushed to the boundaries of our capabilities as a species.by tmountain
4/18/2026 at 9:25:04 AM
That is a genuinely thought provoking idea.by mastermage
4/17/2026 at 7:49:16 PM
Do you know of any other statistical model that can "hallucinate". They clearly have emergent capabilities that come from scale that are absent in any other statistical model we've ever dreamt up.We know that LLMs build complex internal representations of language, logic, and concepts rather than just shallow word-counting.
If you deny that then you probably have an elementary understanding of how they work. Not even Chomsky denies that. The real argument imo is whether those internal representations constitute an actual "understanding" of the world or just flatten out to something much less interesting.
by culi
4/18/2026 at 9:29:50 AM
> Do you know of any other statistical model that can "hallucinate".Actualy most statistical models can "hallucinate", specifically those that are capable of interpolation.
I have witnessed this for example in Gaussian Processes. In my own scientific work.
by mastermage
4/17/2026 at 3:47:40 PM
The Chomsky argument feels like it's moving in a different direction than what's actually useful to know. Whether or not these models have "real" understanding, they're clearly capable of solving problems that were previously considered to require understanding. The more interesting question is whether world models, if they existed, would actually improve the failure modes people care about — like hallucination and planning — or whether we'd just get better stochastic sentence machines with an extra layer of abstraction on top.by 060880
4/16/2026 at 5:59:58 PM
But that you also gave a win to Qwen on flamingo is pretty outrageous! :)Tthe right one looks much better, plus adding sunglasses without prompting is not that great. Hopefully it won't add some backdoor to the generated code without asking. ;)
by cyclopeanutopia
4/16/2026 at 6:13:44 PM
I love how the Chinese models often have an unprompted predilection to add flair.GLM-5.1 added a sparkling earring to a north Virginia opossum the other day and I was delighted: https://simonwillison.net/2026/Apr/7/glm-51/
by simonw
4/16/2026 at 8:57:45 PM
You're running 5.1 locally or hosted?by monksy
4/16/2026 at 9:37:43 PM
I used that one via OpenRouter.by simonw
4/16/2026 at 6:41:12 PM
The flamingo on Qwen's unicycle is sitting on the tire, not the seat. That wins because of sunglasses?by prirun
4/16/2026 at 7:38:22 PM
Can a benchmark meant as a joke not use a fun interpretation of results? The Qwen result has far better style points. Fun sunglasses, a shadow, a better ground, a better sky, clouds, flowers, etc.If we want to get nitty gritty about the details of a joke, a flamingo probably couldn't physically sit on a unicycle's seat and also reach the pedals anyways.
by evilduck
4/16/2026 at 8:37:07 PM
Well, maybe the flamingo is a really good unicyclist...by akavel
4/17/2026 at 5:37:36 AM
Transparency of the wheelStylized gradients on the flamingo
Flowers
Ground/grass has a stylized look and feel
...despite a miss along the Y-axis where it's below the seat, couple oddly organized tail feathers, spokes, the composition overall is much closer to a production quality entity
Opus 4.7 looks like 20 seconds in MS paint.
Qwen3.6 looks incomplete due to the sitting position, but like a WIP I could see on a designer coworkers screen if I walk up and interrupt them. Click and drag it up, adjust tail feathers and spokes, you're there or much closer, to a usable output
by yabutlivnWoods
4/16/2026 at 7:11:44 PM
interesting, I just tried this very model, unsloth, Q8, so in theory more capable than Simon's Q4, and get those three "pelicans". definitely NOT opus quality. lmstudio, via Simon's llm, but not apple/mlx. Of course the same short prompt.Simon, any ideas?
https://ibb.co/FLc6kggm (tried here temperature 0.7 instead of pure defaults)
by rdslw
4/17/2026 at 1:04:00 AM
try Unsloth recommended settings Thinking mode for general tasks: temperature=1.0, top_p=0.95, top_k=20, min_p=0.0, presence_penalty=1.5, repetition_penalty=1.0
Thinking mode for precise coding tasks (e.g. WebDev): temperature=0.6, top_p=0.95, top_k=20, min_p=0.0, presence_penalty=0.0, repetition_penalty=1.0
Instruct (or non-thinking) mode for general tasks: temperature=0.7, top_p=0.8, top_k=20, min_p=0.0, presence_penalty=1.5, repetition_penalty=1.0
Instruct (or non-thinking) mode for reasoning tasks: temperature=1.0, top_p=0.95, top_k=20, min_p=0.0, presence_penalty=1.5, repetition_penalty=1.0
(Please note that the support for sampling parameters varies according to inference frameworks.)
by strobe
4/16/2026 at 8:59:55 PM
Hey I really enjoy your blog. On some things I end up finding a blog post of yours thats a year+ old and at other times, you and I are investigating similar things. I just pulled Qwen3.6 - 35b -A3B (Can't believe thats a A3B coming from 35b).I'm impressed about the reach of your blog, and I'm hoping to get into blogging similar things. I currently have a lot on my backlog to blog about.
In short, keep up the good work with an interesting blog!
by monksy
4/16/2026 at 5:41:56 PM
I've had some really gnarly SVGs from Claude. Here's what I got after many iterations trying to draw a hand: https://imgur.com/a/X4Jqiusby jamwise
4/16/2026 at 6:50:38 PM
Probably because all the training material of humans drawing hands are garbage haha.by giantg2
4/16/2026 at 10:25:56 PM
The real question is what the next truly weird, un-optimized prompt will be. Something involving a sloth debugging a quantum computer in MS Paint?"by jaspanglia
4/16/2026 at 11:37:32 PM
The qwen flamingo looks like it’s smoking’ a doobie.by quietsegfault
4/16/2026 at 7:11:19 PM
Interesting, qwen has the pelican driving on the left lane. Coincidence or has it something to do with the workers providing the RL data?by MeteorMarc
4/16/2026 at 7:24:19 PM
Could be on a bike path where bikes are on the left and pedestrians to the right.by rubiquity
4/17/2026 at 6:43:36 AM
I've been running qwen3.6:35b-a3b-q4_K_M (22.3GB) via Ollama.Is the 20.9GB GGUF version better or negligible in comparison?
by Scrounger
4/16/2026 at 9:00:29 PM
I've been trying the Q4_K_M version, and sometimes it gets stuck in a loop. Gemma 4 doesn’t have this issue.by bwv848
4/16/2026 at 9:33:04 PM
This has happened before with quantizations and other backends (ones not used by the research lab). Give it a week, download latest versions of everything, and try again.by yencabulator
4/17/2026 at 12:22:31 PM
I'm having the same issues, the more I use it. The repetition penalty doesn't seem to help.I get some really amusing 'reflective' responses, but I think it needs a bit more cooking. Maybe I'll try another variant.
by mobiuscog
4/17/2026 at 1:10:27 AM
perhaps increasing repitition_penalty might be helpfulby Readerium
4/17/2026 at 11:18:19 AM
what kind of specs does your laptop have? do you know how many tokens/second you get on it?by logicallee
4/16/2026 at 5:50:28 PM
Oh that is pretty good! And the SVG one!by danielhanchen
4/16/2026 at 5:48:38 PM
How does it do with the "car wash" benchmark? :Dby slekker