2/11/2026 at 5:47:57 PM
Pelican generated via OpenRouter: https://gist.github.com/simonw/cc4ca7815ae82562e89a9fdd99f07...Solid bird, not a great bicycle frame.
by simonw
2/11/2026 at 6:01:59 PM
Thank you for continuing to maintain the only benchmarking system that matters!Context for the unaware: https://simonwillison.net/tags/pelican-riding-a-bicycle/
by btown
2/11/2026 at 7:38:40 PM
It's interesting how some features, such as green grass, a blue sky, clouds, and the sun, are ubiquitous among all of these models' responses.by gabiruh
2/11/2026 at 9:44:37 PM
If you were a pelican, wouldn't you want to go cycling on a sunny day?Do electric pelicans dream of touching electric grass?
by btown
2/12/2026 at 3:37:38 AM
Do electric pelicans dream of touching electric grass?That would be shocking news to me.
by Magniquick
2/12/2026 at 7:49:39 AM
Please leave the Internet :)by davidwritesbugs
2/11/2026 at 9:49:04 PM
It is odd, yeah.I'm guessing both humans and LLMs would tend to get the "vibe" from the pelican task, that they're essentially being asked to create something like a child's crayon drawing. And that "vibe" then brings with it associations with all the types of things children might normally include in a drawing.
by derefr
2/11/2026 at 8:55:29 PM
They will start to max this benchmark as well at some point.by l_eo
2/11/2026 at 10:25:50 PM
It's not a benchmark though, right? Because there's no control group or reference.It's just an experiment on how different models interpret a vague prompt. "Generate an SVG of a pelican riding a bicycle" is loaded with ambiguity. It's practically designed to generate 'interesting' results because the prompt is not specific.
It also happens to be an example of the least practical way to engage with an LLM. It's no more capable of reading your mind than anyone or anything else.
I argue that, in the service of AI, there is a lot of flexibility being created around the scientific method.
by ljm
2/11/2026 at 10:32:56 PM
For 2026 SOTA models I think that is fair.For the last generation of models, and for today's flash/mini models, I think there is still a not-unreasonable binary question ("is this a pelican on a bicycle?") that you can answer by just looking at the result: https://simonwillison.net/2024/Oct/25/pelicans-on-a-bicycle/
by tylervigen
2/12/2026 at 7:58:53 AM
RLHF (reinforcement learning from human feedback) is to a large extent about resolving that ambiguity by simply polling people for their subjective judgement.I've worked one an RLHF project for one of the larger model providers, and the instructions provided to the reviewers were very clear that if there was no objective correct answer, they were still required to choose the best answer, and while there were of course disagreements in the margins, groups of people do tend to converge on the big lines.
by vidarh
2/11/2026 at 10:36:56 PM
So if it can generate exactly what you had in mind based presumably on the most subtle of cues like your personal quirks from a few sentences that could be _terrifying_, right?by interstice
2/12/2026 at 11:19:29 AM
Simon has written a page specifically for you: https://simonwillison.net/2025/nov/13/training-for-pelicans-...by 9dev
2/12/2026 at 1:01:04 AM
This is actually a good benchmark, I use to roll my eyes at it. Then I decided to apply the same idea and ask the models to generate SVG image of "something" not going to put it out there. There was a strong correlation between how good the models are and the image they generated. These were also no vision images, so I don't know if you are serious but this is a decent benchmark.by segmondy
2/12/2026 at 9:48:54 AM
That's a bike that's ergonomically designed for pelicans.It is unreasonable to expect pelicans to ride human bikes, they have different anatomy.
by hasperdi
2/12/2026 at 10:23:51 AM
The next frontier:Draw a pelican on a bicycle ergonomically designed for pelicans.
by MrsPeaches
2/12/2026 at 11:14:06 AM
It may be a joke, but I think this is correct.For reasons, I have tried to get Stable Diffusion to put parrots into spacesuits. Always ended up with the beak coming out where the visor glass should've been, either no wings at all or wings outside the suit, legs and torso just human-shaped.
ChatGPT got the helmet right, but their wings and tail (and sometimes claws) were exposed to vacuum, still very much closer to a human in either a normal or scifi space suit that happens to also be wearing a parrot head inside the space suit, and has tacked some costume wings on the outside.
Essentially, it's got the same category of wrong as fantasy art's approach to what women's armour should look like: aesthetics are great, but it would be instantly lethal if done for real.
by ben_w
2/12/2026 at 7:37:37 PM
My more advanced prompt, for when models do a good job on the original, is this one:> Generate an SVG of a California brown pelican riding a bicycle. The bicycle must have spokes and a correctly shaped bicycle frame. The pelican must have its characteristic large pouch, and there should be a clear indication of feathers. The pelican must be clearly pedaling the bicycle. The image should show the full breeding plumage of the California brown pelican.
by simonw
2/12/2026 at 11:08:32 AM
Thereafter: Design a bike that an actual pelican can learn to ride in real life.by mitjam
2/11/2026 at 6:09:46 PM
Now this is the test that matters, cheers Simon.by _joel
2/11/2026 at 9:23:55 PM
The bird not having wings, but all of us calling it a 'solid bird' is one of the most telling examples of the AI expectations gap yet. We even see its own reasoning say it needs 'webbed feet' which are nowhere to be found in the image.This pattern of considering 90% accuracy (like the level we've seemingly we've stalled out on for the MMLU and AIME) to be 'solved' is really concerning for me.
AGI has to be 100% right 100% of the time to be AGI and we aren't being tough enough on these systems in our evaluations. We're moving on to new and impressive tasks toward some imagined AGI goal without even trying to find out if we can make true Artificial Niche Intelligence.
by RC_ITR
2/12/2026 at 10:15:07 AM
This test is so far beyond AGI. Try to spit out the SVG for a pelican riding a bicycle. You are only allowed to use a simple text editor. No deleting or moving the text cursor. You have 1 minute.by zarzavat
2/12/2026 at 11:49:12 PM
Sorry, is your definition of AGI "doing things worse than humans can do, but way faster?" because that's been true of computers for a long time.by RC_ITR
2/13/2026 at 2:14:22 PM
I mean for this particular benchmark, yes.You'd have to put it in an agentic loop to perform corrections otherwise.
by pixl97
2/12/2026 at 3:50:36 PM
The benchmarks are harder than you might imagine and contain more wrong answers and terrible questions than you would expect.You don't need to take my word for it, try playing MMLU yourself.
https://d.erenrich.net/are-you-smarter-than-an-llm/index.htm...
Its not MMLU-Pro btw, which is considerably harder.
by kingstnap
2/12/2026 at 11:50:21 PM
Sure and AGI will 100% it 100% of the time, even if it is hard.by RC_ITR
2/11/2026 at 9:50:39 PM
MMLU performance caps out around 90% because there are tons of errors in the actual test set. There's a pretty solid post on it here: https://www.reddit.com/r/LocalLLaMA/comments/163x2wc/philip_...As far as I can tell for AIME, pretty much every frontier model gets 100% https://llm-stats.com/benchmarks/aime-2025
by Rudybega
2/12/2026 at 11:44:13 PM
Here's the score for new AIME's, where we know the answers aren't in training.https://matharena.ai/?view=problem&comp=aime--aime_2026
As for MMLU, is your assertion that these AI labs are not correcting for errors in these exams and then self-reporting scores less than 100%?
As implied by the video, wouldn't it then take 1 intern a week max to fix those errors and allow any AI lab to become the first to consistently 100% the MMLU? I can guarantee Moonshot, DeepSeek, or Alibaba would be all over the opportunity to do just that if it were a real problem.
by RC_ITR
2/12/2026 at 12:22:52 AM
It has a wing. Look at the code comments in the SVG!by simonw
2/11/2026 at 8:22:51 PM
This Pelican benchmark has become irrelevant. SVG is already ubiquitous.We need a new, authentic scenario.
by solarized
2/11/2026 at 8:52:41 PM
Like identifying names of skateboard tricks from the description? https://skatebench.t3.gg/by viraptor
2/11/2026 at 9:23:55 PM
I don’t care how practical it may or may not be, this is my new favorite LLM benchmarkby alargemoose
2/11/2026 at 9:38:38 PM
I couldn't find an about page or similar?by stevage
2/11/2026 at 10:00:18 PM
Here's the public sample https://github.com/T3-Content/skatebench/blob/main/bench/tes...I don't think there's a good description anywhere. https://youtube.com/@t3dotgg talks about it from time to time.
by viraptor
2/11/2026 at 9:25:18 PM
o3-pro is better than 5.2 pro! And GPT 5 high is best. Really quite interesting.by hmottestad
2/11/2026 at 9:56:20 PM
1. Take the top ten searches on Google Trends
(on day of new model release)
2. Concatenate
3. SHA-1 hash them
4. Use this as a seed to perform random noun-verb
lookup in an agreed upon large sized dictionary.
5. Construct a sentence using an agreed upon stable
algorithm that generates reasonably coherent prompts
from an immensely deep probability space.
That's the prompt. Every existing model is given that prompt and compared side-by-side.You can generate a few such sentences for more samples.
Alternatively, take the top ten F500 stock performers. Some easy signal that provides enough randomness but is easy to agree upon and doesn't provide enough time to game.
It's also something teams can pre-generate candidate problems for to attempt improvement across the board. But they won't have the exact questions on test day.
by echelon
2/12/2026 at 4:41:50 AM
Have you tried with qwen-coder-next yet?by blurbleblurble
2/11/2026 at 6:25:07 PM
How many pelican riding bicycle SVGs were there before this test existed? What if the training data is being polluted with all these wonky results...by pwython
2/11/2026 at 8:45:19 PM
I'd argue that a models ability to ignore/manage/sift through the noise added to the training set from other LLMs increases in importance and value as time goes on.by bwilliams18
2/11/2026 at 7:36:20 PM
You're correct. It's not as useful as it (ever?) was as a measure of performance...but it's fun and brings me joy.by nerdsniper
2/12/2026 at 2:45:46 AM
Pretty damn great bird, tbh.by brianjking
2/12/2026 at 12:46:48 AM
The idea at the time is that it was obviously not part of the training set, now that it's a metric,it's worthless. Try an elephant smoking s cigar on the beachby TZubiri