4/8/2026 at 1:27:00 PM
I looked into this because part of our pipeline is forced to be chunked. Most advice I've seen boils down to "more contiguity = better", but without numbers, or at least not generalizable ones.My concrete tasks will already reach peak performance before 128 kB and I couldn't find pure processing workloads that benefit significantly beyond 1 MB chunk size. Code is linked in the post, it would be nice to see results on more systems.
by PhilipTrettner
4/12/2026 at 5:32:55 AM
Doesn't it depend what you're doing? xz data compression or some video codecs? Retrograde chess analysis (endgame tablebases)? Number Field Sieve factorization in the linear algebra phase?by throwaway81523
4/11/2026 at 2:43:10 PM
Your results match similar analyses of database systems I’ve seen.64KB-128KB seems like the sweet spot.
by twoodfin