4/12/2026 at 11:49:47 PM
Just spent the last week or so porting TheRock to stagex in an effort to get ROCm built with a native musl/mimalloc toolchain and get it deterministic for high security/privacy workloads that cannot trust binaries only built with a single compiler.It has been a bit of a nightmare and had to package like 30+ deps and their heavily customized LLVM, but got the runtime to build this morning finally.
Things are looking bright for high security workloads on AMD hardware due to them working fully in the open however much of a mess it may be.
by lrvick
4/13/2026 at 2:06:05 AM
I also attempted to package ROCM on musl. Specifically, packaging it for Alpine Linux.It truly is a nightmare to build the whole thing. I got past the custom LLVM fork and a dozen other packages, but eventually decided it had been too much of a time sink.
I’m using llama.cpp with its vulkan support and it’s good enough for my uses. Vulkan so already there and just works. It’s probably on your host too, since so many other things rely on it anyway.
That said, I’d be curious to look at your build recipes. Maybe it can help power through the last bits of the Alpine port.
by WhyNotHugo
4/13/2026 at 3:03:02 AM
Keep an eye out for a stable rocm PR to stagex in the next week or so if all goes well.by lrvick
4/13/2026 at 12:35:44 AM
https://github.com/ROCm/TheRock/issues/3477 makes me quite sad for a variety of reasons. It shouldn't be like this. This work should be usable.by jauntywundrkind
4/13/2026 at 12:53:55 AM
Oh I fully abandoned TheRock in my stagex ROCm build stack. It is not worth salvaging, but it was an incredibly useful reference for me to rewrite it.by lrvick
4/13/2026 at 6:27:04 AM
So much about this confuses me. What do Kitty and ncurses have to do with ROCm? Why is this being built with GCC instead of clang? Why even bother building it yourself when the tarballs are so good and easy to work with?by MrDrMcCoy
4/13/2026 at 8:45:20 AM
The analysis was AI generated. This was Claude brute-forcing itself through building a library.by jeroenhd
4/13/2026 at 7:32:30 AM
On the last one: OP said they were trying to get it working for a musl toolchain, so the tarballs are probably not useful to them (I assume they're built for glibc).Agreed on the others though. Why's it even installing ncurses, surely that's just expected to be on the system?
by CamouflagedKiwi
4/13/2026 at 1:27:10 AM
[flagged]by hackernows_test
4/13/2026 at 2:52:29 AM
Wait ?You don't trust Nvidia because the drivers are closed source ?
I think Nvidia's pledged to work on the open source drivers to bring them closer to the proprietary ones.
I'm hopping Intel can catch up , at 32GB of VRAM for around 1000$ it's very accessible
by 999900000999
4/13/2026 at 8:48:45 AM
Nvidia is opening their source code because they moved most of their source code to the binary blob they're loading. That's why they never made an open source Nvidia driver for Pascal or earlier, where the hardware wasn't set up to use their giant binary blobs.It's like running Windows in a VM and calling it an open source Windows system. The bootstrapping code is all open, but the code that's actually being executed is hidden away.
Intel has the same problem AMD has: everything is written for CUDA or other brand-specific APIs. Everything needs wrappers and workarounds to run before you can even start to compare performance.
by jeroenhd
4/13/2026 at 10:08:41 AM
In the python eco system you can just replace CUDA with DirectML in at least one popular framework and it just runs. You are limited to windows then though.by Asmod4n
4/13/2026 at 3:01:52 AM
Nvidia has been pledging that for years. If it ever actually happens, I am here for it.by lrvick
4/13/2026 at 4:40:00 AM
It happened 2 years ago:https://developer.nvidia.com/blog/nvidia-transitions-fully-t...
by shaklee3
4/13/2026 at 6:47:44 AM
Their userspace is still closed. ROCm is fully open.by cyberax
4/13/2026 at 9:46:10 AM
Provided you happen to have one of those few supported GPUs.Thus being open source isn't of much help without it.
by pjmlp
4/13/2026 at 3:28:12 AM
> IntelFor some workloads, the Arc Pro B70 actually does reasonably well when cached.
With some reasonable bring-up, it also seems to be more usable versus the 32gb R9700.
by cmxch
4/13/2026 at 6:20:59 AM
I have both of those cards. Llama.cpp with SYCL has thus far refused to work for me, and Vulkan is pretty slow. Hoping that some fixes come down the pipe for SYCL, because I have plenty of power for local models (on paper).by MrDrMcCoy
4/13/2026 at 4:16:27 AM
>Just spent the last week or so porting TheRock to stagex in an effort to get ROCm built with a native musl/mimalloc toolchain and get it deterministic for high security/privacy workloads that cannot trust binaries only built with a single compiler....I have a feeling you might not be at liberty to answer, but... Wat? The hell kind of "I must apparently resist Reflections on Trusting Trust" kind of workloads are you working on?
And what do you mean "binaries only built using a single compiler"? Like, how would that even work? Compile the .o's with compiler specific suffixes then do a tortured linker invo to mix different .o's into a combined library/ELF? Are we talking like mixing two different C compilers? Same compiler, two different bootstraps? Regular/cross-mix?
I'm sorry if I'm pushing for too much detail, but as someone whose actually bootstrapped compilers/user spaces from source, your usecase intrigues me just by the phrasing.
by salawat