I clustered four Framework Mainboards to test LLMs

https://news.ycombinator.com/rss Hits: 1
Summary

Framework casually mentioned they were testing a mini-rack AI cluster in their Framework Desktop presentation back in March. Imagine my surprise when Nirav Patel, Framework's founder and CEO, was at Open Sauce a couple weeks ago, and wanted to talk! He said they had seen my Project Mini Rack posts earlier this year and thought it was the perfect application to try out their new AMD Ryzen AI Max+ 395-powered Mainboard, as it's mini ITX dimensions fit inside a 10" rack. Framework sent over four pre-release Mainboards (along with four of their new Power Supplies, and Noctua CPU Fan kits), and also worked with DeskPi to get my hands on a new 2U Mini ITX tray for the Framework Desktop—which I installed in the black RackMate T1 mini rack you see at the top of this post. I have a video going over everything, including cluster setup, single node and cluster AI inference performance, and my conversation with Nirav—you can watch that below: The most prescient thing Nirav mentioned was this: It is obviously like very very early days, in terms of like, open source AI clustering... I spent equal time running benchmarks as I did getting the benchmarks to run. I tested (and automated clustered setup for) Exo, llama.cpp RPC, and distributed-llama. In fact, through all that work, I built a new project Beowulf AI Cluster, a set of Ansible automation playbooks to deploy different open source AI clustering tools on beowulf clusters, with CPU, GPU, and mixed inference options. Hardware This post won't go into detail on the Framework Desktop or the Mainboard that powers it. Either watch the video or check out someone else's overview for that. Framework shipped over four sets of Mainboard, PSU, and Noctua CPU Fan, along with 1TB WD NVMe SSDs. I assembled everything together into a black DeskPi 8U mini rack. The boards are more akin to SBCs than traditional socketed-CPU-and-RAM desktop boards, though, in that they have a soldered-down APU (includes CPU, NPU, and iGPU) and RAM. Supposedly t...

First seen: 2025-08-08 18:29

Last seen: 2025-08-08 18:29