Posts / ai
AMD's In-House Ryzen AI 395 Box: Exciting News or Just Another Mini PC?
So AMD apparently just dropped some news at their AI Dev Day about releasing their own in-house Ryzen AI 395 mini PC box, coming in June. And the tech corners of the internet are… cautiously underwhelmed? Which, honestly, is a pretty reasonable reaction when you dig into what it actually is.
The short version: it’s a 395 with 128GB unified memory. Same as what you can already buy from a dozen different vendors right now. No extra bandwidth, no architectural magic, just AMD putting their own name on the box. One person who was actually at the event confirmed it directly with an engineer on the floor — just a standard 395 system, nothing more.
And yet here I am, finding the whole thing genuinely interesting anyway.
I’ve been following the Strix Halo mini PC space for a while now. There’s something deeply appealing about the idea of a low-power, perma-on AI inference box sitting quietly in the corner, running local LLMs without melting your electricity bill or your wallet. The AMD AI 395 platform has become something of a darling for this use case — people are running genuinely impressive models on these things. MiniMax M2.7 at 229B parameters on a single 128GB unit, Qwen 3.5 122B with decent context windows, all kinds of quantization wizardry to squeeze performance out of the hardware.
The discussion around this announcement got pretty lively pretty quickly. A lot of people are hoping AMD subsidises the price, which would be great for consumers but — as several people pointed out — would essentially be AMD competing directly with the third-party vendors who’ve built businesses selling 395 mini PCs. That’s a thorny one. Current mini PCs have doubled in price from their launch prices, which is frustrating, so part of me thinks a bit of price competition wouldn’t go astray. But I get why the ecosystem partners might feel a bit stabby about it.
The comparison to Nvidia’s DGX Spark keeps coming up, and it’s apt. Nvidia did exactly this — released their own reference hardware and priced it as a premium product. AMD doing the same feels like a natural, if slightly late, move. One comment that stuck with me was the observation that for Australian customers, the Spark and its clones are running at $8,000-$10,000 AUD. If AMD’s box lands at something more reasonable — say $4,000-$5,000 — and comes with proper corporate backing and support, that actually changes the calculus for businesses wanting to deploy local AI infrastructure. That’s a real value proposition, not just a nothingburger.
The AMD software support thread, though. That got spicy. Someone shared a genuinely painful story about buying AMD embedded hardware specifically marketed as “AI Ready,” only to be told by AMD’s own GitHub support to just “get a newer one” when the drivers didn’t work. That kind of thing is maddening. ROCm has been AMD’s Achilles heel in the AI space for years now, and the community sentiment is pretty clear — the hardware is often excellent, but the software stack is inconsistent and under-resourced compared to what Nvidia offers. It’s the kind of thing that makes technically savvy buyers hesitant, even when the specs look great on paper.
The really exciting stuff in the discussion, though, is the forward-looking hardware rumours. AMD Medusa Halo is apparently in the pipeline with potentially 460-691 GB/s memory bandwidth and possibly 256GB of unified memory. That would be a genuine leap. Right now the 395 platform is bottlenecked at 256 GB/s bandwidth, which is fine for a lot of workloads but starts to feel limiting when you’re trying to run massive models at useful speeds with real context windows. Apple Silicon’s M4 Ultra at 800 GB/s bandwidth and 256GB RAM is the benchmark everyone’s measuring against, even if the $12,000+ price tag makes it aspirational rather than practical for most people.
There’s something philosophically interesting happening in this whole space that I keep thinking about. A year ago, running a 200 billion parameter model locally on consumer-adjacent hardware was essentially science fiction. Now people are genuinely debating the best quantization strategies for doing exactly that on a box that costs a few thousand dollars and draws maybe 60-80 watts at idle. The pace of change is genuinely dizzying. I find myself both excited and slightly anxious about it — which is pretty much my standard emotional state regarding AI these days.
For now, the AMD box feels like a “wait and see” situation. If the price is right and AMD uses it as an opportunity to actually invest in better ROCm support — using it internally for their own AI workloads might light a fire under their software team — then it could be quietly significant. If it’s just another mini PC at the same price point, then yeah, nothingburger.
Either way, Medusa Halo can’t come soon enough.