- Joined
- Aug 28, 2019
Strix Halo will do just that. Too big to fit on AM5, up to 16 cores and 40 CUs (RDNA3.5), coping with the bandwidth problem by supporting 256-bit (quad-channel) DDR5 memory.Now, you couldn't practically go this big for a PC SoC, because DDR5 doesn't have the bandwidth to keep a GPU that big busy.
Will there be a mini PC with Strix Halo soldered onto a board, and four DDR5 DIMM slots, actually supporting quad-channel memory? Beats me. We could see fully soldered LPDDR5(X) solutions instead. Maybe 2x CAMM could work, but they can't get to quad-channel with stacking.
(Strix Halo should be considered a laptop-first product. Mini PCs are more of an afterthought.)
XDNA2 may have doubled or tripled die area compared to XDNA1 (on the same TSMC N4 node) to make its big jump to ~45-50 TOPS, and every CPU/APU should have this "consumer-level" of dedicated AI performance eventually because Microsoft demands it. But how much will they improve in the future? Meanwhile, Nvidia advertises "242 AI TOPS" using tensor cores for the RTX 4060. TOPS ratings aren't always comparable (I'm seeing RTX 3080 gets about doubled RTX 4060 Stable Diffusion performance at the same TOPS) so we'll have to see how well these accelerators do in practice. But someone who wants to use AI for more than 15 minutes before getting bored of it can probably afford a discrete GPU.Consumer-level AI runs on the CPU. You can't assume the presence of a dGPU for a mass-market application, which is why AMD, Intel, Apple, and Qualcomm are all putting NPUs on their CPUs.