Comment on Trump’s brand of US capitalism faces ‘socialist’ backlash from conservatives
brucethemoose@lemmy.world 5 days agoOn the contrary, I think inference is going on-device more in the future, but ‘users’ will still need decent CPUs and GPUs. Intel is well set up for this: they have good CPU, GPU, and NPU IP.
Intel can go ARM if they want, no problem, just like AMD can (and almost tried). They could theoretically preserve most of their core design and still switch ISA.
Servers will still need CPUs for a long time.
As for GPU compute, we are both in a bubble, and at several forks in the road:
-
Is bitnet ML going to take off? If it does, that shifts the advantage to almost cyptominer-like ASICs, as expensive matrix multiplication no longer matters for inference.
-
Otherwise, what about NPUs? Huawei is already using them, and even training good production models with them. Intel can try their hand at this game again if loads start shifting away from CUDA.
-
Otherwise, they still have a decent shot at the CUDA ecosystem via ZLUDA and their own frameworks. Training and research will probably forever be Nvidia (and some niches like Cerebra’s), but still.
resipsaloquitur@lemmy.world 5 days ago
Maybe I’ve been in a silo but I’ve never heard “intel” and “npu” in the same breath.
brucethemoose@lemmy.world 5 days ago
Because it’s only used for crappy copilot stuff right now, heh.
But technically the Gaudi 2 processors they bought and developed are server ‘NPUs’. Last I heard, they’re putting them on ice, but they may have integrated the tech into laptop processors anyway, and could scale it back up if Huawei’s NPUs take off.