Comment on Megrez2: 21B latent, 7.5B on VRAM, 3B active—MoE on single 8GB card

<- View Parent
felsiq@piefed.zip ⁨17⁩ ⁨hours⁩ ago

Trying to literally ELI5 so this might be oversimplified a bit:

New AI model using a Mixture of Experts (MoE) approach, which combines different AIs that are optimized for certain things into one AI that’s good at more things. This usually needs a lot of space on graphics cards and requires really high end hardware, but this model fits into 8gb of space on a card, which is a very common amount to have on a modern graphics card, so many more people will be able to use it.

source
Sort:hotnewtop