Damn, your system is insane. Yeah, an RPG maker game is next to nothing compared to that. Still, Dragon Quest I think is 3D. It takes a lot more VRAM than RPG maker.
I have 16GB VRAM, which is a lot for most systems. That’s easily consumed by an LLM. Any model that doesn’t use at least that much tends to perform pretty poorly, in my experience. That’s not mentioning how much heat it generates while running, which has to be removed from the system or it’ll slow down. Even if your system can handle it, it heats up fast. It’s great when I need a heater running, but when I need AC my room gets warm quick.
SabinStargem@lemmy.today 1 day ago
Keep in mind, a 122b (Qwen3.5 family), is high end for consumer machines, but it is likely that DQX would be using a much smaller model. Currently, we have Qwen models that are 8b, 9b, 27b, 35b, 122b, and 397b. Plus, ‘quanting’ can reduce how much memory a model takes up - at a tradeoff, o’course. I am guessing DQX would have multiple local models, and use the player’s hardware metrics to decide which model to deploy.
Alternatively, the Chatty Slime could rely on cloud AI. Depending on Square’s strategy, that could be a freebie or a paid service. If the Chatty Slime gave options to the player - say, trading a potion for a stat seed, or responding to a quiz, Square could sell player behavior data.
…Anyhow, my room has a mini-split AC. One of the best purchases in my life: my room lacked insulation in the first place, so it becomes toasty during summer. The side effect is being able to just run my GPU and not become a human slushy.
Cethin@lemmy.zip 1 day ago
Is your comment written by AI? It seems weird, and we already went over most of what it says.
Also, DQ runs on Nintendo systems. That makes me certain it’s cloud based.
SabinStargem@lemmy.today 1 day ago
No, I didn’t use AI for that. Humans tend to come in many flavors.