Comment on New AI model can hallucinate a game of 1993’s Doom in real time
Even_Adder@lemmy.dbzer0.com 3 months agoI mean, you’ve never seen a purple elephant with a tennis racket. None of that exists in the data set since elephants are neither purple nor tennis players. Exposure to all the individual elements allows for generation of concepts outside the existing data, even though they don’t exit in reality or in the data set.
MentalEdge@sopuli.xyz 3 months ago
Ok.
Try to get an image generator to create an image of a tennis racket, with all racket-like objects removed from the training data.
Explain the concept to it with words alone.
That’s the problem you’ll run into the second you want to realize a new game genre.
Even_Adder@lemmy.dbzer0.com 3 months ago
There are more forms of guidance than just raw words. Just off the top of my head, there’s inpainting, outpainting, controlnets, prompt editing, and embeddings. The researchers who pulled this off definitely didn’t do it with text prompts.
MentalEdge@sopuli.xyz 3 months ago
Obviously.
But at what point does that guidance just become the dataset you removed?
Even_Adder@lemmy.dbzer0.com 3 months ago
The whole point is that it didn’t know the concepts beforehand, and no it doesn’t become the dataset. Observations made of the training data are added to the model’s weights after training, the dataset is never relevant again as the model’s weights are locked in.
Or you could train a more general model. These things happen in steps, research is a process.