Comment on ELI5. Limit of current gen AI/LLMs

greyscale@lemmy.sdf.org ⁨2⁩ ⁨days⁩ ago

An LLM reads the previous prompts and replies, plus any base prompts. This is considered the context window. Don’t ask me why its not infinite.

The machine will then generate text following the previous text that continues the spirit and intent of the previous text, based on other texts previously digested into weights.

Its the same thing as your phones autocomplete but with a few gigabytes of weights instead of a few kilobytes.

If the data its working with is larger than the context, it will lose it. Theres a chance it’ll halucinate anyway because the text generator later in the text is non-deterministic. Say you’re working with insurance data. Maybe your data is familiar enough to data it previously injested data. So now it starts using wrong data, but it “feels” right as far as the LLM is concerned, because its a text generator, not a truth checker.

You can ask it to look again but its just generating fresh tokens while the context gets more polluted.

Just start looking at the volumes of non-trivial psuedo-information it generates and just try to verify some of the facts it states about your data.

source
Sort:hotnewtop