But this takes it back away from understanding how LLMs work to attribute personality. The “decision” isn’t a decision in how beings decide things like that. The rolling of dice on numerous vectors resulted in those words, which were then re-included into the context for another trip through the vector matrix mines to new destination tokens to assemble.
It’s dice rolls where the dies selected are based on what started out, using a bunch of lookup tables. AI proponents like to be smug and say “well you won’t find those words in the model” like “yes a compressed vector map that ends up treating words like multiple tokens, referencing others in chains, gzipped to binary, can’t be searched for strings, you are literally correct in the stupidest, most irrelevant way possible.”
Denjin@feddit.uk 3 days ago
But that’s not a lie. Lying implies that you know what an actual fact is and choose to state something different. An LLM doesn’t care about what anything in its database actually is, it’s just data, it might choose to present something to a user that isn’t what the database suggests but that’s not lying.
Saying stuff like “ooh I’m an evil robot” is just what the model thinks would be what the user wants to see at that particular moment.
REDACTED@infosec.pub 3 days ago
You’re thinking about biological lying. I’m talking about software.
en.wikipedia.org/wiki/Reasoning_system
If the question was to tell it’s darkest secret, but it instead chose to come up with an entertaining story instead of factually answering that question, like other Anthropic LLM models did, then by definition of reasoning system, the system (LLM) decided to lie. There are variables that made it “decide” this route, therefore it is not a static, expected output - at the core of it - it was a system’s choice