This is parrot libel
Comment on I wish I was as bold as these authors.
Tar_alcaran@sh.itjust.works 5 months agoLLMs work differently, statistically predicting the next token (roughly equivalent to a word) based on all those that came before it, and parameters finetuned during training.
Which is what a parrot does.
doubtingtammy@lemmy.ml 5 months ago
ignotum@lemmy.world 5 months ago
You take in some information, combine that with some precious experiences, and then output words
Which is what an LLM does.
WalnutLum@lemmy.ml 5 months ago
Flat epistemological statements like this are why I feel like more STEM people need to take Philosophy.
ignotum@lemmy.world 5 months ago
Big fan of philosophy, so please do tell me how my joke is wrong? Does knowledge and beliefs not come from life experiences?
naevaTheRat@lemmy.dbzer0.com 5 months ago
Yeah this is the exact criticism. They recombine language pieces without really doing language. The end result looks like language, but it lacks any of the important characteristics of language such as meaning and intention.
If I say “Two plus two is four” I am communicating my belief about mathematics.
If an llm emits “two plus two is four” it is outputting a stochastically selected series of tokens linked by probabilities derived from training data. If the statement is true or false then that is accidental.
Hence, stochastic parrot.
ignotum@lemmy.world 5 months ago
If i train an LLM to do math, for the training data i generate
a+b=c
statements, never showing it the same one twice.It would be pointless for it to “memorize” every single question and answer it gets since it would never see that question again. The only way it would be able to generate correct answers would be if it gained a concept of what numbers are, and how the add operation operates on them to create a new number.
Rather than memorizing and parroting it would have to actually understand it in order to generate responses.
It’s called generalization, it’s why large amounts of data is required (if you show the same data again and again then memorizing becomes a viable strategy)
Seems like a pointless distinction, you were told it so you believe it to be the case? Why can’t we say the LLM outputs what it believes is the correct answer? You’re both just making some statement based on your prior experiences which may or may not be true
naevaTheRat@lemmy.dbzer0.com 5 months ago
You’re arguing against a position I didn’t put forward. Also
This is what excessive reduction does to a mfer. That is just such a hysterically absurd take.
artichokecustard@lemmy.world 5 months ago
but, the LLM has faith!
ignotum@lemmy.world 5 months ago
The AI builds some kind of a model of the world in order to better understand the input and improve the output prediction,
You have some mental model of how maths work which you have built up through school and other experiences in your life,
You both are given a maths problem, you both give an answer based on your understanding of mathematics
kogasa@programming.dev 5 months ago
If you fine tune a LLM on math equations, odds are it won’t actually learn how to reliably solve novel problems. Just the same as it won’t become a subject matter expert on any topic, but it’s a lot harder to write simple math that “looks, but is not, correct” than it is to waffle vaguely about a topic. The idea of a LLM creating a robust model of the semantics of the text it’s trained on is, at face value, plausible; it just doesn’t seem to actually happen in practice.
ignotum@lemmy.world 5 months ago
Prompt:
ChatGPT:
It’s trained to generate what is most plausible, but with math, the only plausible response is the correct answer (assuming it has been trained on data where that has been the case)