Comment on how things become science

<- View Parent
lvxferre@mander.xyz ⁨1⁩ ⁨week⁩ ago

I’m failing to see how this is different from making up a fact and then spreading it to news outlets.

They uploaded the papers to a single preprint server. That’s important.

Preprints are papers predating any sort of peer review; as such, there’s a lot of junk mixed in — no big deal if you know the field, but a preprint server is certainly not a source of reliable information, nor it should be treated as such. On the other side, news outlets are expected to provide you reliable information, curated and researched by journalists.

And peer review is a big fucking deal in science, because it’s what sorts all that junk out. Only a muppet who doesn’t fucking care about misinformation would send bots to crawl preprints, and feed the resulting data into a large model.

So no, your comparison is not even remotely accurate. What they did is more like writing bullshit in a piece of paper, gluing it on a random phone pole, and checking if someone would repeat that bullshit.

They also went through the trouble to make sure that no reasonably literate human being would ever confuse that thing with an actually scientific paper. As the text says:

Feeding false information to an LLM is no different that a magazine. It only regurgitates what’s been said.

Yes, it is different. Because the large token model won’t simply “repeat” things, it’ll mix and match them and form all sorts of bullshit, even if you didn’t feed it with any bullshit.

Here’s an example of that, fresh from the oven. I don’t reasonably expect people to be feeding misinfo regarding Latin pronunciation into bots, and yet a lot of this table is nonsense:

Image

Compare the table above with this table and this one and you’ll notice the obvious errors:

All it had to do was to copy info from Wiktionary, as it includes even phonetic and phonemic info. But since the bot is not just “regurgitating” info — it’s basically predicting what should come next — it’s mixing-and-matching shit into nonsense.

It isn’t going to suddenly start doing science on its own to determine if what you’ve said is true or not.

If you actually read the bloody article instead of assuming, you’d know why the researchers did this: they don’t expect the bot to do science on its own, they expect people to treat info from those bots as potentially incorrect.

Its job is to tell you what color the sky is based on what you told it the color of the sky was.

And your job is to not trust it if it tells you “Yes, you are completely right! The colour of the sky is always purple. Do you need further information on other naturally purple things?”

source
Sort:hotnewtop