Betteridge’s law of headlines.
So no.
Submitted 3 months ago by 101@feddit.org to technology@beehaw.org
https://theconversation.com/can-ai-talk-us-out-of-conspiracy-theory-rabbit-holes-238580
Betteridge’s law of headlines.
So no.
That’s just what they want you to think.
Hehe
the better goal is creating new unique conspiracy theories that nobody has heard of with the help of machine learning.
I have two main thoughts on this
LLMs are not at this time reliable sources of factual information. The user may be getting something that was skimmed from factual information, but the output can often be incorrect since the machine can’t “understand” the information it’s outputting.
This could potentially be an excellent way to do real research for people who were not provided research skills by their education. Conspiracy theorists often start off as curious but undisciplined before they fall into the identity aspects of the theories. If a machine using human-like language is able to report factual information quickly, reliably, and without judgement to those who wouldn’t be able to find that info on their own, this could actually be a very useful tool.
halm@leminal.space 3 months ago
According to that research mentioned in the article, the answer is yes. The big caveats are
Butterbee@beehaw.org 3 months ago
It’s not even fundamentally possible with the current LLMs. It’s like saying “Yes, it’s totally possible to do that! We just need to invent something that can do that first!”
halm@leminal.space 3 months ago
I think we agree on the limited capability of (what is currently passed off as) “artificial intelligence”, yes.
CanadaPlus@lemmy.sdf.org 3 months ago
You overestimate how hard it is to get a conspiracy theorist to click on something.
They used a purpose-finetuned GPT-4 model for this study, and it didn’t go off script once.