As long as it has footnoting so I can see where each piece of information was sourced from, AI chat has its use cases. Without that I genuinely do not see the point at all. It’s like when people “ask Google” something and just blindly trust the highlighted “answer” as infallible truth. It’s just a really, really bad habit to develop and I wish more people understood this.
Zworf@beehaw.org 8 months ago
Not infallible truth. But very often it’s something that is just for personal use.
Some things I’ve asked it recently were like “Which torch is smaller out of these 5?”. Once I find which one I want it’s easy to verify. Or “what does this Spanish expression mean?” or "how do I do ".
Not everyone uses it to try and write authoritative stuff.
Ilandar@aussie.zone 8 months ago
All of those questions you asked it return authoritative answers which you take on face value, unless you spend extra time fact checking them yourself.
Zworf@beehaw.org 8 months ago
Yeah but accuracy isn’t a given with the other methods either. If I ask some randos on reddit I won’t get a perfect answer either. If I google specs or reviews online they are often biased (or even literally fraudulent paid reviews) too.
So yeah for me the LLM output is more than good enough.
Ilandar@aussie.zone 8 months ago
I’m sorry, but citing other examples of bad research practices does not magically make AI good. That is a whataboutism.