Comment on Open source projects drown in bad bug reports penned by AI
tal@lemmy.today 1 week ago
So, this isn’t quite the issue being raised by the article – that’s bug reports generated by apparently a bot that they aren’t running.
However, I do feel that there’s more potential with existing LLMs in checking and flagging potential errors than in outright writing code. Like, I’d rather have something like a “code grammar checker” that highlights potential errors for my examination rather than something that generates code from scratch itself and hopes that I will adequately review it.
GenderNeutralBro@lemmy.sdf.org 1 week ago
Agreed. The other good use case I’ve found is as a faster reference for simple things. LLMs are absolutely great for one-liners and generating troublesome (but logically simple) things like complex xpath queries. But I still haven’t seen one generate a good script of even moderate complexity without hand-holding. In some cases I’ve been able to get usable output with a few shots, saving me a bit of time compared to if I’d written the whole darned thing from scratch.
I’ve found LLMs very useful for coding, but they aren’t replacing my actual coding, per se. They replace looking things up, like through man pages, language references, or StackOverflow. Something like ffmpeg, for example, has a million options and it is always a little annoying to sift through the docs manually when I just need to do one specific task.
I’m sure it’ll happen sooner or later. I’m not naive enough to claim that “computers will never be able to do $THING” anymore. I’ll say “not in the next year”, though.
lvxferre@mander.xyz 1 week ago
Your and @tal@lemmy.today’s experiences are basically the same as mine. Except with translation instead of programming.