Comment on One more LLM

<- View Parent
Thorry@feddit.org ⁨1⁩ ⁨day⁩ ago

One of the things I’ve really had AI fanboys going crazy over is by asking them to feed their AI generated code back into the AI and ask for potential issues or mistakes. Without fail it points out very obvious issues and sometimes some less obvious ones as well. If your AI coder is so good, why does it know it fucked up?

This is basically what these new “agent” modes do. Just keep feeding the same thing in on itself till it finds some balance. Often using an external tool, like building the project for example, to determine if it’s done. However I’ve seen this end up in loops a lot. If all of the training data contained the same mistake (or the resulting network always produces that mistake), it can’t fix it. It will just say oh I’ve made a mistake let me fix that over and over again as the same obvious error pops out.

source
Sort:hotnewtop