Comment on GPT-5: Overdue, overhyped and underwhelming. And that’s not the worst of it.
brucethemoose@lemmy.world 21 hours ago
Meanwhile, the chinese and other open models are are killing it. GLM 4.5 is sick. Jamba 1.7 is a great sleeper model for stuff outside coding and STEM. The 32Bs we have like EXAONE and Qwen3 (and finetuned experiments) are mad for 20GB files, and crowding out APIs. There are great little MCP models like Jan too.
Are they AGI? Of course not. They are tools, and that’s what was promised; but the improvements are real.
Hotzilla@sopuli.xyz 15 hours ago
Did you miss that open ai released the oss model few days prior to gpt5?
Larger model: huggingface.co/openai/gpt-oss-120b Smaller model: huggingface.co/openai/gpt-oss-20b
They seem to be quite good
brucethemoose@lemmy.world 24 minutes ago
Nah, I tried them. For the size, they suck, mostly because there’s a high chance they will randomly refuse anything you ask them unless it STEM or Code.
…And there are better models if all you need is STEM and Code.
jbk@discuss.tchncs.de 2 hours ago
mk.absturztau.be/notes/ab3gv6iygjam02uj
WolfLink@sh.itjust.works 14 hours ago
I haven’t played with it too much yet but Queen 3 seems better than GPt-OSS
kopasz7@sh.itjust.works 13 hours ago
In my limitrd testing it seemed relatively trigger-happy on refusals and the results were not impressive either. Maybe on par with 3.5?
Although it is fast at least.