Comment on How close are we to "manually tuning" LLMs?

ndru@lemmy.world ⁨1⁩ ⁨month⁩ ago

I read a series of super interesting set of posts a few months back where someone was exploring the dimensional concept space in LLMs. The jump off point was the discovery of weird glitch tokens which would break GPTs, making them enter a tailspin of nonsense, but the author presented a really interesting deep dive into how concepts are clustered dimensionally. I don’t know if any of that means we’re any where cost to being able to find those conceptual weighting clusters and tune them, but well worth a read for the curious. There’s also a YouTube series which really dives into the nitty gritty of LLMs, much of which goes over my head, but helped me understand at least the outlines of how the magic happens.

(Excuse any confused terminology here, my knowledge level is interested amateur!)

Posts on glitch tokens and exploring how an LLM encodes concepts in multidimensional space. lesswrong.com/…/solidgoldmagikarp-iii-glitch-toke…

YouTube series is by 3Blue1Brown - m.youtube.com/@3blue1brown

source
Sort:hotnewtop