Basically, it’s a calculator that can take letters, numbers, words, sentences, and so on as input.
And produce a mathematically “correct” sounding output, defined by language patterns in the training data.
This core concept is in most if not all “AI” models, not just LLMs, I think.
There’s a part of our brain called the salience network, that continually models and predicts our environment and directs our conscious attention to things it can’t predict. When we talk to each other, most of the formal content is predictable, and the salience network filters it out; the unpredictable part that’s left is the actual meaningful part.
LLMs basically recreate the salience network. They continually model and predict the content of a text stream the same way we do—except instead of modeling someone else’s words so they can recognize the unpredictable/meaningful part, they model their own words so they can keep predicting the next ones.
This raises an obvious issue: when our salience networks process the stream of words coming out of such an LLM, it’s all predictable, so our brains tell us there’s no actual message. When AI developers ran into this, they added a feature called “temperature” that basically injects randomness into the generated text—enough to make it unpredictable, but not obvious nonsense—so our salience networks will get fooled into thinking there’s meaningful content.
This was a great read! I did have a feeling LLMs would be a but boring when setting them to low temperatures and now I understand why.
I have found that LLMs are great for predetermined processes, like generating JSON using a given format and writing code, but they suck at creative tasks, and your amazing explanation now told me why that is.
Thanks again!
This was a great read, Thanks!
I have a new rabbit hole to explore 😝