Basically, it’s a calculator that can take letters, numbers, words, sentences, and so on as input.
And produce a mathematically “correct” sounding output, defined by language patterns in the training data.
This core concept is in most if not all “AI” models, not just LLMs, I think.
This was a great read! I did have a feeling LLMs would be a but boring when setting them to low temperatures and now I understand why.
I have found that LLMs are great for predetermined processes, like generating JSON using a given format and writing code, but they suck at creative tasks, and your amazing explanation now told me why that is.
Thanks again!