Not sure if this is the best community to post in; please let me know if there’s a more appropriate one. AFAIK Aii@programming.dev is meant for news and articles only.
Not sure if this is the best community to post in; please let me know if there’s a more appropriate one. AFAIK Aii@programming.dev is meant for news and articles only.
That’s not true.
How is this untrue? Generative pre-training is literally training the model to predict what might come next in a given text.
That’s not what an LLM is. That’s part of how it works, but it’s not the whole process.
They never claimed that it was the whole thing. Only that it was part of it.