The template is what makes the training process so short for humans. We need minimal data and we can run off of that.
Training is both longer and less effective for the LLM because there is no template.
To give an example suppose it takes just one picture for a human to recognize a dog and it takes 1 million pictures for a ML model to do the same. What I’m saying is that it’s like this because humans come preprogrammed with application specific wetware to do the learning and recognition as a generic operation. That’s why it’s so quick. For AI we are doing it as a one shot operation on something that is not application specific. The training takes longer because of this and is less effective.
I disagree that an LLM has no template, but this is getting away from the point.
Did you look at the post I was replying to? You're talking about LLMs being slower, while that post was impressed by LLMs being "faster".
They're posing it as if LLMs recreate the same templating during their training time, and my core point is disagreeing with that. The two should not be compared so directly.
The better comparison to the templating is all the labor that went into making the LLM, not how long the GPUs run.
Template versus template, or specific training versus specific training. Those comparisons make a lot more sense than going criss-cross.