in my mind(prob oversimplified): making a LLM is: -Get training data - Clean it - Make the model try and predict next token base on preceding tokens. Reward when right. Repeat - Tune the models to human preferences. What part would pretraining of that be?