Right way to load Lora checkpoint to further training #1968
Unanswered
BigDataMLexplorer
asked this question in
Q&A
Replies: 1 comment
-
Hmm, I'm not very familiar with the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
@BenjaminBossan
Hello, can I ask you plase, how should I specify the Trainer if I want to proceed in training from lora checkpoint with same training arguments the model was trained before checkpoint?
I give the example here:
First training:
Now I load model from checkpoint:
PeftModel.from_pretrained(model, checkpoint_path, is_trainable=True)
I need to specify the Trainer again, if I want to continue from checkpoint and use
trainer.train(resume_from_checkpoint)
What I have to write in Trainer? All the arguemnts like in first training?
I am asking because I want the Trainer to continue with exaxtcly same arguments where the Trainer was stopped in training. For example I want to continue with value of learning rate from checkpoint, which decreased linearly during training. Will it overwrite the learning rate, when I write LR again in Trainer? When I put again the same values of learning rate and other things, what will happen?
And on the other hand, if I want to change
Number of eval and save steps, is it ok to change?
Conlusion:
1)I want just to continue the training where it stopped and change eval and save steps.
2)Is not eneough to write again only the model, training and eval dataset in Trainer ?
If so, what will happen to other stable arguments like fp16, weight decay, batch size?
Thank you very much
Beta Was this translation helpful? Give feedback.
All reactions