Replies: 1 comment
-
I don't have much experience with |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I was wondering how you can resume training from a checkpoint with different hyperparameter config when training with
transformers
library. Given the example below, no matter what you change in thetraining_args
, these will be overridden by whatever training args are saved in the checkpoint. Some things like eval, batch_size and save_steps are overridable if you amend the checkpoint's JSON config, but other hyperparameters are not.Given a non-PEFT model, you could just save the entire model from the checkpoitn, load it up and call
trainer.train()
on it to achieve this behaviour, but given a PEFT setup I'm not sure how you can do this?Beta Was this translation helpful? Give feedback.
All reactions