-
Notifications
You must be signed in to change notification settings - Fork 3.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Disable automatic checkpoint loading #368
Comments
@theSoenke the previous checkpoint is only used if the experiment has the same version. Is that what you mean? The reason for this is that the logger object creates an experiment for every version you run. In that folder it stores all checkpoints. When you want to continue training you set the version to the one you want and it just works. If you want to train a different model then you'd need to pass in a different version or let the logger automatically go to the next version. @neggert this is the main reason for this linking. 99% of the time we want to restore a model attached to a specific version. since the versions are handled by the logger, it makes sense to couple them for this purpose. However, it should not limit the flexibility if someone wants to load a different checkpoint. case 1: case 2: case 3: |
Hmm, okay. I think this should work fine for test tube, even with the changes in #413. This has never worked with MLFlow, since MLFlow generates its own "version". I'll have to give some thought as to how to reproduce this behavior in MLFlow. |
@theSoenke has this been fixed on master? @neggert |
Shall we add an option to disable the automatic loading? I sometimes have to run some old version exp, the loading disturbs a lot. |
I kind of want to remove the automatic loading... |
I agree. Loading should be done only on demand IMO. |
Yeah, I'm on board with this - perhaps should still throw a warning if a checkpoint is going to be overwritten? |
For the time being, is there a recommended way to override this functionality? |
Is your feature request related to a problem? Please describe.
The last checkpoint is being automatically restored when a checkpoint exists. This is an issue for me when the model has previously been trained with different settings or I want to train a network from scratch.
If I want to use a checkpoint I would prefer to explicitly pass the checkpoint that should be used.
Describe the solution you'd like
Provide an option to disable the automatic loading of checkpoints to the trainer.
Describe alternatives you've considered
An alternative would be to modify
ModelCheckpoint
class and add that option there so that only that class is responsible for checkpoint creation and loading.The text was updated successfully, but these errors were encountered: