Feature request: resume training without overwriting model dir
See original GitHub issueCurrently it does not seem to be possible to resume a previous training where it left off, if there’s already a model directory.
If overwrite is set to False
, https://github.com/joeynmt/joeynmt/blob/cd6974f862922757129fa7d50b5fd842baa996f0/joeynmt/helpers.py#L34 will simply error out.
If overwrite is set to True
, it will wipe out the existing mdoel directory.
Issue Analytics
- State:
- Created 2 years ago
- Comments:6
Top Results From Across the Web
Issues · joeynmt/joeynmt
Feature request : resume training without overwriting model dir enhancement New feature or request. #162 opened on Dec 14, 2021 by cdleong.
Read more >Keras: Starting, stopping, and resuming training
Learning how to start, stop, and resume training a deep learning model is a super important skill to master — at some point...
Read more >Trainer
The Trainer class provides an API for feature-complete training in PyTorch for most standard use cases. It's used in most of the example...
Read more >Checkpointing — PyTorch Lightning 1.6.3 documentation
Checkpointing your training allows you to resume a training process in case ... use a pre-trained model for inference without having to retrain...
Read more >Create custom training jobs | Vertex AI
Alternatively, if you have already created a Python training application or custom container image, then skip ahead to the Without autopackaging section. With ......
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Trying to think through desired behavior if we create a continue flag, and it now interacts with the overwrite flag.
In related news: https://www.philschmid.de/sagemaker-spot-instance
someday having S3 checkpointing would be cool!