How to train on part of the dataset
See original GitHub issue❓ Questions/Help/Support
Hi @vfdev-5 ,
We have some requirements for very big dataset:
See we have 100 iterations for 1 epoch, we want to call run()
to train on the first 10 iterations data, do some other things for the model and then call run()
to train on the second 10 iterations data, …
Is it possible in ignite now?
I found the the iter for dataloader in ignite is always from beginning?
https://github.com/pytorch/ignite/blob/v0.4.2/ignite/engine/engine.py#L771
Thanks.
Issue Analytics
- State:
- Created 3 years ago
- Comments:17 (8 by maintainers)
Top Results From Across the Web
How to split a Dataset into Train and Test Sets using Python
The simplest way to split the modelling dataset into training and testing sets is to assign 2/3 data points to the former and...
Read more >How to Train to the Test Set in Machine Learning
Train to Test Set for Classification ; from sklearn.neighbors import KNeighborsClassifier. # load the dataset ; print(X.shape, y.shape). # split ...
Read more >Training and Test Sets: Splitting Data | Machine Learning
Figure 1. Slicing a single data set into a training set and test set. · Figure 2. Validating the trained model against test...
Read more >A guide through training dataset in Machine Learning
The training needs to be done with a clean, complete, and accurate set of input data. This learning set is supposed to teach...
Read more >Split Your Dataset With scikit-learn's train_test_split()
The training set is applied to train, or fit, your model. For example, you use the training set to find the optimal weights,...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Hi @Nic-Ma ,
yes, you are right about that 👍 This can be an interesting approach. Thanks ! Currently, a sort of problem with calling multiple times
trainer.run()
is that it will trigger all the times events likeSTARTED
,EPOCH_STARTED
etc which may not what we expect. This is more or less conceptual problem of what therun
is. This is something we are discussing with the team.There is another approach (workaround) to keep control of the dataflow:
Cool, your trick seems very useful, let me make a demo to verify first. Thanks.