Logging inconsistency when using `SplitMNIST` with task labels.
See original GitHub issueEach strategy increments its own training_step_counter
after each step. However, it does not distinguish between new step or new task. Therefore, when using SplitMNIST(n_steps=5, return_task_id=True)
, it logs Step 1 (Task 1)
, Step 2 (Task 2
) and so on.
The expected behavior should be Step 1 (Task 1)
, Step 1 (Task 2)
and so on.
Issue Analytics
- State:
- Created 3 years ago
- Comments:5 (2 by maintainers)
Top Results From Across the Web
A Neural Dirichlet Process Mixture Model for Task-Free ...
Review: Summary: The paper proposes to use a Bayesian nonparametric mixture model for task-free (without explicit task labels) continual learning.
Read more >Task Agnostic Continual Learning Using Online Variational ...
In section 2 we analyze this scenario, and present a method (“labels trick”) to improve the performance of “class learning”. For example, on...
Read more >Split-MNIST - Papers With Code
In this paper, we propose a new method, named Self-Attention Meta-Learner (SAM), which learns a prior knowledge for continual learning that permits learning...
Read more >Custom Neural Network Implementation on MNIST using ...
I wondered where to start with your multiquestion, and I decided to do so with a statement: Your code definitely should not look...
Read more >(PDF) Investigating a Baseline Of Self Supervised Learning Towards ...
MNIST classification results for supervised and self-supervised learning ... carry independent supervised task to train a model using the proxy labels.
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
I don’t think this is obvious or expected in general.
Personally, I prefer to have a
step_id
that increases even between different tasks (Step 1 (Task 1), Step 2 (Task 1), Step 3 (Task 2), ...
). This way, the only difference between MT, MIT and SIT is thetask_id=None
in SIT scenarios during training.We really need to write down this kind of high-level design choices in the documentation, so that we don’t end up making the same discussions over and over.
I’m closing this since we all agree.