GDumb and data augmentation
See original GitHub issueHi! Iβm testing some strategies to reach the accuracy values stated in the original papers. In particular, Gdumb with mem_size=500 reaches 90% of accuracy in the split-mnist benchmark. For the parameters I use the original implementation of GDumb: https://github.com/drimpossible/GDumb
π Describe the bug There is a drop (~30%) in accuracy using data augmentation. This problem doesnβt appear using other strategies (with the same settings: neural network, parameters, regularization etc.) .
π To Reproduce
train_transform = Compose([
#RandomCrop(28, padding=4),
ToTensor(),
Normalize((0.1307,), (0.3081,))
])
eval_transform = Compose([
ToTensor(),
Normalize((0.1307,), (0.3081,))
])
scenario = SplitMNIST(
n_experiences=5, fixed_class_order=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9],
seed=1234, train_transform=train_transform, eval_transform=eval_transform)
π Expected behavior Data augmentation should help to achieve better results despite the strategy.
Issue Analytics
- State:
- Created 2 years ago
- Comments:12 (5 by maintainers)
Top Results From Across the Web
GDumb: A Simple Approach that Questions Our Progress in ...
SGDR, along with standard data augmentation is used. GDumb uses cutmix [46] with p=0.5 and Ξ±=1.0 for regularization on all datasets except MNIST....
Read more >GDumb: A Simple Approach that Questions Our Progress in ...
SGDR, along with standard data augmentation is used. GDumb uses cutmix [46] with p=0.5 and Ξ±=1.0 for regularization on all datasets except MNIST....
Read more >Class-Incremental Learning via Dual Augmentation
Data Augmentation. Literature is rich on data augmentation for improving the generalization of ... Gdumb: A simple approach that questions our.
Read more >GDumb: A Simple Approach that Questions ... - ResearchGate
Termed Greedy Sampler and Dumb Learner [31] randomly samples data ... Augmented Geometric Distillation for Data-Free Incremental Person ReID.
Read more >arXiv:2104.05025v3 [cs.LG] 2 May 2022
GDUMB Prabhu et al. (2020) performs offline training on the buffer with unlimited computation and unrestricted use of data augmentation atΒ ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
@gab709 I did some experiments. I donβt think there is any bug here. The problem is that you are using an MLP. If you use a CNN everything works as expected.
Here are the results using a LeNet-like CNN:
@gab709 if you want to look into this, I can support you. Just ping me on slack if you need help.
Otherwise, I will need the complete script to reproduce the error.