Differences in Validation results when training in Pascal Context Dataset
See original GitHub issueHi Hang Zhang!
First I want to thank you for the amazing repository.
I’m trying to train DeepLabv3 with ResNeSt-101 backbone (DeepLab_ResNeSt101_PContext) for the task of semantic segmentation in Pascal Context Dataset. I’m running the code without any issue, however, I’m still under you results from the pre-trained model that you provide in https://hangzhang.org/PyTorch-Encoding/model_zoo/segmentation.html :
Model | Pix Accuracy | MIoU |
---|---|---|
Mine | 79.1 % | 52.1 % |
Yours | 81.9 % | 56.5 % |
I’m using the exact same hyperparameters as you and using the following training command:
python train.py --dataset pcontext --model deeplab --aux --backbone resnest101
Is there something that I’m missing for reaching you results? I assume that your model is trained using Auxiliary Loss but not Semantic Encoding Loss. Are you using some pretraining data maybe?
Thanks in advance!
Alex.
Issue Analytics
- State:
- Created 3 years ago
- Comments:6 (4 by maintainers)
Hi!
Unfortunately, my GPU does not have enough memory to fit a batch size of 16, so I’m trying to simulate it by using gradient accumulation. I suppose that is the main problem, I was asking in case I have missed something else.
I do use your testing script (https://hangzhang.org/PyTorch-Encoding/model_zoo/segmentation.html#test-pretrained).
So I assume that the only problem is the batch size which is a problem with nearly no solution…
For the experiments in the paper, I used AWS EC2 P3.24dn instance with 8x 32GB V100 gpus, but may not be necessary. 16GB per gpu should be enough for most of the experiments.