Improvement on motion-cnn result: 84.1% on split-1, with VGG-16
See original GitHub issueHi, all
I did some investigation on why the motion-cnn result is much lower than their original paper. After a simple modification, I am able to achieve 84.1% top-1 accuracy. This modification is adding transforms.FiveCrop()
to the transformation. Before this modification, the result is only 80.5%. I use pretrained model fromhttps://github.com/feichtenhofer/twostreamfusion, I think further improvement can be down with transfroms.TenCrop()
.
I think with this modification, it can bridge the gap of performance between twostream model trained on pytorch and other frameworks.
Issue Analytics
- State:
- Created 5 years ago
- Reactions:1
- Comments:42
Top Results From Across the Web
Improvement on motion-cnn result: 84.1% on split-1, with VGG ...
This modification is adding transforms.FiveCrop() to the transformation. Before this modification, the result is only 80.5%. I use pretrained ...
Read more >An Improved VGG16 Model for Pneumonia Image Classification
Thus, this paper proposes IVGG13 (Improved Visual Geometry Group-13), a modified VGG16 model for classification pneumonia X-rays images.
Read more >Expression Recognition Method Using Improved VGG16 ...
18] proposed a facial expression recognition system based on depth CNN. This paper constructs a dual channel CNN model and gathers the channel ......
Read more >The two-stream-action-recognition from jeffreyyihuang
Improvement on motion-cnn result : 84.1% on split-1, with VGG-16. Hi, all. I did some investigation on why the motion-cnn result is much...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
@duygusar yeah, you also need to average for training. Alternatively, you can expand label to match the shape of the model output. But I think it won’t be much variance, you can written down cross entropy loss for these two cases, and compare them.
You can directly refer to official reference, I will just paste it here:
Basically, you just average across ncrops, and you will have the same batch size with label.