Evaluate model performance when using chainer.links.BatchNormalization togather with chainer.config.train=False bug report
See original GitHub issuechainer version: 3.4.0 cupy version: 2.4.0
When I am using resnet-101 to classify images. The resnet-101
model have L.BatchNormalization
layer.
After I trained over, I load model parameter and use chainer.using_config('train', False)
to evaluate the model performance.
I am surprised that even I am using train dataset(not validation dataset), the accuracy is lower than When I was in training procedure observation ( it was 99% in the final iteration of training). It was only 80%
in train split dataset.
I think the L.BatchNormalization
has bug when you set chainer.using_config('train', False)
to evaluate the pretrained model.
Issue Analytics
- State:
- Created 5 years ago
- Comments:18 (8 by maintainers)
Top Results From Across the Web
chainer.links.BatchNormalization
In training mode, it normalizes the input by batch statistics. It also maintains approximated population statistics by moving averages, which can be used...
Read more >Vitis AI User Guide - Xilinx
PetaLinux SDK and Cross compiler tool chain ... Xilinx FPGA devices and evaluation boards supported by the Vitis AI development kit v2.5.
Read more >d2l-en.pdf - Dive into Deep Learning
set), which is held out for evaluation. At the end of the day, we typically report how our models perform on both partitions....
Read more >Deep Learning with PyTorch
11.8 Evaluating the model: Getting 99.7% correct means we're done, right? 308 ... Caffe, Chainer, DyNet, Torch (the Lua-based precursor to PyTorch), MXNet,....
Read more >Dive into Deep Learning
and the test data (which is held out for evaluation), reporting the ... The use of the chain rule (also known as backpropagation)...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
By the way, bad experiment results are not other people’s faults. Please be polite to people who are helping you. It’s your own responsibility to do good research, understand and analyze the experiment results. At least you should read the paper I linked carefully and probably discuss or cite it in your paper.
Deadline is not the reason to be rude.
I raised another issue for this point #5277. As for questions, please post on StackOverflow with chainer tag.