question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Evaluate model performance when using chainer.links.BatchNormalization togather with chainer.config.train=False bug report

See original GitHub issue

chainer version: 3.4.0 cupy version: 2.4.0

When I am using resnet-101 to classify images. The resnet-101 model have L.BatchNormalization layer. After I trained over, I load model parameter and use chainer.using_config('train', False) to evaluate the model performance. I am surprised that even I am using train dataset(not validation dataset), the accuracy is lower than When I was in training procedure observation ( it was 99% in the final iteration of training). It was only 80% in train split dataset. I think the L.BatchNormalization has bug when you set chainer.using_config('train', False) to evaluate the pretrained model.

Issue Analytics

  • State:closed
  • Created 5 years ago
  • Comments:18 (8 by maintainers)

github_iconTop GitHub Comments

1reaction
JirenJincommented, Apr 13, 2018

By the way, bad experiment results are not other people’s faults. Please be polite to people who are helping you. It’s your own responsibility to do good research, understand and analyze the experiment results. At least you should read the paper I linked carefully and probably discuss or cite it in your paper.

Deadline is not the reason to be rude.

0reactions
kmaehashicommented, Aug 28, 2018

We discussed this topic and will document more on modes of BatchNormalization and when to use each of them.

I raised another issue for this point #5277. As for questions, please post on StackOverflow with chainer tag.

Read more comments on GitHub >

github_iconTop Results From Across the Web

chainer.links.BatchNormalization
In training mode, it normalizes the input by batch statistics. It also maintains approximated population statistics by moving averages, which can be used...
Read more >
Vitis AI User Guide - Xilinx
PetaLinux SDK and Cross compiler tool chain ... Xilinx FPGA devices and evaluation boards supported by the Vitis AI development kit v2.5.
Read more >
d2l-en.pdf - Dive into Deep Learning
set), which is held out for evaluation. At the end of the day, we typically report how our models perform on both partitions....
Read more >
Deep Learning with PyTorch
11.8 Evaluating the model: Getting 99.7% correct means we're done, right? 308 ... Caffe, Chainer, DyNet, Torch (the Lua-based precursor to PyTorch), MXNet,....
Read more >
Dive into Deep Learning
and the test data (which is held out for evaluation), reporting the ... The use of the chain rule (also known as backpropagation)...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found