Network does not converge, bad captions
See original GitHub issueHello,
I’ve followed your instructions and started training the network. The loss reaches its minimum value after about 5 epochs and then it starts to diverge again.
After 50 epochs, the generated captions of the best epoch (5th or 6th) look like this:
Predicting for image: 992
2351479551_e8820a1ff3.jpg : exercise lamb Fourth headphones facing pasta soft her soft her soft her soft her soft her dads college soft her dads college soft her her her her her soft her her her her her soft her her her her
Predicting for image: 993
3514179514_cbc3371b92.jpg : fist graffitti soft her soft her Hollywood Fourth Crowd soft her her soft her her her her her soft her her her her her her soft her her her her soft her her her her soft her her her
Predicting for image: 994
1119015538_e8e796281e.jpg : closeout security soft her soft her security fall soft her her her her her fall soft her her her her her her soft her her her her her soft her her her her soft her her her her her
Predicting for image: 995
3727752439_907795603b.jpg : roots college Fourth tree-filled o swing-set places soft her soft her her soft her her soft her her college soft her her her her her her her soft her her her her soft her her her her her her
Any idea what’s wrong?
Issue Analytics
- State:
- Created 6 years ago
- Comments:34 (2 by maintainers)
Top Results From Across the Web
Neural network does not converge with negative symbols
I've created a simple 2-2-1 feedforward ANN to predict an XOR using Keras. The activation function I'm using on all layers is a...
Read more >When Your Network Fails to Converge
Poor selection of weights or bad processing of inputs could leave your network vulnerable to not being able to find a solution. While...
Read more >One neural network, many uses
Let's see what the network captions my own photo: Not bad! It's impressive that the network knows that the image contains a man...
Read more >Optimization for deep learning: theory and algorithms
In this section, we will describe some main tricks needed for training a neural network. 4.1 Possible Slow Convergence Due to Explosion/ ...
Read more >DropEdge: Towards Deep Graph Convolutional Networks ...
Abstract: Over-fitting and over-smoothing are two main obstacles of developing deep Graph Convolutional Networks (GCNs) for node classification.
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
It’s been a while since I worked on this repo. I’ll try to retrain it and reproduce this error sometime next week and see if something needs change.
Meanwhile, @PavlosMelissinos and @MikhailovSergei if you were able to debug this, feel free to update and send a pull request.
I am facing the same issue while using Flickr8k and the captions are not making any sense. Particular words are getting repeated in every sentence. Somehow, it is working better on a subset of 100 images rather than the entire dataset. I have tried changing the batch size but it didn’t help. Could you give any suggestions?