question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Accessing the softmax output of previous RNN state

See original GitHub issue

The recurrence formula in my RNN is h(t) = tanh(W.x(t) + U.h(t-1) + V.O(t-1) + b), where O(t-1) is the classification output (Softmax output) of the RNN at time t-1. How can I access this output at the next time step? I guess I have to write a custom RNN, but I am a bit confused by all the recurrent functions in recurrent.py in Keras and I am not sure which parts should be modified.

Issue Analytics

  • State:closed
  • Created 7 years ago
  • Comments:22 (10 by maintainers)

github_iconTop GitHub Comments

1reaction
monaj07commented, Nov 1, 2016

Does it mean that the number of classes of the output and input_dim must be the same when we want to have readout=True?

0reactions
stale[bot]commented, May 23, 2017

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs, but feel free to re-open it if needed.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Accessing the softmax output of previous RNN state in Keras
The recurrence formula in my RNN is h(t) = tanh(W.x(t) + U.h(t-1) + V.O(t-1) + b) , where O(t-1) is the classification output...
Read more >
Recurrent Neural Networks (RNNs) - Towards Data Science
We can then apply the softmax operation as a post-processing step to obtain a vector ŷ of normalized probabilities over the output. The...
Read more >
9.4. Recurrent Neural Networks - Dive into Deep Learning
Recurrent neural networks (RNNs) are neural networks with hidden states. ... use the hidden layer output of the previous time step in the...
Read more >
Building a Recurrent Neural Network - Step by Step - v1
Each cell takes as input the hidden state from the previous cell (a⟨t−1⟩) and the current time-step's input data (x⟨t⟩). It outputs a...
Read more >
Recurrent Neural Network Guide: a Deep Dive in RNN
The RNNs predict the output from the last hidden state along with output parameter Wy. Prediction is more of a classification task, where...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found