question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Rasa end-to-end test does not evaluate the NLU model on the provided end-to-end stories

See original GitHub issue

Rasa version: 1.2.5

Python version: 3.6

Operating system (windows, osx, …): Ubuntu (through WSL)

Issue: When testing an end-to-end story with rasa test and the --e2e flag, the NLU evaluation is done on the normal training data instead of on the data in my end-to-end story files. The Core evaluation works as expected on my end-to-end stories.

The documentation reads:

Rasa lets you evaluate dialogues end-to-end, running through test conversations and making sure that both NLU and Core make correct predictions.

To do this, you need some stories in the end-to-end format, which includes both the NLU output and the original text.

Therefore, my expectation was to have:

  • The Core evaluation creating the failed_stories.md and story_confmat.pdf for the stories in data_test
  • The NLU evaluation creating the confmat.png, errors.json, and hist.png only for the sentences that are in the data_test/happy.md file. Right now, it evaluates the NLU model on the data in the data/nlu.md file.

FYI, our folder structure is:

data
|_ stories.md
|_ nlu.md
data_test
|_ happy.md
|_ etc.

Please note that I do have the test data in end-to-end format (which we got using the /conversations/{{sender_id}}/story endpoint). Here is a small sample of such an conversation:

## 2373134909417533
* greeting: Hello
    - utter_greet

...

* form: affirm: Yes it does
    - form: form_apartment_interest
    - slot{"apt_interested": true}
    - slot{"requested_slot": "user_email"}
* explain: why do you need this?
    - action_explain
    - action_listen
    - form_apartment_interest
    - slot{"requested_slot": "user_email"}

Did I misunderstand how the end-to-end evaluation works?

Command or request that led to error:

rasa test --stories data_test --e2e --out results

I also tried the following:

rasa test --stories data_test --nlu data_test --e2e --out results

In this case, nothing is returned for the NLU evalutation.

Issue Analytics

  • State:closed
  • Created 4 years ago
  • Comments:6 (3 by maintainers)

github_iconTop GitHub Comments

1reaction
sara-taggercommented, Sep 3, 2019

Thanks for raising this issue, @JustinaPetr will get back to you about it soon✨

Please also check out the docs and the forum in case your issue was raised there too 🤗
0reactions
stale[bot]commented, Dec 12, 2019

This issue has been automatically closed due to inactivity. Please create a new issue if you need more help.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Testing Your Assistant - Rasa
Rasa lets you validate and test dialogues end-to-end by running through test stories. In addition, you can also test the dialogue management ...
Read more >
End-to-End Natural Language Understanding Pipeline for Bangla ...
In this paper, we propose a novel approach to build a Bangla chatbot aimed to be used as a business assistant which can...
Read more >
End-to-End Natural Language Understanding Pipeline ... - arXiv
In this paper, we propose a novel approach to build a Bangla chatbot aimed to be used as a business assistant which can...
Read more >
Rasa OpenSource & Vonage Application integration and automation ...
The story represents the whole conversation in end-to-end format. This can be posted to the '/test/stories' endpoint and used as a test. Vonage...
Read more >
Building an end-to-end Conversational Assistant with Rasa
Rasa is an open-source Conversational AI framework. What I like about Rasa is you are not tied to a pre-built model or use...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found