Question: Does NERModel support sliding_window?
See original GitHub issueHello, I was wondering how does the NERModel deal with long documents. Specifically can we provide the sliding_window
argument to deal with this?
Alternatively, are there any best practices to handle long documents that have not been implemented into this library?
Thanks!
Issue Analytics
- State:
- Created 4 years ago
- Comments:8 (3 by maintainers)
Top Results From Across the Web
Long Sequence for NER Tagging #490 - deepset-ai/FARM
Question Hello, Currently, I have developed a NER Tagging model with BERT using FARM version 0.4.5 When I run the model in Inference...
Read more >Sliding window for long text in BERT for Question Answering
Say I have a text "In June 2017 Kaggle announced that it passed 1 million registered users". Given some stride and max_len ,...
Read more >Classification Models - Simple Transformers
The ClassificationModel class is used for all text classification ... You may use any of these models provided the model_type is supported.
Read more >simpletransformers 0.9.0 - PyPI
An easy-to-use wrapper library for the Transformers library.
Read more >Sliding Window Approach to Time Series Modelling / Training
This is unusual to me, as I am familiar with the training process being used to derive a single set of model parameters....
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
That makes sense…thanks a lot 😄
Because NER is a token level task, I assumed that it would be more sensitive. If a word gets split in the middle, then it certainly changes the assigned NER tag (it would also create two tags for the two pieces of the word). If a sentence gets split, that might potentially change the tags for each word in the sentence as the meaning of the sentence is likely to change (or stop being meaningful altogether).
For classification, this should be less of an issue as classification generally depends on the sequence as a whole.