🐛 [BART] Pipeline OOM
See original GitHub issue🐛 Bug
I try to run BART model myself versus running the model through pipeline
.
Running the BART model myself is fine, but I have OOM on my GPU if I try to run the same model through pipeline.
Please see the following code : https://gist.github.com/Colanim/4fae6ab52c05716062a0f20c4a6b9737
(It assume you have a file cnndm/test.source
with an article on each line)
Run with :
python pipeline_oom.py --model HuggingFace --batch-size 32
(Should not produce OOM on 11G-GPU)
and python pipeline_oom.py --model Pipeline --batch-size 32
(Should produce OOM on 11G-GPU)
Why the pipeline use more memory ?
Issue Analytics
- State:
- Created 3 years ago
- Comments:5 (4 by maintainers)
Top Results From Across the Web
Pipelines - Hugging Face
The pipelines are a great and easy way to use models for inference. These pipelines are objects that abstract most of the complex...
Read more >A pipeline for large raw text preprocessing and model training ...
A recent work based on BART, mBART [41], applies the same model and training procedure of. BART but with multilingual (but not parallel)...
Read more >CUDA error: device-side assert triggered - BART model ...
My code was working fine when I used for another encoder-decoder model (T5), but with bart I am getting this error:
Read more >Package List — Spack 0.20.0.dev0 documentation
bart, perl-extutils-helpers, py-win-unicode-console ... epics-snmp, py-azureml-pipeline-core, r-hardhat ... fastdfs, py-boom-boot-manager, r-kknn.
Read more >pipeline hangs after executing sh step command - Jenkins Jira
JENKINS-45873Subprocess exited with an OOM, pipeline then zombied · Bug. JENKINS-45616Multi-branch pipelines do not interpolate platform ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
OK I figured out the problem Long articles are not getting truncated anymore by pipeline. Will have a look. If you look at the second val.source example it’s 1583 tokens, and pipeline does not truncated it, whereas
Huggingface
does.Related: #4236
Yes I can replicate, sorry for the slow response. I am still trying to figure out why this is happening.