question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

🐛 [BART] Pipeline OOM

See original GitHub issue

🐛 Bug

I try to run BART model myself versus running the model through pipeline.

Running the BART model myself is fine, but I have OOM on my GPU if I try to run the same model through pipeline.

Please see the following code : https://gist.github.com/Colanim/4fae6ab52c05716062a0f20c4a6b9737

(It assume you have a file cnndm/test.source with an article on each line)

Run with : python pipeline_oom.py --model HuggingFace --batch-size 32
(Should not produce OOM on 11G-GPU)

and python pipeline_oom.py --model Pipeline --batch-size 32
(Should produce OOM on 11G-GPU)


Why the pipeline use more memory ?

@sshleifer

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Comments:5 (4 by maintainers)

github_iconTop GitHub Comments

1reaction
sshleifercommented, Jun 18, 2020

OK I figured out the problem Long articles are not getting truncated anymore by pipeline. Will have a look. If you look at the second val.source example it’s 1583 tokens, and pipeline does not truncated it, whereas Huggingface does.

Related: #4236

1reaction
sshleifercommented, Jun 18, 2020

Yes I can replicate, sorry for the slow response. I am still trying to figure out why this is happening.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Pipelines - Hugging Face
The pipelines are a great and easy way to use models for inference. These pipelines are objects that abstract most of the complex...
Read more >
A pipeline for large raw text preprocessing and model training ...
A recent work based on BART, mBART [41], applies the same model and training procedure of. BART but with multilingual (but not parallel)...
Read more >
CUDA error: device-side assert triggered - BART model ...
My code was working fine when I used for another encoder-decoder model (T5), but with bart I am getting this error:
Read more >
Package List — Spack 0.20.0.dev0 documentation
bart, perl-extutils-helpers, py-win-unicode-console ... epics-snmp, py-azureml-pipeline-core, r-hardhat ... fastdfs, py-boom-boot-manager, r-kknn.
Read more >
pipeline hangs after executing sh step command - Jenkins Jira
JENKINS-45873Subprocess exited with an OOM, pipeline then zombied · Bug. JENKINS-45616Multi-branch pipelines do not interpolate platform ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found