question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Inference for pinned model keeps loading

See original GitHub issue

I have pinned the enterprise model: ligolab/DxRoberta This model is pinned for instant start. When I try the Inference API fill-mask task, it responds instantly However, when I try API call to embedding pipeline https://api-inference.huggingface.co/pipeline/feature-extraction/ligolab/DxRoberta , I keep getting the message: {'error': 'Model ligolab/DxRoberta is currently loading', 'estimated_time': 20} and status does not change with time. API call to embedding pipeline was working yesterday when I tested it.

Issue Analytics

  • State:closed
  • Created 2 years ago
  • Comments:8 (4 by maintainers)

github_iconTop GitHub Comments

1reaction
Kvitcommented, May 28, 2021

@Narsil , could you chare snippet of using pipeline_tag in the card? I don’t recall seeing this option it in the documentation https://github.com/huggingface/model_card.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Error executing pinned inference model - Hugging Face Forums
{"error":"Could not load model shaxpir/prosecraft_resumed_ft2 with any of the following classes: (<class 'transformers.models.gptj.modeling_gptj ...
Read more >
How to configure optimized machine learning inference using ...
For Pinned, choose True. Under Additional Parameters, for Read access to /sys directory, choose Enabled. Add your Neo-optimized model ...
Read more >
[13th][TF] MCSAI - No Tiling Model - Inference - Kaggle
PYVIPS INSTALL CODE ONLY WORKS PINNED TO ORIGINAL ENVIRONMENT (2020) ... (not implemented yet) DO_TTA = False N_TTA = 4 # Model loading...
Read more >
how we accelerate inference using ONNX Runtime | Vespa Blog
The file refers to an ONNX model somewhere in the application package. Inputs to the model, while not actually shown here for brevity,...
Read more >
torch.utils.data — PyTorch 1.13 documentation
single- and multi-process data loading,. automatic memory pinning. These options are configured by the constructor arguments of a DataLoader , which has ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found