Torchserve latest-gpu remote docker image does not support Metrics API
See original GitHub issueYour issue may already be reported! Please search on the issue tracker before creating one.
Context
The remote docker image pytorch/torchserve:latest-gpu fails to receive metrics API calls on port 8082. Please see the logs below where I tested:
- Deploying torchserve on EKS with the “latest-gpu” image, metrics API fails.
- Deploying torchserve on EKS with the “latest” image, metrics API succeeds.
- Running torchserve locally with the docker image “latest-gpu”, metrics API fails.
Expected Behavior
Making a request to <torchserve>:8082/metrics should work.
Current Behavior
Making the request results in an error.
Possible Solution
Steps to Reproduce
- Run a docker container using command
docker run --rm -it --gpus all -p 8080:8080 -p 8081:8081 -p 8082:8082 pytorch/torchserve:latest-gpu - Make an inference at least once
curl -X POST http://localhost:8080/predictions/squeezenet1_1 -T kitten.jpg - Query the metrics API endpoint
curl http://localhost:8082/metricsand this will fail
Failure Logs [if any]
Issue Analytics
- State:
- Created 3 years ago
- Comments:5 (1 by maintainers)
Top Results From Across the Web
Image Layer Details - pytorch/torchserve:latest-gpu | Docker Hub
Layer details are not available for this image.
Read more >17. Torchserve Use Cases - PyTorch
Docker - Make sure that MAR file is being copied in volume/directory shared while starting torchserve docker image. Start torchserve with following command ......
Read more >Getting started: Serving PyTorch predictions with a custom ...
This tutorial shows you how to use a custom container to deploy a PyTorch machine learning (ML) model that serves online predictions.
Read more >Deploying PyTorch Model to Production with FastAPI in CUDA ...
The goal is to serve a trained model as a RESTful API inside a docker container with CUDA support, on a remote cloud...
Read more >Serving PyTorch Models Using TorchServe - Supertype
Create a model handler and other additional files for the model; Generate model archive; Serve the model using TorchServe; Monitor and manage the...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found

@jeremiahschung I don’t think the latest gpu image was tagged to
pytorch/torchserve:latest-gpu. You should try withpytorch/torchserve:0.2.0-cuda10.1-cudnn7-runtime. @seemethere Can confirmhttps://hub.docker.com/r/pytorch/torchserve/tags?page=1
Validated this on latest
torchserve:latest-gpuimageClosing.