question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Unable to run TensorRTIS on AKS

See original GitHub issue

I am trying to run TensortRTIS on AKS but get the following failure:

$ kubectl logs trt
trtserver: error while loading shared libraries: libnvidia-ml.so.1: cannot open shared object file: No such file or directory

I tried installing NVIDIA drivers, but it didnt help… (just running the simple-string example for CPU only… not running GPU) https://docs.microsoft.com/en-us/azure/aks/gpu-cluster

My podspec is:

apiVersion: v1
kind: Pod
metadata:
  name: trt
  namespace: default
spec:
  containers:
  - args:
    - --model-store=gs://rakelkar1/trt_sample
    - --allow-poll-model-repository=false
    - --allow-grpc=true
    - --allow-http=true
    - --grpc-port=9000
    - --rest-port=8080
    command:
    - trtserver
    env:
    - name: PORT
      value: "8080"
    image: nvcr.io/nvidia/tensorrtserver:19.05-py3
    imagePullPolicy: IfNotPresent
    name: user-container
    ports:
    - containerPort: 8080
      name: user-port
      protocol: TCP

Issue Analytics

  • State:closed
  • Created 4 years ago
  • Comments:5 (1 by maintainers)

github_iconTop GitHub Comments

2reactions
deadeyegoodwincommented, Jun 14, 2019

TRTIS does support CPUs and should also work on servers that don’t have GPUs or CUDA installed. https://docs.nvidia.com/deeplearning/sdk/tensorrt-inference-server-master-branch-guide/docs/run.html#running-the-inference-server-on-a-system-without-a-gpu

When using kubernetes you need to use only “args” in your spec, like this: https://github.com/NVIDIA/tensorrt-inference-server/blob/master/deploy/single_server/templates/deployment.yaml#L59

Did you try the example Helm chart from the repo to see if that works for you? https://github.com/NVIDIA/tensorrt-inference-server/tree/master/deploy/single_server

1reaction
rakelkarcommented, Jun 16, 2019

works when trtserver is specified in args instead of command - thanks 😃

Read more comments on GitHub >

github_iconTop Results From Across the Web

Unable to run TensorRTIS on AKS · Issue #378 - GitHub
I am trying to run TensortRTIS on AKS but get the following failure: $ kubectl logs trt trtserver: error while loading shared libraries: ......
Read more >
AI & Data Science - NVIDIA Developer Forums
hello guys: I follow the instructions (Installation Guide :: NVIDIA Deep Learning TensorRT Documentation) to install the tensorRT step by step, ...
Read more >
Inference Optimization with NVIDIA TensorRT - YouTube
This tutorial will introduce NVIDIA TensorRT,... ... Your browser can't play this video. ... Inference Optimization with NVIDIA TensorRT.
Read more >
8-bit Inference with TensorRT
TensorRT will: ○ Run inference in FP32 on calibration dataset. ○ Collect required statistics. ○ Run calibration algorithm → optimal ...
Read more >
Unknown desc = failed to fetch anonymous token: unexpected ...
Make sure you login to the right container registry. It's in the name of the docker image you are using. e.g. nvcr.io/nvidia/tensorrt:22.01-py3 ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found