Not able to provide <perf-analyzer-flags> 'shape' for perf_analyzer in config.yaml, results in termination
See original GitHub issueHi,
I have added according to perf-analyzer-flags two flags to the perf_analyzer
as shown below since I have a model with dynamic input shapes. --shape images:3,640,640
works when used with perf_analyzer
separately.
But adding to my config.yaml:
perf_analyzer_flags: percentile: 95 shape: 'images:3,640,640'
or:
perf_analyzer_flags: percentile: 95 shape: images:3,640,640
Results in a termination:
ERROR[perf_analyzer.py:164] perf_analyzer was terminated by signal: SIGABRT
But when I remove the shape flag, this error is gone, but I still need to provide input shape since:
failed with exit status 1 : error: failed to create concurrency manager: input images contain a dynamic shape, provide shapes to send along with the request
So I am not sure how to add images:3,640,640
to the shape
flag.
Issue Analytics
- State:
- Created 2 years ago
- Comments:10 (5 by maintainers)
Top GitHub Comments
@aramesh7 I think this log provides a lot of information about what goes wrong. It looks like the
--shape
is parsed wrong, as you can see in the log below:Which should be:
images:3,640,640
It works perfectly when using performance analyzer directly with the following arguments:
perf_analyzer -m yolov5x_onnx_dynamic -b 1 -u localhost:8000 -i http --concurrency-range=2 --percentile=95 --shape images:3,640,640
Steps to reproduce:
Start a triton server:
docker run --gpus=all --rm -p8000:8000 -p8001:8001 -p8002:8002 -v /model_repository:/models nvcr.io/nvidia/tritonserver:21.09-py3 tritonserver --model-repository=/models --model-control-mode=explicit
Start client docker:
docker run -it --gpus all --net=host nvcr.io/nvidia/tritonserver:21.09-py3-sdk
Use the following config file:
model-analyzer -v profile -f config.yaml
this issue has been resolved with https://github.com/triton-inference-server/model_analyzer/pull/268