Time to get_async_run_results is very slow
See original GitHub issueHi every one, I get result by using async_run method. Time for async_run is 0.007930517196655273, but time for get_async_run_results is 0.11087822914123535. so it is very slow.
while not last_request:
input_batch = []
for idx in range(batch_size):
input_batch.append(image_data[image_idx])
image_idx = (image_idx + 1) % len(image_data)
if image_idx == 0:
last_request = True
batch_size = len(input_batch)
break
request_ids.append(ctx.async_run(
{input_name: input_batch},
{boxes: (InferContext.ResultFormat.RAW),
scores: (InferContext.ResultFormat.RAW)},
batch_size))
time_infer_stop = time.time()
print("v_detection inference time: ", time_infer_stop - time_infer_start)
time_post_start = time.time()
# For async, retrieve results according to the send order
for request_id in request_ids:
results.append(ctx.get_async_run_results(request_id, True))
time_post_stop = time.time()
print("v_detection post processing time: ", time_post_stop - time_post_start)
print("total_time_process: ", time_post_stop - time_infer_start)
how can i reduce get_async_run_results time?
Issue Analytics
- State:
- Created 4 years ago
- Comments:8 (5 by maintainers)
Top Results From Across the Web
Why is first HttpClient.GetAsync call extremely slow in my C# ...
Save this question. Show activity on this post. The first call is taking 10-15 seconds, all successive calls takes less than a second....
Read more >Why Your Computer is Running Slow & How to Fix It
In this guide, we've outlined a number of potential issues that could be slowing down your computer and have covered how to fix...
Read more >9 Best Tips to Troubleshoot Windows 7 Is Running Slow
Method 1. Generally speaking, insufficient disk space or system memory will cause your computer to slow or lag. If you open too many...
Read more >Solved: Spark transformation becomes very slow at times.
The problem is that the cluster speed is changing during the day, which means that at times the Spark transformation finishes in minutes...
Read more >GP7 - Slow startup time : r/GuitarPro - Reddit
Hey, So GP7 is taking a lot of time starting up each time (about 20 seconds) and thats when my PC is unoccupied...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
The inference time is determined by the model and the framework. Using the async API doesn’t speed up the actual inference time on the server… it just allows that client thread to do something else instead of just waiting for the response (which is what the thread does in the non-async API).
Have you looked at perf_client: https://docs.nvidia.com/deeplearning/sdk/tensorrt-inference-server-master-branch-guide/docs/client.html#performance-example-application
There is also some information in the blog post listed in the README that talks about using features of TRTIS to get better performance (mostly throughput improvements as latency is more a function of the model and framework).
Closing. Reopen if you have some more information to report.