You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We have successfully created batch version of the model using onnx and trt. We are trying this on a A10 GPU, here is what we have observed: for a batch of 16 we get 96ms inference time and if we run that in non-batch mode we are getting 224ms for 16 images.
I wanted to check these numbers with you and see if they make sense. Also, do you have a batch implementation that we can compare against?
The text was updated successfully, but these errors were encountered:
Hi,
We have successfully created batch version of the model using onnx and trt. We are trying this on a A10 GPU, here is what we have observed: for a batch of 16 we get 96ms inference time and if we run that in non-batch mode we are getting 224ms for 16 images.
I wanted to check these numbers with you and see if they make sense. Also, do you have a batch implementation that we can compare against?
The text was updated successfully, but these errors were encountered: