Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batch inference time #15

Open
omidb opened this issue Feb 6, 2024 · 1 comment
Open

Batch inference time #15

omidb opened this issue Feb 6, 2024 · 1 comment

Comments

@omidb
Copy link

omidb commented Feb 6, 2024

Hi,

We have successfully created batch version of the model using onnx and trt. We are trying this on a A10 GPU, here is what we have observed: for a batch of 16 we get 96ms inference time and if we run that in non-batch mode we are getting 224ms for 16 images.

I wanted to check these numbers with you and see if they make sense. Also, do you have a batch implementation that we can compare against?

@jaybdub
Copy link
Contributor

jaybdub commented Feb 9, 2024

Hi @omidb ,

Thanks for reaching out.

I haven't tested a batch implementation yet, but would let you know if that changes.

Best,
John

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants