You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A default optimization that LitServe can provide users is to map the decode_request function in case of dynamic batching using a ThreadPool. This can be useful for cases like image loading which is IO based.
I did a quick test with a ResNet-152 model for image classification and observed the following throughput (Requests per second) performance gain with threadpool:
Motivation
Pitch
Alternatives
Additional context
The text was updated successfully, but these errors were encountered:
🚀 Feature
A default optimization that LitServe can provide users is to map the
decode_request
function in case of dynamic batching using a ThreadPool. This can be useful for cases like image loading which is IO based.I did a quick test with a ResNet-152 model for image classification and observed the following throughput (Requests per second) performance gain with threadpool:
Motivation
Pitch
Alternatives
Additional context
The text was updated successfully, but these errors were encountered: