Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Do optimum-benchmark support deepspeed_mii #288

Open
neozhang307 opened this issue Oct 10, 2024 · 2 comments
Open

Do optimum-benchmark support deepspeed_mii #288

neozhang307 opened this issue Oct 10, 2024 · 2 comments

Comments

@neozhang307
Copy link

Thank you very much for your efforts in this work.

I found that you support deepspeed. But I think for inference workload, deepspeed-mii has better performance, as claimed in this repo: https://github.com/microsoft/DeepSpeed-MII.

I am curious about its performance. I wonder how hard it is to support this backend?

@IlyasMoutawwakil
Copy link
Member

nope we don't but you can submit a PR following the example of in #250

@neozhang307
Copy link
Author

Let me see if I can implement it myself. But I don't know where to start.
The basic API for mii is:

import mii
pipe = mii.pipeline("mistralai/Mistral-7B-v0.1")
response = pipe(["DeepSpeed is", "Seattle is"], max_new_tokens=128)
print(response)

Can you point out some backend code that I can start from?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants