Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FT] Support llama.cpp inference #402

Open
JoelNiklaus opened this issue Nov 22, 2024 · 3 comments
Open

[FT] Support llama.cpp inference #402

JoelNiklaus opened this issue Nov 22, 2024 · 3 comments
Labels
feature request New feature/request

Comments

@JoelNiklaus
Copy link
Contributor

Issue encountered

Currently, inference of open models on my Mac device is quite slow since vllm does not support mps.

Solution/Feature

Llama.cpp does support mps and would significantly speed up local evaluation of open models.

Posssible alternatives

Allowing the use of the mps device in other ways of loading models would also work.

@JoelNiklaus JoelNiklaus added the feature request New feature/request label Nov 22, 2024
@clefourrier
Copy link
Member

Hi! Feel free to open a PR for this if you need it fast as our roadmap for EOY is full :)

@JoelNiklaus
Copy link
Contributor Author

Sounds good. Might do at some point, for now it is not a priority for me.

@julien-c
Copy link
Member

would be an awesome feature IMO! cc @gary149

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature/request
Projects
None yet
Development

No branches or pull requests

3 participants