Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CPU vs GPU #100

Closed
Selbyl opened this issue Mar 28, 2023 · 7 comments
Closed

CPU vs GPU #100

Selbyl opened this issue Mar 28, 2023 · 7 comments

Comments

@Selbyl
Copy link

Selbyl commented Mar 28, 2023

Is there an option to change this platform to use CUDA or ROCM instead of CPU?

@gamersalpha
Copy link

for now no

i waiting to ...

he is tooo slowly with CPU

@Ionaut
Copy link

Ionaut commented Mar 30, 2023

Yeah, CUDA gpu support would be great. That and leaving the model in the vram for the duration of the conversation.

@SurvivaLlama
Copy link

So.. Slow for everyone?

@gaby
Copy link
Member

gaby commented Apr 2, 2023

This is a llama.cpp issue, not Serge. May be worth creating this issue on that repo.

@lee-b
Copy link

lee-b commented Apr 10, 2023

llama.cpp is specifically intended by be a CPU-only implementation. They've said before that they won't support GPU. Someone would have to submit a patch so good that they're willing to accept it anyway, or else I suppose this project would have to find a similar, but GPU-based version of it.

@gaby gaby closed this as completed Apr 10, 2023
@jpaodev
Copy link

jpaodev commented Feb 27, 2024

@lee-b @gaby since llama.cpp does support GPU offloading and solutions such as CLBlast or ROCm now, is this up for debate? Thanks in advance!

@gaby
Copy link
Member

gaby commented Feb 27, 2024

@jpaodev In progress #944

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants