Replies: 4 comments 1 reply
-
Its a good idea. I am testing them and also looking at adding support for different embeddings. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Just saw the new accelerate update. I think it will now be possible. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Support for GPTQ models would be awesome |
Beta Was this translation helpful? Give feedback.
0 replies
-
You can use GPTQ now. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
7B HF models take more than 12Gb of memory, so is there a way to use GPTQ ones?
Beta Was this translation helpful? Give feedback.
All reactions