More on the model support #733
Replies: 2 comments
-
yes, CodeLlama sample configurations are here https://github.com/OpenAccess-AI-Collective/axolotl/tree/main/examples/code-llama |
Beta Was this translation helpful? Give feedback.
-
@winglian, thanks for the answer, it helps. However, the GPU memory requirements of the smaller model (both for fine-tuning and inference) are well written. However, I do not see the same for bigger models like 34B or 70B clearly. For example consider this:
Here, it is saying that to use a 34 B model, I would require 40 +GB of VRAM. Is this for inference (only) or fine-tuning too? Also, does it also handle the multi-GPU training (I Believe yes, because I saw the deepspeed configs)? |
Beta Was this translation helpful? Give feedback.
-
So, I have these two models to fine-tune
CodeLlama derives the same architecture as of Llama. And Mistral is fairly new in the community. So, my question is can I use axolotl to fine-tune these two above models?
Beta Was this translation helpful? Give feedback.
All reactions