-
Notifications
You must be signed in to change notification settings - Fork 295
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add Mistral-7B-Instruct-v0.1 from huggingface. #2010
Conversation
The A10G pipeline is failing even after disabling the test there. I can't even see the logs. |
@xuzhao9 has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
@xuzhao9 - is there anything that needs a fix here? |
@pranavsharma The CPU test exceeds time limit (5 min), can you also help disable the CPU test? |
@xuzhao9 - it's still failing with OOM. |
'phi_1_5' : (512, 512, 'AutoConfig.from_pretrained("microsoft/phi-1_5", trust_remote_code=True)', 'AutoModelForCausalLM') | ||
'phi_1_5' : (512, 512, 'AutoConfig.from_pretrained("microsoft/phi-1_5", trust_remote_code=True)', 'AutoModelForCausalLM'), | ||
# as per this page https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1 trust_remote_code=True is not required | ||
'mistral_7b_instruct' : (512, 512, 'AutoConfig.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")', 'AutoModelForCausalLM') |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
reduce these numbers to avoid OOM, you're likely hitting the OOM because of how large the activations are
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Reducing it to 128 doesn't work either.
@msaroufim @xuzhao9 - how should we make progress on this? It's been pending for a while now. |
Hi @pranavsharma , after 2 runs it still OOMs on A100 40GB. We need to either 1) slice/tune the model so that it will not OOM on A100 40GB, or 2) disable the A100 test, essentially not testing this model in our CI. |
How do I disable A100 test? |
train_benchmark: false | ||
train_deterministic: false | ||
not_implemented: | ||
- device: NVIDIA A10G |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
- device: NVIDIA A10G | |
- device: NVIDIA A10G | |
- device: NVIDIA A100-SXM4-40GB |
@pranavsharma Add the device name in the |
3bfd811
to
a88756b
Compare
@xuzhao9 - does this look good? |
@@ -35,6 +35,8 @@ | |||
'llama_v2_13b' : (512,512, 'AutoConfig.from_pretrained("meta-llama/Llama-2-13b-hf")', 'AutoModelForCausalLM'), | |||
'llama_v2_70b' : (512, 512, 'AutoConfig.from_pretrained("meta-llama/Llama-2-70b-hf")', 'AutoModelForMaskedLM'), | |||
'phi_1_5' : (512, 512, 'AutoConfig.from_pretrained("microsoft/phi-1_5", trust_remote_code=True)', 'AutoModelForCausalLM'), | |||
# as per this page https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1 trust_remote_code=True is not required | |||
'mistral_7b_instruct' : (128, 128, 'AutoConfig.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")', 'AutoModelForCausalLM') |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
A trailing comma is needed
'mistral_7b_instruct' : (128, 128, 'AutoConfig.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")', 'AutoModelForCausalLM') | |
'mistral_7b_instruct' : (128, 128, 'AutoConfig.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")', 'AutoModelForCausalLM'), |
806656b
to
fd9b63f
Compare
Moved it to canary. |
@xuzhao9 has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
Add Mistral-7B-Instruct-v0.1 from huggingface. See https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1