-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inconsistency in Padding-Free Benchmarks with Different Transformers Versions #70
Comments
DescriptionWe are again observing some variation with paddingfree in benchmarks for #78 comparing against published benches, there are some instances the improvement in training speed is significantly lesser. Using FLAN (6000 samples)For FLAN, we are expecting both pretokenized and untokenized to be similar to the currently published values but notice current runs for both pretokenized and untokenized datasets show lesser improvement than before.
Orca-Math (2000 samples)For OrcaMath, we also notice different degrees of improvements between pretokenized and untokenized datasets than before.
Update: new benchmarks with larger 8000 sample subset of ORCA pushed to #78 with slightly more consistent values, but PF improvement of FOAK still shows quite minmal improvement. |
|
Description
We observe no improvement with PaddingFree on QLoRA and GPTQ-LoRA when running benchmarks on OrcaMath.
However,
Mistral7B (OrcaMath) with
transformers==4.42.4
Mistral7B (FLAN) with
transformers==4.42.4
NOTE: There are some variability between the transformers versions, when transformers is upgraded
Mistral7B (OrcaMath) with
transformers==4.44.0
The text was updated successfully, but these errors were encountered: