You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Understood, the -sfp models are 8 bit weights, but I understand people are interested in more aggressive quantization.
BTW for just decreasing the memory footprint there was a commit that makes the kv cache preallocation smaller + configurable 129e66a - but I get aggressive quantization benefits go beyond that.
Working on a list of priorities + call-for-contributions, will post more soon.
FYI we do support an experimental 4.5 bit quantization method (NUQ), but those weights are not available on Kaggle.
We can more easily support this once we are able to ingest other weight formats (#11).
An update on this, we do have the ability to import from pytorch weights. Work is still ongoing on evaluating the nonuniform 4.5-bit format.
I'm increasingly concerned about uniform integer quantization in the style of k quants. Recent work such as https://arxiv.org/pdf/2407.03211 points out that human raters detect much more harm than automated metrics, especially in non-English languages, even for int8. Another paper also reports concerns after human evals, apparently also with int8.
It would be awesome if the repo supported quantization methods.
Reference: k-quants
The text was updated successfully, but these errors were encountered: