Skip to content

Latest commit

 

History

History
13 lines (13 loc) · 1.63 KB

NEWS_OR_UPDATE.md

File metadata and controls

13 lines (13 loc) · 1.63 KB

News or Update

  • 2023-06-05 - (Update) - Integrate with 🤗 peft to use gptq quantized model to train adapters, support LoRA, AdaLoRA, AdaptionPrompt, etc.
  • 2023-05-30 - (Update) - support download/upload quantized model from/to 🤗 Hub.
  • 2023-05-27 - (Update) - Support quantization and inference for gpt_bigcode, codegen and RefineWeb/RefineWebModel(falcon) model types.
  • 2023-05-04 - (Update) - Support using faster cuda kernel when not desc_act or group_size == -1
  • 2023-04-29 - (Update) - Support loading quantized model from arbitrary quantize_config and model_basename.
  • 2023-04-28 - (Update) - Support CPU offload and quantize/inference on multiple devices, support gpt2 type models.
  • 2023-04-26 - (Update) - Using triton to speed up inference is now supported.
  • 2023-04-25 - (News&Update) - MOSS is an open-source tool-augmented conversational language model from Fudan University, quantization is now supported in AutoGPTQ.
  • 2023-04-23 - (Update) - Support evaluation on multiple (down-stream) tasks such as: language-modeling, text-classification, text-summarization.
  • 2023-04-22 - (News) - qwopqwop200's AutoGPTQ-triton provides faster speed to integrate with quantized model, for everyone who can access to triton, try and enjoy yourself!
  • 2023-04-20 - (News) - AutoGPTQ is automatically compatible with Stability-AI's newly released gpt_neox type model family StableLM.
  • 2023-04-16 - (Update) - Support quantization and inference for bloom, gpt_neox, gptj, llama and opt.