Skip to content

Latest commit

 

History

History
5 lines (3 loc) · 670 Bytes

File metadata and controls

5 lines (3 loc) · 670 Bytes

Highlights

This project demonstrates the process of fine-tuning the Starcoder2-3B model, a code-generating LLM, on proprietary code, which we could imagine as a company's internal codebase, to better align with internal coding standards and leverage specialized libraries. Given the substantial size of these models, traditional fine-tuning approaches can be excessively demanding on computational resources. However, we'll introduce techniques to effectively fine-tune these models on just a single GPU using QLoRA, PEFT, and Bits and Bytes, ensuring a more practical approach for resource-limited environments.