Nigeria No1. Music site And Complete Entertainment portal for Music Promotion WhatsApp:- +2349077287056
Friday, 1 December 2023
Show HN: 80% faster, 50% less memory, 0% loss of accuracy Llama finetuning https://bit.ly/46KgNMO
Show HN: 80% faster, 50% less memory, 0% loss of accuracy Llama finetuning Hi HN! I'm just sharing a project I've been working on during the LLM Efficiency Challenge - you can now finetune Llama with QLoRA 5x faster than Huggingface's original implementation on your own local GPU. Some highlights: 1. Manual autograd engine - hand derived backprop steps. 2. QLoRA / LoRA 80% faster, 50% less memory. 3. All kernels written in OpenAI's Triton language. 4. 0% loss in accuracy - no approximation methods - all exact. 5. No change of hardware necessary. Supports NVIDIA GPUs since 2018+. CUDA 7.5+. 6. Flash Attention support via Xformers. 7. Supports 4bit and 16bit LoRA finetuning. 8. Train Slim Orca fully locally in 260 hours from 1301 hours (5x faster). 9. Open source version trains 5x faster or you can check out Unsloth Pro and Max codepaths for 30x faster training! https://bit.ly/46G3tcj... has more info about Unsloth! Hopefully you can try it out! Wrote a blog post at https://bit.ly/3uNphFz if you want to learn more about our manual hand derived backprop or Triton kernels and stuff! Thanks once again! https://bit.ly/46VObR7 December 1, 2023 at 03:42PM
Labels:
Hacker News
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment