Llama 3.1/3.2 LLM Toolbox
This toolbox already includes 16 Jupyter notebooks specially optimized for Llama 3.1 and Llama 3.2 LLMs. The logs of successful runs are also provided. More notebooks will be regularly added.
Once you've subscribed to The Kaitchup Pro or purchased access, you can also request repository access here.
To run the code in the toolbox, CUDA 12.4 and PyTorch 2.4 are recommended. PyTorch 2.5 might already work but I didn't test it yet.
Toolbox content
- Supervised Fine-Tuning with Chat Templates (6 notebooks)
- Full fine-tuning
- LoRA fine-tuning
- LoRA fine-tuning (with Llama 3.1/3.2 Instruct)
- QLoRA fine-tuning with Bitsandbytes quantization
- QLoRA fine-tuning with AutoRound quantization
- LoRA and QLoRA fine-tuning with Unsloth
- Preference Optimization (2 notebooks)
- DPO training with LoRA (TRL and Transformers)
- ORPO training with LoRA (TRL and Transformers)
- Quantization (3 notebooks)
- AWQ
- AutoRound
- GGUF for llama.cpp
- Inference (4 notebooks)
- Transformers with and without a LoRA adapter
- vLLM offline and online inference
- Ollama (not released yet)
- llama.cpp
- Merging (3 notebooks)
- Merge a LoRA adapter into the base model
- Merge a QLoRA adapter into the base model
- Merge several Llama 3.1/3.2 models into one with mergekit (not released yet)
You can find all the toolbox and more content created by The Kaitchup here:
https://newsletter.kaitchup.com/p/ai-toolboxes
Note: If you are a subscriber to The Kaitchup Pro, you already have access to the repository and all the other toolboxes. If you just subscribed, you will also receive an access token in a few hours. Contact The Kaitchup (https://newsletter.kaitchup.com/) if you don't receive it within the 24 hours following your subscription.
All the notebooks and access to the repository