In this video series, you will learn how to train and fine-tune Llama 3 model from scratch.
The goal is to code LLaMA 3 from scratch in PyTorch to create models with sizes 3B, 6B, 22B, 45B, 35B and 45BM params. In this second video, you'll learn about continous pretraining, LLM benchmarks and you'll also get to see the results.
🤖 Models:
Llama-3-6B-v0.1: huggingface.co...
Llama-3-6B-v0.1 adapters: huggingface.co...
Llama-3-6B-v0 (Untrained): huggingface.co...
📚Papers:
LoRA: Low-Rank Adaptation of Large Language Models: arxiv.org/abs/...
QLoRA: Efficient Finetuning of Quantized LLMs
: arxiv.org/abs/...
💻 To follow along you can use this colab notebook:
github.com/Bla...
🎥 Coding Llama 3 from scratch video series
Part 1: • Coding Llama 3 from sc...
15 окт 2024