Full-text tutorial (requires MLExpert Pro): www.mlexpert.io/bootcamp/fine-tuning-llama-3-llm-for-rag What performance did you get with your fine-tuned model?
Thank you in have watch several videos on llama fine tuning With lots of differences you are the best ! Especially anout dataset and how you formatted it
I’m trying to fine-tune the LLaMA 3.1 8B model without quantization, but when I try to use the pipeline with the unquantized model, I encounter this error: Trying to set a tensor of shape torch.Size([128256, 4096]) in 'weight' (which has shape torch.Size([128264, 4096])), this looks incorrect. Do you know why this is happening? maybe i'm using the wrong pad token idk. Thanks in advance
If we apply the same approach to a dataset with another language like France or Arabic, will the approach change? Assume same columns structure, and names like you showed in the video.