Go behind the screen of the Large Language Models like ChatGPT, Claude, Gemini and CoPilot to understand how this technology actually works!
This video walks you through creating your own Generative Pretrained Transformer (GPT) language model. First we explore the background science, then we will look at tokens, the atomic units of these large language models. Finally we will move on to data preparation, training a model and ultimately, generating text with the model.
Class Jupyter Notebook:
* github.com/jas...
Script to generate text:
* github.com/jas...
Other References:
* Standford Univesity - CS231n: Deep Learning for Computer Vision - cs231n.github....
* Visual Transformer, Explained - poloclub.githu...
* Cousera: Generative AI with LLMs - www.coursera.o...
* Attention is All You Need by Vaswani et al. in 2017 - arxiv.org/abs/...
* The Illustrated Transformer by Jay Alammar - jalammar.githu...
* Visualizing Attention, a Transformer's Heart - www.3blue1brow...
* Let's build GPT: from scratch, in code, spelled out. - by Andrej Karpathy - • Let's build GPT: from ...
* nanoGPT by Andrej Karpathy - github.com/kar...
* OpenAI GPT-2 - github.com/ope...
17 сен 2024