Тёмный

Understanding the Llama 3 Tokenizer | Llama for Developers 

AI at Meta
Подписаться 20 тыс.
Просмотров 7 тыс.
50% 1

Опубликовано:

 

3 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 15   
@loabrasumente2283
@loabrasumente2283 3 месяца назад
TLDR - from llama 2 to llama3 they switched from sentencepiece to tiktoken - vocab size 32k -> 128k - ~15% fewer tokens for english, ~50% fewer for "some other languages"
@prasannakumar7035
@prasannakumar7035 2 месяца назад
adding to the list compression ration played a vital role in reducing the token length
@parvesh-rana
@parvesh-rana 3 месяца назад
Aston please explain the attention mechanism , Actually I am stuck in the chapter "Attention and transformer" of your book d2l
@therobotocracy
@therobotocracy 28 дней назад
My instinct is tokenization is underestimate in importance. Usually the hard and boring fundamentals are where the magic happens. Across all fields.
@anirbansen7132
@anirbansen7132 2 месяца назад
Informative
@kaushilkundalia2197
@kaushilkundalia2197 4 дня назад
Is the Llama 3 paper out yet? He mentions it @ 24:02
@AIatMeta
@AIatMeta День назад
Yes! You can read the Llama 3 research paper here: ai.meta.com/research/publications/the-llama-3-herd-of-models/
@stephennfernandes
@stephennfernandes 2 месяца назад
could someone from the meta LLaMa 3 team please explain how to train my very own tiktoken tokenizer like you guys did for llama 3. there is no opensource steps to recreate this
@Sashvinth
@Sashvinth 2 месяца назад
You develop Tamil language for Tamil users
@HamedSoheili-q4r
@HamedSoheili-q4r 2 месяца назад
so this guy is payied to use open sourced tiktoken
@maksymkyiv1111
@maksymkyiv1111 3 месяца назад
ok.
@Windowsmakes
@Windowsmakes 2 месяца назад
x
@inteist
@inteist Месяц назад
Classic example of a provably smart guy not being able to express his thoughts... 5 minutes of pain is all I managed to force myself to watch. A shame.
@user-wr4yl7tx3w
@user-wr4yl7tx3w 3 месяца назад
i don't think this format works unless the intent is to discuss at a high level.
Далее
How Many Twins Can You Spot?
00:17
Просмотров 11 млн
`const` was a mistake
31:50
Просмотров 136 тыс.