Тёмный

LLM Chronicles #5.3: Fine-tuning DistilBERT for Sentiment Analysis (Lab) 

Donato Capitella
Подписаться 7 тыс.
Просмотров 734
50% 1

In this lab we'll see how to fine-tune DistilBERT for analyzing the sentiment of restaurant reviews. We'll look at how to do this from scratch, adding the specific layers for classification by hand. We'll conclude by looking at how to use the HuggingFace Transformers library for this.
🖹 Lab Notebook: colab.research...
🕤 Timestamps:
- 00:50 - Knowledge Distillation
- 02:15 - Restaurant Review Dataset
- 04:44 - PyTorch Dataset and Dataloader
- 09:40 - Fine-tuning DistilBERT for Classification Tasks
- 14:29 - Evaluation
- 18:30 - Freezing Weights of Base Model
- 20:00 - Using Hugging Face Transformers for Fine-Tuning (Trainer class)
References:
- "DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter", arxiv.org/abs/...
- "HuggingFace Transformers", huggingface.co...
- "Knowledge distillation", www.analyticsv...

Наука

Опубликовано:

 

29 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 5   
@sameersayyad6170
@sameersayyad6170 2 месяца назад
One day you'll get your flowers sir!
@sameersayyad6170
@sameersayyad6170 2 месяца назад
The best!❤
@Sluo1947
@Sluo1947 3 месяца назад
Love this!! Thank you
@donatocapitella
@donatocapitella 3 месяца назад
Glad you enjoyed it!
@user-kl6xq5we4x
@user-kl6xq5we4x 5 месяцев назад
A huge Love and big Thank You ! from Pakistan. I like your name Donato Capitella.
Далее
LoRA & QLoRA Fine-tuning Explained In-Depth
14:39
Просмотров 37 тыс.
Oxxxymiron, ooes - журавли
00:19
Просмотров 57 тыс.
SIGMA ENVY IS UNTOUCHABLE 🔥 #insideout2
00:10
Просмотров 4,3 млн
Has Generative AI Already Peaked? - Computerphile
12:48
Волшебный ластик на iPhone
0:36
Просмотров 136 тыс.
It's time for change, it's time for Linux.
10:53
Просмотров 517 тыс.