Тёмный

BERT vs GPT 

CodeEmporium
Подписаться 121 тыс.
Просмотров 65 тыс.
50% 1

#machinelearning #shorts #deeplearning #chatgpt #neuralnetwork #datascience

Опубликовано:

 

18 янв 2023

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 23   
@darshantank554
@darshantank554 Год назад
One is for Natural language understanding and another is for Natural language generation
@Ram_jagat
@Ram_jagat Месяц назад
true
@borntodoit8744
@borntodoit8744 11 дней назад
Remember it as INPUT>MODEL >OUTPUT MODEL INPUT (NLU) - txt recognition, vision recognition/image/movie, sound recognition/voice, MODEL OUTPUT (NLG) -txt generation, image/movie generation, sound/voice generation + tool integration MODEL PROCESSING -Basic (classification summeration extraction) advanced (reasoning, planning, orchestration)
@VarunTulsian
@VarunTulsian Год назад
this is very useful. Just wanted to add that the gpt decoder doesn't have the cross attention in the transformer block.
@Tech_kenya
@Tech_kenya 10 месяцев назад
What is cross attention
@methylphosphatePOET
@methylphosphatePOET 9 месяцев назад
@@Tech_kenya It's when word vectors reference other word vectors as opposed to just referencing themselves.
@imran7TW
@imran7TW 17 дней назад
@@methylphosphatePOET so kinda the opposite of self attention?
@cs101-qm2ud
@cs101-qm2ud 5 месяцев назад
Wonderfully put.
@CodeEmporium
@CodeEmporium 5 месяцев назад
Thanks a lot :)
@JillRhoads
@JillRhoads 7 месяцев назад
I hadnt known that BERT was an acronym and had been wondering why the Sweden LLM was called Bert. I wonder if this is why. Thanks for the info!
@maninzn
@maninzn Год назад
Great explanation. For eg, if I have to read all the client emails and understand their requirements and auto create tasks based on that prediction, which model should I go for? BERT or GPT?
@vladislavkorecky618
@vladislavkorecky618 11 месяцев назад
What if I stack both encoders and decoders? Do I get some BERTGPT hybrid?
@davronsherbaev9133
@davronsherbaev9133 10 месяцев назад
there is also Whiser model, that has similar facebook BART decoder part, but has audio decoder.
@nicholaszustak6299
@nicholaszustak6299 10 месяцев назад
So BERT doesn’t have a decoder? Did I misunderstand
@contactdi8426
@contactdi8426 3 месяца назад
Can you please explain about their training process?
@Dr_Larken
@Dr_Larken Год назад
Bert also Drives a trans am! Jokes aside I do appreciate your videos!
@saimadhaviyalla5682
@saimadhaviyalla5682 9 месяцев назад
transformer models are usually parallelly run right?
@eugeneku3239
@eugeneku3239 8 месяцев назад
Not when it's decoding. No.
@hubgit9556
@hubgit9556 11 месяцев назад
good
@usama57926
@usama57926 Год назад
I love you ❤
@obieda_ananbeh
@obieda_ananbeh Год назад
Awesome 👏
@CodeEmporium
@CodeEmporium Год назад
Thanks So much
Далее
Watching Neural Networks Learn
25:28
Просмотров 1,1 млн
Embeddings - EXPLAINED!
12:58
Просмотров 4,8 тыс.
How ChatGPT Works Technically | ChatGPT Architecture
7:54
AI Agents explained in 2 minutes!
1:53
Просмотров 30
What are Transformer Models and how do they work?
44:26
Informer: Time Series Attention Architecture
19:23
Просмотров 1,2 тыс.
Let's build GPT: from scratch, in code, spelled out.
1:56:20
GPT function calling in a nutshell
15:36
Просмотров 35 тыс.