Тёмный

Jörg Tiedemann: Releasing the MAMMOTH - a framework for large-scale modular multilingual NLP models 

Finnish Center for Artificial Intelligence FCAI
Подписаться 2 тыс.
Просмотров 115
50% 1

Abstract: Neural language models have been grown in size and importance over the past years. We address two challenging aspects in the field of NLP: The support of a wide variety of languages and the runtime efficiency of such models. We focus on encoder-decoder models and modular architectures that balance between task-specific components and parameter sharing. In particular, we want to achieve effective cross-lingual transfer learning while keeping language-specific modules that can operate independently. The latter is important for efficient inference reducing computational costs and energy consumption at runtime, a crucial task for modern NLP.
There are several ways of implementing multilingual NLP systems but little consensus as to whether different approaches exhibit similar effects. Are the trends that we observe when adding more languages the same as those we observe when sharing more parameters? MAMMOTH (github.com/Helsinki-NLP/Mammoth) is a flexible framework for training various types of modular architectures making it possible to systematically compare different approaches.
Special care is taken to optimize the scalability in multinode training on large HPC clusters such as LUMI. I will report the current stage of our research including initial results, our efforts on hyper-parameter tuning, the optimization of modular architectures, scalability benchmarks and the final goal of training a large-scale multilingual translation model with massively parallel data sets.
Speaker: Jörg Tiedemann works as professor of language technology at the Department of Digital Humanities at the University of Helsinki. His main research interest is in cross-lingual NLP and machine translation.

Наука

Опубликовано:

 

26 ноя 2023

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии    
Далее
The moment we stopped understanding AI [AlexNet]
17:38
Просмотров 810 тыс.
Do you think that ChatGPT can reason?
1:42:28
Просмотров 25 тыс.
Why Fine Tuning is Dead w/Emmanuel Ameisen
50:07
Просмотров 26 тыс.
The Future of Knowledge Assistants: Jerry Liu
16:55
Просмотров 47 тыс.
LoRA & QLoRA Fine-tuning Explained In-Depth
14:39
Просмотров 33 тыс.
Data Scientist vs. AI Engineer
10:39
Просмотров 157 тыс.