Тёмный

Machine Learning Seminar : Multilevel and Domain Decomposition Methods for Training Neural Networks 

Legato Team
Подписаться 622
Просмотров 247
50% 1

Speaker: Rolf Krause (Università della Svizzera italiana, Switzerland)
Title: Multilevel and Domain Decomposition Methods for Training Neural Networks.
Time: Wednesday, 2023.10.11, 10:00 a.m. (CET)
Place: fully virtual (contact Jakub Lengiewicz to register)
Format: 30 min. presentation + 30 min. discussion
Abstract: The constantly increasing sizes of DNNs, measured in terms of parameters and training data available, is putting high demands on the training process, as the minimization process itself and the identification of suitable hyper-parameters are becoming more and more time consuming. A natural demand at this point is to devise scalable, parallel, and efficient training methodologies. We will present non-linear domain decomposition methods and multi-level techniques, which allow for scalability, convergence control, and automatic identification of training related hyper-parameters. We discuss the presented methods and demonstrate their convergence and scalability properties through benchmark problems.
Prof. Rolf Krause holds a chair of advanced scientific computing in the faculty of informatics and is the director of the institute of computational science (ICS) at the Università della Svizzera italiana. He is also the Co-director of the Center for Computational Medicine in Cardiology (CCMC) at USI. His research focuses on numerical simulation, machine learning, optimization, and data driven approaches. The complexity of real world applications constitutes a challenge for model and data based prediction, turning the development of models and of solution methods into a challenging task. In addition to a well balanced combination of methodological and mathematical knowledge, it also requires experience in dealing with subtle aspects of the implementation.
The Machine Learning Seminar is a regular weekly seminar series aiming to harbour presentations of fundamental and methodological advances in data science and machine learning as well as to discuss application areas presented by domain specialists. The uniqueness of the seminar series lies in its attempt to extract common denominators between domain areas and to challenge existing methodologies. The focus is thus on theory and applications to a wide range of domains, including Computational Physics and Engineering, Computational Biology and Life Sciences, Computational Behavioural and Social Sciences. More information about the ML Seminar, together with video recordings from past meetings you will find here: www.jlengineer.eu/ml-seminar/
Featured playlist

Хобби

Опубликовано:

 

12 ноя 2023

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 1   
@mathisallyouneed6187
@mathisallyouneed6187 7 месяцев назад
Looks like block coordinate descent, where block corresponds to weights of the sub-problems defined by the speaker. Moreover, quasi newton can still be applied.
Далее
Survive 100 Days In Nuclear Bunker, Win $500,000
32:21
Это iPhone 16
00:52
Просмотров 630 тыс.
Why Does Diffusion Work Better than Auto-Regression?
20:18
The moment we stopped understanding AI [AlexNet]
17:38
Просмотров 854 тыс.
Transformer Neural Networks Derived from Scratch
18:08
Просмотров 132 тыс.
делали такой маникюр??😁
1:00
Просмотров 4,7 млн
А вы бы как с ним поступили?
0:22