Тёмный

Understanding Coordinate Descent 

Machine Learning TV
Подписаться 37 тыс.
Просмотров 7 тыс.
50% 1

Course link: www.coursera.o...
let's just have a little aside on the coordinate decent algorithm, and then we're gonna describe how to apply coordinate descent to solving our lasso objective. So, our goal here is to minimize sub function g. So, this is the same objective that we have whether we are talking about our closed form solution, gradient descent, or this coordinate descent algorithm. But, let me just be very explicit, where. We're saying we wanna minimize over all possible w some g(w), where here, we're assuming g(w) is function of multiple variables. Let's call it g(w0,w1,...,wD). So this W we are trying to write in some bold font here. And often, minimizing over a large set of variables can be a very challenging problem. But in contrast, often it's possible to think about optimizing just a single dimension, keeping all of the other dimensions fixed. So easy for each coordinate when keeping others fixed, because that turns into just a 1D optimization problem. And so, that's the motivation behind coordinate decent, where the coordinate descent algorithm, it's really intuitive.

Опубликовано:

 

8 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 6   
@homeycheese1
@homeycheese1 2 месяца назад
will coordinate descent always converge using LASSO even if the ratio of number of features to number of observations/samples is large?
@lara6893
@lara6893 Год назад
Emily and Carlos rock, heck yeah!!
@pnachtwey
@pnachtwey 3 месяца назад
This works ok on nice functions like g(x,y)=x^2+y^2 but real data often looks more like Grand Canyon where the path is very narrow and very windy.
@NayeliGC
@NayeliGC Год назад
Should'nt that be argmin w instead of just min? since we want to return one of the arguments?
@jurohd
@jurohd Год назад
I think the left arrow
@polinba
@polinba Год назад
Min and argmin are just different ways to name it, in this case they are similar
Далее
Solve any equation using gradient descent
9:05
Просмотров 53 тыс.
AdaBoost, Clearly Explained
20:54
Просмотров 753 тыс.
Mark Rober vs Dude Perfect- Ultimate Robot Battle
19:00
▼ЮТУБ ВСЁ, Я НА ЗАВОД 🚧⛔
30:49
Просмотров 421 тыс.
Unigrama, Bigrama, Trigrama e Ngrama
10:59
Просмотров 429
Proximal Gradient Descent Algorithms
14:58
Просмотров 12 тыс.
Gradient Boosting : Data Science's Silver Bullet
15:48
5.1 Proximal and Projected Gradient Descent
35:04
Просмотров 19 тыс.
Understanding Metropolis-Hastings algorithm
9:49
Просмотров 70 тыс.
The Most Important Algorithm in Machine Learning
40:08
Просмотров 411 тыс.
CoordinateDescent
5:00
Просмотров 4,8 тыс.