DeepMind's AlphaGo Zero algorithm beat the best Go player in the world by training entirely by self-play. It played against itself repeatedly, getting better over time with no human gameplay input. AlphaGo Zero was a remarkable moment in AI history, a moment that will always be remembered. Move 37 in particular is worthy of many philosophical debates. You'll see what I mean and get a technical overview of its neural components (code + animations) in this video. Enjoy!
Code for this video:
github.com/Zet...
Please Subscribe! And like. And comment. That's what keeps me going.
Want more education? Connect with me here:
Twitter: / sirajraval
instagram: / sirajraval
Facebook: / sirajology
There are 2 errors in this video:
1. At the top of the residual network, it says value layer twice. One should say 'policy' layer.
2 The residual network is 40 layers, i say 20.
This video is apart of my Machine Learning Journey course:
github.com/llS...
More Learning Resources:
deepmind.com/b...
/ alphago-zero-explained...
hackernoon.com...
web.stanford.e...
tim.hibal.org/b...
www.jessicayung...
Join us in the Wizards Slack channel:
wizards.herokua...
Sign up for the next course at The School of AI:
www.theschool.ai
And please support me on Patreon:
www.patreon.co...
#AlphaGoZero #Deepmind #SirajRaval
Signup for my newsletter for exciting updates in the field of AI:
goo.gl/FZzJ5w
Hit the Join button above to sign up to become a member of my channel for access to exclusive content! Join my AI community: chatgptschool.io/ Sign up for my AI Sports betting Bot, WagerGPT! (500 spots available):
www.wagergpt.co
19 сен 2024