Notes on Deep Learning Theory Заметки ШАД (англ.)





Заметки по Deep Learning ШАД.

Evgenii (Eugene) Golikov

Neural Networks and Deep Learning lab. Moscow Institute of Physics and Technology Moscow, Russia [email protected]

December 11, 2020

These are the notes for the lectures that I was giving during Fall 2020 at the Moscow Institute of Physics and Tech- nology (MIPT) and at the Yandex School of Data Analysis (YSDA). The notes cover some aspects of initialization, loss landscape, generalization, and a neural tangent kernel theory. While many other topics (e.g. expressivity, a mean-field theory, a double descent phenomenon) are missing in the current version, we plan to add them in future revisions.


Комментариев пока нет. Вы можете стать первым!  
Добавить комментарий