This textbook establishes a theoretical framework for understanding deep learning models of practical relevance. With an approach that borrows from theoretical physics, Roberts and Yaida provide clear and pedagogical explanations of how realistic deep neural networks actually work. To make results from the theoretical forefront accessible, the authors eschew the subject’s traditional emphasis on intimidating formality without sacrificing accuracy. Straightforward and approachable, this volume balances detailed first-principle derivations of novel results with insight and intuition for theorists and practitioners alike. This self-contained textbook is ideal for students and researchers interested in artificial intelligence with minimal prerequisites of linear algebra, calculus, and informal probability theory, and it can easily fill a semester-long course on deep learning theory. For the first time, the exciting practical advances in modern artificial intelligence capabilities can be matched with a set of effective principles, providing a timeless blueprint for theoretical research in deep learning.
- Detailed step-by-step explanations for all equations and clear exposition of both old and new concepts in deep learning theory make the book accessible to readers with a minimal prerequisite of linear algebra, calculus, and informal probability theory
- Many novel results that appear for the first time in the literature, taking readers to the forefront of deep learning theory
- Provides a unique approach that bridges deep learning and theoretical physics, demonstrating to the ML community how a theoretical physics approach can be useful, while also teaching techniques that are valuable for theoretical physicists
Conditions of Use
This book is licensed under a Creative Commons License (CC BY-NC-SA). You can download the ebook The Principles of Deep Learning Theory for free.
- Title
- The Principles of Deep Learning Theory
- Subtitle
- An Effective Theory Approach to Understanding Neural Networks
- Publisher
- Cambridge University Press
- Author(s)
- Boris Hanin, Daniel A. Roberts, Sho Yaida
- Published
- 2022-05-26
- Edition
- 1
- Format
- eBook (pdf, epub, mobi)
- Pages
- 472
- Language
- English
- ISBN-10
- 1316519333
- ISBN-13
- 9781316519332
- License
- CC BY-NC-SA
- Book Homepage
- Free eBook, Errata, Code, Solutions, etc.
Preface 0 Initialization 1 Pretraining 2 Neural Networks 3 Effective Theory of Deep Linear Networks at Initialization 4 RG Flow of Preactivations 5 Effective Theory of Preactivations at Initialization 6 Bayesian Learning 7 Gradient-Based Learning 8 RG Flow of the Neural Tangent Kernel 9 Effective Theory of the NTK at Initialization 10 Kernel Learning 11 Representation Learning The End of Training Epilogue: Model Complexity from the Macroscopic Perspective A Information in Deep Learning B Residual Learning References Index