Shine a spotlight into the deep learning “black box”. This comprehensive and detailed guide reveals the mathematical and architectural concepts behind deep learning models, so you can customize, maintain, and explain them more effectively.
Inside Math and Architectures of Deep Learning you will
Math, theory, and programming principles side by side Linear algebra, vector calculus and multivariate statistics for deep learning The structure of neural networks Implementing deep learning architectures with Python and PyTorch Troubleshooting underperforming models Working code samples in downloadable Jupyter notebooks The mathematical paradigms behind deep learning models typically begin as hard-to-read academic papers that leave engineers in the dark about how those models actually function. Math and Architectures of Deep Learning bridges the gap between theory and practice, laying out the math of deep learning side by side with practical implementations in Python and PyTorch. Written by deep learning expert Krishnendu Chaudhury, you’ll peer inside the “black box” to understand how your code is working, and learn to comprehend cutting-edge research you can turn into practical applications.
Foreword by Prith Banerjee.
About the technology
Discover what’s going on inside the black box! To work with deep learning you’ll have to choose the right model, train it, preprocess your data, evaluate performance and accuracy, and deal with uncertainty and variability in the outputs of a deployed solution. This book takes you systematically through the core mathematical concepts you’ll need as a working data vector calculus, linear algebra, and Bayesian inference, all from a deep learning perspective.
About the book
Math and Architectures of Deep Learning teaches the math, theory, and programming principles of deep learning models laid out side by side, and then puts them into practice with well-annotated Python code. You’ll progress from algebra, calculus, and statistics all the way to state-of-the-art DL architectures taken from the latest research.
What's inside
The core design principles of neural networks Implementing deep learning with Python and PyTorch Regularizing and optimizing underperforming models About the reader
Readers need to know Python and the basics of algebra and calculus.
About the author
Krishnendu Chaudhury is co-founder and CTO of the AI startup Drishti Technologies. He previously spent a decade each at Google and Adobe.
Table of Contents
1 An overview of machine learning and deep learning 2 Vectors, matrices, and tensors in machine learning 3 Classifiers and vector calculus 4 Linear algebraic tools in machine learning 5 Probability distributions in machine learning 6 Bayesian tools for machine learning 7 Function How neural networks model the world 8 Training neural Forward propagation and backpropagation 9 Loss, optimization, and regularization 10 Convolutions in neural networks 11 Neural networks
A good, summarized reference of all the maths behind Machine Learning techniques. It is not the most thorough but is a great refresher for vectors calculus, probability, Bayesian inference, etc. - and how all of these come together in ML in general and Deep Learning in particular. I found the "math" part more lucid than the "architecture" part - Ian Goodfellow's book explains the Deep Learning architecture much better.