**The Matrix Calculus You Need For Deep Learning**

by Terence Parr, Jeremy Howard

**Publisher**: arXiv.org 2018**Number of pages**: 33

**Description**:

This paper is an attempt to explain all the matrix calculus you need in order to understand the training of deep neural networks. We assume no math knowledge beyond what you learned in calculus 1, and provide links to help you refresh the necessary math where needed.

Download or read it online for free here:

**Download link**

(740KB, PDF)

## Similar books

**Neural Networks and Deep Learning**

by

**Michael Nielsen**

Neural networks and deep learning currently provide the best solutions to many problems in image recognition, speech recognition, and natural language processing. This book will teach you the core concepts behind neural networks and deep learning.

(

**9746**views)

**Deep Learning**

by

**Yoshua Bengio, Ian Goodfellow, Aaron Courville**-

**MIT Press**

This book can be useful for the university students learning about machine learning and the practitioners of machine learning, artificial intelligence, data-mining and data science aiming to better understand and take advantage of deep learning.

(

**16619**views)

**Deep Learning Tutorial**

by

**LISA lab**-

**University of Montreal**

This book will introduce you to some of the most important deep learning algorithms and show you how to run them using Theano. Theano is a python library that makes writing deep learning models easy, and gives the option of training them on a GPU.

(

**8040**views)

**Deep Learning: Technical Introduction**

by

**Thomas Epelbaum**-

**arXiv.org**

This note presents in a technical though hopefully pedagogical way the three most common forms of neural network architectures: Feedforward, Convolutional and Recurrent. For each network, their fundamental building blocks are detailed.

(

**5372**views)