mlatoz

ANN Intuition

Plan Of Attack

What we will learn in this section:


The Neuron


The Neuron
The Neuron


Additional Reading

Efficient BackProp

By Yann LeCun et al. (1998)

Efficient BackProp

Efficient BackProp - Yann LeCun


The Activation Function

Threshold Function

Sigmoid Function

where
x = Sum of weights

Rectifier Function

Hyperbolic Tangent Function


Additional Reading

Deep Sparse Rectifier Neural Networks

By Xavier Glorot et al. (2011)

Deep Sparse Rectifier Neural Networks

Deep Sparse Rectifier Neural Networks - Xavier Glorot


How do Neural Networks Learn?

   where
     y^ = y-cap


Additional Reading

A List of Cost Functions used in Neural Networks, Alongside Applications

CrossValidated (2015)

A List of Cost Functions used in Neural Networks

A List of Cost Functions used in Neural Networks, Alongside Applications - CrossValidated


Gradient Descent

      wt = wt-1 - a . dwt

where
  w = Weight Vector
  dw = Gradient of w
  a = Learning Rate
  t = Iteration Number


Stochastic Gradient Descent

where
  w(t) represents the model parameters (weights and biases)
  at iteration t.

  η is the learning rate (a hyperparameter that controls the step size).

  ∇J(w(t), x(i), y(i)) is the gradient of the loss function J,
  evaluated at the current parameters w(t) using a randomly selected
  data point (x(i), y(i)).


Additional Reading

A Neural Network in 11 Lines of Python (Part 1)

Andrew Trask (2015)

A Neural Network in 11 Lines of Python

A Neural Network in 11 Lines of Python - Andrew Trask


Additional Reading

A Neural Network in 13 Lines of Python (Part 2 - Gradient Descent)

Andrew Trask (2015)

A Neural Network in 13 Lines of Python

A Neural Network in 13 Lines of Python - Andrew Trask


Backpropagation


Additional Reading

Neural Networks and Deep Learning

Michael Nielson (2015)

A Neural Network in 13 Lines of Python

Neural Networks and Deep Learning - Michael Nielson


Training the ANN with Stochastic Gradient Descent


«Previous