TrisZaska's Machine Learning Blog

Overview about Multi-layer Neural Network

1. Introduction
2. History and Overview about Artificial Neural Network
3. Single neural network
4. Multi-layer neural network
5. Install and using Multi-layer Neural Network to classify MNIST data
6. Summary
7. References

Multi-layer Neural Network

We went through many kinds of stuff about Single Neural Network, hopefully, it makes sense to you, we also already discussed generally Multi-layer Neural Network in the sub-section 2.2.2. In this section, we'll go more detail about this model to discover why Multi-layer Neural Network can solve the problem of Perceptron, so after this section you'll get deeper understanding, prepare for the final part where we actually install real Multi-layer Neural Network into real world problem, use it to classify handwritten MNIST data . So, firstly look at the image below,
Let's briefly reviews some the concepts, as we knew, Multi-layer Neural Network has 3 layers,
  • Input layer (Layer 1) used to receive information so-called features or variables from data, the number of neurons in input layer belongs to the number of features you want to feed in plus 1 bias unit.
  • Hidden layer (Layer 2) is considered to process information from input layer, we can have more than one layers with many neurons in each layer in Hidden layer (called Deep Neural Networks and it's out of scope of this papers).
  • Output layer (Layer 3) used to return the output, similarly, the number of neurons in output layer belong to what are outputs you want. Example, if your Neural Networks used for binary classification, output layer has just two neurons represented 0 or 1, and we already implemented Neural Networks for binary classification in the previous section, in the next section we will install Multi-layer for multi-classification.
There are three big steps in Multi-layer Neural Network that quite similar to Single Neural Network,
  1. Activating Multi-layer Neural Network by forwarding propagation, it means we feed the network starting from input layer, information go from left to right to generate the output.
  2. Based on the output, we calculate the error, using it to minimize the cost function we will discuss later.
  3. Update the weights using back-propagation, we discuss very detail about it later, too.
Of course, we want to optimize the weights of the net, so we must repeat these steps in multiple epochs to achieve better and better model.

No comments :

Post a Comment

Leave a Comment...