Fundamental Concepts of Neural Nets: Perceptron


Neural Networks are extremely efficient tools for analyzing past and present data. They are made such that they reflect a biological network of neurons at a very fundamental level. However, even this fundamental image of our mind has done wonders with the help of computation resources like high processing power and voluminous storage systems; and with each passing day, we are getting nearer to building closer representations of the mind. The amalgamation of the human mind and external machines has the capability to not only change the world we live in, but also create new ones. Decades or even centuries of research goes into such initiatives, but here also comes the question of mass extinction before the achievement of something so close to a miracle.

Coming back to basics, the study of neural networks is called Deep Learning and it expands into several interesting sub-topics.

The origins of the neural network can be credited to the initial design of the Perceptron in the year 1957 by Frank Rosenblatt. Understanding the functioning of a Perceptron is the most basic step that must be taken to grasp the advanced concepts of deep learning. It is the simplest neural network model which very closely follows the model of Logistic Regression, only with a slight difference in the loss function. In fact, a Perceptron can be treated as a singled layered neural net.

The Intuition of a Perceptron



Image Courtesy: Wikipedia

The Perceptron is a system which takes in a set of inputs (x1 to xn) and then applies respective weights on each incoming input (w1 to wn). Thereafter, it uses a simple summation function on the weighted inputs and passes the value through an activation function. In simple terms, it is just like a calculator with only three functions:


  • -          Multiplication (Weights * Inputs)
  • -          Summation (Adding the weighted Inputs)
  • -          Activation Function (Could be any depending upon the requirement)

The activation function can be any function which processes the summed value and provides a certain outcome. For instance, a perceptron might use the Rectified Linear Unit or ReLU function. It is one of the most used functions in neural networks due to its simplicity, efficiency and ease of computation. It outputs the input directly if it is positive and otherwise it outputs zero.

The entire model of a perceptron was loosely inspired by a single biological neuron that forms the base of the complex network of the human mind.

To make it simpler, one can compare the perceptron with a logistic regression model. They are essentially the same as already mentioned, except for the loss function.

Y = sigmoid(wx + b)

This is the equation represents a logistic regression model where x is the known variable, w is the assigned weight and b is the constant. The sigmoid function is used to weigh the probability outputs either towards 0 or a 1. This is a complex function due to computations involving inverse of logit functions.

On the other hand, the Perceptron follows the exact equation, but finally, uses a very simple activation function. It is often the ReLU function which helps to speed up any neural network manifold.


Image Courtesy: towardsdatascience

Multilayer Perceptron


Researchers started wondering if it is also possible to replicate the network-like structure of the human brain. Thus, they formulated the idea that if multiple layers of neurons are inserted in a perceptron, it might start giving improved results. The concept of multi-layer perceptron was thereby created.


Image Courtesy: Wikipedia


A tremendous amount of research work went into transforming this basic idea into a success story, and a most celebrated scientist and major believer in the concept, Jeff Hinton, took the lead.

In the year 1986, Jeff Hinton and his team came up with the milestone idea of Back Propagation which turned out to be the backbone of major neural network structures that we use today.
Backpropagation is the simple idea of chain rule of differentiation, but it served as a major founding ground for robust Artificial Intelligence initiatives.


For over twenty years, the research under Jeff Hinton and his team continued to build what most of the world had little hope from. The idea that machines can replicate the human mind and perform miraculous tasks. Even during the AI Winter when the funding for neural network research dried up since there were computational limitations and people moved over to more feasible techniques like SVMs and Gradient Boosting, Hinton kept the lead on ample research work in the field of neural networks.

In 2006, he finally came up with an impactful paper which talked of deep neural networks with a lot of hidden layers. It was eminent that multiple layers improved results significantly (also demonstrated by the ImageNet competition in 2012 where Deep Neural Nets beat other results by a huge margin). This encouraged powerful companies in the technical sector to invest heavily in neural network research since they had the volume of data necessary and sufficient computation resources.

Today, neural networks have taken the field of artificial intelligence to unimaginable heights and there is still so much to do!

[ Watch this space for a detailed discussion on multi-layered perceptron. ]

Comments

Brands Worked with or Featured On

Brands Worked with or Featured On

Popular Posts