1. Getting Started
Humans have drawn inspiration from nature and surroundings. I feel the greatest true source of our excellence has been our capacity to learn and draw various rules and inferences from our habits, traits, behaviours. The fascination of humans to invent something that can think, behave and do things almost as quickly as us, I feel that’s the back ground from where the artificial neural network (ANN) draws its inspiration from.
I reckon the attempt of machine learning and AI is to understand perhaps how a robot can learn from us and walk on the streets, or pack and deliver goods, how a driverless car can ride us home while abiding the traffic rules almost as we do. Neural network opens doors of interesting possibilities for humans. Deep learning draws its inspiration from the biology of the functioning of a human brain.
I am no biologist yet let me attempt to give you a biological back ground of ANN.
Human brains under a microscope divulge a well-constructed network, below image, can be considered a pictorial representation of the same. It’s a network of neurons all well connected, transferring signals neuron to neuron for every decision we make or action we humans take.
Over 100,000 people subscribe to our newsletter.
See stories of Analytics and AI in your inbox.
If we were to zoom into it, we would observe each neuron connection looks like below. Each neuron has a bunch of short little tentacles extending in various directions called dendrites. It also has long tentacles called axon. The axon ends on branches called synapse which in turn hooks up to the dendrites of the next neuron.
Our brain literally consists of billions and billions of networks of neurons.
How do they function? These dendrites happen to be sensitive to a certain kind of chemicals (pardon my knowledge of biology). When there are a certain variety of chemical signals around these dendrites, they absorb them. When a certain amount of chemical is accumulated at the cell and it gets overwhelmed (let me call it crossing the threshold), it fires. It translates to generating an electrical pulse all the way down the tail – axon, into the synapse and these are picked up as a signal by the dendrites of the next cell. So essentially we have a system which looks as to how many chemicals I have around me and if it crosses my threshold I shoot. Boom! These are then picked up by the next cell.
The 2 interesting components we find:
- Accumulator (∑)
- Trigger (∫)
I am not a biologist, so I seek an apology in advance if I may have trampled basic biology while attempting to be able to back my mathematical models.
2. Revisiting Logistic Regression
Logistic decision making will not seem unknown to the folks already from ML/AI field of study. Mathematician David Cox formulated the famous logit function, where the probability of occurrence or dichotomous categorical outcome was represented as a linear combination of a function.
The function has the following distribution and is also coined sigmoid function due to its (∫) shape.
We must already be familiar with the “Titanic” dataset for predicting “survivors” based on predictors
Essentially this forms the fundamental of logistic regression.
We get inputs from 6 predictors into what is called an accumulator, the β’s let’s call them weights W with some bias. The accumulator function calls for summation oft = β0 + β1Class + β2Gender + β3Age + β4NoOfTravellingSiblings + β5NoOfTravellingParents + β6EmbarkedFrom
∑WnXn + B
The accumulator ∑ spits out a probability number, the sigmoid function (∫) based on the threshold decides to call the output as “Survived” or “Will not survive”. For instance setting a threshold or 0.5 or 50%, when the summation function spits any value greater than 0.5 the sigmoid function calls it results as “Will survive” the titanic disaster. Essentially we can say the sigmoid is triggered only once the threshold of 0.5 is reached by the inputs.
We can draw a clear correlation with the explanation of the human brain cell. Just to reiterate the dendrites absorb the electrical pulse and keep accumulating them until the threshold is reached. An electrical pulse is triggered once the threshold is reached. This forms the basis of the intuition for the neural networks. The decision boundary of the logistic model is a single line which is divided by the threshold point into “Survived” and “Not survived” classes.
For the function above if b were to be a –ive real number the threshold point would be such that w.x weight times the input is a number greater than bias b. According to our intuition so far this will result in the electrical pulse to be triggered and result in 1 as output.
3. Perceptron And History
Neural networks are quite an interesting area of machine learning, simultaneously one of the oldest and newest areas of research. Late in 1940 researchers tried to model the human brain. We have revisited the logistic regression to be able to infer the intuition of perceptron. Basing perceptron researchers solved a number of problems. But soon it turned out that the perceptron was limited in capabilities to be able to solve only certain kinds of problems (linear classification problems to be specific.). It couldn’t solve simple things like solving a XOR problem. Soon after that the research in neural networks basically died. With a lot of hype around it, turned out it couldn’t solve basic and simple XOR problems.
They were re-born, yes in 1980’s researchers figured out that putting multiple perceptron together into a network could solve complex classification & prediction problems. The breakthrough came with the development of backward propagation, symmetric networks. This led to a great deal of excitement. These models were now able to solve all kinds of learning problems but the learning needed time or compute power. The computing power with only as much in the period, neural networks were unable to evolve and flourish, to say the least.
5 years back with cloud computing, big data and GPUs of processing power making a revolutionary entry into the technological world. Neural networks made a re-entry the 2nd time. It’s now the back bone of all the research for some of the very complex ML problems right from the driverless car, speech and image recognition.
4. Artificial Neural Network Model
The basic unit of ANN is a perceptron or single neuron.
The output is just the squashed function of the summation of inputs. It would literally mean the linear classifier or the output from the logistic regression. Or conversely, the logistic classifier is a neural network with 1 neuron.
But when we begin to arrange a layer of neurons together into a network we get a neural net. Especially when we have the inputs feeding into several neurons and they form inputs in turn to neurons in one direction we get a feed forward neural network.
They are used for classification or regressions. It’s called feed forward because the information from 1 layer of units is passed to the next layer of units and then they compute and pass on until the output layer. The output classifies or regresses the class label or value respectively based on the weights and biases at each layer neurons. The layers if more than one between input and output conventionally are referred as the deep hidden layer.
Solving AND with Neural Network –
The NN has quite a representational power, NN with a single layer we can represent any bounded continuous function to quite a degree of accuracy. And to start with a slightly simpler Boolean function we can represent virtually any Boolean function into a neural net with 100 % accuracy.
The AND Boolean function case for 2 inputs can be solved with a hidden layer of 4 nodes. We can represent an input value of 0 as GREEN and 1 as RED. The network with the Weight vectors of W1, W2, W3 and W4 on to output Y will optimize the solution for AND gate.
And this network can not only solve for AND but for any Boolean function. Now that’s quite a statement and power of NN. The hidden unit here encode all possible combinations for 1s and 0s and solves for the desired function.
5. XOR GATE Problem – Solved by Neural Network
XOR gate is a digital logic gate that sets the signal as high only when 1 of the inputs are high. Compared against an OR gate XOR is also called as TRUE OR.
A linear classifier can’t solve this. To visualize this on a graph, it would look something like below. Now a straight line cannot classify one from the other class label
There is no way to solve this with a single neuron, however with 2 layers neurons we can solve this problem.
On layer 2 interestingly we redefine our axis and plot the outputs from layer 1 neuron 1 and the output from layer 1 Neuron 2. Interestingly we began with solving a 4 point classification now leaves us with only 3 points on the board. As (1, 0) are common outputs from both 1 & 2 neurons.
At the second layer, the classification becomes easy as the linear classifier will now be able to separate them into 2 classes.
6. Backward Propagation
Back propagation can be considered as an optimization problem. Basically, it entails what the hidden units (neurons) should look like to achieve the results closely explaining and/or learning from the training data. The figure briefly explains the back propagated neural network with hidden layer and nodes. It’s an online algorithm which basically starts with a set of weights for each unit and receives the inputs and the output data point at a time and then computes the values for each node on the hidden layers. It results in an output Y, this output may be quite different to the Y* (the output in the training dataset). Hence Y-Y* would be the Error ἐ. In this each case it’s just running the sigmoid and forward feed.
ἐ Error Y-Y* on the output will now be pushed back to adjust or optimize for the backward propagation algorithm, to come as close possible to the true label or output. This acts as a synthetic training example for each layer backwards.
Derivative solves this optimization problem. Basically, ∂ἐ/∂k, where k is the node on a hidden layer, tells whether the weight needs to be higher or lower to the current. On each layer and node, based on the error ἐ, the derivative of the error w.r.t the weight tells us whether the value needs to be increased or decreased and overall this just keeps relaying until we solve over iterations and get an optimized value as close to the output of the training data as possible. The final weights determine the model and perform the predictions.
For complex problems, the training time could be quite lengthy and GPUs of compute may be required to solve the massive scale problems.
7. Multiclass Neural Network
A striking question arises, how do we solve the multi-class prediction problems? Well, in that case, let’s pick up the famous IRIS dataset. Just to revisit we have a dataset of 3 species of IRIS flowers namely “Sentosa”, “Virginica”, “Versicolor” and their sepal length and width and petal length and width. A sample of 5 records looks as below.
Plotting the petal details on R the graph looks as below.
To be able to solve this classification problem and train a Neural Network to be able to predict species based on sepal and petal details we flatten the results.
The code to do so in R
We flatten the output by transforming the Species attribute with the output species classes. For the records with output as “Sentosa” the data is flagged as 1 and respectively is for the rest.
Next, we sample the data into training and test sample in 80:20 ratio and train neural net with 1 hidden layer of 3 nodes.
Upon training the plot looks as below.
As we observe the trained neural network has 4 inputs of the attributes and 1 hidden layer for their respective weights (indicated in black) and biases (indicated in blue). And output predictions.
Based on a set of input the NN basically predicts a set of 3 values i.e. the output species (Sentosa, Versicolor or a Virginica) chances for a set of input values. A sample predictions code and output.
For each record, the max value is considered the winner or the predicted class.
Wrapping up, with the humongous leap in the computing capabilities, neural network and deep learning open up quintessential possibilities of very interesting inventions ahead. As Theo Jansen aptly put –