Biological and Artificial Neural Networks

In machine learning, the artificial neural networks are a category of algorithms that are inspired by the biological neural networks in the brain, and designed to carry out both supervised and unsupervised learning tasks, such as classification and clustering. To understand how such neural network algorithms work, we first consider some basic concepts in biological neural system.

The human brain consists of $10^{11}$ neurons interconnected through about $10^{14}$ to $10^{15}$ synaptic junctions to form millions of neural networks. Hundreds specialized cortical areas are formed based on these networks for different information processing tasks.

Functionally, a neuron consists of the following three parts:

neuron.png

The function of a neuron can be modeled mathematically. Each neuron, modeled as a node in the neural network, receives input signal or stimulus from $n$ neurons and its activation or the net input is the weighted sum of all such inputs:

$\displaystyle a=\sum_{j=1}^d w_j x_j+b$ (1)

where $b$ is the offset or bias, $x_j$ is the input signal from the jth node, $w_j$ is the synaptic connectivity to the jth input node:

$\displaystyle w_j\;\left\{ \begin{array}{ll}
> 0 & \mbox{excitatory input} \\
< 0 & \mbox{inhibitory input} \\
= 0 & \mbox{no connection}
\end{array} \right.$ (2)

Same as in the case of linear regression, we define $x_0=1$ and $w_0=b$, so that both the weight and pattern vectors are augmented to become ${\bf x}=[x_0=1,\,x_1,\cdots,x_d]^T$ and ${\bf w}=[w_0=b,\,,w_1,\cdots,w_d]^T$, and Eq. (1) above can now be conveniently written as

$\displaystyle a=\sum_{j=1}^d w_j x_j+b=\sum_{j=0}^d w_j x_j={\bf w}^T{\bf x}$ (3)

The output signal or response of the neuron is a function of its activation:

$\displaystyle y=g(a)=g\left(\sum_{j=0}^d w_j x_j+b\right)=g({\bf w}^T{\bf x})$ (4)

neuronModel.png

Here $g(x)$ is an activation function, which typically take one of the following forms:

ReLU.png

The function of a neural network can be modeled mathematically as a hierarchical structure shown below containing multiple layers of neurons, called nodes in the context of artificial neural networks:

The purpose is to train the network according to certain mathematical rules, the learning rules or learning laws, by modifying the weights of a network iteratively based on the inputs (and the desired outputs if the learning is supervised), so that given the input of the network as the stimulus, the network will produce the desired output as the response.

threelayernet.gif

The learning paradigms of the neural networks are listed below, depending on the interpretations of the input and output of the neural network.

  1. Pattern Associator

    This is the most general form of neural networks that learns and stores the associative relationship between two sets of patterns represented by vectors.

    Human memory is associative in the sense that given one pattern, some associated pattern(s) may be produced. Examples include: (Evolution, Darwin), (Einstein, $E=mc^2$), (food, sounding bell, salivation).

  2. Auto-associator

    As a special pattern associator, auto-associator associates a prestored pattern to an incomplete or noisy version of the pattern.

  3. Regression

    This is another special kind of pattern associator which takes a vector input ${\bf x}\in{\cal R}^d$ and produces a real value $y\in{\cal R}$ as a multivariable function $y=f({\bf x})$ at its only output node.

  4. Classification

    This is a variation of the pattern associator of which the output patterns are a set of categorical symbols representing different classes $\{C_1,\cdots,C_K\}$, i.e., each input pattern is classified by the network into one of the classes

    $\displaystyle f: {\bf x} \in {\cal R}^d \Longrightarrow y \in \{C_1,\cdots,C_K\}$ (10)

  5. Regularity Detector

    This is an unsupervised learning process. The network discovers automatically the regularity in the inputs so that similar patterns are automatically detected and grouped together in the same cluster or class.