Backpropagation in Multilayer Perceptrons
POLYTECHNIC UNIVERSITY Department of Computer and Information Science
Backpropagation in Multilayer Perceptrons
K. Ming Leung Abstract: A training algorithm for multilayer perceptrons known as backpropagation is discussed.
Directory
? Table of Contents ? Begin Article
Copyright c 2008 mleung@poly.edu Last Revision Date: March 3, 2008
Table of Contents
1. Introduction
2. Multilayer Perceptron
3. Backpropagation Algorithm
4. Variations of the Basic Backpropagation Algorithm 4.1. Modified Target Values 4.2. Other Transfer Functions 4.3. Momentum 4.4. Batch Updating 4.5. Variable Learning Rates 4.6. Adaptive Slope
5. Multilayer NN as Universal Approximations
Section 1: Introduction
3
1. Introduction Single-layer networks are capable of solving only linearly separable classification problems. Researches were aware of this limitation and have proposed multilayer networks to overcome this. However they were not able to generalize their training algorithms to these multilayer networks until the thesis work of Werbos in 1974. Unfortunately this work was not known to the neural network community until after it was rediscovered independently by a number of people in the middle 1980s. The training algorithm, now known as backpropagation (BP), is a generalization of the Delta (or LMS) rule for single layer perceptron to include differentiable transfer function in multilayer networks. BP is currently the most widely used NN.
2. Multilayer Perceptron
We want to consider a rather general NN consisting of L layers (of course not counting the input layer). Let us consider an arbitrary layer, say , which has N neurons X1( ), X2( ), . . . , XN( ), each with a transfer function f ( ). Notice that the transfer function may be dif-
Toc
Back
Doc Doc
Section 2: Multilayer Perceptron
4
ferent from layer to layer. As in the extended Delta rule, the transfer function may be given by any differentiable function, but does not need to be linear. These neurons receive signals from the neurons in the preceding layer, - 1. For example, neuron Xj( ) receives a signal from Xi( -1) with a weight factor wi(j). Therefore we have an N -1 by N weight matrix, W( ), whose elements are given by Wi(j ), for i = 1, 2, . . . , N -1 and j = 1, 2, . . . , N . Neuron Xj( ) also has a bias given by b(j ), and its activation is a(j ).
To simplify the notation, we will use n(j )(= yin,j) to denote the net input into neuron Xj( ). It is given by
N -1
n(j ) =
a(i -1)wi(j) + b(j ),
i=1
j = 1, 2, . . . , N .
Toc
Back
Doc Doc
Section 2: Multilayer Perceptron
5
a(10)
X1(0)
X1(1)
X1( -1)
X1( )
X1(L-1)
X1(L)
a(1L)
Xi( -1)
Wi(j )
Xj( )
a(N00)
XN(00)
XN(11)
XN(
-1)
-1
XN( )
X (L-1)
NL-1
XN(LL)
a(NLL)
Figure 1: A general multilayer feedforward neural network.
Toc
Back
Doc Doc
................
................
In order to avoid copyright disputes, this page is only a partial summary.
To fulfill the demand for quickly locating and searching documents.
It is intelligent file search solution for home and business.
Related download
- module 3 artificial neural networks
- multi layer networks and backpropagation algorithm
- questions bank
- derivation of backpropagation
- a tutorial on backward propagation through time bptt in
- my attempt to understand the backpropagation algorithm for
- introduction to multi layer feed forward neural networks
- understanding belief propagation and its generalizations
- stock price prediction using back propagation neural
- backpropagation in multilayer perceptrons
Related searches
- backpropagation algorithm pdf
- neural network backpropagation code
- backpropagation learning algorithm
- backpropagation algorithm neural network
- backpropagation algorithm code
- python backpropagation code
- backpropagation algorithm python
- neural network backpropagation explained
- neural networks backpropagation tutorial
- neural network backpropagation algorithm
- how backpropagation works
- backpropagation in deep learning