Title: Neural Networks. R
1Neural Networks. R G Chapter 8
- 8.1 Feed-Forward Neural Networks
- otherwise known as
- The Multi-layer Perceptron
- or
- The Back-Propagation Neural Network
2A diagramatic representation of a Feed-Forward NN
x1
x2
y
x3
Inputs and outputs are numeric.
3Inputs and outputs
- Must be numeric, but can have any range in
general. - However, R G prefer to consider constraining to
(0-1) range inputs and outputs.
4Neural Network Input FormatReal input data
values are standardized (scaled) so that they all
have ranges from 0 1.
5Categorical input format
- We need a way to convert categores to numberical
values. - For hair-colour we might have values red,
blond, brown, black, grey. - 3 APPROACHES
- 1. Use of (5) Dummy variables (BEST)
- Let XR1 if hair-colour red, 0 otherwise, etc
- 2. Use a binary array 3 binary inputs can
represent 8 numbers. Hence let red (0,0,0),
blond, (0,0,1), etc - However, this sets up a false associations.
- 3. VERY BAD red 0.0, blond 0.25, , grey
1.0 - Converts nominal scale into false interval scale.
6Calculating Neuron OutputThe neuron threshhold
function. The following sigmoid function, called
the standard logistic function, is often used to
model the effect of a neuron.
Consider node i, in the hidden layer. It has
inputs x1, x2, and x3, each with a
weight-parameter.
Then calculate the output from the following
function
7Note the output values are in the range
(0,1). This is fine if we want to use our output
to predict a probability of an event happening.
.
8Other output types
- If we have a categorical output with several
values, then we can use dummy output notes for
each value of the attribute. - E.g. if we were predicting one of 5 hair-colour
classes, we would have 5 output nodes, with 1
being certain yes, and 0 being certain no.. - If we have a real output variable, with values
outside the range (0-1), then another
transformation would be needed to get realistic
real outputs. Usually the inverse of the scaling
transformation. i.e.
9Training the Feed-forward net
- The performance parameters of the feed-forward
neural network are the weights. - The weights have to be varied so that the
predicted output is close to the true output
value corresponding to the inpute values. - Training of the ANN (Artificial Neural Net) is
effected by - Starting with artibrary wieghts
- Presenting the data, instance by instance
- adapting the weights according the error for
each instance. - Repeating until convergence.
108.2 Neural Network Training A Conceptual View
11Supervised Learning/Training with Feed-Forward
Networks
- Backpropagation Learning
- Calculated error of each instance is used to
ammend weights. - Least squares fitting.
- All the errors for all instances are squared and
summed (ESS). All weights are then changed to
lower the ESS. - BOTH METHODS GIVE THE SAME RESULTS.
- IGNOR THE R G GENETIC ALGORITHM STUFF.
12Unsupervised Clustering with Self-Organizing Maps
13(No Transcript)
14r
n
n
n n r(x-n)
x
Data Instance
158.3 Neural Network Explanation
- Sensitivity Analysis
- Average Member Technique
168.4 General Considerations
- What input attributes will be used to build the
network? - How will the network output be represented?
- How many hidden layers should the network
contain? - How many nodes should there be in each hidden
layer? - What condition will terminate network training?
17Neural Network Strengths
- Work well with noisy data.
- Can process numeric and categorical data.
- Appropriate for applications requiring a time
element. - Have performed well in several domains.
- Appropriate for supervised learning and
unsupervised clustering.
18Weaknesses
- Lack explanation capabilities.
- May not provide optimal solutions to problems.
- Overtraining can be a problem.
19Building Neural Networks with iDA
209.1 A Four-Step Approach for Backpropagation
Learning
- Prepare the data to be mined.
- Define the network architecture.
- Watch the network train.
- Read and interpret summary results.
21Example 1 Modeling the Exclusive-OR Function
22(No Transcript)
23(No Transcript)
24Step 1 Prepare The Data To Be Mined
25(No Transcript)
26Step 2 Define The Network Architecture
27(No Transcript)
28(No Transcript)
29Step 3 Watch The Network Train
30(No Transcript)
31Step 4 Read and Interpret Summary Results
32(No Transcript)
33(No Transcript)
34Example 2 The Satellite Image Dataset
35Step 1 Prepare The Data To Be Mined
36(No Transcript)
37Step 2 Define The Network Architecture
38(No Transcript)
39Step 3 Watch The Network Train
40Step 4 Read And Interpret Summary Results
41(No Transcript)
42(No Transcript)
439.2 A Four-Step Approach for Neural Network
Clustering
44Step 1 Prepare The Data To Be Mined
45Step 2 Define The Network Architecture
46(No Transcript)
47Step 3 Watch The Network Train
48(No Transcript)
49Step 4 Read And Interpret Summary Results
50(No Transcript)
51(No Transcript)
529.3 ESX for Neural Network Cluster Analysis