Download 23 Sep - 04 Oct, 2002 WSC7 7

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
Canadian Weather Analysis
Using Connectionist Learning
Paradigms
Imran Maqsood*, Muhammad Riaz Khan, Ajith Abraham
*Environmental Systems Engineering Program, Faculty of Engineering, University of Regina,
Regina, Saskatchewan S4S 0A2, Canada, E-mail: [email protected]
Partner
Technologies Incorporated, 1155 Park Street, Regina, Saskatchewan S4N 4Y8,
Canada, E-mail:[email protected]
Faculty of Information Technology, School of Business Systems, Monash University,
Clayton 3800, Australia, E-mail: [email protected]
7th Online World Conference on Soft Computing in Industrial Applications (on WWW), September 23 - October 4, 2002
CONTENTS
•
•
•
•
•
Introduction
MLP, ERNN and RBFN Background
Experimental Setup of a Case Study
Test Results
Conclusions
23 Sep - 04 Oct, 2002
WSC7
2
1. INTRODUCTION
• Weather forecasts provide critical information
about future weather
• Weather forecasting remains a complex business,
due to its chaotic and unpredictable nature
• Combined with threat by the global warming and
green house gas effect, impact of extreme weather
phenomena on society is growing costly, causing
infrastructure damage, injury and the loss of life.
23 Sep - 04 Oct, 2002
WSC7
3
• Accurate weather forecast models are important to
the countries, where the entire agriculture depends
upon weather
• Previously, several artificial intelligence
techniques have been used in the past for
modeling chaotic behavior of weather
• However, several of them use simple feed-forward
neural network training methods using
backpropagation algorithm
23 Sep - 04 Oct, 2002
WSC7
4
Study Objectives
• To develop an accurate and reliable predictive
models for forecasting the weather of Vancouver,
BC, Canada.
• To compare performance of multi-layered
perception (MLP) neural networks, Elman
recurrent neural networks (ERNN) and radial basis
function network (RBFN) for the weather
analysis.
23 Sep - 04 Oct, 2002
WSC7
5
2. ANN BACKGROUND
INFORMATION
•
•
•
•
•
•
ANN Advantages
An ability to solve complex and non-linear
problems
Quick response
Self-organization
Real time operation
Fault tolerance via redundant information coding
Adaptability and generalization
23 Sep - 04 Oct, 2002
WSC7
6
(a) Multi-Layered Perceptron (MLP) Networks
• network is arranged in layers of
neurons
• every neuron in a layer computes
sum of its inputs and passes this
sum through a nonlinear function
as its output.
• Each neuron has only one output,
but this output is multiplied by a
weighting factor if it is to be used
as an input to another neuron (in a
next higher layer)
• There are no connections among
neurons in the same layer.
23 Sep - 04 Oct, 2002
WSC7
wjk
Wij
I1
O1
I2
O2



In




k



Om
Oi
Vj
Input
Layer
Hidden
Layer
Output
Layer
Figure: Architecture of 3-layered MLP network
for weather forecasting
7
(b) Elman Recurrent Neural Networks (ERNN)
• ERNN are a subclass of
recurrent networks
• They are multilayer perceptron
networks augmented with one
or more additional context
layers storing output values of
one of the layers delayed by one
step and used for activating this
or some other layer in the next
time step
• The Elman network can learn
sequences that cannot be
learned with other recurrent
neural network
23 Sep - 04 Oct, 2002
Feedback D-1
I1
I2
I3










O1
Om
In
Feedback D-1
Input Layer
Hidden Layer
Output Layer
Figure: Architecture of 3-layered ERNN
WSC7
8
(c) Radial Basis Function Network (RBFN)
• network consists of 3-layers:
input layer, hidden layer, and
output layer
• The neurons in hidden layer are
of local response to its input
and known as RBF neurons,
while the neurons of the output
layer only sum their inputs and
are called linear neurons
• network is inherently well
suited for weather prediction,
because it naturally uses
unsupervised learning to cluster
the input data.
23 Sep - 04 Oct, 2002
Output =  wi i(x)
Pure linear

W0
W1
Adjustable weights wi
w0 = bias
Wn
W2
1
I1
2



n
I2
Adjustable centers ci
Adjustable spreads i
Im
Input
Layer
Figure: Architecture of RBFN
WSC7
9
3. EXPERIMENTAL SETUP
– Minimum Temperature (oC)
– Maximum Temperature(oC)
– Wind-Speed (km/hr)
Min. temperature
20.0
10.0
Aug
Jul
Jun
May
Apr
Mar
Feb
Jan
Dec
-10.0
Nov
0.0
Oct
• Observed Parameters
(most important):
Max. temperature
Sep
Sep 2000 – Aug 2001
Daily Temperature (oC)
• Vancouver, BC, Canada
• 1-yr data:
30.0
1 3031 6061 9091 120
121 150
151 180
181 210
211 240
241 270
271 300
301 330
331 360
361
Time of the Year
100
Wind speed (km/hr)
Weather Data:
80
60
40
20
Time of the Year
23 Sep - 04 Oct, 2002
WSC7
10
Aug
Jul
Jun
May
Apr
Mar
Feb
Jan
Dec
Nov
Oct
Sep
0
Training and Testing Datasets
• Dataset 1: MLP and ERNN
– Testing dataset = 11-20 January 2001
– Training dataset = remaining data
• Dataset 2: RBFN, MLP and ERNN
– Testing dataset = 01-15 April 2001
– Training dataset = remaining data
• We used this above method to ensure that there is no bias
on the training and test datasets
23 Sep - 04 Oct, 2002
WSC7
11
Simulation System Used
• Pentium-III, 1GHz processor 256 MB RAM
• all the experiments were simulated using MATLAB
Steps taken before starting the training process:
• Error level was set to a value (10-4)
• The hidden neurons were varied (10-80) and the
optimal number for each network were then decided.
23 Sep - 04 Oct, 2002
WSC7
12
4. TEST RESULTS
Training Convergence of MLP and ERNN
1
1
10
10
0
0
10
10
-1
-1
10
Mean Square Error
Mean Square Error
10
-2
10
-3
10
Goal = 0.0004
-4
-5
-3
10
Goal = 0.0004
-4
10
10
-2
10
10
Levenberg-Marquardt
Iterations = 7
One-Step-Secant
Iterations = 1,015
-5
10
-6
Levenberg-Marquardt
Iterations = 11
One-Step-Secant
Iterations = 673
-6
10
10
Number of Iterations
Number of Iterations
Convergence of the LM and OSS training
algorithms using MLP network
23 Sep - 04 Oct, 2002
Convergence of the LM and OSS training
algorithms using ERNN
WSC7
13
Comparison of Actual vs. 10-day ahead Forecasting
using OSS and LM approaches
(a) Minimum Temperature (11-20 Jan 2001)
ERNN
MLP network
15
Minimum Temperature Forecasting
10
Actual
MLP-OSS
MLP-LM
5
0
-5
Temperature (oC)
Temperature (oC)
15
Minimum Temperature Forecasting
10
Actual
ERNN-OSS
ERNN-LM
5
0
-5
0
2
4
6
Days of the Month
Performance evaluation
parameters (min. temperature)
Mean absolute % error (MAPE)
Root mean square error (RMSE)
Mean absolute deviation (MAD)
Correlation coefficient
Training time (minutes)
23 Sep - 04 Number
Oct, 2002
of iterations (epochs)
8
10
0
2
4
6
Days of the Month
MLP Network
OSS
LM
0.0221
0.0202
OSS
0.0182
LM
0.0030
0.0199
0.7651
0.9657
0.3
WSC7
1015
0.0199
0.7231
0.9826
0.3
673
0.0031
0.1213
0.9998
7
11
0.0199
0.8411
0.9940
1
7
8
10
ERNN
14
(b) Maximum Temperature (11-20 Jan 2001)
ERNN
MLP network
15
Maximum Temperature Forecasting
Temperature (oC)
Temperature (oC)
15
10
5
0
Actual
MLP-OSS
Maximum Temperature Forecasting
10
5
0
Actual
MLP-LM
-5
ERNN-OSS
ERNN-LM
-5
0
2
4
6
Days of the Month
8
Performance evaluation
parameters
10
0
2
4
6
Days of the Month
8
Mean absolute % error (MAPE)
MLP Network
OSS
LM
0.0170
0.0087
ERNN
OSS
LM
0.0165
0.0048
Root mean square error (RMSE)
Mean absolute deviation (MAD)
Correlation coefficient
Training time (minutes)
Number of iterations (epochs)
0.0200
0.8175
0.964
0.4
850
0.0199
0.7944
0.945
1.8
1135
23 Sep - 04 Oct, 2002
WSC7
0.0099
0.4217
0.999
30
7
10
0.0067
0.2445
0.982
30
10
15
(c) Maximum Wind-Speed (11-20 Jan 2001)
ERNN
MLP network
50
Maximum Wind-Speed Forecasting
40
Actual
MLP-OSS
MLP-LM
30
20
10
0
Wind-Speed (km/h)
Wind-Speed (km/h)
50
Maximum Wind-Speed Forecasting
40
Actual
ERNN-OSS
ERNN-LM
30
20
10
0
0
2
4
6
Days of the Month
Performance evaluation
parameters (wind-speed)
8
10
0
2
4
6
Days of the Month
Mean absolute % error (MAPE)
MLP Network
OSS
LM
0.0896
0.0770
OSS
0.0873
LM
0.0333
Root mean square error (RMSE)
Mean absolute deviation (MAD)
Correlation coefficient
Training time (minutes)
Number of iterations (epochs)
0.1989
0.8297
0.9714
0.3
851
0.0199
0.7618
0.9886
0.5
1208
0.0074
0.3126
0.9995
8
12
23 Sep - 04 Oct, 2002
WSC7
0.0162
0.6754
0.9974
1
8
8
10
ERNN
16
Comparison of Relative Percentage Error
between Actual and Forecasted Parameters
MLP network
5
0
0
10
20
30
-5
Time (days)
-10
0
0
10
5
0
0
10
20
30
-5
Time (days)
5
0
0
10
40
40
0
0
23 Sep - 04 Oct, 2002
10
20
-20
Time (days)
WSC7
30
Relative Error (%)
Relative Error (%)
-10
20
20
30
-5
-10
-40
30
-5
Time (days)
Wind-Speed
20
10
Relative Error (%)
Relative Error (%)
5
-10
10
Maximum
Temperature
ERNN
10
Relative Error (%)
Minimum
Temperature
Relative Error (%)
10
Time (days)
20
0
0
10
20
30
-20
-40
Time (days)
17
Comparison of Training of Connectionist Models
Network model
Number of
hidden neurons
Number of
hidden layers
Activation function used
in hidden layer
Activation function used in
output layer
MLP
45
1
Log-sigmoid
Pure linear
ERNN
45
1
Tan-sigmoid
Pure linear
RBFN
180
2
Gaussian function
Pure linear
23 Sep - 04 Oct, 2002
WSC7
18
15
o
Temperature ( C)
Comparison among three Neural Networks
Techniques for 15-day ahead Forecasting (1-15 Apr 2001)
Maximum
Temperature
10
5
Actual value
RBFN
MLP
RNN
0
0
3
6
Minimum
Temperature
Temperature ( oC)
15
9
12
15
Days of the Month
10
5
Wind-Speed
Wind Speed (km/h)
0
60 0
3
6
9
12
15
Days of the Month
40
20
0
0
3
6
9
12
15
Days of the Month
23 Sep - 04 Oct, 2002
WSC7
19
Performance Evaluation of
RBFN, MLP and ERNN Techniques
Model
RBFN
Performance
Evaluation
Parameters
MAP
MAD
Correlation Coefficient
MLP
MAP
MAD
ERNN
MAP
MAD
23 Sep - 04 Oct, 2002
Correlation Coefficient
Correlation Coefficient
Maximum
Temperature
Minimum
Temperature
Wind Speed
3.821
0.420
0.987
3.622
1.220
0.947
4.135
0.880
0.978
6.782
1.851
0.943
6.048
1.898
0.978
6.298
1.291
0.972
5.802
0.920
0.946
5.518
0.464
0.965
5.658
0.613
0.979
WSC7
20
5. CONCLUSIONS
• In this paper, we developed and compared the
performance of multi-layered perceptron (MLP)
neural network, Elman recurrent neural network
(ERNN) and radial basis functions network
(RBFN).
• It can be inferred that ERNN could yield more
accurate results, if good data selection strategies,
training paradigms, and network input and output
representations are determined properly.
23 Sep - 04 Oct, 2002
WSC7
21
• Levenberg-Marquardt (LM) approach appears to
be the best learning algorithm. However, it
requires more memory and is computationally
complex while compared to one-step-secant (OSS)
algorithm.
• Empirical results clearly demonstrate that
compared to MLP neural network and ERNN,
RBFN are much faster and more reliable for the
weather forecasting problem considered.
• A comparison of the neurocomputing
techniques with other statistical techniques
would be another future research topic.
23 Sep - 04 Oct, 2002
WSC7
22
THANK YOU !
23 Sep - 04 Oct, 2002
WSC7
23
Related documents