Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Introduction to Bio-Inspired Models During the last three decades, several efficient machine learning tools have been inspired in biology and nature: Artificial Neural Networks (ANN) are inspired in the brain to automatically learn and generalize (model) observed data. Evolutive and genetic algorithms offer a solution to standard optimization problems when no much information about the function to optimize is available. Artificial ant colonies offer an alternative solution for optimization problems. All these methods share some common properties: They are inspired in nature (not in human logical reasoning). They are automatic (no human intervention) and nonlinear. They provide efficient solutions to some hard NP-problems. 1 Introduction to Artificial Neural Networks Artificial Neural Networks are inspired in the structure and functioning of the brain, which is a collection of interconnected neurons (the simplest computing elements performing information processing): Each neuron consists of a cell body, that contains a cell nucleus. There are number of fibers, called dendrites, and a single long fiber called axon branching out from the cell body. The axon connects one neuron to others (through the dendrites). The connecting junction is called synapse. 2 Functioning of a “Neuron” • • • • The synapses releases chemical transmitter substances. The chemical substances enter the dendrite, raising or lowering the electrical potential of the cell body. When the potential reaches a threshold, an electric pulse or action potential is sent down to the axon affecting other neurons. (Therefore, there is a nonlinear activation). Excitatory and inhibitory synapses. weights (+ or -, excitatory or inhibitory) neuron potential: mixed input of neighboring neurons nonlinear activation function (threshold) 3 Multilayer perceptron. Backpropagation algorithm Inputs Outputs Gradient descent 1. Init the neural weight with random values 2. Select the input and output data and train it 3. Compute the error associate with the output The neural activity (output) is given by a no linear function. 4. Compute the error associate with the hidden neurons 5. Compute f ( x) 1 1 e cx and update the neural weight according to these values 4 Time Series Modeling and Forecast Sometimes the chaotic time series have a stochastic look difficult to predict An example is Henon map xn 1 1 1.4 xn2 0.3 xn 1 5 Example: Supervised Fitting and Prediction Given a time series with 2000 points (T=20), generated from a Lorenz system (chaotic behavior). To check modeling power different parameters are tested . 3:6:3x n+1 Three variables (x,y,z) (xn,yn,zn) y1 Continuous System Neural Network 3:k:3 yi 3:k:3 zi Wik h1 (xn+1,yn+1,zn+1) yn+1 zn+1 h2 hk 3:15:3 wkj x2 xn x3 yn xj zn 6 Dynamical Behavior A simple model doesn’t capture the complete structure of the system , then the dynamics of the system is not reproduce. A complex system it’s overfitting the problem and the dynamics of the system is not reproduce 3:6:3 3:15:3 Only a intermediate model with an appropriate amount of parameters can model the functional structure of the system and the dynamics 7 Time series from a infrared laser. Xi Infrared laser intensity is modeled using a neural network. Only time lagged intensities are used. Net 6:5:5:1 The Neural network reproduces laser behavior Xi-1 Xi-2 Xi-3 Xi-j The Neural Network can be synchronized with the time series obtained from the laser. 8 Structural Learning: Modular Neural Networks With the aim of giving some flexibility to the network topology, modular neural networks combine different neural blocks into a global topology. Fully-connected topology (too many parameters). Combining several blocks (parameter reduction). 2*4+4*4+4*1+9= 37 weights Assigning different subnets to specific tasks we can simplify the complexity of the model. 2(2*2)+2(2*2)+4*1+9= 29 weights In most of the cases, block division is a heuristic task !!! How to obtain an optimal “block division” for a given problem ? 9 Functional Networks Functional networks are a generalization of neural networks which combine both qualitative domain knowledge and data. I x Qualitative knowledge: x 3 = F (x 1, x 2), F F F F y u z I Initial Topology Theorem. The simplest functional form is: Simplified Topology x f y f z f + f-1 u This is the optimal “block division” for this problem !!! Data: (x1i , x 2i , x 3i ), i=1,2,... Learning (least squares): {f1, ..., fn} {a1, ..., an} 10 Some FN Architectures Associative Model: F(x,y) is an associative operator. 1 F( x, y) h ( f (x) g(y)) Separable Model: A simple topology. F( x, y) f (x) g(y) Sliced-Conditioned Model: F( x0 , y) c x * f (y) T F( x, y0 ) c y * ( x)T where f and y are covenient basis for the x- and y-constant slices. F( x, y) fi (x) * gi (y) i 11 A First Example. Functional vs Neural u x x 2 y 2 y3 100 points of Training Data with Uniform Noise in (-0.01,0.01). 25x25 points from the exact surface for Validation. Neural Network 2:2:2:1 MLP 15 parameters RMSE=0.0074 2:3:3:1 MLP 25 parameters RMSE=0.0031 Functional Network (separable model) F = {1,x ,x2 ,x3} 12 parameters RMSE=0.0024 Knowledge of the network structure (separable). Appropriate family of functions (polynomial). Non-parametric approach to learn the neuron functions !!!! 12 Functional Nets & Modular Neural Nets Advantages and shortcomings of Neural Nets Black-box topology with no problem connection. Efficient non-parametric models for approximating functions. Functional Nets Parametric learning techniques (supply basis functions). Model driven optimal topology. The topology of the network is obtained from the Functional network. The neuron functions are Approximated using MLPs. Hybrid functional-neural networks (Modular networks) 13 Another example. Nonlinear Time Series Time series modeling and forecasting is an important problem with many practical applications. Nonlinear time series modeling is a difficult task because: Sensitivity to initial conditions Trajectories starting at very close initial points split away after a few iterates. Goal: predicting the future using past values. x1, x2,…, xn ¿¿¿ xn+1 ??? xn ,yn ,zn X1=0.8 0.8 0.6 Modeling methods: X1=0.8 + 10-3 0.4 X1=0.8 - 10-3 0.2 xn+1 = F(x1, x2,…, xn) There are many well-known techniques for linear time series (ARMA, etc.). 0 0 10 20 30 40 n Fractal geometry Evolve in a irregular fractal space. Nonlinear time series may exhibit complex seemingly stochastic behavior. Nonlinear Maps (the Lozi model) xn 1 1.7 xn1 0.5xn2 14 Functional Models (separation) x n f (x n1 ) g(x n2 ) 11.7 x n1 0.5x n 2 500 training points 1000 validation points. Separable Functional Net: ¿which basis family? F={sin(x),…,sin(mx), cos(x),…,cos(mx)} With 4*m parameters Symmetric Modular Functional Neural Net: 1:m:1 1:m:1 f1 m=11 (44 pars) RMSE=5.3e-3 + FN f2 m=7 (42 pars) RMSE=1.5e-3 + MFNN 1 With 6*m parameters x n 1 1.7 x n1 0.5xn2 Asymmetric Modular Functional Neural Net: 1:2m:1 1:2:1 With 2*m-2 parameters m=7 (42 pars) RMSE=4.0e-4 + MFNN 2 15 Minimum Description Length Description Length for a model The Minimum Description Length (MDL) algorithm has proved to be simple and efficient in several problems about Model Selection. 16