* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download NEURAL NETWORKS AND FUZZY SYSTEMS
Orthogonal matrix wikipedia , lookup
Singular-value decomposition wikipedia , lookup
System of linear equations wikipedia , lookup
Non-negative matrix factorization wikipedia , lookup
Four-vector wikipedia , lookup
Perron–Frobenius theorem wikipedia , lookup
Gaussian elimination wikipedia , lookup
Cayley–Hamilton theorem wikipedia , lookup
NEURONAL DYNAMICS 2: ACTIVATION MODELS 姓 名:王莹 学 号:0620110201 时 间:2006-10-09 3.1 NEURONAL DYNAMICAL SYSTEM Neuronal activations change with time. The way they change depends on the dynamical equations as following: x g(FX ,FY , ) (3-1) y h(FX ,FY , ) (3-2) 2006.10.9 3.2 ADDITIVE NEURONAL DYNAMICS first-order passive decay model In the absence of external or neuronal stimuli, the simplest activation dynamics model is: x i xi y j yj (3-3) (3-4) 2006.10.9 3.2 ADDITIVE NEURONAL DYNAMICS since for any finite initial condition xi (t ) xi (0)e t The membrane potential decays exponentially quickly to its zero potential. 2006.10.9 3.2 ADDITIVE NEURONAL DYNAMICS Passive Membrane Decay Passive-decay rate Ai 0 scales the rate to the membrane’s resting potential. xi Ai xi solution : xi(t) xi( 0 )e -Ai t Passive-decay rate measures: the cell membrane’s resistance or “friction” to current flow. 2006.10.9 property Pay attention to Ai property The larger the passive-decay rate,the faster the decay--the less the resistance to current flow. 3.2 ADDITIVE NEURONAL DYNAMICS Membrane Time Constants The membrane time constant C i scales the time variable of the activation dynamical system. The multiplicative constant model: C i x i -A i x i (3-8) 2006.10.9 Solution and property solution x i (t ) xi (0)e Ai t Ci property The smaller the capacitance ,the faster things change As the membrane capacitance increases toward positive infinity,membrane fluctuation slows to stop. 3.2 ADDITIVE NEURONAL DYNAMICS Membrane Resting Potentials Definition Define resting Potential Pi as the activation value to which the membrane potential equilibrates in the absence of external or neuronal inputs: C i x i -A i x i Pi (3-11) Solutions x i (t) x i (0)e - Ai t Ci Pi (1 - e Ai - Ai t Ci ) (3-12) 2006.10.9 Note The time-scaling capacitance dose not affect the asymptotic or steady-state solution. The steady-state solution does not depend on the finite initial condition. In resting case,we can find the solution quickly. 3.2 ADDITIVE NEURONAL DYNAMICS Additive External Input Add input Apply a relatively constant numeral input to a neuron. x i -x i I i (3-13) solution x i (t) x i (0)e-t I i (1- e -t ) (3-14) Meaning of the input Input can represent the magnitude of directly experiment sensory information or directly apply control information. The input changes slowly,and can be assumed constant value. 3.3 ADDITIVE NEURONAL FEEDBACK Neurons do not compute alone. Neuron modify their state activations with external input and with the feedback from one another. This feedback takes the form of path-weighted signals from synaptically connected neurons. 3.3 ADDITIVE NEURONAL FEEDBACK Synaptic Connection Matrices n neurons in field FX p neurons in field FY The ith neuron axon in FX jth neurons in FY a synapse m ij m ij is constant,can be positive,negative or zero. Meaning of connection matrix The synaptic matrix or connection matrix M is an n-by-p matrix of real number whose entries are the synaptic efficacies m ij.the ijth synapse is excitatory if m ij 0 inhibitory if m ij 0 The matrix M describes the forward projections from neuron field FX to neuron field FY The matrix N describes the feedforward projections from neuron field FY to neuron field F X The neural network can be specified by the 4-tuple (M, N, FX , FY ) 3.3 ADDITIVE NEURONAL FEEDBACK Bidirectional and Unidirectional connection Topologies Bidirectional networks M and N have the same or approximately the same structure. N M T M NT Unidirectional network A neuron field synaptically intraconnects to itself. BAM : Bidirectional associative memories M is symmetric, M M T the unidirectional network is BAM 2006.10.9 Augmented field and augmented matrix Augmented field FX FY FZ FX FY M connects FX to FY ,N connects FY to FX then the augmented field FZ intraconnects to itself by the square block matrix B 0 B N M 0 2006.10.9 Augmented field and augmented matrix In the BAM case,when N M then B BT hence a BAM symmetries an arbitrary rectangular matrix M. T In the general case, P C N M Q P is n-by-n matrix. Q is p-by-p matrix. T P PT Q QT the neurons If and only if, N M in FZ are symmetrically intraconnected C CT 2006.10.9 3.4 ADDITIVE ACTIVATION MODELS Define additive activation model n+p coupled first-order differential equations defines the additive activation model p j 1 n x i -Ai x i S j ( y j )n ji I i (3-15) y j -Aj y j Si ( xi )mij J j (3-16) j 1 2006.10.9 additive activation model define The additive autoassociative model correspond to a system of n coupled first-order differential equations n x i -Ai x i S j ( x j )m ji I i (3-17) j 1 2006.10.9 additive activation model define A special case of the additive autoassociative model xi Ci x i Ri j x j xi rij Ii xi ' S j ( x j )mij I i Ri j where Ri' is n 1 1 1 ' Ri Ri j rij (3-18) (3-19) (3-20) rij measures the cytoplasmic resistance between neurons i and j. 2006.10.9 Hopfield circuit and continuous additive BAM Hopfield circuit arises from if each neuron has a strictly increasing signal function and if the synaptic connection matrix is symmetric xi Ci xi ' S j ( x j )mij I i Ri j (3-21) continuous additive bidirectional associative memories p j 1 n x i -Ai x i S j ( y j )mij I i y j -A j y j Si ( xi )mij J j (3-22) (3-23) i 1 2006.10.9 3.5 ADDITIVE BIVALENT MODELS Discrete additive activation models correspond to neurons with threshold signal function The neurons can assume only two value: ON and OFF. ON represents the signal value +1. OFF represents 0 or –1. Bivalent models can represent asynchronous and stochastic behavior. Bivalent Additive BAM BAM-bidirectional associative memory Define a discrete additive BAM with threshold signal functions, arbitrary thresholds and inputs,an arbitrary but constant synaptic connection matrix M,and discrete time steps k. p xik 1 S j ( y kj )mij I i (3-24) j 1 p y kj 1 Si ( xik )mij I j (3-25) i 1 2006.10.9 Bivalent Additive BAM Threshold binary signal functions 1 k Si ( xi ) Si ( xik 1 ) 0 1 k S j ( y j ) S j ( y kj 1 ) 0 if if xik U i xik U i if xik U i if if y kj V j y kj V j if y kj V j (3-26) (3-27) For arbitrary real-value thresholds U U1 , , U n for neurons FX V V1 , , V p for neurons FY 2006.10.9 A example for BAM model Example A 4-by-3 matrix M represents the forward synaptic projections from FX to FY . A 3-by-4 matrix MT represents the backward synaptic projections from FY to FX . 2 3 0 1 2 0 M 0 3 2 2 1 1 3 1 0 2 T M 0 2 3 1 2 0 2 1 A example for BAM model Suppose at initial time k all the neurons inFY are ON. So the signal state vectorS (Yk ) at time k corresponds to S (Yk ) (1 1 1) Input X k ( x , x , x , x ,) (5 2 3 1) k 1 k 2 k 3 k 4 Suppose Ui V j 0 2006.10.9 A example for BAM model is: first:at time k+1 through synchronous operation,the result S ( X k ) (1 0 1 1) next:at time k+1 ,these FX signals pass “forward” through the filter M to affect the activations of the FY neurons. The three neurons compute three dot products,or correlations.The signal state vector S ( X k ) multiplies each of the three columns of M. 2006.10.9 A example for BAM model the result is: 4 S ( X k ) M ( i 1 Si ( xik )mi1 , (5 ( y1k 1 4 3) y2k 1 4 i 1 Si ( xik )mi 2 , k S ( x i i )mi3 ) i 1 y3k 1 ) Yk 1 4 synchronously compute the new signal state vector S (Yk 1 ): S (Yk 1 ) (0 1 1) A example for BAM model the signal vector passes “backward” through the synaptic filter S (Yk 1 ) at time k+2: S (Yk 1 )M T (2 2 5 0) ( x1k 2 X k 2 x2k 2 x3k 2 x4k 2 ) synchronously compute the new signal state vector S ( X k 2 ) (1 0 1 1) S ( X k ) : A example for BAM model since S ( X k 2 ) S ( X k ) then S (Yk 3 ) S (Yk 1 ) conclusion These same two signal state vectors will pass back and forth in bidirectional equilibrium forever-or until new inputs perturb the system out of equilibrium. A example for BAM model asynchronous state changes may lead to different bidirectional equilibrium keep the first FY neurons ON,only update the second and third FY neurons. At k,all neurons are ON. Yk 1 S ( X k ) M ( 5 4 3) new signal state vector at time k+1 equals: S (Yk 1 ) (1 1 1) A example for BAM model new FX activation state vector equals: X k 2 S (Yk 1 ) M T ( 1 1 5 2) synchronously thresholds S ( X k 2 ) (0 0 1 0) passing this vector forward to FY gives Yk 3 S ( X k 2 ) M (0 3 2) S (Yk 3 ) (1 1 1) S (Yk 1 ) A example for BAM model Similarly, S ( X k 4 ) S ( X k 2 ) (0 0 1 0) for any asynchronous state change policy we apply to the neurons FX The system has reached a new equilibrium,the binary pair (0 0 1 0), (1 1 1) represents a fixed point of the system. Conclusion Different subset asynchronous state change policies applied to the same data need not product the same fixedpoint equilibrium. They tend to produce the same equilibria. All BAM state changes lead to fixed-point stability. Bidirectional Stability Definition A BAM system ( Fx , F y , M ) is Bidirectional stable if all inputs converge to fixed-point equilibria. A denotes a binary n-vector in B denotes a binary p-vector in 0,1n 0,1p Bidirectional Stability Represent a BAM system equilibrates to bidirectional fixed point Af ) as M MT M MT M Af ( Af , B f A A' A' A '' MT B B B' B' Bf Bf Lyapunov Functions Lyapunov Functions L maps system state variables to real numbers and decreases with time. In BAM case,L maps the Bivalent product space to real numbers. Suppose L is sufficiently differentiable to apply the chain rule: L n i L dxi xi dt i L xi xi (3-28) Lyapunov Functions The quadratic choice of L 1 1 T L xIx 2 2 x i2 (3-29) i Suppose the dynamical system describes the passive decay system. xi xi (3-30) The solution xi (t ) xi (0)e t (3-31) Lyapunov Functions The partial derivative of the quadratic L: L xi xi L i (3-32) xi2 (3-33) or L i 2 xi (3-34) (3-35) In either case L 0 (3-36) At equilibrium L0 This occurs if and only if all velocities equal zero xi 0 Conclusion A dynamical system is stable if some Lyapunov Functions L decreases along trajectories. A dynamical system is asymptotically stable if it strictly decreases along trajectories Monotonicity of a Lyapunov Function provides a sufficient condition for stability and asymptotic stability. Linear system stability For symmetric matrix A and square matrix B,the quadratic T L xAx form behaves as a strictly decreasing Lyapunov function for any linear dynamical system x xB if and only if the matrix ABT BA is negative definite. T L xA x x AxT xAB T x T xBAx T x[ AB T BA]x T The relations between convergence rate and eigenvalue sign A general theorem in dynamical system theory relates convergence rate and ergenvalue sign: A nonlinear dynamical system converges exponetially quickly if its system Jacobian has eigenvalues with negative real parts. Locally such nonlinear system behave as linearly. Bivalent BAM theorem The average signal energy L of the forward pass of theFX Signal state vector S ( X ) through M,and the backward pass Of the FY signal state vector S (Y ) through M T : S ( X ) MS (Y ) T S (Y ) MS ( X ) T L (3-46) 2 since S (Y ) M T S ( X ) T [S (Y ) M T S ( X )T ]T T S ( X )M S (Y ) L S ( X )M S (Y ) T n p S ( x )S ( y )m i i (3-47) j i j j ij (3-48) Lower bound of Lyapunov function The signal is Lyapunov function clearly bounded below. For binary or bipolar,the matrix coefficients define the attainable bound: L mij i j The attainable upper bound is the negative of this expression. Lyapunov function for the general BAM system The signal-energy Lyapunov function for the general BAM system takes the form L S ( X ) MS(Y )T S ( X )[I U ]T S (Y )[J V ]T Inputs I [ I1 , , I N ] and J [ J 1 , , J P ] and constant vectors of thresholds U [U1 , , U N ] V [V1 , , VN ] the attainable bound of this function is. L m [ I ij i j i i Ui ] [ J j j V j ] Bivalent BAM theorem Bivalent BAM theorem.every matrix is bidrectionally stable for synchronous or asynchronous state changes. Proof consider the signal state changes that occur from time k to time k+1,define the vectors of signal state changes as: S (Y ) S (Yk 1 ) S (Yk ) S1 ( y1 ), , S p ( y p ) , S ( X ) S ( X k 1 ) S ( X k ) S1 ( x1 ), , S n ( x n ) , Bivalent BAM theorem define the individual state changes as: S j ( y j ) S j ( y j k 1 ) S j ( y j k ) Si ( xi ) Si ( xi k 1 ) S i ( xi ) k We assume at least one neuron changes state from k to time k+1. Any subset of neurons in a field can change state,but in only one field at a time. For binary threshold signal functions if a state change is nonzero, Bivalent BAM theorem Si ( xi ) 1 0 1 Si ( xi ) 0 1 1 For bipolar threshold signal functions S i ( xi ) 2 Si ( xi ) 2 The “energy”change L Lk 1 Lk L Differs from zero because of changes in field FX or in field FY Bivalent BAM theorem L S ( X ) MS(Yk )T S ( X )[I U ]T S ( X )[S (Yk ) M T [ I U ]]T S ( x ) I S ( x )U S ( x ) S ( y ) m S ( x ) I S ( x )U S ( x )[ S ( y ) m I U ] Si ( xi )[ xik 1 U i ] S i ( xi ) S j ( y kj ) T mij i j i i i i i i i k T i i i ij j j i i 0 j i k T i j j j i i i i i i ij i i i i i Bivalent BAM theorem Suppose S i ( xi ) 0 Then Si ( xi ) Si ( xi k 1 ) Si ( xi k ) 1 0 k 1 This implies xi U i so the product is positive: Si ( xi )[xik 1 U i ] 0 Another case suppose S i ( xi ) 0 Si ( xi ) Si ( xi k 1 0 1 ) Si ( xi ) k Bivalent BAM theorem k 1 This implies xi Ui so the product is positive: Si ( xi )[xik 1 U i ] 0 So Lk 1 Lk 0 for every state change. Since L is bounded,L behaves as a Lyapunov function for the additive BAM dynamical system defined by before. Since the matrix M was arbitrary,every matrix is bidirectionally stable. The bivalent Bam theorem is proved. Property of globally stable dynamical system Two insights about the rate of convergence First,the individual energies decrease nontrivially.the BAM system does not creep arbitrary slowly down the toward the nearest local minimum.the system takes definite hops into the basin of attraction of the fixed point. Second,a synchronous BAM tends to converge faster than an asynchronous BAM.In another word, asynchronous updating should take more iterations to converge.