* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download divergent plate boundary
Donald O. Hebb wikipedia , lookup
Neuroesthetics wikipedia , lookup
Process tracing wikipedia , lookup
Neuroinformatics wikipedia , lookup
Neural oscillation wikipedia , lookup
Optogenetics wikipedia , lookup
Artificial intelligence wikipedia , lookup
Neural coding wikipedia , lookup
Holonomic brain theory wikipedia , lookup
Gene expression programming wikipedia , lookup
Channelrhodopsin wikipedia , lookup
Neural modeling fields wikipedia , lookup
Time series wikipedia , lookup
Biological neuron model wikipedia , lookup
Central pattern generator wikipedia , lookup
Neuroeconomics wikipedia , lookup
Synaptic gating wikipedia , lookup
Neuropsychopharmacology wikipedia , lookup
Pattern recognition wikipedia , lookup
Catastrophic interference wikipedia , lookup
Neural binding wikipedia , lookup
Metastability in the brain wikipedia , lookup
Neural engineering wikipedia , lookup
Artificial neural network wikipedia , lookup
Development of the nervous system wikipedia , lookup
Convolutional neural network wikipedia , lookup
Nervous system network models wikipedia , lookup
Neural Network and Earthquake Prediction Professor Sin-Min Lee What is Data Mining? • Process of automatically finding the relationships and patterns, and extracting the meaning of enormous amount of data. • Also called “knowledge discovery” Objective • Extracting the hidden, or not easily recognizable knowledge out of the large data… Know the past • Predicting what is likely to happen if a particular type of event occurs … Predict the future Application • Marketing example – Sending direct mail to randomly chosen people – Database of recipients’ attribute data (e.g. gender, marital status, # of children, etc) is available – How can this company increase the response rate of direct mail? Application (Cont’d) • Figure out the pattern, relationship of attributes that those who responded has in common • Helps making decision of what kind of group of people the company should target • Data mining helps analyzing large amount of data, and making decision…but how exactly does it work? • One method that is commonly used is decision tree Decision Tree • One of many methods to perform data mining - particularly classification • Divides the dataset into multiple groups by evaluating attributes • Decision tree can be explained a series of nested if-then-else statements. Decision Tree (Cont’d) • Each non-leaf node has a predicate associated, testing an attribute of data • Leaf node represents a class, or category • To classify a data, start from root node and traverse down the tree by testing predicates and taking branches Example of Decision Tree Advantage of Decision Tree • simple to understand and interpret • require little data preparation • able to handle nominal and categorical data. • perform well with large data in a short time • the explanation for the condition is easily explained by boolean logic. Advantages of Decision Tree • Easy to visualize the process of classification – Can easily tell why the data is classified in a particular category - just trace the path to get to the leaf and it explains the reason • Simple, fast processing – Once the tree is made, just traverse down the tree to classify the data Decision Tree is for… • Classifying the dataset which – The predicates return discrete values – Does not have an attributes that all data has the same value CMT catalog: Shallow earthquakes, 1976-2005 INDIAN PLATE MOVES NORTH COLLIDING WITH EURASIA Gordon & Stein, 1992 COMPLEX PLATE BOUNDARY ZONE IN SOUTHEAST ASIA Northward motion of India deforms all of the region Many small plates (microplates) and blocks Molnar & Tapponier, 1977 India subducts beneath Burma microplate at about 50 mm/yr Earthquakes occur at plate interface along the Sumatra arc (Sunda trench) These are NOAA IN DEEP OCEAN tsunami has long wavelength, travels fast, small amplitude - doesn’t affect ships AS IT APPROACHES SHORE, it slows. Since energy is conserved, amplitude builds up - very damaging TSUNAMI WARNING Deep ocean buoys can measure wave heights, verify tsunami and reduce false alarms Because seismic waves travel much faster (km/s) than tsunamis, rapid analysis of seismograms can identify earthquakes likely to cause major tsunamis and predict when waves will arrive HOWEVER, HARD TO PREDICT EARTHQUAKES recurrence is highly variable Sieh et al., 1989 Extend earthquake history with geologic records paleoseismology M>7 mean 132 yr s 105 yr Estimated probability in 30 yrs 7-51% EARTHQUAKE RECURRENCE AT SUBDUCTION ZONES IS COM PLICATED In many subduction zones, thrust earthquakes have patterns in space and time. Large earthquakes occurred in the Nankai trough area of Japan approximately every 125 years since 1498 with similar fault areas In some cases entire region seems to have slipped at once; in others slip was divided into several events over a few years. Repeatability suggests that a segment that has not slipped for some time is a gap due for an earthquake, but it’s hard to use this concept well because of variability GAP? NOTHING YET Ando, 1975 1985 MEXICO EARTHQUAKE • SEPTEMBER 19, 1985 • M8.1 • A SUBDUCTION ZONE QUAKE • ALTHOUGH LARGER THAN USUAL, THE EARTHQUAKE WAS NOT A “SURPRISE” • A GOOD, MODERN BUILDING CODE HAD BEEN ADOPTED AND IMPLEMENTED 1985 MEXICO EARTHQUAKE • EPICENTER • 400 BUILDINGS LOCATED 240 KM COLLAPSED IN FROM MEXICO CITY OLD LAKE BED ZONE OF MEXICO CITY • SOIL-STRUCTURE RESONANCE IN OLD LAKE BED ZONE WAS A MAJOR FACTOR 1985 MEXICO EARTHQUAKE: ESSENTIAL STRUCTURES-SCHOOLS 1985 MEXICO EARTHQUAKE: STEEL FRAME BUILDING 1985 MEXICO EARTHQUAKE: POUNDING 1985 MEXICO EARTHQUAKE: NUEVA LEON APARTMENT BUILDINGS 1985 MEXICO EARTHQUAKE: SEARCH AND RESCUE • Definition • Characteristics • Project:California Earthquake Prediction) Neural Networks • AIMA – Chapter 19 • Fundamentals of Neural Networks : Architectures, Algorithms and Applications. L, Fausett, 1994 • An Introduction to Neural Networks (2nd Ed). Morton, IM, 1995 Neural Networks • McCulloch & Pitts (1943) are generally recognised as the designers of the first neural network • Many of their ideas still used today (e.g. many simple units combine to give increased computational power and the idea of a threshold) Neural Networks • Hebb (1949) developed the first learning rule (on the premise that if two neurons were active at the same time the strength between them should be increased) Neural Networks • During the 50’s and 60’s many researchers worked on the perceptron amidst great excitement. • 1969 saw the death of neural network research for about 15 years – Minsky & Papert • Only in the mid 80’s (Parker and LeCun) was interest revived (in fact Werbos discovered algorithm in 1974) Neural Networks Neural Networks • We are born with about 100 billion neurons • A neuron may connect to as many as 100,000 other neurons Neural Networks • Signals “move” via electrochemical signals • The synapses release a chemical transmitter – the sum of which can cause a threshold to be reached – causing the neuron to “fire” • Synapses can be inhibitory or excitatory The First Neural Neural Networks McCulloch and Pitts produced the first neural network in 1943 Many of the principles can still be seen in neural networks of today What is neural network ? • Def 1: Imitate the brain, and surpass the brain to manage both pattern processing problem and symbolic problem. • Example: learning and self-organization What is neural network ? (cont.) Def 2: Complex-valued neural networks are the network that deal with complexvalued information by using complex-valued parameters and variables. Example: 1. Good dish: color, smell, taste 2. Prediction: seismic history, ground water, abnormal behavior, nearby seismic activities What is neural network ? (cont.) Def 3: brain artificial brain artificial intelligence neural network Example: information processing in the real world should be flexible enough to deal with unexpectedly (geo figure) and dynamically (fore/main/after-shock) changing environment. A new sort of computer • What are (everyday) computer systems good at... and not so good at? Good at Rule-based systems: doing what the programmer wants them to do Not so good at Dealing with noisy data Dealing with unknown environment data Massive parallelism Fault tolerance Adapting to circumstances Neural networks to the rescue • Neural network: information processing paradigm inspired by biological nervous systems, such as our brain • Structure: large number of highly interconnected processing elements (neurons) working together • Like people, they learn from experience (by example) Neural networks to the rescue • Neural networks are configured for a specific application, such as pattern recognition or data classification, through a learning process • In a biological system, learning involves adjustments to the synaptic connections between neurons same for artificial neural networks (ANNs) Where can neural network systems help • when we can't formulate an algorithmic solution. • when we can get lots of examples of the behavior we require. ‘learning from experience’ • when we need to pick out the structure from existing data. Inspiration from Neurobiology • A neuron: many-inputs / one-output unit • output can be excited or not excited • incoming signals from other neurons determine if the neuron shall excite ("fire") • Output subject to attenuation in the synapses, which are junction parts of the neuron Synapse concept • The synapse resistance to the incoming signal can be changed during a "learning" process [1949] Hebb’s Rule: If an input of a neuron is repeatedly and persistently causing the neuron to fire, a metabolic change happens in the synapse of that particular input to reduce its resistance Mathematical representation The neuron calculates a weighted sum of inputs and compares it to a threshold. If the sum is higher than the threshold, the output is set to 1, otherwise to -1. Non-linearity A simple perceptron • It’s a single-unit network • Change the weight by an amount proportional to the difference between the desired output and Input the actual output. Δ Wi = η * (D-Y).Ii Learning rate Actual output Desired output Perceptron Learning Rule Example: A simple single unit adaptive network • The network has 2 inputs, and one output. All are binary. The output is – 1 if W0I0 + W1I1 + Wb > 0 – 0 if W0I0 + W1I1 + Wb ≤ 0 • We want it to learn simple OR: output a 1 if either I0 or I1 is 1. Learning • From experience: examples / training data • Strength of connection between the neurons is stored as a weight-value for the specific connection • Learning the solution to a problem = changing the connection weights Operation mode • Fix weights (unless in online learning) • Network simulation = input signals flow through network to outputs • Output is often a binary decision • Inherently parallel • Simple operations and threshold: fast decisions and real-time response Characteristics • Distributedness and parallelism • Locality • Weighted sum and activation function with nonlinearity • Plasticity • Generalization Characteristics (cont.) • 1. Distributedness and parallelism: process information distributedly in parallel and trouble in a single neuron does not give rise to fatal impact on the brain function. • CAEP: if one or some of related area becomes inactive recently, it is not going to affect the area movement as whole. (normal fault, strike-slip fault, thrust fault) Characteristics (cont.) Characteristics (cont.) • 2. Locality: information transferred by a neuron is limited by its nearby neurons. • CAEP: short term earthquake prediction is highly influenced by it’s geologic figure locally. Characteristics (cont.) • 3. Weighted sum and activation function with nonlinearity: input signal is weighted at the synoptic connection by a connection weight. • CAEP: nearby location will be weighted with each activation function. Characteristics (cont.) • 4. Plasticity: connection weights change according to the information fed to the neuron and the internal state. This plasticity of the connection weights leads to learning and self-organization. The plasticity realizes the adaptability against the continuously varying environment. • CAEP: calculate the stress of focused point according to the seismic wave history in the around area Characteristics (cont.) • 5. Generalization: A neural network constructs its own view of the world by inferring an optimal action on the basis of previously learned events by interpolation, and extrapolation. • CAEP: get a view of one area from past experience by pattern representation Prediction. Basic Function of CSEP • Neuron: list of locations along San Andreas Fault, and two of the associated faults— Hayward and Calaveras. Basic Function of CSEP (cont.) • Neuron’s parameters: magnitude, date, latitude, longitude, depth, location, ground water, observation, etc.