Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Bridging the Gap between Applications and Tools: Modeling Multivariate Time Series X Liu, S Swift & A Tucker Department of Computer Science Birkbeck College University of London MTS Applications at Birkbeck Screening Forecasting Explanation Forecasting Predicting Visual Field Deterioration of Glaucoma Patients Function Prediction for Novel Proteins from Multiple Sequence/Structure Data Explanation Input (observations): t-0 t-3 : Tail Gas Flow in_state 0 : Reboiler Temperature in_state 1 Output (explanation): t - 7 : Top Temperature in_state 0 with probability=0.92 t - 54 : Feed Rate in_state 1 with probability=0.71 t - 75 : Reactor Temperature in_state 0 with probability=0.65 The Gaps Screening Automatic / Semi- Automatic Analysis of Outliers Forecasting Analysing Short Multivariate Time Series Explanation Coping with Huge Search Spaces The Problem - What/Why/How Short-Term Forecasting of Visual Field Progression Using a Statistical MTS Model The Vector Auto-Regressive Process - VAR(P) There Could be Problems if the MTS is Short A Modified Genetic Algorithm (GA) can be Used VARGA The Prediction of Visual Field Deterioration Plays an Important Role in the Management of the Condition Background - The Dataset The interval between tests is about 6 months 76 75 18 19 74 73 72 15 16 17 71 70 69 68 11 12 13 14 67 66 65 64 63 6 7 8 9 10 62 61 60 59 58 1 2 3 4 5 43 42 41 40 39 20 21 22 23 24 48 47 46 45 44 25 26 27 28 29 52 51 50 49 30 31 32 33 55 54 53 34 35 36 57 56 37 38 Typically, 76 points are measured Values Range Between 60 =very good, 0 = blind The number of tests can range between 10 and 44 x x Points used in this paper (Right Eye) Usual Position of Blind Spot (Right Eye) Background - The VAR Process Vector Auto-Regressive Process of Order P: VAR(P) x(t) VF Test for Data Points at Time t (K1) Ai Parameter Matrix at Lag i (KK) x(t-i) VF Test for Data Points at lag i from t (K1) (t) Observational Noise at time t (K1) The Genetic Algorithm “A Search/Optimisation method that solves a problem through maintaining and improving a population of suitable candidate solutions using biological metaphors” Generate a Population of random Chromosomes (Solutions) Repeat for a number of Generations Cross Over the current Population Mutate the current Population Select the Fittest for the next Population Loop The best solution to the problem is the Chromosome in the last generation which has the highest Fitness GAs - Chromosome Example X Y 0-127 0-31 0000000-1111111 00000-11111 0000000.00000-1111111.11111 GAs - Mutation Each Bit (gene) of a Chromosome is Given a Chance MP of inverting A ‘1’ becomes a ‘0’, and a ‘0’ becomes a 1’ 01101101 These Ones! 00101111 GAs - Crossover (2) A B 01011101 11101010 X=4 11101101 01011010 C D VARGA - Representation Chromosome a111 … … … … a1ij … … a1KK A1 a211 … … … … a2ij … … A2 a1KK ... am11 … … … … amij … … amKK Am ... ap11 … … … … apij … … Ap apKK VARGA - The Genetic Algorithm GA With Extra Mutation Order Mutation After Gene Mutation Parents and Children Mutate (Both) Genes are Bound Natural Numbers Fitness is -ve Forecast Error Minimisation Problem - Roulette Wheel Run for EACH Patient Evaluation - Methods for Comparison SPlus: Yule Walker Equations, AIC and Whittles Recursion, NK(P+1), Standard Package Holt-Winters Univariate Forecasting Method, Is the Data Univariate? (GA Solution) Pure Noise Model, VAR(0), Worst Case Forecast, (Non-Differenced = 0) 54 out of the Possible 82 Patients VF Records Could not be Used : SPlus Implementation Results - Graph Comparison Scores for Cases 0 to 6 2000 1500 HW S-Plus VARGA Noise Score 1000 500 0 0 1 2 3 4 5 6 Case Number The Lower the Score - the Better Score is the One Step Ahead Forecast Error Results - Table Summary Method VARGA S-Plus HW Noise Order Average (number of order) Score 26 of 1, 2 of 2 12 of 0, 14 of 1, 1 of 2, 1 of 3 N/A 28 of 0 559.82 616.12 683.79 816.53 Average = The Average One Step Forecast Error For the 28 Patients (Both GA’s Fitness) (The Lower - The Better) Conclusion - Results VARGA Has a Better Performance VARGA Can Model Short MTS The Visual Field Data is Definitely Multivariate Data Has a High Proportion of Noise Conclusion - Remarks Non-Linear Methods and Transformations Performance Enhancements for the GA Improve Crossover Irregularly Spaced Methods Space-Time Series Methods Time Dependant Relationships Between Variables Generating Explanations in MTS Useful to know probable explanations for a given set of observations within a time series E.g. Oil Refinery: ‘Why a temperature has become high whilst a pressure has fallen below a certain value?’ Possible paradigm which facilitates Explanation is the Bayesian Network Evolutionary Methods to learn BNs Extend work to Dynamic Bayesian Networks Dynamic Bayesian Networks Static BNs repeated over t time slices Contemporaneous / Non-Contemporaneous Links Used for Prediction / Diagnosis within dynamic systems n P( X 1... X n ) P( X i | i ) i 1 Assumptions - 1 Assume all variables take at least one time slice to impose an effect on another. The more frequently a system generates data, the more likely this will be true. Contemporaneous Links can be excluded from the DBN Each variable at time, t, will be considered independent of one another Representation P pairs of the form (ParentVar, TimeLag) Each pair represents a link from a node at a previous time slice to the node in question at time t. Examples : Variable 1: { (1,1); (2,2); (0,3)} Variable 4: { (4,1); (2,5)} Search Space Given the first assumption and proposed representation the Search Space for each variable will be: 2 NMaxLag Algorithm Structure Search : Evolutionary Algorithms, Hill Climbing etc. Multivariate Time Series Parameter Calculation given structure Dynamic Bayesian Network Library for Different Operating States User Explanation Algorithm (e.g. using Stochastic Simulation) Generating Synthetic Data (1) (2) Oil Refinery Data Data recorded every minute Hundreds of variables Selected 11 interrelated variables Discretised each variable into k states Large Time Lags (up to 120 minutes between some variables) Different Operating States Results SOT FF TGF TT RinT Explanations - using Stochastic Simulation Explanations - using Stochastic Simulation 1 0.9 0.8 0.7 SOF-SP SOT TT BPF-SP BPF P(y=1) 0.6 0.5 0.4 0.3 0.2 0.1 0 1 7 13 19 25 31 37 43 49 55 61 67 73 79 85 91 97 time-x Explanation Input (observations): t-0 t-3 : Tail Gas Flow in_state 0 : Reboiler Temperature in_state 1 Output (explanation): t - 7 : Top Temperature in_state 0 with probability=0.92 t - 54 : Feed Rate in_state 1 with probability=0.71 t - 75 : Reactor Temperature in_state 0 with probability=0.65 Future Work Exploring the use of different searches and metrics Improving accuracy (e.g. different discretisation policies, continuous DBNs) Using the library of DBNs in order to quickly classify the current state of a system Automatically Detecting Changing Dependency Structure Acknowledgements BBSRC BP-AMOCO British Council for Prevention of Blindness EPSRC Honeywell Hi-Spec Solutions Honeywell Technology Center Institute of Opthalmology Moorfields Eye Hospital MRC Intelligent Data Analysis X Liu Department of Computer Science Birkbeck College University of London Intelligent Data Analysis An interdisciplinary study concerned with effective analysis of data Intelligent application of data analytic tools Application of “intelligent” data analytic tools IDA Requires Careful thinking at every stage of an analysis process (strategic aspects) Intelligent application of relevant domain knowledge Assessment and selection of appropriate analysis methods IDA Conferences IDA-95, Baden-Baden IDA-97, London IDA-99, Amsterdam IDA-2001, Lisbon IDA in Medicine and Pharmacology IDAMAP-96, Budapest IDAMAP-97, Nagoya IDAMAP-98, Brighton IDAMAP-99, Washington DC IDAMAP-2000, Berlin Other IDA Activities IDA Journal (Elsevier 1997) Journal Special Issues (1997 -) Introductory Books (Springer 1999) The Dagstuhl Seminar (Germany 2000) European Summer School (Italy 2000) Special Sessions at Conferences Concluding Remarks Strategies for data analysis and mining Strategies for human-computer collaboration in IDA Principles for exploring and analysing “big data” Benchmarking interesting real-world datasets as well as computational methods A long term interdisciplinary effort The Screening Architecture Results from a GP Clinic