Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
DNA vaccination wikipedia , lookup
Hygiene hypothesis wikipedia , lookup
Immune system wikipedia , lookup
Adoptive cell transfer wikipedia , lookup
Innate immune system wikipedia , lookup
Monoclonal antibody wikipedia , lookup
Molecular mimicry wikipedia , lookup
Adaptive immune system wikipedia , lookup
Cancer immunotherapy wikipedia , lookup
Immunosuppressive drug wikipedia , lookup
Artificial Immune Systems Andrew Watkins Why the Immune System? • Recognition – Anomaly detection – Noise tolerance • • • • • • • • Robustness Feature extraction Diversity Reinforcement learning Memory Distributed Multi-layered Adaptive Definition AIS are adaptive systems inspired by theoretical immunology and observed immune functions, principles and models, which are applied to complex problem domains (de Castro and Timmis) Some History • Developed from the field of theoretical immunology in the mid 1980’s. – Suggested we ‘might look’ at the IS • 1990 – Bersini first use of immune algos to solve problems • Forrest et al – Computer Security mid 1990’s • Hunt et al, mid 1990’s – Machine learning How does it work? Immune Pattern Recognition BCR or Antibody B-cell Receptors (Ab) Epitopes Antigen B-cell • The immune recognition is based on the complementarity between the binding region of the receptor and a portion of the antigen called epitope. • Antibodies present a single type of receptor, antigens might present several epitopes. – This means that different antibodies can recognize a single antigen Immune Responses Antibody Concentration Cross-Reactive Response Secondary Response Primary Response Lag Lag Response to Ag1 Lag Response to Ag1 ... ... Antigen Ag1 Antigens Ag1, Ag2 ... Response to Ag1 + Ag3 Response to Ag2 ... Antigen Ag1 + Ag3 Time Clonal Selection Clonal deletion (negative selection) Self-antigen Proliferation (Cloning) M M Antibody Memory cells Selection Differentiation Plasma cells Foreign antigens Self-antigen Clonal deletion (negative selection) Immune Network Theory • Idiotypic network (Jerne, 1974) • B cells co-stimulate each other – Treat each other a bit like antigens • Creates an immunological memory Paratope Suppression Negative response Ag 1 2 Idiotope 3 Antibody Activation Positive response Shape Space Formalism • Repertoire of the immune system is complete (Perelson, 1989) • Extensive regions of complementarity • Some threshold of recognition Ve e V Ve e Ve e Self/Non-Self Recognition • Immune system needs to be able to differentiate between self and non-self cells • Antigenic encounters may result in cell death, therefore – Some kind of positive selection – Some element of negative selection General Framework for AIS Solution Immune Algorithms Affinity Measures Representation Application Domain Representation – Shape Space • Describe the general shape of a molecule •Describe interactions between molecules •Degree of binding between molecules •Complement threshold Define their Interaction • Define the term Affinity • Affinity is related to distance – Euclidian D L 2 ( Ab Ag ) i i i 1 • Other distance measures such as Hamming, Manhattan etc. etc. • Affinity Threshold Basic Immune Models and Algorithms • • • • • Bone Marrow Models Negative Selection Algorithms Clonal Selection Algorithm Somatic Hypermutation Immune Network Models Bone Marrow Models • Gene libraries are used to create antibodies from the bone marrow • Use this idea to generate attribute strings that represent receptors • Antibody production through a random concatenation from gene libraries An individual genome corresponds to four libraries: Library 1 A1 A2 A3 A4 A5 A6 A7 A8 A3 Library 2 Library 3 B1 B2 B3 B4 B5 B6 B7 B8 Library 4 C1 C2 C3 C4 C5 C6 C7 C8 B2 D1 D2 D3 D4 D5 D6 D7 D8 C8 A3 B2 C8 D5 A3 B2 C8 D5 Expressed Ab molecule = four 16 bit segments = a 64 bit chain D5 Negative Selection Algorithms • Forrest 1994: Idea taken from the negative selection of T-cells in the thymus • Applied initially to computer security • Split into two parts: – Censoring – Monitoring Detector Set (R) Self strings (S) Generate random strings (R0) Match No Yes Reject Detector Set (R) Protected Strings (S) Match Yes Non-self Detected No Clonal Selection Algorithm (de Castro & von Zuben, 2001) Randomly initialise a population (P) For each pattern in Ag Determine affinity to each Ab in P Select n highest affinity from P Clone and mutate prop. to affinity with Ag Add new mutants to P endFor Select highest affinity Ab in P to form part of M Replace n number of random new ones Until stopping criteria Immune Network Models (Timmis & Neal, 2001) Initialise the immune network (P) For each pattern in Ag Determine affinity to each Ab in P Calculate network interaction Allocate resources to the strongest members of P Remove weakest Ab in P EndFor If termination condition met exit else Clone and mutate each Ab in P (based on a given probability) Integrate new mutants into P based on affinity Repeat Somatic Hypermutation • Mutation rate in proportion to affinity • Very controlled mutation in the natural immune system • The greater the antibody affinity the smaller its mutation rate • Classic trade-off between exploration and exploitation How do AIS Compare? • Basic Components: – AIS B-cell in shape space (e.g. attribute strings) • Stimulation level – ANN Neuron • Activation function – GA chromosome • fitness Comparing • Structure (Architecture) – AIS and GA fixed or variable sized populations, not connected in population based AIS – ANN and AIS • Do have network based AIS • ANN typically fixed structure (not always) • Learning takes place in weights in ANN Comparing • Memory – AIS in B-cells • Network models in connections – ANN In weights of connections – GA individual chromosome Comparing • • • • • • Adaptation Dynamics Metadynamics Interactions Generalisation capabilities Etc. many more. Where are they used? • • • • • • Dependable systems Scheduling Robotics Security Anomaly detection Learning systems Artificial Immune Recognition System (AIRS): An Immune-Inspired Supervised Learning Algorithm AIRS: Immune Principles Employed • Clonal Selection • Based initially on immune networks, though found this did not work • Somatic hypermutation – Eventually • Recognition regions within shape space • Antibody/antigen binding AIRS: Mapping from IS to AIS • Antibody • Recognition Ball (RB) • Antigens • Immune Memory Feature Vector Combination of feature vector and vector class Training Data Memory cells—set of mutated Artificial RBs Classification • Stimulation of an ARB is based not only on its affinity to an antigen but also on its class when compared to the class of an antigen • Allocation of resources to the ARBs also takes into account the ARBs’ classifications when compared to the class of the antigen • Memory cell hyper-mutation and replacement is based primarily on classification and secondarily on affinity AIRS Algorithm • Data normalization and initialization • Memory cell identification and ARB generation • Competition for resources in the development of a candidate memory cell • Potential introduction of the candidate memory cell into the set of established memory cells Memory Cell Identification A Memory Cell Pool ARB Pool MCmatch Found A 1 Memory Cell Pool MCmatch ARB Pool ARB Generation A 1 Memory Cell Pool MCmatch Mutated Offspring 2 ARB Pool Exposure of ARBs to Antigen 1 A Memory Cell Pool MCmatch 3 Mutated Offspring 2 ARB Pool Development of a Candidate Memory Cell 1 A Memory Cell Pool MCmatch 3 Mutated Offspring 2 ARB Pool Comparison of MCcandidate and MCmatch 1 A Memory Cell Pool MCmatch 3 Mutated Offspring 4 2 A MC candidate ARB Pool Memory Cell Introduction 1 A Memory Cell Pool MCmatch 3 Mutated Offspring 2 4 5 A MCcandidate ARB Pool Memory Cells and Antigens Memory Cells and Antigens AIRS: Performance Evaluation Fisher’s Iris Data Set Pima Indians Diabetes Data Set Ionosphere Data Set Sonar Data Set Iris 1 Grobian (rough) 2 SSV 3 Ionosphere 100% Diabetes Sonar 3-NN + simplex 98.7% Logdisc 77.7% TAP MFT Bayesian 92.3% 98.0% 3-NN 96.7% IncNet 77.6% Naïve MFT Bayesian 90.4% C-MLP2LN 98.0% IB3 96.7% DIPOL92 77.6% SVM 90.4% 4 PVM 2 rules 98.0% MLP + BP 96.0% Linear Discr. Anal. 77.5%77.2% Best 2-layer MLP + BP, 12 hidden 90.4% 5 PVM 1 rule 97.3% AIRS 94.9 SMART 76.8% MLP+BP, 12 hidden 84.7% 6 AIRS 96.7 C4.5 94.9% GTO DT (5xCV) 76.8% MLP+BP, 24 hidden 84.5% 7 FuNe-I 96.7% RIAC 94.6% ASI 76.6% 1-NN, Manhatten 84.2% 8 NEFCLASS 96.7% SVM 93.2% Fischer discr. anal 76.5% AIRS 84.0 9 CART 96.0% Non-linear perceptron 92.0% MLP+BP 76.4% MLP+BP, 6 hidden 83.5% 10 FUNN 95.7% FSM + rotation 92.8% LVQ 75.8% FSM methodology? 83.6% 11 1-NN 92.1% LFC 75.8% 1-NN Euclidean 82.2% 12 DB-CART 91.3% RBF 75.7% DB-CART, 10xCV 81.8% 13 Linear perceptron 90.7% NB 75.573.8% CART, 10xCV 67.9% 14 OC1 DT 89.5% kNN, k=22, Manh 75.5% 15 CART 88.9% MML 75.5% … ... 22 AIRS 74.1 23 C4.5 73.0% 11 others reported with lower scores, including Bayes, Kohonen, kNN, ID3 … AIRS: Observations • ARB Pool formulation was over complicated – Crude visualization – Memory only needs to be maintained in the Memory Cell Pool • Mutation Routine – Difference in Quality – Some redundancy AIRS: Revisions • Memory Cell Evolution – Only Memory Cell Pool has different classes – ARB Pool only concerned with evolving memory cells • Somatic Hypermutation – Cell’s stimulation value indicates range of mutation possibilities – No longer need to mutate class Comparisons: Classification Accuracy • Important to maintain accuracy AIRS1: Accuracy AIRS2: Accuracy Iris 96.7 96.0 Ionosphere 94.9 95.6 Diabetes 74.1 74.2 Sonar 84.0 84.9 • Why bother? Comparisons: Data Reduction • Increase data reduction—increased efficiency Training Set Size AIRS1: Memory Cells AIRS2: Memory Cells Iris 120 42.1 / 65% 30.9 / 74% Ionosphere 200 140.7 / 30% 96.3 / Diabetes 691 470.4 / 32% 273.4 / Sonar 192 144.6 / 25% 177.7 / 7% 52% 60% Features of AIRS • No need to know best architecture to get good results • Default settings within a few percent of the best it can get • User-adjustable parameters optimize performance for a given problem set • Generalization and data reduction More Information • http://www.cs.ukc.ac.uk/people/rpg/abw5 • http://www.cs.ukc.ac.uk/people/staff/jt6 • http://www.cs.ukc.ac.uk/aisbook