* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Predicting Genetic Regulatory Response Using Classification
Polycomb Group Proteins and Cancer wikipedia , lookup
Pharmacogenomics wikipedia , lookup
Pathogenomics wikipedia , lookup
Genetic engineering wikipedia , lookup
Quantitative trait locus wikipedia , lookup
Gene desert wikipedia , lookup
Essential gene wikipedia , lookup
Nutriepigenomics wikipedia , lookup
Therapeutic gene modulation wikipedia , lookup
Public health genomics wikipedia , lookup
History of genetic engineering wikipedia , lookup
Site-specific recombinase technology wikipedia , lookup
Genome evolution wikipedia , lookup
Genomic imprinting wikipedia , lookup
Ridge (biology) wikipedia , lookup
Minimal genome wikipedia , lookup
Artificial gene synthesis wikipedia , lookup
Epigenetics of human development wikipedia , lookup
Microevolution wikipedia , lookup
Designer baby wikipedia , lookup
Genome (book) wikipedia , lookup
Gene expression profiling wikipedia , lookup
Predicting Genetic Regulatory Response Using Classification Us v. Them (“Them” being Manuel Middendorf, Anshul Kundaje, Chris Wiggins, Yoav Freund, and Christina Leslie, in “Predicting Genetic Regulatory Response Using Classification” (2004) The Problem • Current studies of gene transcription tend to be descriptive • Need for a predictive system – the ability to predict gene regulation for new experiments • Rather than determining patterns in sets of genes and conditions, look at underyling causes of those patterns The Important Parts of Genes & Experiments • Regulation is determined by binding sites (motifs) and regulators (parents) • The significance of experiments, then, is how they affect regulators • The significance of genes is what motifs they contain Binding Sites and Regulation • Discretize gene response into only up-regulated (1), down-regulated(-1), or unchanged (0) • A motif is either present (1) or absent (0) • A parent is either up-regulated (1), downregulated (-1), or unchanged (0) • Assume (and we need to check with someone who actually knows something about biology on this) that things only happen if motif is present and parent is either up- or down-regulated What our matrix really looks like • • • • • • g = # of genes e = # of experiments p = # of parents m = # of motifs Then we have g*e response values For each response, we have p*m parent/motif combinations • For each parent/motif combination, there are three possibilities – present and up-regulated, present and down-regulated, or all those other possibilities where nothing happens • Represent these possibilities as a pair of binary variables, one for up and one for down M1, P1 + M1, P1 - M1, P2 + M1, P2 - … M1, Pp + … Mm, Pp - G1, E1 M1, P1 + M1, P1 - M1, P2 + M1, P2 - … M1, Pp + … Mm, Pp - G1, E2 M1, P1 + M1, P1 - M1, P2 + M1, P2 - … M1, Pp + … Mm, Pp - … M1, P1 + M1, P1 - M1, P2 + M1, P2 - … M1, Pp + … Mm, Pp - G1, Ee M1, P1 + M1, P1 - M1, P2 + M1, P2 - … M1, Pp + … Mm, Pp - … M1, P1 + M1, P1 - M1, P2 + M1, P2 - … M1, Pp + … Mm, Pp - Gg, Ee Some Numbers • In the paper, the initial dataset had 6110 genes and 173 experiments • 354 motifs are considered • 475 regulators are considered • Set of genes to consider is reduced to only 1411 genes of interest Some More Numbers • Only train on genes that are up- or downregulated • Approx. 8% of gene/experiment pairs from the overall sample appear to be, so, assuming this holds true in the reduced sample, we have 19,632 gene / experiment pairs to train on • For each of these values we have 2*354*475 = 336,300 predictor variables Some Problems • 19,632 by 336,300 is an awfully large matrix to want to do any calculations with • We have far more variables than observations Possible Solutions • Random Projection: – Pro: we can reduce our dimensionality – Con: it seems like a somewhat silly approach – Con: there’ll still be a lot of calculations just to make the projection Possible Solutions • Variable Selection or PCA – Pro: we could reduce our dimensionality in a more informed way – Con: computationally painful Possible Solutions • Random Forest • From Breiman and Cutler: • Grow a number of classification trees, and take the vote of the classifications of all trees • For each tree, if we have n cases, sample, with replacement, n cases, using some number of randomly chosen variables much smaller than the full number of variables – Pro: allows for flexibility in reduction of dimensions being considered – Pro: dimension is reduced without computation problems – Con: ? Their Solution • Alternating decision tree: • A tree with alternating levels of prediction nodes and splitter nodes • Combines a set of weak prediction rules (the splitter nodes) to make one strong rule Alternating Decision Trees v. Random Forests • Both the same rough concept – take a vote from many weak rules to get a strong rule • In ADTs, rules are based on single variables, and may be conditioned on values of other variables • In random forests, rules can be based on multiple variables, and are only marginal over all values of other variables • ADTs are fairly straightforward to read and interpret Alternating Decision Trees • Pro: computationally kind • Pro: works better than Random Forests (Creamer & Freund, 2004) • Con: we didn’t come up with it Other Ideas?