Survey							
                            
		                
		                * Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
COMPARING ASSOCIATION RULES AND DECISION TREES FOR DISEASE PREDICTION Carlos Ordonez MOTIVATION  Three main issues about mining association rules in medical datasets: 1. 2. 3.  A significant fraction of association rules is irrelevant Most relevant rules with high quality metrics appear only at low support # of discovered rules becomes extremely large at low support Search constraints:   Find only medically significant association rules Make search more efficient MOTIVATION   Decision tree  a well-known machine learning algorithm Association rules vs. Decision tree Accuracy  Interpretability  Applicability  ASSOCIATION RULES Support  Confidence  Lift  confidence( x  y )  lift ( x  y )  sup port ( x  y ) sup port ( x) confidence( x  y ) sup port ( y ) Lift quantifies the predictive power of x  y  Rules such that lift(xy) > 1 are interesting!  CONSTRAINED ASSOCIATION RULES  Transforming Medical Data Set  Data must be transformed to binary dimensions Numeric attributes  intervals, each interval is mapped to an item.  Categorical attributes each categorical value is an item  If an attribute has negation add that as an item  Each item is corresponds to the presence or absence of one categorical value or one numeric interval CONSTRAINED ASSOCIATION RULES  1. Search Constraints Max itemset size (k)  2. Group    3. Reduces the combinatorial explosion of large itemsets and helps finding simple rules gi >0  Aj belongs to a group gi =0  Aj is not group-constrained at all This avoids finding trivial or redundant rules Antecedent/Consequent ci = 1  Ai is an antecedent ci = 2  Ai is a consequent Binned at 40(adult) and 60(old) Percentage of vessel narrowing LAD, LCX and RCA are binned at 70% and 50% LM is binned at 30% and 50% 9 heart regions ( 2 ranges with 0.2 as cutoff) Patients 655 attributes 25 Binned at 200 and 250 PARAMETERS k=4  Min support = 1% ≈ 7  Min confidence = 70%  Min lift = 1.2    To get rules where there is stronger implication dependence between X and Y Rules with conf ≥ 90 and lift ≥ 2, with 2 or more items in the consequent were considered medically significant. HEALTHY ARTERIES 9,595 associations  771 rules  DISEASED ARTERIES Several unneeded items were filtered out ( with values in lower (healthy) ranges)  10,218 associations  552 rules  PREDICTIVE RULES FROM DECISION TREES CN4.5  using gain ratio  CART  similar results  Threshold for the height of the tree to produce simple rules  Percentage of patients (ls)   Fraction of patients where the antecedent appears Confidence factor (cf)  Focus on predicting LDA disease  PREDICTIVE RULES FROM DECISION TREES 1. All measurements without binning as independent variables, numerical variables are automatically split  Without any threshold on height: 181 node  90% accuracy  height = 14  most rules more than 5 attributes  except 5 rules, other involve less than 2% of the patients  More than 80% of rules refer to less than 1% of patients  Many rules involve attributes with missing information  Many rules had the same variable being split several times  Few rules with cf = 1 but splits included borderline cases and involves few patients  PREDICTIVE RULES FROM DECISION TREES  With threshold = 10 on height 83 nodes  77% accuracy  Most rules have repeated attributes  More than 5 attributes  Perfusion cutoffs higher than 0.5  Low cf and involved less than 1% of the population   With threshold = 3 on height 65% accuracy  Simpler rules  RELATED WORK PREDICTIVE RULES FROM DECISION TREES 2. Items (binary variables) as independent variables like association rules are used With threshold = 3 on height    Most of the rules were much closer to the prediction requirements 10 nodes DISCUSSION  Decision trees       are not as powerful as association rules in this case Do not work well with combinations of several target variables Fail to identify many medically relevant combinations of independent numeric variable ranges and categorical values Tend to find complex and long rules, if the height is unlimited Find few predictive rules with reasonably sized (>1%) sets of patients in such cases Rules some times repeat the same attribute DISCUSSION - ALTERNATIVES  build many decision trees with different independent attributes   Create a family of small trees, each tree has a weight   It’s error-prone, difficult to interpret, slow for higher # of attributes Each tree becomes similar to a small set of association rules Constraints for association rules can be adopted to decision trees (future work) DISCUSSION – DECISION TREE ADVANTAGES    DT partitions the data set, ARs on the same target attributes may refer to overlap DT represents a predictive model of data set, ARs are disconnected among themselves DT is guaranteed to have at least 50% prediction accuracy and generally above 80% for binary target variables, ARs require trial and error to find the best threshold