Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
159.410/710 User Interface Design © Paul Lyons 2010 Epistemology Approaches to knowledge Engineers Holistic Constructionist Scientists Reductionist Analyst/synthesists Arts Complexity Subjectivists I hear and I forget. I see and I remember. I do and I understand. Confucius (attributed) ~ 3 ~ 159.410/710 User Interface Design Epistemology Types of HCI research Development of interaction widgets Usability – efficiency, enjoyability Internet & web Social applications Mobile applications – shoehorning complex applications onto tiny screens ~ 4 ~ 159.410/710 User Interface Design Characteristics of HCI research HCI research has a focus on people what computers can do is not the main point what computers can help people do is variety of contributing disciplines sociology psychology statistics computer science observation techniques controlled experiments handing noisy data developing (genuinely) new interface paradigms rigorous research methodologies are required it isn’t enough to develop a new interface or a new interface component does the new interface make things better? how do you know? are you sure? ~ 5 ~ 159.410/710 User Interface Design Characteristics of HCI research Things to measure time to complete a task number of tasks completed in a standard time performance measures largely industry-driven accuracy of performing a task accuracy of performing a task enjoyment emotional wellbeing why people choose to spend discretionary time using computers difficult to measure in a laboratory setting e.g. contributing to Wikipedia why people choose to stop using applications people’s usage patterns of mobile computing devices and social apps ~ 6 ~ 159.410/710 User Interface Design Characteristics of HCI research Replication of results Multiple studies that reach the same or similar conclusion Triangulation by different research methods if a single method produces identical results repeatedly, the reason may be a flawed method Results may change over time reasons for using a computer 1980s vs. 2000s finding information – searching and tagging vs. hierarchical directories ~ 7 ~ 159.410/710 User Interface Design Characteristics of HCI research Tradeoffs speed vs accuracy (Fitt’s Law) better interface vs. familiar interface more efficient keyboard vs. QWERTY keyboard iPad is cool and new it’s the coolth that persuades people to adopt it how do you measure that? security vs. usability eye-scans and fingerprints? a revolutionary, undeniably better computer vs. environmental costs of computer disposal ~ 8 ~ 159.410/710 User Interface Design Characteristics of HCI research HCI is an interdisciplinary discipline in the past human factors engineering psychology all suit experimental design in the present in the future ubiquitous? Virtual Reality? mind-activated? library science information science art and design competition with judges (cf. architecture)? reductionist widely accepted: statistical tests control groups reliable more holistic more subjective less trusted (not less trustworthy) ~ 9 ~ 159.410/710 User Interface Design Epistemology Types of model generative produce principles and guidelines or actual systems (e.g. Colour Harmoniser) prescriptive suggest ways of building things (e.g. patterns) predictive allow us to plan for the future explanatory explain what causes the data that have been observed descriptive generalisations about the data – allow us to see order amidst chaos ~ 10 ~ 159.410/710 User Interface Design Experimental Research Usability Testing The goal of usability testing is simply to find flaws in a specific interface A small number of users may take part … it can be structured or unstructured. … there is no claim that the results can be generalised. The goal is simply to find flaws and help the developers improve the interface. If that involves jumping in and helping a user or changing the task mid-process, that is acceptable. Lazar, Feng and Hocheiser Research Methods in HCI 2010 ~ 11 ~ 159.410/710 User Interface Design Experimental Research HCI research – 57 varieties observations field studies surveys usability studies interviews focus groups controlled experiments rich, not reproducible reproducible, reductionist descriptive research; observations – may be quantitative and accurate relational research; establishes correlations between factors – does not establish causality typing speed correlated with hours spent gaming does time spent gaming improve typing? are good typists successful gamers? experimental research; can establish causality allocate users to two groups randomly expose one group to games, the other not measure typing ability of groups after a suitable interval ~ 12 ~ 159.410/710 User Interface Design Experimental Research Null and alternative hypotheses H0 nochange effect no in speed no change in user satisfaction the widgetcauses… causes… thenew treatment the mutual exclusion seesaw H1 some an effect change in speed and also some change in user satisfaction However… testing multiple hypotheses can complicate controls and variables a good hypotheses is clear and unambiguous clearly distinguishes between independent and dependent variables is testable in a single experiment clearly identifies control groups and conditions of experiment generally derives from preliminary observational studies each combination of independent variables is a condition ~ 13 ~ 159.410/710 User Interface Design Experimental Research Independent and dependent variables “cause” “effect” independent variable dependent variable variations in value are under the experimenter’s control variations in value are observed Null hypothesis: there is no speed change between the original widget and the new widget experimenter measures this it’s the dependent variable if the experimental results are plotted on a graph independent variable goes on the x-axis dependent variable goes on the y-axis experimenter makes choice of widget it’s the independent variable dependent variable independent variable ~ 14 ~ 159.410/710 User Interface Design Experimental Research Typical independent variables Typical dependent variables Technology Efficiency typing vs. speech mouse vs. joystick, touchpad etc time to complete a task, speed Accuracy Design error rate pull-down vs. pop-up menu colour scheme layout Subjective Satisfaction Demographic Ease of learning and retention rate Likert scale ratings gender, age, experience, education time to learn, loss after a week, a month Context Cognitive Demand lighting, noise, seated vs standing, other people in the vicinity time before onset of fatigue ~ 15 ~ 159.410/710 User Interface Design Experimental Research Components of an experiment Treatments randomisation is often necessary things we want to compare (cf. medical treatments) compare two splines A and B for a CAD tool use a within-subjects design measure time-to-complete task with A, then B flaw: subjects learnt the task, so B is best solution; randomise order of tasks. Units “things” that treatment is applied to (normally human subjects) comparing two treatments using a between-subjects design allocate subjects to treatment A, till enough then allocate subjects to treatment B Assignment method how subjects are assigned to treatments ~ 16 ~ flaw: A is applied to early birds, B to late sleepers Solution: randomise allocation to the treatments 159.410/710 User Interface Design Experimental Research Significance tests this approach depends on being able to distinguish between an effect and no effect how do we decide whether or not an effect is real? we measure the probability that it occurred by chance if that probability is sufficiently low, we say that there is a significant effect. p < 0.05 0.005 says that the probability that the observed behaviour occurred by chance is less than 5% 0.5% or that the probability that the effect is real exceeds 95% 99.5% whether that’s good enough depends on the application for a new drug, a significance level of p < 0.05 is not good enough if the null hypothesis is “the standard dose is not fatal” ~ 17 ~ 159.410/710 User Interface Design Experimental Research Type I errors & Type II errors (aka “false positive”) (aka “false negative”) study concludes widget no better widget is no different widget is better study concludes widget is better type I (gullibility) error type II (blindness) error probability of Type I error = α probability of Type II error = β generally aim for p < 0.05 probability that effect occurred by chance, p-value = α probability of correctly rejecting an incorrect null hypothesis, statistical power of a test = 1- β α and β are related; the less gullible you are, the more likely you are to be blind to improvements keep β low by using large sample sizes probability of finding an effect that does exist ~ 18 ~ 159.410/710 User Interface Design Experimental Research Limitations of experimental research controlled experiments are a very powerful technique but hypothesis must be well-defined number of variables must be limited, preferably orthogonal HCI problems can be difficult to define many, interrelated factors may be involved factors other than independent variables may not affect dependent variables e.g. difficult to factor out familiarity with technology in an age-related study prescreen to ensure homogeneity between subject groups use statistical techniques designed to filter out confounding factors (analysis of covariables) subjects’ behaviour in a lab differs from behaviour in real world ~ 19 ~ 159.410/710 User Interface Design Experimental Design True Experiments <x> is an intuitive interface a testable hypothesis subjects will be able to use <x> correctly it in under 1 minute all of them? two conditions (one treatment, one control) 50% > 75% sometimes more quasi-experiment random assignment of subjects no? non-experiment quantitative measurements e.g. not ethical to randomly assign children to parents to study effect of single-parent upbringing significance tests attention to bias elimination replicable ~ 20 ~ 159.410/710 User Interface Design Experimental Design Other types of experiment quasi-experiments (subjects not randomly assigned) may be necessary for practical or ethical reasons can still produce useful results but more susceptible to confounding factors non-experiments (no control group) insufficient subjects – use what’s available researcher lacks influence (modified Word interface) may be necessary for practical or ethical reasons can still produce useful results but even more susceptible to confounding factors e.g. usability trials – aim is to detect problems formal experiments are designed to detect subtle effects to factor out researcher bias researcher’s specialist knowledge may trump population’s preferences (e.g. user surveys for Xerox showed little demand for such a device) is demonstrating that it is possible to build something a valid experiment? engineering research often stops at this point ~ 21 ~ 159.410/710 User Interface Design Experimental Design Important considerations number of independent variables Hypothesis: There is no difference between target selection speed when using a mouse, a joystick, or a trackball to select icons of different sizes (small, medium, large) How many independent variables? number of conditions 3x3=9 number of dependent variables 1 • type of pointing device • icon size measurement may need careful thought e.g. is typing speed wpm or error-free wpm? is speech recogniser error rate definitive? ~ 22 ~ 159.410/710 User Interface Design Experimental Design Structure of an experiment basic design factorial design withingroup 1 independent variable >1 independent variable only one group but subjects experience multiple conditions eliminates individual differences smaller population required learning and fatigue may cause effects one group per condition each subject experiences only one condition betweengroup splitplot no learning effect less fatigue effect susceptible to differences between groups mix of within-group and between-group ~ 23 ~ Order tasks using a Latin square to factor out fatigue Subj1 1 2 3 Subj2 2 3 1 Subj3 3 1 2 randomise order of conditions and/or provide preliminary training tasks with large diffs between individuals e.g. cognitively complex tasks suits small subject pools requires big groups, randomly selected cognitively simple tasks, no learning effect (inter-subject diffs increase with complexity) tasks where subject difference is indep var. effect of using GPS (binary, within-group) on three age-groups (between-groups) 159.410/710 User Interface Design Experimental Design Watch out for interaction effects if the effect of a variable depends on the value of the other variable, the variables interact the variables are (should be) independent, but their effects interact task duration complex task simple task Office 2003 genuinely independent variables Office 2007 ~ 24 ~ 159.410/710 User Interface Design Experimental Design Reliability of experimental results random errors research involving human subjects is noisy noise value actual value observed value = actual value + random error (noise) sample size with increased sample size, actual values add, relative size of noise tends to 0 systematic errors the same each time – not cancelled by large sample more deleterious than noise ~ 25 ~ 159.410/710 User Interface Design Experimental Design Systematic errors instrumentation errors can often replace instruments (e.g. stopwatch) with software experimental procedure non-random task condition order allows learning & fatigue to have an effect opposite systematic errors instructions may introduce errors complete the task as fast as possible vs. take your time, no rush produced different results instructions from different members of research team may differ subjects under time stress were slower! trivial details data entry on a PDA – holding PDA in hand produced different results from sitting PDA on table randomise conditions and tasks when using within-group design use identical instructions for all participants – written or recorded run pilot studies beforehand to detect potential biases don’t want to realise half-way through the experiment that all the results are compromised you have overlooked something use real participants from target population ~ 26 ~ 159.410/710 User Interface Design Experimental Design Systematic errors instrumentation errors can often replace instruments (e.g. stopwatch) with software participants experimental procedure non-random task condition order allows learning & fatigue to have an effect opposite systematic errors instructions may introduce errors complete the task as fast as possible vs. take your time, no rush produced different results instructions from different members of research team may differ trivial details data entry on a PDA – holding PDA in hand produced different results from sitting PDA on table randomise conditions and tasks when using within-group design use identical instructions for all participants – written or recorded run pilot studies beforehand to detect potential biases don’t want to realise half-way through the experiment that all the results are compromised you have overlooked something use real participants from target population ~ 27 ~ 159.410/710 User Interface Design Experimental Design Systematic errors instrumentation errors experimental procedure participants age bias education bias (particularly prevalent in university studies) interest in the product (or its domain) recruit a set of participants representative of target population may be quite skewed - e.g. for elder-care systems don’t stress the participants explain that the system is under test, not them any result they produce is good organise schedule conservatively so participants aren’t inconvenienced it’s polite, and it produces better results! ~ 28 ~ 159.410/710 User Interface Design Experimental Design Systematic errors instrumentation errors experimental procedure participants age bias education bias (particularly prevalent in university studies) interest in the product (or its domain) recruit a set of participants representative of target population may be quite skewed - e.g. for elder-care systems don’t stress the participants explain that the system is under test, not them any result they produce is good organise schedule conservatively so participants aren’t inconvenienced it’s polite, and it produces better results! ~ 29 ~ 159.410/710 User Interface Design Experimental Design Systematic errors instrumentation errors experimental procedure participants experimenter behaviour express no opinion about the system maintain noncommittal body language be ready to start on time use the same experimenter each time, if possible, or a recorded protocol if multiple experimenters are necessary, require them to follow a written experimental protocol ~ 30 ~ 159.410/710 User Interface Design Experimental Design Systematic errors instrumentation errors experimental procedure participants experimenter behaviour express no opinion about the system maintain noncommittal body language be ready to start on time use the same experimenter each time, if possible, or a recorded protocol if multiple experimenters are necessary, require them to follow a written experimental protocol ~ 31 ~ 159.410/710 User Interface Design Experimental Design Systematic errors instrumentation errors experimental procedure participants experimenter behaviour environmental factors physical environment: noise, temperature , humidity, lighting, vibration social environment: people nearby, power relationships of participants and people nearby, interruptions quiet room suitable lighting comfortable furniture non-distracting environment observation by CCTV or from behind 1-way mirror, if possible for field studies, visit the location beforehand to check for problems ~ 32 ~ 159.410/710 User Interface Design Experimental Design Experimental Procedures 1. 2. 3. 4. 5. 6. 7. Identify a research hypothesis Design the study Run a pilot Recruit participants Run data collection sessions Analyse the data Report the conclusions 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. Set up the experimental environment/equipment Greet participants Outline the purpose of the study and the procedures Obtain participants’ consent Assign participants to experimental condition Participants complete pre-survey (if any) Participants complete training task Participants complete survey task Participants complete post-survey (if any) Debrief (can be more useful than formal survey) ~ 33 ~ 159.410/710 User Interface Design Analysing the Data There are many analytical tools independent samples t-test paired-samples (e.g. before/after tests) one-way ANOVA factorial repeated measures correlation regression Chi-square ~ 34 ~ 159.410/710 User Interface Design Analysing the Data Data Preparation error checking and correction incorrect grouping of survey forms computing experience > age impossible age paper forms needs checking survey software (or Excel) could check at data collection time age: 23½, 23 years and 7 months, nearly twenty-four if data can’t be corrected, may need to be thrown away e.g. because subjects are anonymous may need pre-processing coding text as numbers (e.g. 1, 2, 3, for no degree, bachelors, P/G) extracting general themes from individual interviews coding interaction events (e.g. click(100, 250) “select book”) consistency may need to be verified if > 1 coder analysis may require data to be restructured related information in different surveys (pre & post trial, for example) analysis software requires specific formatting SPSS independent samples and paired samples t-tests use same data in 1 column and 2 parallel columns ~ 35 ~ 159.410/710 User Interface Design Analysing the Data Start with an exploratory analysis good 50 forthinitial percentile comparison - good for between most skewed popular groups (e.g. Pareto) distribution and when outliers may be errors Get a feel for the data mean, median, mode range, variance, standard deviation Box-and-whisker plots Histograms measures of central tendency measures of spread datamax – data estimates spread – likely around to increase mean: sign-independent, with sample size &but sensitive has different to outliers) units from samples min (crude more complex measures of spread assume a normal distribution may be necessary to modify data to conform variance n ∑ (xi – x)2 i=1 ~ 36 ~ (n – 1) 159.410/710 User Interface Design Analysing the Data Start with an exploratory analysis Get a feel for the data mean, median, mode range, variance, standard deviation measures of central tendency measures of spread estimates spread around mean: sign-independent, but has different units from samples more complex measures of spread assume a normal distribution may be necessary to modify data to conform variance n ∑ (xi – x)2 i=1 ~ 37 ~ (n – 1) 159.410/710 User Interface Design Analysing the Data Start with an exploratory analysis Get a feel for the data mean, median, mode range, variance, standard deviation measures of central tendency measures of spread same units as samples more complex measures of spread assume a normal distribution may be necessary to modify data to conform √ s n ∑ (xi – x)2 i=1 (n – 1) standard deviation measures mean deviation of samples from mean of samples ~ 38 ~ 159.410/710 User Interface Design Analysing the Data Mean differences treatments are different two groups, same task comparing two search engines differences between means difference between treatments one group, two tasks Δ=5 Δ=5 significance tests are necessary to determine probability that difference is due to chance are these means different? are these means different? test IVs conditions per IV between-groups 1 1 ≥2 2 ≥3 ≥2 independent-samples t-test 1-way ANOVA factorial ANOVA within-group 1 1 ≥2 2 ≥3 ≥2 paired-samples t-test repeated measures ANOVA repeated measures ANOVA mixed ≥2 ≥2 split-plot ANOVA ~ 39 ~ 159.410/710 User Interface Design Analysing the Data To compare 2 means use a t-test null hypothesis task completion times for subjects using word-prediction software do not differ from task-completion times for subjects who do not use the software signal x1 – x2 t= noise = √ s12 + s 22 n1 n2 signal noise s2 is, of course, the variance remember: same as p-value generally say there’s a significant effect if α ≤ 0.05 for 2 gps however, significance of a particular t depends on size of subject groups specifically degrees of freedom, df = total participants – number of groups = n1 + n2 -2 consult published tables showing α value for particular (t, df) combinations statistical software usually outputs α from builtin tables ~ 40 ~ 159.410/710 User Interface Design Analysing the Data To compare 2 means use a t-test null hypothesis task completion times for subjects using word-prediction software do not differ from task-completion times for subjects who do not use the software signal x1 – x2 t= noise = √ s12 + s 22 n1 n2 signal noise generally say there’s a significant effect if α ≤ 0.05 however, significance of a particular t depends on size of subject groups specifically degrees of freedom, df = total participants – number of groups = n1 + n2 -2 consult published tables showing α value for particular (t, df) combinations statistical software usually outputs α from builtin tables ~ 41 ~ 159.410/710 User Interface Design Analysing the Data To compare 2 means use a t-test null hypothesis task completion times for subjects using word-prediction software do not differ from task-completion times for subjects who do not use the software for unrelated samples use independent-samples t-test for a single group use paired-samples t-test times for group using word-prediction software times for group uses conventional software times for subject using word-prediction software and for same subject uses conventional software SPSS t-test data comprises times and group membership SPSS t-test data comprises times with software and times without software t-value t-value high t-value high P(null hypothesis false) ~ 42 ~ 159.410/710 User Interface Design Analysing the Data What if hypothesis predicts sign of difference? if we know that sign of effect will be + or – instruct analysis software to use a 1-tailed t-test α = 0.1 indicates same level of confidence as α = 0.05 for 2-tailed test Do NOT use a one-tailed t-test because 2-tailed test indicates no significance test should be hypothesis-driven, not data-driven! ~ 43 ~ 159.410/710 User Interface Design Analysing the Data ANOVA: within-gp variances vs. population variance null hypothesis: sample sets A, B, C & D belong to 1 population if smeans for means of sample sets A - D > scombined population there is more than 1 population F= found variation in averages/expected variation in averages F=1 supports null hypothesis x1i - x1 Σ(x i )2 x1 SS1 Sum of Squares1 x2 SS2 ~ 44 ~ 159.410/710 User Interface Design Analysing the Data ANOVA: within-gp variances vs. population variance null hypothesis: sample sets A, B, C & D belong to 1 population if smeans for means of sample sets A - D > scombined population there is more than 1 population F= found variation in averages/expected variation in averages ~ 45 ~ 159.410/710 User Interface Design Analysing the Data ANOVA: within-gp variances vs. population variance null hypothesis: sample sets A, B, C & D belong to 1 population if smeans for means of sample sets A - D > scombined population there is more than 1 population F= found variation in averages/expected variation in averages within-groups variability (aka error variance) variability due to differences between means (aka effect) if effect variance is large w.r.t. error variance treated group & untreated groups act as different populations (treatment has an effect) Group 1 Group 2 Observation 1 Observation 2 Observation 3 2 3 1 6 7 5 Mean Sums of Squares (SS) 2 2 6 2 Overall Mean Total Sums of Squares 4 28 much larger differences between means than in the diagram MAIN EFFECT ANOVA determines p taking df into account ~ 46 ~ SS Effect Error 24.0 4.0 df 1 4 MS F p 24.0 1.0 24.0 .008 159.410/710 User Interface Design Analysing the Data Use ANOVA (aka F-test) to compare means of ≥ 2 groups the parameter actually generated by the calculation (cf t-test t) we’ve already seen the special case of ANOVA for comparing 2 means: the t-test design IVs conditions 1-way ANOVA between-group 1 ≥3 factorial ANOVA between-group ≥2 repeated measures ANOVA within-group split-plot ANOVA between-group and within-group ~ 47 ~ (F = t2) 159.410/710 User Interface Design Analysing the Data Use ANOVA (aka F-test) to compare means of ≥ 2 groups we’ve already seen the special case of ANOVA for comparing 2 means: the t-test design IVs conditions 1-way ANOVA between-group 1 ≥3 factorial ANOVA between-group ≥2 repeated measures ANOVA within-group split-plot ANOVA between-group and within-group ~ 48 ~ 159.410/710 User Interface Design Analysing the Data Use ANOVA (aka F-test) to compare means of ≥ 2 groups we’ve already seen the special case of ANOVA for comparing 2 means: the t-test 1-way ANOVA design IVs conditions between-group 1 ≥3 SPSS data input for 1-way ANOVA (pared down to the minimum) 245 236 321 . . 246 213 265 . . 178 289 222 . 0 0 0 . . 1 1 1 . . 2 2 2 . task durations code SPSS output from the analysis standard text entry (control group) sum of sqs between-group within-group df Mean sq F significance 7842.250 2 3921.125 2.174 0.139 37880.375 21 1803.827 text-prediction So, how would we summarise this thesis or report? the statistical calculation greater than 0.05in F-value not significant same calculation produces 2produces sets of aresults – these aren’t relevant dictation significance obtained by table lookup A 1-way ANOVA analysis with text-entry method as independent variable and task completion time as dependent variable suggests there is no significant difference between the three conditions: (F(2, 21) = 2.174, n.s.) 24 samples in 3 groups gives df = 21 ~ 49 ~ 159.410/710 User Interface Design Analysing the Data Use ANOVA (aka F-test) to compare means of ≥ 2 groups we’ve already seen the special case of ANOVA for comparing 2 means: the t-test 1-way ANOVA design IVs conditions between-group 1 ≥3 SPSS data input for 1-way ANOVA (pared down to the minimum) 245 236 321 . . 246 213 265 . . 178 289 222 . 0 0 0 . . 1 1 1 . . 2 2 2 . task durations code SPSS output from the analysis standard text entry (control group) sum of sqs between-group within-group df Mean sq F significance 7842.250 2 3921.125 2.174 0.139 37880.375 21 1803.827 text-prediction So, how would we summarise this in a thesis or report? dictation A 1-way ANOVA analysis with text-entry method as independent variable and task completion time as dependent variable suggests there is no significant difference between the three conditions: (F(2, 21) = 2.174, n.s.) 24 samples in 3 groups gives df = 21 ~ 50 ~ 159.410/710 User Interface Design Analysing the Data Use ANOVA (aka F-test) to compare means of ≥ 2 groups we’ve already seen the special case of ANOVA for comparing 2 means: the t-test design IVs conditions 1-way ANOVA between-group 1 ≥3 factorial ANOVA between-group ≥2 SPSS data entry format Q: does nature of the task (composition or transcription) affect performance? data entry 2 method 1 0 SPSS fn is called Univariate analysis task time task type data entry method task type 0 1 245 236 … 246 213 … 178 289 … 256 269 … 265 232 … 189 321 gp1 gp2 dictation gp3 gp4 predictive gp5 gp6 standard ~ 51 ~ 0 0 … 0 0 … 0 0 … 1 1 … 1 1 … 1 1 0 0 … 1 1 … 2 2 … 0 0 … 1 1 … 2 2 159.410/710 User Interface Design Analysing the Data Use ANOVA (aka F-test) to compare means of ≥ 2 groups we’ve already seen the special case of ANOVA for comparing 2 means: the t-test design IVs conditions 1-way ANOVA between-group 1 ≥3 factorial ANOVA between-group ≥2 SPSS data entry format Q: does nature of the task (composition or transcription) affect performance? data entry 2 method 1 0 SPSS fn is called Univariate analysis task time task type data entry method task type 0 1 245 236 … 246 213 … 178 289 … 256 269 … 265 232 … 189 321 gp1 gp2 dictation gp3 gp4 predictive gp5 gp6 standard ~ 52 ~ 0 0 … 0 0 … 0 0 … 1 1 … 1 1 … 1 1 0 0 … 1 1 … 2 2 … 0 0 … 1 1 … 2 2 159.410/710 User Interface Design Analysing the Data Use ANOVA (aka F-test) to compare means of ≥ 2 groups we’ve already seen the special case of ANOVA for comparing 2 means: the t-test design IVs conditions 1-way ANOVA between-group 1 ≥3 factorial ANOVA between-group ≥2 Q: does nature of the task (composition or transcription) affect performance? SPSS output task type entry method interaction task * entry error IVs sum of sq df mean square 2745.188 1 2745.188 1.410 0.242 17564.625 2 8782.313 4.512 0.017 114.875 2 57.437 0.030 0.971 81751.625 42 1946.467 F significance task type caused no significant effect: F(1, 42) = 1.41, n.s entry method had a significant effect: F(2, 42) = 4.51, p < 0.05 there is no significant interaction between task and entry ~ 53 ~ 159.410/710 User Interface Design Analysing the Data Use repeated measures ANOVA for within-group studies previous between-groups design requires lots of participants (72, if 12 subjects/group) what about a within-groups design? specially if only some are eligible – e.g. disabled to study effect of 1 IV, use 1-way repeated measures ANOVA 3 data points from each participant, all in the same row 245 236 321 246 213 265 278 289 222 to study effect of >1 IV, use multi-level repeated measures ANOVA for 3 x 2 factorial study, 6 data points per participant per row transcription composition standard predictive dictation standard predictive dictation participant1 245 246 178 256 265 189 participant2 236 213 289 269 232 321 within-groups design faster less fatigue can control for learning smaller sample ~ 54 ~ 159.410/710 User Interface Design Analysing the Data Assumptions of t tests and F tests no systematic errors e.g. different instructors, with different sets of instructions correlation between errors of participants in each instructor’s group will systematically skew results homogeneity of variance (identical distribution of errors) populations should have comparable variances x1 x2 significantly do these distributions have^different means? not easy to say, either for people or for software normal distribution of errors may be violated if data is highly skewed (non-normal distribution) ~ 55 ~ 159.410/710 User Interface Design Analysing the Data Use Pearson’s r to identify correlations is factora related to factorb? determine Pearson’s product moment correlation coefficient (r) r varies from -1 to 1 -1: perfect negative linear relationship 0: no relationship +1: perfect positive linear relationship computer experience 12 6 3 19 can determine r values for time with time with standard predictive software software 245 236 321 212 246 213 265 189 experience * standard s/ware experience * predictive s/ware standard * predictive experience r experience timestd timepred 1 -0.723 -0.468 -0.723 0.043 1 -0.468 0.325 0.243 0.325 0.432 1 significance timestd timepred r significance r significance (experience, timestd) has significant –ve correlation time with std software decreases with computer experience no other significant correlations ~ 56 ~ 159.410/710 User Interface Design Analysing the Data Use Pearson’s r to identify correlations r2 represents percentage of variance in X that can be explained by variable Y represents percentage of variance in Y that can be explained by variable X but beware: correlation does not imply causation e.g. negative correlation between income and speed of internet search does earning more make you worse at using the internet? or does higher income imply greater age and less familiarity with the internet? higher income age less internet experience lower performance ~ 57 ~ 159.410/710 User Interface Design