* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download 235_lecture6_080219
Birthday problem wikipedia , lookup
Corecursion wikipedia , lookup
Generalized linear model wikipedia , lookup
Regression analysis wikipedia , lookup
Least squares wikipedia , lookup
Fisher–Yates shuffle wikipedia , lookup
Hardware random number generator wikipedia , lookup
Probability box wikipedia , lookup
Psyc 235: Introduction to Statistics DON’T FORGET TO SIGN IN FOR CREDIT! (and check out the lunar eclipse on Thursday from about 9-10) Graded Assessment AL1: Feb 25th & BL1: Feb 27th • In Psych 289: anytime between 9-5 • Can sign up for guaranteed spot:  9am, 11:30am, 2pm, 4:30pm  Sign up in lab or Office Hours (Thurs Rm 25) • Bring ID and notes. Graded Assessment • ALEKS will be unavailable:  AL1: 8am Mon - 11:59pm Wed  BL1: 8am Wed - 11:59pm Fri • Conflict/Makeup exams:  must be within that window  let us know ASAP (as in TODAY) Quiz on this Thursday • Use this quiz as practice exam for the assessment. • Get your notes ready beforehand. • Complete like an assessment • Make note of trouble areas, additional notes you would like, etc. • Can do the quiz in office hours and then ask Jason questions Questions? Intersection & Union • Intersection:  P(A  B) = P(A)*P(B)  (If mutually exclusive = 0) • Union:  P(A U B) =P(A)+P(B)- P(AB) • Compliment:  p(A)=1-p(A) Independent vs. Dependent Events • Independent Events: unrelated events that intersect at chance levels given relative probabilities of each event • Dependent Events: events that are related in some way Concepts of union and intersection are the same  However, P(AB)  P(A)*P(B) • Do you think mutually exclusive events are dependent or independent? Conditional Probability p(B|A) = p(BA) p(A) Conceptually this means: A B Baye’s Theorem p(B|A) = p(A|B)p(B) p(A|B)p(B) + p(A|B)p(B) • Can we break this down a little to understand it better? • p(A|B)*p(B)=p(AB) A • p(A|B)*p(B) + p(A|B)*p(B) = p(AB) + p(AB) = p(A) p(B|A) = •So, this is just: p(BA) p(A) B Law of Total Probabilities A B • p(A) = p(AB) + p(AB) • p(A) = p(A|B)p(B) + p(A|B)p(B) _ B Random Variables • Where are we? • In set theory, we were talking about theoretical variables that only took on two values: either a 0 or 1. They were in the group or not. • Now we’re going to talk about variables that can take on multiple values. Random Variables • But wait, didn’t we already talk about variables that had multiple values? • When we were talking about central tendancy and dispersion, we were talking about specific distributions of data…now we’re going to start discussing theoretical distributions. Data World vs. Theory World • Theory World: Idealization of reality (idealization of what you might expect from a simple experiment)  Theoretical probability distribution • Data World: data that results from an actual simple experiment  Frequency distribution But First… • Before we get into random variables, we need to spend a little bit of time thinking about:  the kinds of values variables can take on  what those values mean  how we can combine them 4 Standard Scales • Categorical (Nominal) Scale  Numbers serve only as labels  Only relevant info is frequency • Ordinal Scale  Things that are ranked  Numbers give you order of items, but not distance between/relation between • Interval Scale  Scale with arbitrary 0 point and arbitrary units  However, units give you proportional relationship between values • Ratio Scale  Scale has an absolute 0 point  Intervals between units is constant What kind of scale is this? • • • • • • Temperature Grades Number Scale Terror Alert Scale Class Rank What are other scales you are familiar with? Discrete vs. Continuous Random Variables • Discrete  Finite number of outcomes  (x = sum of dice)  Countable infinite number of outcomes  Numbers from 1 to infinity • Continuous  Uncountably Infinite  (x=number of flips to get a head)  (Convergent series: the sum of 1-infinity approaches some value) Probability Density Distributions • Discrete: draw on board  Probability mass function • Continuous  (x= spot where pointer lands)  Probability mass funtion Next: • Now that we know more about random variables, we can apply everything that we’ve learned so far. • Graphing and displaying data • Central tendency & dispersion • Transformations of mean and variance • Contingency Tables Central Tendency in Random Variables • E(x)=∑(X*p(x)) • Var(x)=∑((X-E(x))2*p(x)) Properties of Expectation • • • • • E(a)=a E(aX)=a*E(X) E(X+Y)=E(X)+E(Y) E(X+a)=E(X)+a E(XY)=E(X) * E(Y) Properties of Variance • • • • • Var(aX)=a2Var(X) Var(X+a)=Var(X) Var(X-a)=Var(X) Var(X+Y)= Var(X) + Var(Y) Var(X2)=E(X)+Var(X)2 Contingency Tables for 2 random variables A(yes) not A (no) B (yes) p(a) p(b) p(a+b) not B(no) p(c) p(d) p(c+d) p(a+c) p(b+d) 1 • A is facilitative of B when p(B|A)>P(B) • A is inhibitory for B when p(B|A)<P(B) Remember • 1st Exam Feb 25/27  Sign up for exam timeslots in lab Wed or Office Hours Thurs  (or also first-come-first-served on exam day) • Quiz on Thursday
 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
									 
                                             
                                             
                                             
                                             
                                             
                                             
                                             
                                             
                                             
                                             
                                            