Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Chapter 7 Point Estimation of Parameters Learning Objectives • Explain the general concepts of estimating • Explain important properties of point estimators • Know how to construct point estimators using the method of maximum likelihood • Understand the central limit theorem • Explain the important role of the normal distribution Statistical Inference • Used to make decisions or to draw conclusions about a population • Utilize the information contained in a sample from the population • Divided into two major areas – parameter estimation – hypothesis testing • Use sample data to compute a number • Called a point estimate Statistic and Sampling Distribution • Obtain a point estimate of a population parameter • Observations are random variables • Any function of the observation, or any statistic, is also a random variable • Sample mean and sample variance are statistics • Has a probability distribution • Call the probability distribution of a statistic a sampling distribution Definition of the Point Estimate • Suppose we need to estimate the mean of a single population by a sample mean – Population mean, , is the unknown parameter – Estimator of the unknown parameter is the sample mean X – X is a statistic and can take on any value • Convenient to have a general symbol • Symbols are used in parameter estimation – Unknown population parameter id denoted by – Point estimate of this parameter by ˆ – Point estimator is a statistic and is denoted by ̂ General Concepts of Point Estimation • Unbiased Estimator – Estimator should be “close” to the true value of the unknown parameter – Estimator is unbiased when its expected value is equal to the parameter of interest ˆ E () – Bias is zero • Variance of a Point Estimator – Considering all unbiased estimators, the one with the smallest variance is called the minimum variance unbiased estimator (MVUE) – MVUE is most likely estimator that gives a close value to the true value of the parameter of interest Standard Error • Measure of precision can be indicated by the standard error • Sampling from a normal distribution with mean and variance 2 • Distribution of X is normal with mean and variance 2/n • Standard error of X x n • Not know , we will substitute the s into the above equation s ̂ x n Mean Square Error (MSE) • It is necessary to use a biased estimator • Mean square error of the estimator can be used • Mean square error of an estimator is difference between the estimator and the unknown parameter ˆ ) E ( ˆ )2 MSE ( Eq.7-3 • An estimator is an unbiased estimator – If the MSE of the estimator is equal to the variance of the estimator – Bias is equal to zero Relative Efficiency • Suppose we have two estimators of a parameter with their corresponding mean square errors • Defined as ˆ MSE ( 1 ) ˆ ) MSE ( 2 • If this relative efficiency is less than 1 • Conclude that the first estimator give us a more efficient estimator of the unknown parameter than the second estimator • Smaller mean square error Example • Suppose we have a random sample of size 2n from a population denoted by X, and E(X)= and V(X)= 2 • Let be two estimators of • Which is the better estimator of ? Explain your choice. Solution • Expected values are 2n Xi 1 2n 1 2n E X 1 E i 1 E X i 2n 2n i 1 2n n Xi 1 n 1 E X 2 E i 1 E X i n n n i 1 n • X and X are unbiased estimators of • Variances are 1 2 V X 1 • MSE 2 2n V X 2 2 n MSE (ˆ1 ) 2 / 2n n 1 2 MSE (ˆ2 ) / n 2n 2 • Conclude thatX is the “better” estimator with the smaller variance 1 Methods of Point Estimation • Definition of unbiasness and other properties do not provide any guidance about how good estimators can be obtained • Discuss the method of maximum likelihood • Estimator will be the value of the parameter that maximizes the probability of occurrence of the sample values Definition • Let X be a random variable with probability distribution f(x;) – is a single unknown parameter – Let x1, x2, …, xn be the observed values in a random sample of size n – Then the likelihood function of the sample L()= f(x1:). f(x2:). … f(xn:) Sampling Distribution of Mean • Sample mean is a statistic – Random variable that depends on the results obtained in each particular sample • Employs a probability distribution • Probability distribution of X is a sampling distribution – Called sampling distribution of the mean Sampling Distributions of Means • Determine the sampling distribution of the sample mean X • Random sample of size n is taken from a normal population with mean and variance 2 • Each observation is a normally and independently distributed random variable with mean and variance 2 Sampling Distributions of MeansCont. • By the reproductive property of the normal distribution x X 1 X 2 .... X n n • X-bar has a normal distribution with mean x ... n • Variance x 2 2 2 .... 2 n 2 2 n Central Limit Theorem • Sampling from an unknown probability distribution • Sampling distribution of the sample mean will be approximately normal with mean and variance 2/n • Limiting form of the distribution of X Z X / n • Most useful theorems in statistics, called the central limit theorem • If n 30, the normal approximation will be satisfactory regardless of the shape of the population Two Independent Populations • Consider a case in which we have two independent populations – First population with mean 1 and variance 21 and the second population with mean 2 and variance 22 – Both populations are normally distributed – Linear combinations of independent normal random variables follow a normal distribution • Sampling distribution of and variance X1 X 2 is normal with mean x x x x 1 2 1 2 2 x1 x2 1 2 2 x1 2 x2 12 n1 22 n2