Parameter: Characteristics that are used to describe the population.
Statistic: a function of the observable random variables in a sample which does not include any unknown quantities.
Estimator: A statistic that is used to estimate an unknown parameter.
Point Estimation Cont…
Parameter
Estimator
Population mean \(\mu\)
Sample mean \(\bar{x}\)
Population variance \(\sigma^2\)
Sample variance \(s^2\)
Population proportion \(p\)
Sample proportion \(\hat{p}\)
Maximum Likelihood Estimators
The point in the parameter space that maximizes the likelihood function.
Likelihood function is given by; \[𝐿(𝑥,\theta)=\prod_{𝑖=1}^𝑛𝑓(𝑥_i,\theta)\]
The idea of maximum likelihood estimation is to first assume our data come from a known family of distributions that contain parameters.
Then the maximum likelihood estimates (MLEs) of the parameters will be the parameter values that are most likely to have generated our data.
Example 1
Consider a simple coin-flipping example. Let’s say we flipped a coin 200 times and observed 103 heads and 97 tails. If the probability of “success” (i.e. getting a head) is \(p\),
Define a function that will calculate the likelihood function for a given value of \(p\); then
Search for the value of \(p\) that results in the highest likelihood.
Example 2
Suppose we have data points representing the weight (in kg) of students in a class.
This dataset appears to follow a normal distribution. Find the MLEs for the mean and standard deviation for this distribution?
Normal distribution - Maximum Likelihood Estimation
The MLE of \(\mu\) is defined as \(\hat{\mu}_{MLE}=argmax(x_1,...,x_n|\mu,\sigma^2)\); where \(\hat{\mu}_{MLE}\) is the value of \(\mu\) that maximizes the likelihood function.
If we maximise the above likelihood function, we get \(\hat{\mu}_{MLE}=\bar{x}.\)
Since the MLE of \(\mu\) is the sample mean, computing the MLE in R becomes straightforward.
Interval Estimation
Point estimators are often use as sample measures for population parameters.
It is also helpful to know how reliable this estimate is, that is, how much sampling uncertainty is associated with it.
A useful way to express this uncertainty is to calculate an interval estimate or confidence interval for the population parameter
In other words, the confidence interval is of the form “point estimate ± uncertainty”
Confidence Interval for Mean
Case 1: When data is normal/ large sample and \(\sigma\) is known.
\[\bar{x}\pm z_{\alpha/2}\sigma/\sqrt{n}\]
Case 2: When data is normal/ large samples and \(\sigma\) is unknown.
\[\bar{x}\pm t_{n-1,\alpha/2}\sigma/\sqrt{n}\]
Case 3: When data is non-normal/ small samples
For this, bootstrap approach is used as follows.
CONFIDENCE INTERVALS FOR Difference of Means
Case 1: Sampling from two independent normal distributions with known variances.