MANE 6313
Week 3, Module A
Student Learning Outcome
Analyze simple comparative experiments and experiments with a single factor.
Module Learning Outcome
Review probability distributions.
Simple Comparative Experiments
-
Chapter two only focuses on simple experiments to compare two conditions. Later, we will extend these techniques.
-
Largely review of statistics and hypothesis testing from previous classes.
-
Statistics focuses on two emphases: Descriptive statistics and inferential statistics.
-
In inferential statistics we are concerned with estimation and hypothesis testing.
-
Differences or fluctuations in outcomes from the same treatment are called noise in the results
-
Often this noise is referred to as experimental error or simply error
-
We classify statistical error as those errors that arise from variation that is uncontrolled and generally unavoidable.
-
The presence of noise implies that the response variable is a random variable
Review of Probability Distributions
-
The probability structure of a random variable is defined by its probability distribution.
-
For a discrete random variable
- For a continuous random variable
- The mean of a probability distribution is defined to be
- The expected value operator is \(\(E(y)\equiv \mu\)\)
- The variance of a probability distribution is defined to be
-
The variance can be written in terms of the expectation \(\sigma^2=E\left[(y-\mu)^2\right]\)
-
We will often use the variance operator \(V(y)\equiv E\left[(y-\mu)^2\right] =\sigma^2\)
Common Probability Distributions1
Functions of Random Variables
-
\(E(c)=c\)
-
\(E(y)=\mu\)
-
\(E(cy)=cE(y)=c\mu\)
-
\(V(c)=0\)
-
\(V(y)=\sigma^2\)
-
\(V(cy)=c^2V(y)=c^2\sigma^2\)
For two random variables with the following properties \(E(y_1)=\mu_1\), \(V(y_1)=\sigma_1^2\), \(E(y_2)=\mu_2\) and \(V(y_2)=\sigma_2^2\)
-
\(E(y_1+y_2)=E(y_1)+E(y_2)=\mu_1+\mu_2\)
-
\(V(y_1+y_2)=V(y_1)+V(y_2)+2Cov(y_1,y_2)\) where \(Cov(y_1,y_2)=E\left[(y_1-\mu_1)(y_2-\mu_2)\right]\)
-
\(V(y_1-y_2)=V(y_1)+V(y_2)-2Cov(y_1,y_2)\)
If \(y_1\) and \(y_2\) are independent, then
-
\(V(y_1\pm y_2)=V(y_1)+V(y_2)=\sigma_1^2+\sigma_2^2\)
-
\(E(y_1\cdot y_2)=E(y_1)\cdot E(y_2)=\mu_1\cdot\mu_2\)
In general note that
-
For two random variables to be independent we need to examine the joint distribution function \(f_{y_1,y_2}(y_1,y_2)\) and the two marginal distribution functions \(f_{y_1}(y_1)\) and \(f_{y_2}(y_2)\)
-
We need to show that \(f_{y_1,y_2}(y_1,y_2)=f_{y_1}(y_1)\cdot f_{y_2}(y_2)\)
-
For normally distributed random variables (and only normally distributed r.v.) if \(Cov(y_1,y_2)=0\) then \(y_1\) and \(y_2\) are independent.
Sampling
- We will use the following statistics as point estimators for \(\mu\) and \(\sigma^2\) respectively
$$ \overline{y}=\frac{\sum_{i=1}^ny_i}{n} $$ and $$ S^2=\frac{\sum_{i=1}^n(y_i-\overline{y})^2}{n-1} $$ - Properties of estimators
-
Properties of point estimators
-
Unbiased
-
Minimum variance
-
-
The normal distribution is one of the most important sampling distributions
Central Limit Theorem
If \(y_1,y_2,\ldots,y_n\) is a sequence of \(n\) independent and identically distributed random variables with \(E(y_i)=\mu\) and \(V(y_i)=\sigma^2\) (both finite) and \(x=y_1+y_2+\cdots +y_n\) then $$ z_n=\frac{x-n\mu}{\sqrt{n\sigma^2}} $$ has an approximate \(N(0,1)\) distribution
- What is the implication here?