Chapter 7 - Central Limit Theorem - University of Kentucky
[Pages:10]Topics
Part 1 ? Single measurement
1.
Basic stuff (Chapter 1 and 2)
2.
Propagation of uncertainties (Chapter 3)
Part 2 ? Multiple measurements as independent results
1.
Mean and standard deviation (Chapter 4)
2.
Basic on probability distribution function (not in text explicitly)
3.
The Binomial distribution (Chapter 10)
4.
The Poisson distribution (Chapter 11)
5.
Normal distribution (first half of Chapter 5)
6.
2 test ? how well does the data fit the distribution model? (Chapter 12)
Part 3 ? Multiple measurements as one sample
1.
Central limit theorem (not in text explicitly)
2.
Normal distribution (second half of Chapter 5)
3.
Propagation of error (Chapter 3)
4.
Rejection of data (Chapter 6)
5.
Merging two sets of data together (Chapter 7)
Part 4 - Dependent variables
1.
Curve fitting (Chapter 8)
2.
Covariance and correlation (Chapter 9)
One Single Experiment of n Measurements
In one single experiment, you measure a random variable (x) N times and obtain n results, x1, x2, ......., xN. We know how to: (i) Calculate the mean of these N readings (but do not know why). (ii) Calculate the sample standard deviation of these N readings (but do not know why). (iii) Calculate the population standard deviation to see how spread is these N readings. (iv) Make a histogram and expect it will follow some kind of probability density function.
In here, we just focus on calculating the mean of these N readings. i.e. the mean of these N readings will itself become a random variable if we repeat the experiment.
Many Experiments of n Measurements for Each Experiment
Now we perform the same experiment as in the last slide for n times. In
each experiment, we measure a random variable (x) N times and obtain
n results, x1, x2, ......., xN. We then calculate the mean x of these N readings.
So after N experiments, we will have performed N2 measurements and have N means:
Experiment 1: x11, x21, ......., xN1 Experiment 2: x12, x22, ......., xN2 :
mean x1 mean x2
Experiment N: x1N, x2N, ......., xNN mean xN
New random variable
Central Limit Theorem (Unified Theory of Data Analysis)
For the N readings in each experiment x1, x2, ......., xN they may follow some kind of distribution function, but the mean x of these N experiment will always follow the Normal distribution GX,(x) if N is large (ideally N). Note that the mean of each experiment is now the random variable of this Normal distribution.
Experiment 1: x11, x21, ......., xN1 Experiment 2: x12, x22, ......., xN2 :
mean x1 mean x2
Experiment N: x1N, x2N, ......., xNN mean xN
Follow some kind of distribution function
Follow Normal distribution G X,(x)
But what are the values of X and ?
If you know the distribution function of x1, x2, ......., xN (unlikely)
But what are the values of X and in the Normal
distribution G X,(x)?
Suppose we know the distribution function of x1, x2, ......., xN, say, P(x). This distribution has its own mean x and
variance x2 : x = xP(x) dx
2 x
=
(x - x )2 P(x) dx
Central limit theorem state that:
X = x
2
=
2 x
N
Note: divided by N.
Unfortunately we rarely know the distribution function P(x).
If you don't know the distribution function of x1, x2, ......., xN (likely)
We can guess! Suppose we have x1, x2, ......., xN in one experiment:
Best Estimate of X = x = x1 + x2 +Lx N N
(mean of the set of data)
(Eq. 5.42 of text)
Best Estimate of x (or N ) =
(x1 - x)2 + (x2 - x)2 + L(x N - x)2 (N -1)
(sample standard deviation of the set of data)
(Eq. 5.45 of text)
Note what we are estimating: parameters X and of the normal distribution.
Best Estimate of X (slide 1)
Suppose we have x1, x2, ......., xn in one experiment. For simplicity, we assume (not necessary) the distribution
function of these data is P(x) Gx ,x (x). According to the Central Limit Theorem, the mean of these data point with follow (necessary) Normal distribution function GX , (x).
The theorem also requires x = X and x/ N = .
Pr ob(x1) = x
1
2
e-
(x1 -x)2 2 x 2
dx1
Pr ob(x2 ) = x
1
2
e-
(x2 -x)2 2 x 2
dx 2
M
Pr ob(x N ) = x
1
2
e-
(x 2 -x)2 2 x 2
dx N
Best Estimate of X (slide 2)
Pr ob (x1, x2 ,L, x N ) = Pr ob(x1) Pr ob(xN )LPr ob(xN )
=
x
1
2
e-
(x1 -x) 2 x 2
2
dx 1
x
1
2
e-
(x2 -x) 2 x 2
2
dx
2
L
x
1
2
e-
(x 2 -x)2 2 x 2
dx
N
( ) =
1
N x
2
e dx dx Ldx -
(x1
-
x)2
+
(x
2 -x)2 2 x 2
+L(x
N
-
x)
2
N
12
N
2
Principle of maximum likelihood :
x
Pr
ob
(x 1
,
x
2
, L,
x
N
)
=
0
2(x 1
-
x)
+
2(x2 - x) +L+
2 x 2
2(x N
-
x)
=
0
(x1 - x) + (x2 - x) +L+ (x N - x)
Nx = x1 + x2 +L + x N
x = x1 + x2 +L+ xN N
Best estimate of X = x = x1 + x2 +L+ x N N
................
................
In order to avoid copyright disputes, this page is only a partial summary.
To fulfill the demand for quickly locating and searching documents.
It is intelligent file search solution for home and business.
Related download
- 72mw available in key central u s market global data centers
- central texas neurology consultants 512 218 1222
- department of consumer affairs for the bureau of automotive repair
- physics phy ccm
- chapter 7 central limit theorem university of kentucky
- june 2018 amer group
- reviews palwogeogra phy and diastrophism
- estimating the actual death rate caused by prescription opioid
- najla amer elyounsi
- central america weather hazards assessment
Related searches
- university of kentucky report card
- university of kentucky employee email
- university of kentucky football ranking
- university of kentucky softball schedule
- university of kentucky graduate school
- university of kentucky construction companies
- university of kentucky graduate apply
- university of kentucky core requirements
- university of kentucky master programs
- university of kentucky apply
- university of kentucky masters degree
- university of kentucky graduate program