Statistics Distribution Of Normal Population

STATISTICS DISTRIBUTION OF NORMAL POPULATION

Ethan

IN LOVING MEMORY OF MAMBA DAY 4.13, 2016

In probability and statistics, a statistic is a function of samples. Theoretically, it is also a random variable with a probability distribution, called statistics distribution. As a statistic is
a basis of inference of population distribution and characteristics, to determine a distribution of a statistic is one of the most fundamental problems in statistics. Generally speaking, it is rather complicate to determine a precise distribution of a statistic.
But, for normal population, there are a few effective methods. This work will introduce some common statistics distributions of normal population and derive some important related theorems.

Digression

Let’s consider the probability distribution of
= X2, where X ~ N(0,1). The probability density of
Y is given by

Therefore, Y
~ Gamma(1/2, 1/2), where

By using Gamma function (See Appendix I), one can determine the expectation and variance of
Y, E(Y)
= alpha/lambda,
Var(Y) =
alpha/lambda2.

The characteristic function of Gamma probability density is given by

where X
~ Gamma(alpha,
lambda).

Consider the distribution of a summation,
Z, of two independent random variables, X ~ Gamma(alpha1,
lambda) and Y ~ Gamma(alpha2, lambda). The characteristic function of
Z is given by

Thus, Z
~ Gamma(alpha1+alpha2,
lambda).

ChiDistribution

Suppose a population
X ~ N(0, 1), X1, …, Xn are iid, then

It is easy to check that E(Y)=n,
D(Y)=2n.

Two important properties of
Chidistribution:

Proof:

The first one is trivial.

To see the second one, let
Yi = 2(lambda)Xi, then

Therefore,

This completes the proof.
#

t-Distribution

Suppose X ~
N(0,1), Y ~ Chi2(n), and X and
Y are independent, then

of which the probability density is given by

One can check that, when
n>2, E(T)=0,
D(T)=n/(n-2).

F-Distribution

Suppose X ~
Chi2(m), ~ Chi2(n), and
X and Y are independent, then

of which the probability density is given by

Theorem 1

Suppose X1, …,
Xn are samples extracted from a normal population N(musigma2), with a sample mean and a sample variance,

then, (1)(2)(3)(4)

Proof:

Formula (1) can be easily checked by performing Gaussian normalization.

For (2) and (3),

where

Now construct a finite-dimensional linear transformation
A from
Y to
Z,

One can show that
A is an orthonormal matrix. Further,

Therefore,
Zi’s are also iid ~
N(0, 1).

Consequently, U and
V are independent, where

Since V ~
Chi2(n-1),
then

To see (4), now that
U’ and
V’ are independent, satisfying

by the definition of
t-distribution, we have

This completes the proof.
#

Theorem 2

Suppose X1, …,
Xm and Y1,…, Yn are samples taken from two normal populations
N(mu1, sigma2), N(mu2,
sigma2), respectively, and they are all independent, then,

Hint: This can be shown by applying Theorem 1.

Theorem 3

Suppose X1, …,
Xm and Y1,…, Yn are samples taken from two normal populations
N(mu1, sigma12),
N
(mu2, sigma22), respectively, and they are all independent, then,

Hint: This can be shown by using the definition of
F-distribution.

APPENDIX

I. Gamma Function

Gamma function is defined as

which has following properties

Proof:

This completes the proof.
#

II. Probability Density after Transformation

Suppose random vectors
X, Y are in Rand a bijective operator T is defined as

If the probability density of
X is
pX(x),
then that of Y is given by

Proof:

The cumulative probability function of
Y is given by

Hence,

This completes the proof.
#

This theorem can be used to determine the probability density typical of the form “Z=X/Y”, “Z=X+Y”, etc., by introducing an irrelevant variable
and integrating the joint distribution of (Z,U)to get the marginal distribution of
Z, as long as T is bijective.

时间: 2024-10-02 02:53:20

Statistics Distribution Of Normal Population的相关文章

概率论与数理统计常用英文词汇对照

Probability Theory 概率论 Trial 试验 intersection交 union 并 frequency 频率 difference 差 additivity 可加性complementation 对立 contain 包含 equivalent 等价 mean 均值 convolution [,k?nv?'lu:??n] 卷积variance 方差 covariance 协方差 correlated 相关 standard deviation 标准差Random expe

Variance

http://mathworld.wolfram.com/Variance.html Variance For a single variate having a distribution with known population mean , the population variance , commonly also written , is defined as (1) where is the population mean and denotes the expectation v

coursera机器学习公开课笔记15: anomaly-detection

Note This personal note is written after studying the opening course on the coursera website, Machine Learning by Andrew NG . And images, audios of this note all comes from the opening course. 01_density-estimation In this next set of videos, I'd lik

echarts的基本使用

echarts的基本使用 官网:http://echarts.baidu.com/index.html ECharts,一个使用 JavaScript 实现的开源可视化库,可以流畅的运行在 PC 和移动设备上,兼容当前绝大部分浏览器(IE8/9/10/11,Chrome,Firefox,Safari等),底层依赖轻量级的矢量图形库 ZRender,提供直观,交互丰富,可高度个性化定制的数据可视化图表. echarts能做什么? 报表,各种饼状图柱状图.... 示例:http://echarts.

一篇文章看懂spark 1.3+各版本特性

Spark 1.6.x的新特性Spark-1.6是Spark-2.0之前的最后一个版本.主要是三个大方面的改进:性能提升,新的 Dataset API 和数据科学功能的扩展.这是社区开发非常重要的一个里程碑.1. 性能提升根据 Apache Spark 官方 2015 年 Spark Survey,有 91% 的用户想要提升 Spark 的性能.Parquet 性能自动化内存管理流状态管理速度提升 10X 2. Dataset APISpark 团队引入了 DataFrames,新型Datase

第17章 标准库特殊设施

17.1类型tuple tuple是类是pair的模板.每个pair都不相同,但是都有两个成员.而一个tuple类型会有任意数量的成员.(定义在tuple头文件中) tuple<T1,T2,...,Tn> t; t是一个tuple,成员数位n,第i个成员的类型为Ti.所有成员都进行值初始化 tuple<T1,T2,...,Tn> t(v1,v2,...,vn); t是一个tuple,成员类型为T1,T2,...,Tn,每个成员用对应的初始值vi进行初始化.此构造函数是explici

常见的机器学习&amp;数据挖掘知识点

常见的机器学习&数据挖掘知识点 转载请说明出处 Basis(基础): MSE(Mean Squared Error, 均方误差) RMSE(Root Mean Squared Error, 均方根误差) RRSE(Root Relative Squared Error, 相对平方根误差) MAE(Mean Absolute Error, 平均绝对误差) RAE(Root Absolute Error, 平均绝对误差平方根) LSM(Least Mean Squared, 最小均方) LSM(Le

应用概率统计模板

apsart.cls 1 % !Mode:: "TeX:UTF-8" 2 %% 3 %% This is file `APSart.cls', 4 %% 5 %% Copyright 2006-2012 6 %% 7 %% ---------------------------------------------------------------------- 8 %% 9 %% It may be distributed and/or modified under the 10 %

NEURAL NETWORKS, PART 3: THE NETWORK

NEURAL NETWORKS, PART 3: THE NETWORK We have learned about individual neurons in the previous section, now it’s time to put them together to form an actual neural network. The idea is quite simple – we line multiple neurons up to form a layer, and co