Applied Nonparametric Statistics-lec6

Ref: https://onlinecourses.science.psu.edu/stat464/print/book/export/html/8



前面都是对一两个样本的检查,现在考虑k个样本的情况,我们的假设是:

  • Analysis of Variance (ANOVA)

assumptions are:

  1. Groups are independent
  2. Distributions are Normally distributed
  3. Groups have equal variances

那么我们的假设就是:

H0:μ1=μ2=μ3
H1:at least one not equal

R里面使用anova函数,具体可以参见以前的代码。(计算α多样性指数中的Simpson Index)

simpsonBox = read.csv("simpsonIndex1.csv")
Group = factor(c(rep(1,21), rep(22,21), rep(43,20)), labels = c("A", "B", "C"))
simpsonData = data.frame(simpsonIndex = simpsonBox$x, group = Group)
# 非参数检验,检查方差是否相同
fligner.test(shannonIndex ~ group, data = shannonData)
# 正态分布的数据,检查方差是否相同
bartlett.test(simpsonIndex ~ group, data = simpsonData)
# anova
simpsonAov <- aov(simpsonIndex ~ group, data = simpsonData)
summary(simpsonAov)

  

  • 非参数的方法则是kruskal test

kruskal.test



上面的anova分析之后,如果我们拒绝了原假设,知道几个组的均值是不同的,那么两两组之间,它们差异的显著性如何?

就像之前做过的TukeyHSD(要求数据正态分布),我们还可以做

  • Bonferroni Adjustment

The Bonferroni adjustment simply divides the Type I error rate (.05) by the number of tests (in this case, three).

pairwise.t.test(simpsonData$simpsonIndex,simpsonData$group,p.adjust="bonferroni")

  我们可以比较一下,它和TukeyHSD在结果上的差别:

其实结果是一致的,都说明C组与A组的差异显著。一般认为Bonferroni更保守。我们使用的函数可以参照:

https://stat.ethz.ch/R-manual/R-devel/library/stats/html/pairwise.t.test.html


  • Holm Adjustment

这篇文章指出,Holm Adjustment比Bonferroni更好,同样使用pairwise.t.test函数。

Ref: http://rtutorialseries.blogspot.jp/2011/03/r-tutorial-series-anova-pairwise.html

  • The Fisher Least Significant Difference (LSD) method essentially does not correct for the Type I error rate

for multiple comparisons and is generally not recommended relative to other options.

library(agricolae)
LSD.test()

  

时间: 2024-08-08 02:47:50

Applied Nonparametric Statistics-lec6的相关文章

Applied Nonparametric Statistics

参考网址: https://onlinecourses.science.psu.edu/stat464/node/2 Binomial Distribution Normal Distribution 将正态分布标准化.这也就是Z-score Confidence Interval 在上面的前提下,假设σ^2已知,现在构造μ的置信区间: 利用上面Z-score的公式,且 套入公式,解出μ.注意此处的标准差用的是σ/根号n.最终解出: 当σ^2=Var(X)不知道时,我们可以用样本的标准差,计算Z

Applied Nonparametric Statistics-lec8

Ref:https://onlinecourses.science.psu.edu/stat464/print/book/export/html/11 additive model value = typical value + row effect + column effect + residual predicate value = typical value + row effect + column effect 其中value是我们关注的值,typical value是overall

Applied Nonparametric Statistics-lec4

Ref: https://onlinecourses.science.psu.edu/stat464/print/book/export/html/5 Two sample test 直接使用R的t-test t.test(n, t, alternative="two.sided", var.equal=T) permutation test 当我们判断两个样本的均值或者中值是否相等时,如果样本数量足够大,可以使用t-test. 但是,当两个样本的数量都很小时,它们的分布可能是有偏的,

Applied Nonparametric Statistics-lec9

Ref:https://onlinecourses.science.psu.edu/stat464/print/book/export/html/12 前面我们考虑的情况是:response是连续的,variable是离散的.举例:如果打算检查GPA的中位数是否与学生坐在教室的位置有关, 那么GPA的中位数是连续的,是响应变量:学生坐的位置(前中后)是离散的,是解释变量. 现在考虑解释变量也是连续的情况,即检查两个连续变量之间的因果关系.其中,我们最关心的是关系的强弱和方向. 首先,我们考虑线性

Applied Nonparametric Statistics-lec2

Ref: https://onlinecourses.science.psu.edu/stat464/print/book/export/html/3 The Binomial Distribution in R: # return PMF. prob is the probability of success . x can be a list dbinom(x, size, prob) # CDF pbinom(x, size, prob) # returns a value for a p

Applied Nonparametric Statistics-lec3

Ref: https://onlinecourses.science.psu.edu/stat464/print/book/export/html/4 使用非参数方法的优势: 1. 对总体分布做的假设少,所以总体分布未知也可以: 2. 容易做: 3. 一般对离群值更具鲁棒性robust: 4. 适用于数据中包含ranks, ordinal or categorical的. In a skewed distribution, the population median, η, is a bette

psu online course

https://onlinecourses.science.psu.edu/statprogram/programs Graduate Online Course Overviews Printer-friendly versionPrinter-friendly version Picture of Thomas Building where the Eberly College of Science and the Department of Statistics resides.The D

Machine and Deep Learning with Python

Machine and Deep Learning with Python Education Tutorials and courses Supervised learning superstitions cheat sheet Introduction to Deep Learning with Python How to implement a neural network How to build and run your first deep learning network Neur

数学类杂志SCI2013-2014影响因子

ISSN Abbreviated Journal Title Full Title Category Subcategory Country total Cites IF        2013-2014 IF 2012-2013 IF 2011-2012 IF 2010-2011 IF 2009-2010 IF 2008-2009 IF 2007-2008 5-Year Impact Factor Immediacy Index Articles Cited Half-Life Eigenfa