Dean Adams, Iowa State University
Statistical methods are often used to test hypotheses and make inferences
Requires parametric theory to estimate parameters & CI
Numerous distributions of expected values have been generated from theory for different types of data and hypotheses
Each have assumptions about the behavior of the underlying data
Power \(\small\downarrow\) as dimensionality \(\small\uparrow\), and eventually computations cannot be completed (the ‘curse’ of dimensionality)
Alternative mechanisms for evaluating hypotheses are required
Schematic of randomization procedure for \(t\)-test
Complete enumeration of all possible permutations
Provides exact probability of \(E_{obs}\)
## D.obs P-value
## 2.445696 0.001000
Note the null distribution from our previous example:
\(\small\uparrow\) # of iterations improves precision of estimates of significance
With higher computing power, large numbers of iterations are feasible
Resample dataset many times with replacement
Each bootstrap iteration contains \({N}\) objects, but some are represented multiple times, and others not at all
The bootstrap is very useful for estimating confidence intervals, because summary measures derived from bootstrap samples approximate those of the original population distribution
Generate boostrap datasets
Estimate summary statistic from each \(E_{boot}\)
Bootstrap CI are: upper and lower \(\small\alpha/2\) values of \(E_{boot}\) sample (usually 0.025 and 0.975)
Sometimes the the bootstrap distribution is skewed, such that \(\small\mu_{E_{boot}}\neq{E_{obs}}\)
Bias is alleviated by finding fraction (Fr) of bootstrap values above \(\small{E}_{obs}\)
Adjust as \(\small{CI}=\Phi[2\Phi^{-1}(Fr)\pm{Z}_{\alpha/2}]\) where \(\small\Phi\) is the cumulative normal distribution
‘Leave one out’ resampling: each iteration contains \(\small{N-1}\) objects
Investigate the precision of \(\small{E}_{obs}\) and how sensitive it is to specific values in a dataset
Useful to measure bias, standard error or CI of test statistic
\(\small{Bias}(E_{obs})=E_{obs}-\mu_{E_{jack}}\)
Devising a proper permutation test requires several components:
1: Identifying the null hypothesis \(\small{H}_{0}\)
2: Determining whether there is a known expected value under \(\small{H}_{0}\)
3: Identifying which values may be permuted and how to estimate expected distribution under \(\small{H}_{0}\)
Devising a proper permutation test requires several components:
1: Identifying the null hypothesis \(\small{H}_{0}\)
2: Determining whether there is a known expected value under \(\small{H}_{0}\)
3: Identifying which values may be permuted and how to estimate expected distribution under \(\small{H}_{0}\)
Essentially, one must determine:
1: What to permute?
2: How to permute it?
Permutation tests generate empirical sampling distributions under \(\small{H}_{0}\)
How do we accomplish this?
A simple logic flow:
Define \(\small{H}_{0}\) and \(\small{H}_{1}\)
Identify what differs between \(\small{H}_{0}\) and \(\small{H}_{1}\) (i.e., what does \(\small{H}_{1}\) quantify relative to \(\small{H}_{0}\)?)
Permute the data which ‘breaks up’ the signal in \(\small{H}_{1}\) relative to \(\small{H}_{0}\)
More formally, we seek to identify the correct exchangeable units under \(\small{H}_{0}\)
Exchangeable units are those values such that the permuted distribution is the same as that of the original (Commanges. J. Nonparam. Stat. 2003)
For linear models, we seek to retain two properties:
1st moment exchangeability: the expected value remains constant
2nd moment exchangeability: the variance remains constant
Incorrectly assigning exchangeable units can result in elevated type I error rates and incorrect inferences
Here, permuting phylogenetically independent contrasts is incorrect, because these values contain information from both the response (\(\small\mathbf{Y}\)) data as well as the phylogeny among taxa (see PCM lecture)
For simple linear models: \(\small\mathbf{Y}=\mathbf{X}\mathbf{\beta } + \mathbf{E}\), permuting \(\small\mathbf{Y}\) relative to \(\small\mathbf{X}\) is often proposed
This is sufficient for \(t\)-test & correlation tests; and for simple linear models (e.g., single-factor models)
For more complex models with Multiple explanatory factors:
Factorial model: \(\small\mathbf{Y}=\mathbf{X_{A}}\mathbf{\beta_{A}} +\mathbf{X_{B}}\mathbf{\beta_{B}} +\mathbf{X_{AB}}\mathbf{\beta_{AB}}+\mathbf{E}\)
Permuting \(\small\mathbf{Y}\) is possible, but seems inadequate
An alternative is to restrict the resampling to sub-strata of data (strata based on levels within factors)
Permute \(\small\mathbf{Y}\) within levels of A then within levels of B
Permits evaluation of \(\small{SS_{A}}\) and \(\small{SS_{B}}\) but NOT \(\small{SS_{AB}}\)
The key to identifying exchangeable units lies with the \(\small{H}_{0}\):
Factorial models \(\small\mathbf{Y}=\mathbf{X_{A}}\mathbf{\beta_{A}} +\mathbf{X_{B}}\mathbf{\beta_{B}} +\mathbf{X_{AB}}\mathbf{\beta_{AB}}+\mathbf{E}\) are a set of sequential hypotheses comparing full (\(\small\mathbf{X}_{F}\)) and reduced (\(\small\mathbf{X}_{R}\)) models
Testing each \(\small\mathbf{X}_{F}\) requires appropriate permutation procedure for each \(\small\mathbf{X}_{R}\)
Residual randomization provides proper exchangeable units under each \(\small\mathbf{X}_{R}\)
Permute residuals \(\mathbf{E}_{R}\) from reduced model \(\small\mathbf{X}_{R}\), rather than original values
Evaluates \(\small{SS}_{\mathbf{X}_{F}}\) while holding effects of \(\small\mathbf{X}_{R}\) constant
Must specify full and reduced models (Type I SS used in example)
Mathematical justification:
1: For any \(\small\mathbf{X}_{R}\): \(\small{SS}_{\mathbf{X}_{F}}=0\)
2: Under \(\small\mathbf{X}_{R}\), \(\mathbf{E}_{R}\) represent those components of SS NOT explained by \(\small\mathbf{X}_{R}\) (includes \(\small{RSS}\) of \(\small\mathbf{X}_{F}\) plus SS from term(s) not in \(\small\mathbf{X}_{R}\))
3: Thus, permuting \(E_{R}\) precisely embodies \(\small{H}_{R}\) of: \(\small{SS}_{\mathbf{X}_{F}}=0\)
1: Fit \(\small\mathbf{X}_{F}\) for each term in model; obtain coefficients and summary statistics (e.g., \(\small{SS}_{X}\))
2: Fit \(\small\mathbf{X}_{R}\) for each \(\small\mathbf{X}_{F}\); Estimate \(\small\hat{\mathbf{Y}}_{R}\) and \(\mathbf{E}_{R}\)
3: Permute, \(E_{R}\): obtain pseudo values as: \(\small\mathbf{\mathcal{Y}} = \mathbf{\hat{Y}}_{R} + \mathbf{E}_{R}\)
4: Fit \(\small\mathbf{X}_{F}\) using \(\small\mathbf{\mathcal{Y}}\): obtain coefficients and summary statistics
5: Repeat
NOTE: for single-factor models, permuting \(\small\mathbf{Y}\) is equivalent to RRPP
Reason: \(\small{H}_{0}\) for single-factor model is \(\mathbf{Y}\)~1 (intercept model)
Residuals of \(\small{H}_{0}\) are simply deviations from the \(\small\mu_{Y}\)
And since \(\small\mu_{Y}\) is a constant, permutation distribution will be identical
OLS Factorial Model (Adams & Collyer, unpubl.)
Full Randomization, Restricted Randomization, and RRPP are all fine (but not full-model residuals)
GLS Factorial Model (Adams & Collyer, unpubl.)
Only RRPP is appropriate
RRPP is unaltered for multivariate data
Shuffle ROWS of \(\mathbf{E}_{R}\)
The rest of the procedure is unchanged
Mantel tests (see Matrix Covariation) shuffle rows AND columns of distance matrices
Prefered approach: RRPP
1: PCoA of distance matrix to obtain coordinates for \(\small{Y}\)
2: Fit model: \(\small \mathbf{Y}=\mathbf{X}\mathbf{\beta } +E\)
3: RRPP of rows of \(\mathbf{E}_{R}\)
Does pupfish body shape differ between populations (marsh vs. sinkhole) or between the sexes?
This is a factorial MANOVA: Y~ Pop + Sex + Pop:Sex
## Df SS MS Rsq F Z Pr(>F)
## Sex 1 0.015780 0.0157802 0.28012 28.209 4.7773 0.001 **
## Pop 1 0.009129 0.0091294 0.16206 16.320 4.7097 0.001 **
## Sex:Pop 1 0.003453 0.0034532 0.06130 6.173 3.7015 0.001 **
## Residuals 50 0.027970 0.0005594 0.49651
## Total 53 0.056333
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
PCA of data and predicted values (see GLM lecture)
RRPP appropriate for ‘standard’ OLS (ordinary least squares) models
ANOVA designs
Regression designs
Factorial models
ANCOVA models
More generally RRPP is also appropriate for GLS (generalized least squares) models
These are models where covariance between objects is not zero (e.g., phylogenetic non-independence, spatial non-independence, temporal non-independence, etc.).
OLS is a special case of GLS (see LM lecture)
Resampling methods are of primary importance in multivariate analysis
Methods are flexible, and may be used with univariate or high-dimensional data
RRPP is the most general approach (and the only approach that is appropriate for GLS models)