Abstract
When a linear model is chosen by searching for the best subset among a set of candidate predictors, a fixed penalty such as that imposed by the Akaike information criterion may penalize model complexity inadequately, leading to biased model selection. We study resampling-based information criteria that aim to overcome this problem through improved estimation of the effective model dimension. The first proposed approach builds upon previous work on bootstrap-based model selection. We then propose a more novel approach based on cross-validation. Simulations and analyses of a functional neuroimaging data set illustrate the strong performance of our resampling-based methods, which are implemented in a new R package.
| Original language | American English |
|---|---|
| Pages (from-to) | 1161-1186 |
| Number of pages | 26 |
| Journal | Annals of the Institute of Statistical Mathematics |
| Volume | 64 |
| Issue number | 6 |
| DOIs | |
| State | Published - Dec 2012 |
| Externally published | Yes |
Keywords
- Adaptive model selection
- Covariance inflation criterion
- Cross-validation
- Extended information criterion
- Functional connectivity
- Overoptimism
All Science Journal Classification (ASJC) codes
- Statistics and Probability
Fingerprint
Dive into the research topics of 'Resampling-based information criteria for best-subset regression'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver