Lower confidence bounds for prediction accuracy in high dimensions via AROHIL Monte Carlo

@article{Dobbin2011LowerCB,
  title={Lower confidence bounds for prediction accuracy in high dimensions via AROHIL Monte Carlo},
  author={Kevin K. Dobbin and Stephanie Cooke},
  journal={Bioinformatics},
  year={2011},
  volume={27 22},
  pages={
          3129-34
        }
}
MOTIVATION Implementation and development of statistical methods for high-dimensional data often require high-dimensional Monte Carlo simulations. Simulations are used to assess performance, evaluate robustness, and in some cases for implementation of algorithms. But simulation in high dimensions is often very complex, cumbersome and slow. As a result, performance evaluations are often limited, robustness minimally investigated and dissemination impeded by implementation challenges. This… 

References

SHOWING 1-10 OF 31 REFERENCES
A method for constructing a confidence bound for the actual error rate of a prediction rule in high dimensions.
TLDR
An alternative to McLachlan's method is presented that can be applied when p >> n, with an additional adjustment in the presence of feature selection, and coverage probabilities of the new method are shown to be nominal or conservative over a wide range of scenarios.
A Shrinkage Approach to Large-Scale Covariance Matrix Estimation and Implications for Functional Genomics
TLDR
This work proposes a novel shrinkage covariance estimator that exploits the Ledoit-Wolf (2003) lemma for analytic calculation of the optimal shrinkage intensity and applies it to the problem of inferring large-scale gene association networks.
Calculating Confidence Intervals for Prediction Error in Microarray Classification Using Resampling
TLDR
An extensive study of existing confidence interval methods and a simple bias reduction on the BCCV percentile interval is proposed, which provides mildly conservative inference under all circumstances studied and outperforms the other methods in microarray applications with small to moderate sample sizes.
Stochastic simulation
  • B. Ripley
  • Computer Science
    Wiley series in probability and mathematical statistics : applied probability and statistics
  • 1987
TLDR
Brian D. Ripley's Stochastic Simulation is a short, yet ambitious, survey of modern simulation techniques, and three themes run throughout the book.
Optimally splitting cases for training and testing high dimensional classifiers
TLDR
A non-parametric algorithm for determining an optimal splitting proportion that can be applied with a specific dataset and classifier algorithm is developed and applied to any dataset, using any predictor development method, to determine the best split.
A faster circular binary segmentation algorithm for the analysis of array CGH data
TLDR
A hybrid approach to obtain the P-value of the test statistic in linear time is presented and it is shown that the substantial gain in speed with only a negligible loss in accuracy and that the stopping rule further increases speed.
On partial least squares dimension reduction for microarray-based classification: a simulation study
A Paradigm for Class Prediction Using Gene Expression Profiles
TLDR
The prediction paradigm will serve as a good framework for comparing different prediction methods and may accelerate the development of molecular classifiers that are clinically useful.
Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images
  • S. Geman, D. Geman
  • Physics
    IEEE Transactions on Pattern Analysis and Machine Intelligence
  • 1984
TLDR
The analogy between images and statistical mechanics systems is made and the analogous operation under the posterior distribution yields the maximum a posteriori (MAP) estimate of the image given the degraded observations, creating a highly parallel ``relaxation'' algorithm for MAP estimation.
...
1
2
3
4
...