您好,欢迎光临本网站![请登录][注册会员]  
文件名称: Computer Age Statistical Inference: Algorithms,Evidence,and Data Science.
  所属分类: 机器学习
  开发工具:
  文件大小: 4mb
  下载次数: 0
  上传时间: 2017-06-03
  提 供 者: luxiao******
 详细说明: 1 Algorithms and Inference 3 1.1 A Regression Example 4 1.2 Hypothesis Testing 8 1.3 Notes 11 2 Frequentist Inference 12 2.1 Frequentism in Practice 14 2.2 Frequentist Optimality 18 2.3 Notes and Details 20 3 Bayesian Inference 22 3.1 Two Examples 24 3.2 Uninfor mative Prior Distributions 28 3.3 Flaws in Frequentist Inference 30 3.4 A Bayesian/Frequentist Comparison List 33 3.5 Notes and Details 36 4 Fisherian Inference and Maximum Likelihood Estimation 38 4.1 Likelihood and Maximum Likelihood 38 4.2 Fisher Information and the MLE 41 4.3 Conditional Inference 45 4.4 Permutation and Randomization 49 4.5 Notes and Details 51 5 Parametric Models and Exponential Families 53 ix x Contents 5.1 Univariate Families 54 5.2 The Multivariate Normal Distribution 55 5.3 Fisher’s Information Bound for Multiparameter Families 59 5.4 The Multinomial Distribution 61 5.5 Exponential Families 64 5.6 Notes and Details 69 Part II Early Computer-Age Methods 73 6 Empirical Bayes 75 6.1 Robbins’ Formula 75 6.2 The Missing-Species Problem 78 6.3 A Medical Example 84 6.4 Indirect Evidence 1 88 6.5 Notes and Details 88 7 James–Stein Estimation and Ridge Regression 91 7.1 The James–Stein Estimator 91 7.2 The Baseball Players 94 7.3 Ridge Regression 97 7.4 Indirect Evidence 2 102 7.5 Notes and Details 104 8 Generalized Linear Models and Regression Trees 108 8.1 Logistic Regression 109 8.2 Generalized Linear Models 116 8.3 Poisson Regression 120 8.4 Regression Trees 124 8.5 Notes and Details 128 9 Survival Analysis and the EM Algorithm 131 9.1 Life Tables and Hazard Rates 131 9.2 Censored Data and the Kaplan–Meier Estimate 134 9.3 The Log-Rank Test 139 9.4 The Proportional Hazards Model 143 9.5 Missing Data and the EM Algorithm 146 9.6 Notes and Details 150 10 The Jackknife and the Bootstrap 155 10.1 The Jackknife Estimate of Standard Error 156 10.2 The Nonparametric Bootstrap 159 10.3 Resampling Plans 162 Contents xi 10.4 The Parametric Bootstrap 169 10.5 Influence Functions and Robust Estimation 174 10.6 Notes and Details 177 11 Bootstrap Confidence Intervals 181 11.1 Neyman’s Construction for One-Parameter Problems 181 11.2 The Percentile Method 185 11.3 Bias-Corrected Confidence Intervals 190 11.4 Second-Order Accuracy 192 11.5 Bootstrap-t Intervals 195 11.6 Objective Bayes Intervals and the Confidence Distribution 198 11.7 Notes and Details 204 12 Cross-Validation and Cp Estimates of Prediction Error 208 12.1 Prediction Rules 208 12.2 Cross-Validation 213 12.3 Covariance Penalties 218 12.4 Training, Validation, and Ephemeral Predictors 227 12.5 Notes and Details 230 13 Objective Bayes Inference and MCMC 233 13.1 Objective Prior Distributions 234 13.2 Conjugate Prior Distributions 237 13.3 Model Selection and the Bayesian Information Criterion 243 13.4 Gibbs Sampling and MCMC 251 13.5 Example: Modeling Population Admixture 256 13.6 Notes and Details 261 14 Postwar Statistical Inference and Methodology 264 Part III Twenty-First-Century Topics 269 15 Large-Scale Hypothesis Testing and FDRs 271 15.1 Large-Scale Testing 272 15.2 False-Discovery Rates 275 15.3 Empirical Bayes Large-Scale Testing 278 15.4 Local False-Discovery Rates 282 15.5 Choice of the Null Distribution 286 15.6 Relevance 290 15.7 Notes and Details 294 16 Sparse Modeling and the Lasso 298 xii Contents 16.1 Forward Stepwise Regression 299 16.2 The Lasso 303 16.3 Fitting Lasso Models 308 16.4 Least-Angle Regression 309 16.5 Fitting Generalized Lasso Models 313 16.6 Post-Selection Inference for the Lasso 317 16.7 Connections and Extensions 319 16.8 Notes and Details 321 17 Random Forests and Boosting 324 17.1 Random Forests 325 17.2 Boosting with Squared-Error Loss 333 17.3 Gradient Boosting 338 17.4 Adaboost: the Original Boosting Algorithm 341 17.5 Connections and Extensions 345 17.6 Notes and Details 347 18 Neural Networks and Deep Learning 351 18.1 Neural Networks and the Handwritten Digit Problem 353 18.2 Fitting a Neural Network 356 18.3 Autoencoders 362 18.4 Deep Learning 364 18.5 Learning a Deep Network 368 18.6 Notes and Details 371 19 Support-Vector Machines and Kernel Methods 375 19.1 Optimal Separating Hyperplane 376 19.2 Soft-Margin Classifier 378 19.3 SVM Criterion as Loss Plus Penalty 379 19.4 Computations and the Kernel Trick 381 19.5 Function Fitting Using Kernels 384 19.6 Example: String Kernels for Protein Classification 385 19.7 SVMs: Concluding Remarks 387 19.8 Kernel Smoothing and Local Regression 387 19.9 Notes and Details 390 20 Inference After Model Selection 394 20.1 Simultaneous Confidence Intervals 395 20.2 Accuracy After Model Selection 402 20.3 Selection Bias 408 20.4 Combined Bayes–Frequentist Estimation 412 20.5 Notes and Details 417 Contents xiii 21 Empirical Bayes Estimation Strategies 421 21.1 Bayes Deconvolution 421 21.2 g-Modeling and Estimation 424 21.3 Likelihood, Regularization, and Accuracy 427 21.4 Two Examples 432 21.5 Generalized Linear Mixed Models 437 21.6 Deconvolution and f -Modeling 440 21.7 Notes and Details 444 Epilogue 446 References 453 Author Index 463 Subject Index 467 ...展开收缩
(系统自动生成,下载前可以参看下载内容)

下载文件列表

相关说明

  • 本站资源为会员上传分享交流与学习,如有侵犯您的权益,请联系我们删除.
  • 本站是交换下载平台,提供交流渠道,下载内容来自于网络,除下载问题外,其它问题请自行百度
  • 本站已设置防盗链,请勿用迅雷、QQ旋风等多线程下载软件下载资源,下载后用WinRAR最新版进行解压.
  • 如果您发现内容无法下载,请稍后再次尝试;或者到消费记录里找到下载记录反馈给我们.
  • 下载后发现下载的内容跟说明不相乎,请到消费记录里找到下载记录反馈给我们,经确认后退回积分.
  • 如下载前有疑问,可以通过点击"提供者"的名字,查看对方的联系方式,联系对方咨询.
 相关搜索: statistical inference data science
 输入关键字,在本站1000多万海量源码库中尽情搜索: