Details

Rank-Based Methods for Shrinkage and Selection


Rank-Based Methods for Shrinkage and Selection

With Application to Machine Learning
1. Aufl.

von: A. K. Md. Ehsanes Saleh, Mohammad Arashi, Resve A. Saleh, Mina Norouzirad

109,99 €

Verlag: Wiley
Format: PDF
Veröffentl.: 04.03.2022
ISBN/EAN: 9781119625414
Sprache: englisch
Anzahl Seiten: 480

DRM-geschütztes eBook, Sie benötigen z.B. Adobe Digital Editions und eine Adobe ID zum Lesen.

Beschreibungen

<b>Rank-Based Methods for Shrinkage and Selection</b> <p><b>A practical and hands-on guide to the theory and methodology of statistical estimation based on rank</b> <p>Robust statistics is an important field in contemporary mathematics and applied statistical methods. <i>Rank-Based Methods for Shrinkage and Selection: With Application to Machine Learning</i> describes techniques to produce higher quality data analysis in shrinkage and subset selection to obtain parsimonious models with outlier-free prediction. This book is intended for statisticians, economists, biostatisticians, data scientists and graduate students. <p><i>Rank-Based Methods for Shrinkage and Selection</i> elaborates on rank-based theory and application in machine learning to robustify the least squares methodology. It also includes: <ul><li>Development of rank theory and application of shrinkage and selection</li> <li>Methodology for robust data science using penalized rank estimators</li> <li>Theory and methods of penalized rank dispersion for ridge, LASSO and Enet</li> <li>Topics include Liu regression, high-dimension, and AR(p)</li> <li>Novel rank-based logistic regression and neural networks</li> <li>Problem sets include R code to demonstrate its use in machine learning</li></ul>
<p><b>1 Introduction to Rank-based Regression </b><b>1</b></p> <p>1.1 Introduction 1</p> <p>1.2 Robustness of the Median 1</p> <p>1.2.1 Mean vs. Median 1</p> <p>1.2.2 Breakdown Point 4</p> <p>1.2.3 Order and Rank Statistics 5</p> <p>1.3 Simple Linear Regression 6</p> <p>1.3.1 Least Squares Estimator (LSE) 6</p> <p>1.3.2 Theil’s Estimator 7</p> <p>1.3.3 Belgium Telephone Data Set 7</p> <p>1.3.4 Estimation and Standard Error Comparison 9</p> <p>1.4 Outliers and their Detection 11</p> <p>1.4.1 Outlier Detection 12</p> <p>1.5 Motivation for Rank-based Methods 13</p> <p>1.5.1 Effect of a Single Outlier 13</p> <p>1.5.2 Using Rank for the Location Model 16</p> <p>1.5.3 Using Rank for the Slope 19</p> <p>1.6 The Rank Dispersion Function 20</p> <p>1.6.1 Ranking and Scoring Details 23</p> <p>1.6.2 Detailed Procedure for R-estimation 25</p> <p>1.7 Shrinkage Estimation and Subset Selection 30</p> <p>1.7.1 Multiple Linear Regression using Rank 30</p> <p>1.7.2 Penalty Functions 32</p> <p>1.7.3 Shrinkage Estimation 34</p> <p>1.7.4 Subset Selection 36</p> <p>1.7.5 Blended Approaches 39</p> <p>1.8 Summary 39</p> <p>1.9 Problems 41</p> <p><b>2 Characteristics of Rank-based Penalty Estimators </b><b>47</b></p> <p>2.1 Introduction 47</p> <p>2.2 Motivation for Penalty Estimators 47</p> <p>2.3 Multivariate Linear Regression 49</p> <p>2.3.1 Multivariate Least Squares Estimation 49</p> <p>2.3.2 Multivariate R-estimation 51</p> <p>2.3.3 Multicollinearity 51</p> <p>2.4 Ridge Regression 53</p> <p>2.4.1 Ridge Applied to Least Squares Estimation 53</p> <p>2.4.2 Ridge Applied to Rank Estimation 55</p> <p>2.5 Example: Swiss Fertility Data Set 56</p> <p>2.5.1 Estimation and Standard Errors 59</p> <p>2.5.2 Parameter Variance using Bootstrap 60</p> <p>2.5.3 Reducing Variance using Ridge 61</p> <p>2.5.4 Ridge Traces 62</p> <p>2.6 Selection of Ridge Parameter 𝜆2 65</p> <p>2.6.1 Quadratic Risk 65</p> <p>2.6.2 K-fold Cross-validation Scheme 68</p> <p>2.7 LASSO and aLASSO 71</p> <p>2.7.1 Subset Selection 71</p> <p>2.7.2 Least Squares with LASSO 71</p> <p>2.7.3 The Adaptive LASSO and its Geometric Interpretation 73</p> <p>2.7.4 R-estimation with LASSO and aLASSO 77</p> <p>2.7.5 Oracle Properties 78</p> <p>2.8 Elastic Net (Enet) 82</p> <p>2.8.1 Naive Enet 82</p> <p>2.8.2 Standard Enet 83</p> <p>2.8.3 Enet in Machine Learning 84</p> <p>2.9 Example: Diabetes Data Set 85</p> <p>2.9.1 Model Building with R-aEnet 85</p> <p>2.9.2 MSE vs. MAE 88</p> <p>2.9.3 Model Building with LS-Enet 91</p> <p>2.10 Summary 94</p> <p>2.11 Problems 95</p> <p><b>3 Location and Simple Linear Models </b><b>101</b></p> <p>3.1 Introduction 101</p> <p>3.2 Location Estimators and Testing 104</p> <p>3.2.1 Unrestricted R-estimator of 𝜃 104</p> <p>3.2.2 Restricted R-estimator of 𝜃 107</p> <p>3.3 Shrinkage R-estimators of Location 108</p> <p>3.3.1 Overview of Shrinkage R-estimators of 𝜃 108</p> <p>3.3.2 Derivation of the Ridge-type R-estimator 113</p> <p>3.3.3 Derivation of the LASSO-type R-estimator 114</p> <p>3.3.4 General Shrinkage R-estimators of 𝜃 114</p> <p>3.4 Ridge-type R-estimator of 𝜃 117</p> <p>3.5 Preliminary Test R-estimator of 𝜃 118</p> <p>3.5.1 Optimum Level of Significance of PTRE 121</p> <p>3.6 Saleh-type R-estimators 122</p> <p>3.6.1 Hard-Threshold R-estimator of 𝜃 122</p> <p>3.6.2 Saleh-type R-estimator of 𝜃 123</p> <p>3.6.3 Positive-rule Saleh-type (LASSO-type) R-estimator of 𝜃 125</p> <p>3.6.4 Elastic Net-type R-estimator of 𝜃 127</p> <p>3.7 Comparative Study of the R-estimators of Location 129</p> <p>3.8 Simple Linear Model 132</p> <p>3.8.1 Restricted R-estimator of Slope 134</p> <p>3.8.2 Shrinkage R-estimator of Slope 135</p> <p>3.8.3 Ridge-type R-estimation of Slope 135</p> <p>3.8.4 Hard-Threshold R-estimator of Slope 136</p> <p>3.8.5 Saleh-type R-estimator of Slope 137</p> <p>3.8.6 Positive-rule Saleh-type (LASSO-type) R-estimator of Slope 138</p> <p>3.8.7 The Adaptive LASSO (aLASSO-type) R-estimator 138</p> <p>3.8.8 nEnet-type R-estimator of Slope 139</p> <p>3.8.9 Comparative Study of R-estimators of Slope 140</p> <p>3.9 Summary 141</p> <p>3.10 Problems 142</p> <p><b>4 Analysis of Variance (ANOVA) </b><b>149</b></p> <p>4.1 Introduction 149</p> <p>4.2 Model, Estimation and Tests 149</p> <p>4.3 Overview of Multiple Location Models 150</p> <p>4.3.1 Example: Corn Fertilizers 151</p> <p>4.3.2 One-way ANOVA 151</p> <p>4.3.3 Effect of Variance on Shrinkage Estimators 153</p> <p>4.3.4 Shrinkage Estimators for Multiple Location 156</p> <p>4.4 Unrestricted R-estimator 158</p> <p>4.5 Test of Significance 161</p> <p>4.6 Restricted R-estimator 162</p> <p>4.7 Shrinkage Estimators 163</p> <p>4.7.1 Preliminary Test R-estimator 163</p> <p>4.7.2 The Stein–Saleh-type R-estimator 164</p> <p>4.7.3 The Positive-rule Stein–Saleh-type R-estimator 165</p> <p>4.7.4 The Ridge-type R-estimator 167</p> <p>4.8 Subset Selection Penalty R-estimators 169</p> <p>4.8.1 Preliminary Test Subset Selector R-estimator 169</p> <p>4.8.2 Saleh-type R-estimator 170</p> <p>4.8.3 Positive-rule Saleh Subset Selector (PRSS) 171</p> <p>4.8.4 The Adaptive LASSO (aLASSO) 173</p> <p>4.8.5 Elastic-net-type R-estimator 177</p> <p>4.9 Comparison of the R-estimators 178</p> <p>4.9.1 Comparison of URE and RRE 179</p> <p>4.9.2 Comparison of URE and Stein–Saleh-type R-estimators 179</p> <p>4.9.3 Comparison of URE and Ridge-type R-estimators 179</p> <p>4.9.4 Comparison of URE and PTSSRE 180</p> <p>4.9.5 Comparison of LASSO-type and Ridge-type R-estimators 180</p> <p>4.9.6 Comparison of URE, RRE and LASSO 181</p> <p>4.9.7 Comparison of LASSO with PTRE 181</p> <p>4.9.8 Comparison of LASSO with SSRE 182</p> <p>4.9.9 Comparison of LASSO with PRSSRE 182</p> <p>4.9.10 Comparison of nEnetRE with URE 183</p> <p>4.9.11 Comparison of nEnetRE with RRE 183</p> <p>4.9.12 Comparison of nEnetRE with HTRE 183</p> <p>4.9.13 Comparison of nEnetRE with SSRE 184</p> <p>4.9.14 Comparison of Ridge-type vs. nEnetRE 184</p> <p>4.10 Summary 185</p> <p>4.11 Problems 185</p> <p><b>5 Seemingly Unrelated Simple Linear Models </b><b>191</b></p> <p>5.1 Introduction 191</p> <p>5.1.1 Problem Formulation 193</p> <p>5.2 Signed and Signed Rank Estimators of Parameters 194</p> <p>5.2.1 General Shrinkage R-estimator of 𝛽 198</p> <p>5.2.2 Ridge-type R-estimator of 𝛽 199</p> <p>5.2.3 Preliminary Test R-estimator of 𝛽 201</p> <p>5.3 Stein–Saleh-type R-estimator of 𝛽 202</p> <p>5.3.1 Positive-rule Stein–Saleh R-estimators of 𝛽 202</p> <p>5.4 Saleh-type R-estimator of 𝛽 203</p> <p>5.4.1 LASSO-type R-estimator of the 𝛽 205</p> <p>5.5 Elastic-net-type R-estimators 206</p> <br clear="all" /> <p> </p> <p>5.6 R-estimator of Intercept When Slope Has Sparse Subset 207</p> <p>5.6.1 General Shrinkage R-estimator of Intercept 207</p> <p>5.6.2 Ridge-type R-estimator of 𝜃 209</p> <p>5.6.3 Preliminary Test R-estimators of 𝜃 209</p> <p>5.7 Stein–Saleh-type R-estimator of 𝜃 210</p> <p>5.7.1 Positive-rule Stein–Saleh-type R-estimator of 𝜃 211</p> <p>5.7.2 LASSO-type R-estimator of 𝜃 213</p> <p>5.8 Summary 213</p> <p>5.8.1 Problems 214</p> <p><b>6 Multiple Linear Regression Models </b><b>215</b></p> <p>6.1 Introduction 215</p> <p>6.2 Multiple Linear Model and R-estimation 215</p> <p>6.3 Model Sparsity and Detection 218</p> <p>6.4 General Shrinkage R-estimator of 𝛽 221</p> <p>6.4.1 Preliminary Test R-estimator 222</p> <p>6.4.2 Stein–Saleh-type R-estimator 224</p> <p>6.4.3 Positive-rule Stein–Saleh-type R-estimator 225</p> <p>6.5 Subset Selectors 226</p> <p>6.5.1 Preliminary Test Subset Selector R-estimator 226</p> <p>6.5.2 Stein–Saleh-type R-estimator 228</p> <p>6.5.3 Positive-rule Stein–Saleh-type R-estimator (LASSO-type) 229</p> <p>6.5.4 Ridge-type Subset Selector 231</p> <p>6.5.5 Elastic Net-type R-estimator 231</p> <p>6.6 Adaptive LASSO 232</p> <p>6.6.1 Introduction 232</p> <p>6.6.2 Asymptotics for LASSO-type R-estimator 233</p> <p>6.6.3 Oracle Property of aLASSO 235</p> <p>6.7 Summary 238</p> <p>6.8 Problems 239</p> <p><b>7 Partially Linear Multiple Regression Model </b><b>241</b></p> <p>7.1 Introduction 241</p> <p>7.2 Rank Estimation in the PLM 242</p> <p>7.2.1 Penalty R-estimators 246</p> <p>7.2.2 Preliminary Test and Stein–Saleh-type R-estimator 248</p> <p>7.3 ADB and ADL2-risk 249</p> <p>7.4 ADL2-risk Comparisons 253</p> <p>7.5 Summary: L2-risk Efficiencies 260</p> <p>7.6 Problems 262</p> <p><b>8 Liu Regression Models </b><b>263</b></p> <p>8.1 Introduction 263</p> <p>8.2 Linear Unified (Liu) Estimator 263</p> <p>8.2.1 Liu-type R-estimator 266</p> <p>8.3 Shrinkage Liu-type R-estimators 268</p> <p>8.4 Asymptotic Distributional Risk 269</p> <p>8.5 Asymptotic Distributional Risk Comparisons 271</p> <p>8.5.1 Comparison of SSLRE and PTLRE 272</p> <p>8.5.2 Comparison of PRSLRE and PTLRE 274</p> <p>8.5.3 Comparison of PRLRE and SSLRE 276</p> <p>8.5.4 Comparison of Liu-Type Rank Estimators With Counterparts 277</p> <p>8.6 Estimation of d 279</p> <p>8.7 Diabetes Data Analysis 280</p> <p>8.7.1 Penalty Estimators 281</p> <p>8.7.2 Performance Analysis 284</p> <p>8.8 Summary 288</p> <p>8.9 Problems 288</p> <p><b>9 Autoregressive Models </b><b>291</b></p> <p>9.1 Introduction 291</p> <p>9.2 R-estimation of 𝜌 for the AR(𝑝)-Model 292</p> <p>9.3 LASSO, Ridge, Preliminary Test and Stein–Saleh-type R-estimators 294</p> <p>9.4 Asymptotic Distributional L2-risk 296</p> <p>9.5 Asymptotic Distributional L2-risk Analysis 299</p> <p>9.5.1 Comparison of Unrestricted vs. Restricted R-estimators 300</p> <p>9.5.2 Comparison of Unrestricted vs. Preliminary Test R-estimator 300</p> <p>9.5.3 Comparison of Unrestricted vs. Stein–Saleh-type R-estimators 300</p> <p>9.5.4 Comparison of the Preliminary Test vs. Stein–Saleh-type R-estimators 302</p> <p>9.6 Summary 303</p> <p>9.7 Problems 304</p> <p><b>10 High-Dimensional Models </b><b>307</b></p> <p>10.1 Introduction 307</p> <p>10.2 Identifiability of 𝛽∗ and Projection 309</p> <p>10.3 Parsimonious Model Selection 309</p> <p>10.4 Some Notation and Separation 311</p> <p>10.4.1 Special Matrices 311</p> <p>10.4.2 Steps Towards Estimators 312</p> <p>10.4.3 Post-selection Ridge Estimation of 𝛽∗ 𝒮1 and 𝜷∗ 𝒮2 312</p> <p>10.4.4 Post-selection Ridge R-estimators for 𝛽∗ 𝒮1 and 𝜷∗ 𝒮2 313</p> <p>10.5 Post-selection Shrinkage R-estimators 315</p> <p>10.6 Asymptotic Properties of the Ridge R-estimators 316</p> <p>10.7 Asymptotic Distributional L2-Risk Properties 321</p> <p>10.8 Asymptotic Distributional Risk Efficiency 324</p> <p>10.9 Summary 326</p> <p>10.10 Problems 327</p> <p><b>11 Rank-based Logistic Regression </b>329</p> <p>11.1 Introduction 329</p> <p>11.2 Data Science and Machine Learning 329</p> <p>11.2.1 What is Robust Data Science? 329</p> <p>11.2.2 What is Robust Machine Learning? 332</p> <p>11.3 Logistic Regression 333</p> <p>11.3.1 Log-likelihood Setup 334</p> <p>11.3.2 Motivation for Rank-based Logistic Methods 338</p> <p>11.3.3 Nonlinear Dispersion Function 341</p> <p>11.4 Application to Machine Learning 342</p> <p>11.4.1 Example: Motor Trend Cars 344</p> <p>11.5 Penalized Logistic Regression 347</p> <p>11.5.1 Log-likelihood Expressions 347</p> <p>11.5.2 Rank-based Expressions 348</p> <p>11.5.3 Support Vector Machines 349</p> <p>11.5.4 Example: Circular Data 353</p> <p>11.6 Example: Titanic Data Set 359</p> <p>11.6.1 Exploratory Data Analysis 359</p> <p>11.6.2 RLR vs. LLR vs. SVM 365</p> <p>11.6.3 Shrinkage and Selection 367</p> <p>11.7 Summary 370</p> <p>11.8 Problems 371</p> <p><b>12 Rank-based Neural Networks </b><b>377</b></p> <p>12.1 Introduction 377</p> <p>12.2 Set-up for Neural Networks 379</p> <p>12.3 Implementing Neural Networks 381</p> <p>12.3.1 Basic Computational Unit 382</p> <p>12.3.2 Activation Functions 382</p> <p>12.3.3 Four-layer Neural Network 384</p> <p>12.4 Gradient Descent with Momentum 386</p> <p>12.4.1 Gradient Descent 386</p> <p>12.4.2 Momentum 388</p> <p>12.5 Back Propagation Example 389</p> <p>12.5.1 Forward Propagation 390</p> <p>12.5.2 Back Propagation 392</p> <p>12.5.3 Dispersion Function Gradients 394</p> <p>12.5.4 RNN Algorithm 395</p> <p>12.6 Accuracy Metrics 396</p> <p>12.7 Example: Circular Data Set 400</p> <p>12.8 Image Recognition: Cats vs. Dogs 405</p> <p>12.8.1 Binary Image Classification 406</p> <p>12.8.2 Image Preparation 406</p> <p>12.8.3 Over-fitting and Under-fitting 409</p> <p>12.8.4 Comparison of LNN vs. RNN 410</p> <p>12.9 Image Recognition: MNIST Data Set 414</p> <p>12.10 Summary 421</p> <p>12.11 Problems 421</p> <p>Bibliography 433</p> <p>Author Index 443</p> <p>Subject Index445</p>
<p><b>A. K. Md. Ehsanes Saleh, PhD,</b> is a Professor Emeritus and Distinguished Professor in the School of Mathematics and Statistics, Carleton University, Ottawa, Canada. He is Fellow of IMS, ASA and Honorary member of SSC, Canada.</p> <p><b>Mohammad Arashi, PhD,</b> is an Associate Professor at Ferdowsi University of Mashhad in Iran and Extraordinary Professor and C2 rated researcher at University of Pretoria, Pretoria, South Africa. He is an elected member of ISI.</p> <p><b>Resve A. Saleh, M.Sc, PhD (Berkeley),</b> is a Professor Emeritus in the Department of ECE at the University of British Columbia, Vancouver, Canada, and formerly with University of Illinois and Stanford University. He is the author of 4 books and Fellow of the IEEE.</p> <p><b>Mina Norouzirad, PhD,</b> is a post-doctoral researcher at the Center for Mathematics and Applications (CMA) of Nova University of Lisbon, Portugal.</p>
<p><b>A practical and hands-on guide to the theory and methodology of statistical estimation based on rank</b></p> <p>Robust statistics is an important field in contemporary mathematics and applied statistical methods. <i>Rank-Based Methods for Shrinkage and Selection: With Application to Machine Learning</i> describes techniques to produce higher quality data analysis in shrinkage and subset selection to obtain parsimonious models with outlier-free prediction. This book is intended for statisticians, economists, biostatisticians, data scientists and graduate students. <p><i>Rank-Based Methods for Shrinkage and Selection</i> elaborates on rank-based theory and application in machine learning to robustify the least squares methodology. It also includes: <ul><li>Development of rank theory and application of shrinkage and selection</li> <li>Methodology for robust data science using penalized rank estimators</li> <li>Theory and methods of penalized rank dispersion for ridge, LASSO and Enet</li> <li>Topics include Liu regression, high-dimension, and AR(p)</li> <li>Novel rank-based logistic regression and neural networks</li> <li>Problem sets include R code to demonstrate its use in machine learning</li></ul>

Diese Produkte könnten Sie auch interessieren:

Statistics for Microarrays
Statistics for Microarrays
von: Ernst Wit, John McClure
PDF ebook
90,99 €