Details

Regression Analysis By Example Using R


Regression Analysis By Example Using R


Wiley Series in Probability and Statistics 6. Aufl.

von: Ali S. Hadi, Samprit Chatterjee

106,99 €

Verlag: Wiley
Format: EPUB
Veröffentl.: 11.10.2023
ISBN/EAN: 9781119830894
Sprache: englisch
Anzahl Seiten: 480

DRM-geschütztes eBook, Sie benötigen z.B. Adobe Digital Editions und eine Adobe ID zum Lesen.

Beschreibungen

<p><i><b>Regression Analysis By Example Using R</b></i></p> <p><b>A STRAIGHTFORWARD AND CONCISE DISCUSSION OF THE ESSENTIALS OF REGRESSION ANALYSIS</b></p> <p>In the newly revised sixth edition of <i>Regression Analysis By Example Using R</i>, distinguished statistician Dr Ali S. Hadi delivers an expanded and thoroughly updated discussion of exploratory data analysis using regression analysis in R. The book provides in-depth treatments of regression diagnostics, transformation, multicollinearity, logistic regression, and robust regression.</p> <p>The author clearly demonstrates effective methods of regression analysis with examples that contain the types of data irregularities commonly encountered in the real world. This newest edition also offers a brand-new, easy to read chapter on the freely available statistical software package R.</p> <p>Readers will also find:</p> <ul> <li>Reorganized, expanded, and upgraded exercises at the end of each chapter with an emphasis on data analysis</li> <li>Updated data sets and examples throughout the book</li> <li>Complimentary access to a companion website that provides data sets in xlsx, csv, and txt format</li> </ul> <p>Perfect for upper-level undergraduate or beginning graduate students in statistics, mathematics, biostatistics, and computer science programs, <i>Regression Analysis By Example Using R </i>will also benefit readers who need a reference for quick updates on regression methods and applications.</p>
<p>Preface xiv</p> <p><b>1 Introduction 1</b></p> <p>1.1 What Is Regression Analysis? 1</p> <p>1.2 Publicly Available Data Sets 2</p> <p>1.3 Selected Applications of Regression Analysis 3</p> <p>1.3.1 Agricultural Sciences 3</p> <p>1.3.2 Industrial and Labor Relations 4</p> <p>1.3.3 Government 5</p> <p>1.3.4 History 5</p> <p>1.3.5 Environmental Sciences 6</p> <p>1.3.6 Industrial Production 6</p> <p>1.3.7 The Space Shuttle Challenger 7</p> <p>1.3.8 Cost of Health Care 7</p> <p>1.4 Steps in Regression Analysis 7</p> <p>1.4.1 Statement of the Problem 9</p> <p>1.4.2 Selection of Potentially Relevant Variables 9</p> <p>1.4.3 Data Collection 9</p> <p>1.4.4 Model Specification 10</p> <p>1.4.5 Method of Fitting 12</p> <p>1.4.6 Model Fitting 13</p> <p>1.4.7 Model Criticism and Selection 14</p> <p>1.4.8 Objectives of Regression Analysis 15</p> <p>1.5 Scope and Organization of the Book 16</p> <p><b>2 A Brief Introduction to R 19</b></p> <p>2.1 What Is R and RStudio? 19</p> <p>2.2 Installing R and RStudio 20</p> <p>2.3 Getting Started With R 21</p> <p>2.3.1 Command Level Prompt 21</p> <p>2.3.2 Calculations Using R 22</p> <p>2.3.3 Editing Your R Code 24</p> <p>2.3.4 Best Practice: Object Names in R 25</p> <p>2.4 Data Values and Objects in R 25</p> <p>2.4.1 Types of Data Values in R 25</p> <p>2.4.2 Types (Structures) of Objects in R 28</p> <p>2.4.3 Object Attributes 34</p> <p>2.4.4 Testing (Checking) Object Type 34</p> <p>2.4.5 Changing Object Type 34</p> <p>2.5 R Packages (Libraries) 35</p> <p>2.5.1 Installing R Packages 35</p> <p>2.5.2 Name Spaces 36</p> <p>2.5.3 Updating R 37</p> <p>2.5.4 Datasets in R Packages 37</p> <p>2.6 Importing (Reading) Data into R Workspace 37</p> <p>2.6.1 Best Practice: Working Directory 38</p> <p>2.6.2 Reading ASCII (Text) Files 38</p> <p>2.6.3 Reading CSV Files 40</p> <p>2.6.4 Reading Excel Files 40</p> <p>2.6.5 Reading Files from the Internet 41</p> <p>2.7 Writing (Exporting) Data to Files 42</p> <p>2.7.1 Diverting Normal R Output to a File 42</p> <p>2.7.2 Saving Graphs in Files 42</p> <p>2.7.3 Exporting Data to Files 43</p> <p>2.8 Some Arithmetic and Other Operators 43</p> <p>2.8.1 Vectors 43</p> <p>2.8.2 Matrix Computations 45</p> <p>2.9 Programming in R 50</p> <p>2.9.1 Best Practice: Script Files 50</p> <p>2.9.2 Some Useful Commands or Functions 50</p> <p>2.9.3 Conditional Execution 51</p> <p>2.9.4 Loops 53</p> <p>2.9.5 Functions and Functionals 54</p> <p>2.9.6 User Defined Functions 55</p> <p>2.10 Bibliographic Notes 60</p> <p><b>3 Simple Linear Regression 65</b></p> <p>3.1 Introduction 65</p> <p>3.2 Covariance and Correlation Coefficient 65</p> <p>3.3 Example: Computer Repair Data 69</p> <p>3.4 The Simple Linear Regression Model 72</p> <p>3.5 Parameter Estimation 73</p> <p>3.6 Tests of Hypotheses 77</p> <p>3.7 Confidence Intervals 82</p> <p>3.8 Predictions 83</p> <p>3.9 Measuring the Quality of Fit 84</p> <p>3.10 Regression Line Through the Origin 88</p> <p>3.11 Trivial Regression Models 89</p> <p>3.12 Bibliographic Notes 90</p> <p><b>4 Multiple Linear Regression 97</b></p> <p>4.1 Introduction 97</p> <p>4.2 Description of the Data and Model 97</p> <p>4.3 Example: Supervisor Performance Data 98</p> <p>4.4 Parameter Estimation 100</p> <p>4.5 Interpretations of Regression Coefficients 101</p> <p>4.6 Centering and Scaling 104</p> <p>4.6.1 Centering and Scaling in Intercept Models 104</p> <p>4.6.2 Scaling in No-Intercept Models 105</p> <p>4.7 Properties of the Least Squares Estimators 106</p> <p>4.8 Multiple Correlation Coefficient 107</p> <p>4.9 Inference for Individual Regression Coefficients 108</p> <p>4.10 Tests of Hypotheses in a Linear Model 111</p> <p>4.10.1 Testing All Regression Coefficients Equal to Zero</p> <p>4.10.2 Testing a Subset of Regression Coefficients Equal to 113</p> <p>4.10.3 Testing the Equality of Regression Coefficients</p> <p>4.10.4 Estimating and Testing of Regression Parameters 118</p> <p>4.11 Predictions 121</p> <p>4.12 Summary 122</p> <p><b>5 Regression Diagnostics: Detection of Model Violations 131</b></p> <p>5.1 Introduction 131</p> <p>5.2 The Standard Regression Assumptions 132</p> <p>5.3 Various Types of Residuals 134</p> <p>5.4 Graphical Methods 136</p> <p>5.5 Graphs Before Fitting a Model 139</p> <p>5.5.1 One-Dimensional Graphs 139 5.5.2 Two-Dimensional Graphs 140</p> <p>5.5.3 Rotating Plots 142</p> <p>5.5.4 Dynamic Graphs 142</p> <p>5.6 Graphs After Fitting a Model 143</p> <p>5.7 Checking Linearity and Normality Assumptions 143</p> <p>5.8 Leverage, Influence, and Outliers 144</p> <p>5.8.1 Outliers in the Response Variable 146</p> <p>5.8.2 Outliers in the Predictors 146</p> <p>5.8.3 Masking and Swamping Problems 147</p> <p>5.9 Measures of Influence 148</p> <p>5.9.1 Cook’s Distance 150</p> <p>5.9.2 Welsch and Kuh Measure 151</p> <p>5.9.3 Hadi’s Influence Measure 151</p> <p>5.10 The Potential-Residual Plot 152</p> <p>5.11 Regression Diagnostics in R 154 5.12 What to Do with the Outliers? 155</p> <p>5.13 Role of Variables in a Regression Equation 156</p> <p>5.11.1 Added-Variable Plot 156</p> <p>5.11.2 Residual Plus Component Plot 157</p> <p>5.14 Effects of an Additional Predictor 159</p> <p>5.15 Robust Regression 161</p> <p><b>6 Qualitative Variables as Predictors 167</b></p> <p>6.1 Introduction 167</p> <p>6.2 Salary Survey Data 168</p> <p>6.3 Interaction Variables 171</p> <p>6.4 Systems of Regression Equations 175</p> <p>6.4.1 Models with Different Slopes and Different Intercepts 176</p> <p>6.4.2 Models with Same Slope and Different Intercepts 183</p> <p>6.4.3 Models with Same Intercept and Different Slopes 184</p> <p>6.5 Other Applications of Indicator Variables 185</p> <p>6.6 Seasonality 186</p> <p>6.7 Stability of Regression Parameters Over Time 187</p> <p><b>7 Transformation of Variables 195</b></p> <p>7.1 Introduction 195</p> <p>7.2 Transformations to Achieve Linearity 197</p> <p>7.3 Bacteria Deaths Due to X-Ray Radiation 199</p> <p>7.3.1 Inadequacy of a Linear Model 200</p> <p>7.3.2 Logarithmic Transformation for Achieving Linearity 201</p> <p>7.4 Transformations to Stabilize Variance 203</p> <p>7.5 Detection of Heteroscedastic Errors 208</p> <p>7.6 Removal of Heteroscedasticity 210</p> <p>7.7 Weighted Least Squares 211</p> <p>7.8 Logarithmic Transformation of Data 212</p> <p>7.9 Power Transformation 213</p> <p>7.10 Summary 216</p> <p><b>8 Weighted Least Squares 223</b></p> <p>8.1 Introduction 223</p> <p>8.2 Heteroscedastic Models 224</p> <p>8.2.1 Supervisors Data 224</p> <p>8.2.2 College Expense Data 226</p> <p>8.3 Two-Stage Estimation 227</p> <p>8.4 Education Expenditure Data 229</p> <p>8.5 Fitting a Dose-Response Relationship Curve 237</p> <p><b>9 The Problem of Correlated Errors 241</b></p> <p>9.1 Introduction: Autocorrelation 241</p> <p>9.2 Consumer Expenditure and Money Stock 242</p> <p>9.3 Durbin-Watson Statistic 245</p> <p>9.4 Removal of Autocorrelation by Transformation 246</p> <p>9.5 Iterative Estimation with Autocorrelated Errors 249</p> <p>9.6 Autocorrelation and Missing Variables 250</p> <p>9.7 Analysis of Housing Starts 251</p> <p>9.8 Limitations of the Durbin-Watson Statistic 253</p> <p>9.9 Indicator Variables to Remove Seasonality 255</p> <p>9.10 Regressing Two Time Series 257</p> <p><b>10 Analysis of Collinear Data 261</b></p> <p>10.1 Introduction 261</p> <p>10.2 Effects of Collinearity on Inference 262</p> <p>10.3 Effects of Collinearity on Forecasting 267</p> <p>CONTENTS</p> <p>10.4 Detection of Collinearity 271</p> <p>10.4.1 Simple Signs of Collinearity 271</p> <p>10.4.2 Variance Inflation Factors 274</p> <p>10.4.3 The Condition Indices 276</p> <p><b>11 Working With Collinear Data 283</b></p> <p>11.1 Introduction 283</p> <p>11.2 Principal Components 283</p> <p>11.3 Computations Using Principal Components 287</p> <p>11.4 Imposing Constraints 289</p> <p>11.5 Searching for Linear Functions of the β's 292</p> <p>11.6 Biased Estimation of Regression Coefficients 295</p> <p>11.7 Principal Components Regression 296</p> <p>11.8 Reduction of Collinearity in the Estimation Data 298</p> <p>11.9 Constraints on the Regression Coefficients 300</p> <p>11.10 Principal Components Regression: A Caution 301</p> <p>11.11 Ridge Regression 303</p> <p>11.12 Estimation by the Ridge Method 305</p> <p>11.13 Ridge Regression: Some Remarks 308</p> <p>11.14 Summary 311</p> <p>11.15 Bibliographic Notes 311</p> <p><b>12 Variable Selection Procedures 321</b></p> <p>12.1 Introduction 321</p> <p>12.2 Formulation of the Problem 322</p> <p>12.3 Consequences of Variables Deletion 322</p> <p>12.4 Uses of Regression Equations 324</p> <p>12.4.1 Description and Model Building 324</p> <p>12.4.2 Estimation and Prediction 324</p> <p>12.4.3 Control 324</p> <p>12.5 Criteria for Evaluating Equations 325</p> <p>12.5.1 Residual Mean Square 325</p> <p>12.5.2 Mallows Cp 326</p> <p>12.5.3 Information Criteria 327</p> <p>12.6 Collinearity and Variable Selection 328</p> <p>12.7 Evaluating All Possible Equations 328</p> <p>12.8 Variable Selection Procedures 329</p> <p>12.8.1 Forward Selection Procedure 329</p> <p>12.8.2 Backward Elimination Procedure 330</p> <p>12.8.3 Stepwise Method 330</p> <p>12.9 General Remarks on Variable Selection Methods 331</p> <p>12.10 A Study of Supervisor Performance 332</p> <p>12.11 Variable Selection with Collinear Data 336</p> <p>12.12 The Homicide Data 336</p> <p>12.13 Variable Selection Using Ridge Regression 339</p> <p>12.14 Selection of Variables in an Air Pollution Study 339</p> <p>12.15 A Possible Strategy for Fitting Regression Models 345</p> <p>12.16 Bibliographic Notes 347</p> <p><b>13 Logistic Regression 353</b></p> <p>13.1 Introduction 353</p> <p>13.2 Modeling Qualitative Data 354</p> <p>13.3 The Logit Model 354</p> <p>13.4 Example: Estimating Probability of Bankruptcies 356</p> <p>13.5 Logistic Regression Diagnostics 358</p> <p>13.6 Determination of Variables to Retain 359</p> <p>13.7 Judging the Fit of a Logistic Regression 362</p> <p>13.8 The Multinomial Logit Model 364</p> <p>13.8.1 Multinomial Logistic Regression 364</p> <p>13.8.2 Example: Determining Chemical Diabetes 365</p> <p>13.8.3 Ordinal Logistic Regression 368</p> <p>13.8.4 Example: Determining Chemical Diabetes Revisited 368</p> <p>13.9 Classification Problem: Another Approach 370</p> <p><b>14 Further Topics 375</b></p> <p>14.1 Introduction 375</p> <p>14.2 Generalized Linear Model 375</p> <p>14.3 Poisson Regression Model 376</p> <p>14.4 Introduction of New Drugs 377</p> <p>14.5 Robust Regression 378</p> <p>14.6 Fitting a Quadratic Model 379</p> <p>14.7 Distribution of PCB in U.S. Bays 381</p> <p>Exercises 384</p> <p>References 385</p> <p>Index</p>
<p><b>Ali S. Hadi, PhD, </b>Fellow ASA (1997), Member ISI (1998), Fellow AAS (2019) is Distinguished University Professor and former Chair of the Department of Mathematics and Actuarial Science at the American University in Cairo (AUC). He is also the Founder of the Actuarial Science Program at AUC (2004), the Founder of the Data Science Program at AUC (2019), and the former Vice Provost and Director of Graduate Studies and Research at AUC. Dr. Hadi is also a Stephen H. Weiss Presidential Fellow and Professor Emeritus at Cornell University, USA. He is the author and co-author of four other books and numerous articles. For more info, see his Website at: www1.aucegypt.edu/faculty/hadi.
<p><B>A STRAIGHTFORWARD AND CONCISE DISCUSSION OF THE ESSENTIALS OF REGRESSION ANALYSIS</B> <p>In the newly revised sixth edition of <i>Regression Analysis By Example Using R</i>, distinguished statistician Dr Ali S. Hadi delivers an expanded and thoroughly updated discussion of exploratory data analysis using regression analysis in R. The book provides in-depth treatments of regression diagnostics, transformation, multicollinearity, logistic regression, and robust regression. <p>The author clearly demonstrates effective methods of regression analysis with examples that contain the types of data irregularities commonly encountered in the real world. This newest edition also offers a brand-new, easy to read chapter on the freely available statistical software package R. <p>Readers will also find: <ul><li>Reorganized, expanded, and upgraded exercises at the end of each chapter with an emphasis on data analysis</li> <li>Updated data sets and examples throughout the book</li> <li>Complimentary access to a companion website that provides data sets in xlsx, csv, and txt format</li></ul> <p>Perfect for upper-level undergraduate or beginning graduate students in statistics, mathematics, biostatistics, and computer science programs, <i>Regression Analysis By Example Using R </i>will also benefit readers who need a reference for quick updates on regression methods and applications.

Diese Produkte könnten Sie auch interessieren:

Statistics for Microarrays
Statistics for Microarrays
von: Ernst Wit, John McClure
PDF ebook
90,99 €