Details

Responsible Data Science


Responsible Data Science


1. Aufl.

von: Grant Fleming, Peter C. Bruce

25,99 €

Verlag: Wiley
Format: PDF
Veröffentl.: 13.04.2021
ISBN/EAN: 9781119741770
Sprache: englisch
Anzahl Seiten: 304

DRM-geschütztes eBook, Sie benötigen z.B. Adobe Digital Editions und eine Adobe ID zum Lesen.

Beschreibungen

<p><b>Explore the most serious prevalent ethical issues in data science with this insightful new resource</b></p> <p>The increasing popularity of data science has resulted in numerous well-publicized cases of bias, injustice, and discrimination. The widespread deployment of “Black box” algorithms that are difficult or impossible to understand and explain, even for their developers, is a primary source of these unanticipated harms, making modern techniques and methods for manipulating large data sets seem sinister, even dangerous. When put in the hands of authoritarian governments, these algorithms have enabled suppression of political dissent and persecution of minorities. To prevent these harms, data scientists everywhere must come to understand how the algorithms that they build and deploy may harm certain groups or be unfair.</p> <p><i>Responsible Data Science</i> delivers a comprehensive, practical treatment of how to implement data science solutions in an even-handed and ethical manner that minimizes the risk of undue harm to vulnerable members of society. Both data science practitioners and managers of analytics teams will learn how to:</p> <ul> <li>Improve model transparency, even for black box models</li> <li>Diagnose bias and unfairness within models using multiple metrics</li> <li>Audit projects to ensure fairness and minimize the possibility of unintended harm</li> </ul> <p>Perfect for data science practitioners, <i>Responsible Data Science</i> will also earn a spot on the bookshelves of technically inclined managers, software developers, and statisticians.</p>
<p>Introduction xix</p> <p><b>Part I Motivation for Ethical Data Science and Background Knowledge 1</b></p> <p><b>Chapter 1 Responsible Data Science 3</b></p> <p>The Optum Disaster 4</p> <p>Jekyll and Hyde 5</p> <p>Eugenics 7</p> <p>Galton, Pearson, and Fisher 7</p> <p>Ties between Eugenics and Statistics 7</p> <p>Ethical Problems in Data Science Today 9</p> <p>Predictive Models 10</p> <p>From Explaining to Predicting 10</p> <p>Predictive Modeling 11</p> <p>Setting the Stage for Ethical Issues to Arise 12</p> <p>Classic Statistical Models 12</p> <p>Black-Box Methods 14</p> <p>Important Concepts in Predictive Modeling 19</p> <p>Feature Selection 19</p> <p>Model-Centric vs. Data-Centric Models 20</p> <p>Holdout Sample and Cross-Validation 20</p> <p>Overfitting 21</p> <p>Unsupervised Learning 22</p> <p>The Ethical Challenge of Black Boxes 23</p> <p>Two Opposing Forces 24</p> <p>Pressure for More Powerful AI 24</p> <p>Public Resistance and Anxiety 24</p> <p>Summary 25</p> <p><b>Chapter 2 Background: Modeling and the Black-Box Algorithm 27</b></p> <p>Assessing Model Performance 27</p> <p>Predicting Class Membership 28</p> <p>The Rare Class Problem 28</p> <p>Lift and Gains 28</p> <p>Area Under the Curve 29</p> <p>AUC vs. Lift (Gains) 31</p> <p>Predicting Numeric Values 32</p> <p>Goodness-of-Fit 32</p> <p>Holdout Sets and Cross-Validation 33</p> <p>Optimization and Loss Functions 34</p> <p>Intrinsically Interpretable Models vs. Black-Box Models 35</p> <p>Ethical Challenges with Interpretable Models 38</p> <p>Black-Box Models 39</p> <p>Ensembles 39</p> <p>Nearest Neighbors 41</p> <p>Clustering 41</p> <p>Association Rules 42</p> <p>Collaborative Filters 42</p> <p>Artificial Neural Nets and Deep Neural Nets 43</p> <p>Problems with Black-Box Predictive Models 45</p> <p>Problems with Unsupervised Algorithms 47</p> <p>Summary 48</p> <p><b>Chapter 3 The Ways AI Goes Wrong, and the Legal Implications 49</b></p> <p>AI and Intentional Consequences by Design 50</p> <p>Deepfakes 50</p> <p>Supporting State Surveillance and Suppression 51</p> <p>Behavioral Manipulation 52</p> <p>Automated Testing to Fine-Tune Targeting 53</p> <p>AI and Unintended Consequences 55</p> <p>Healthcare 56</p> <p>Finance 57</p> <p>Law Enforcement 58</p> <p>Technology 60</p> <p>The Legal and Regulatory Landscape around AI 61</p> <p>Ignorance Is No Defense: AI in the Context of Existing Law and Policy 63</p> <p>A Finger in the Dam: Data Rights, Data Privacy, and Consumer Protection Regulations 64</p> <p>Trends in Emerging Law and Policy Related to AI 66</p> <p>Summary 69</p> <p><b>Part II The Ethical Data Science Process 71</b></p> <p><b>Chapter 4 The Responsible Data Science Framework 73</b></p> <p>Why We Keep Building Harmful AI 74</p> <p>Misguided Need for Cutting-Edge Models 74</p> <p>Excessive Focus on Predictive Performance 74</p> <p>Ease of Access and the Curse of Simplicity 76</p> <p>The Common Cause 76</p> <p>The Face Thieves 78</p> <p>An Anatomy of Modeling Harms 79</p> <p>The World: Context Matters for Modeling 80</p> <p>The Data: Representation Is Everything 83</p> <p>The Model: Garbage In, Danger Out 85</p> <p>Model Interpretability: Human Understanding for Superhuman Models 86</p> <p>Efforts Toward a More Responsible Data Science 89</p> <p>Principles Are the Focus 90</p> <p>Nonmaleficence 90</p> <p>Fairness 90</p> <p>Transparency 91</p> <p>Accountability 91</p> <p>Privacy 92</p> <p>Bridging the Gap Between Principles and Practice with the Responsible Data Science (RDS) Framework 92</p> <p>Justification 94</p> <p>Compilation 94</p> <p>Preparation 95</p> <p>Modeling 96</p> <p>Auditing 96</p> <p>Summary 97</p> <p><b>Chapter 5 Model Interpretability: The What and the Why 99</b></p> <p>The Sexist Résumé Screener 99</p> <p>The Necessity of Model Interpretability 101</p> <p>Connections Between Predictive Performance and Interpretability 103</p> <p>Uniting (High) Model Performance and Model Interpretability 105</p> <p>Categories of Interpretability Methods 107</p> <p>Global Methods 107</p> <p>Local Methods 113</p> <p>Real-World Successes of Interpretability Methods 113</p> <p>Facilitating Debugging and Audit 114</p> <p>Leveraging the Improved Performance of Black-Box Models 116</p> <p>Acquiring New Knowledge 116</p> <p>Addressing Critiques of Interpretability Methods 117</p> <p>Explanations Generated by Interpretability Methods Are Not Robust 118</p> <p>Explanations Generated by Interpretability Methods Are Low Fidelity 120</p> <p>The Forking Paths of Model Interpretability 121</p> <p>The Four-Measure Baseline 122</p> <p>Building Our Own Credit Scoring Model 124</p> <p>Using Train-Test Splits 125</p> <p>Feature Selection and Feature Engineering 125</p> <p>Baseline Models 127</p> <p>The Importance of Making Your Code Work for Everyone 129</p> <p>Execution Variability 129</p> <p>Addressing Execution Variability with Functionalized Code 130</p> <p>Stochastic Variability 130</p> <p>Addressing Stochastic Variability via Resampling 130</p> <p>Summary 133</p> <p><b>Part III EDS in Practice 135</b></p> <p><b>Chapter 6 Beginning a Responsible Data Science Project 137</b></p> <p>How the Responsible Data Science Framework Addresses the Common Cause 138</p> <p>Datasets Used 140</p> <p>Regression Datasets—Communities and Crime 140</p> <p>Classification Datasets—COMPAS 140</p> <p>Common Elements Across Our Analyses 141</p> <p>Project Structure and Documentation 141</p> <p>Project Structure for the Responsible Data</p> <p>Science Framework: Everything in Its Place 142</p> <p>Documentation: The Responsible Thing to Do 145</p> <p>Beginning a Responsible Data Science Project 151</p> <p>Communities and Crime (Regression) 151</p> <p>Justification 151</p> <p>Compilation 154</p> <p>Identifying Protected Classes 157</p> <p>Preparation—Data Splitting and Feature Engineering 159</p> <p>Datasheets 161</p> <p>COMPAS (Classification) 164</p> <p>Justification 164</p> <p>Compilation 166</p> <p>Identifying Protected Classes 168</p> <p>Preparation 169</p> <p>Summary 172</p> <p><b>Chapter 7 Auditing a Responsible Data Science Project 173</b></p> <p>Fairness and Data Science in Practice 175</p> <p>The Many Different Conceptions of Fairness 175</p> <p>Different Forms of Fairness Are Trade-Offs with Each Other 177</p> <p>Quantifying Predictive Fairness Within a Data Science Project 179</p> <p>Mitigating Bias to Improve Fairness 185</p> <p>Preprocessing 185</p> <p>In-processing 186</p> <p>Postprocessing 186</p> <p>Classification Example: COMPAS 187</p> <p>Prework: Code Practices, Modeling, and Auditing 187</p> <p>Justification, Compilation, and Preparation Review 189</p> <p>Modeling 191</p> <p>Auditing 200</p> <p>Per-Group Metrics: Overall 200</p> <p>Per-Group Metrics: Error 202</p> <p>Fairness Metrics 204</p> <p>Interpreting Our Models: Why Are They Unfair? 207</p> <p>Analysis for Different Groups 209</p> <p>Bias Mitigation 214</p> <p>Preprocessing: Oversampling 214</p> <p>Postprocessing: Optimizing Thresholds</p> <p>Automatically 218</p> <p>Postprocessing: Optimizing Thresholds Manually 219</p> <p>Summary 223</p> <p><b>Chapter 8 Auditing for Neural Networks 225</b></p> <p>Why Neural Networks Merit Their Own Chapter 227</p> <p>Neural Networks Vary Greatly in Structure 227</p> <p>Neural Networks Treat Features Differently 229</p> <p>Neural Networks Repeat Themselves 231</p> <p>A More Impenetrable Black Box 232</p> <p>Baseline Methods 233</p> <p>Representation Methods 233</p> <p>Distillation Methods 234</p> <p>Intrinsic Methods 235</p> <p>Beginning a Responsible Neural Network Project 236</p> <p>Justification 236</p> <p>Moving Forward 239</p> <p>Compilation 239</p> <p>Tracking Experiments 241</p> <p>Preparation 244</p> <p>Modeling 245</p> <p>Auditing 247</p> <p>Per-Group Metrics: Overall 247</p> <p>Per-Group Metrics: Unusual Definitions of “False Positive” 248</p> <p>Fairness Metrics 249</p> <p>Interpreting Our Models: Why Are They Unfair? 252</p> <p>Bias Mitigation 253</p> <p>Wrap-Up 255</p> <p>Auditing Neural Networks for Natural Language Processing 258</p> <p>Identifying and Addressing Sources of Bias in NLP 258</p> <p>The Real World 259</p> <p>Data 260</p> <p>Models 261</p> <p>Model Interpretability 262</p> <p>Summary 262</p> <p><b>Chapter 9 Conclusion 265</b></p> <p>How Can We Do Better? 267</p> <p>The Responsible Data Science Framework 267</p> <p>Doing Better As Managers 269</p> <p>Doing Better As Practitioners 270</p> <p>A Better Future If We Can Keep It 271</p> <p>Index 273</p>
<p><b>GRANT FLEMING</b> is a Data Scientist at Elder Research Inc. His professional focus is on machine learning for social science applications, model interpretability, civic technology, and building software tools for reproducible data science.</p><p><b>PETER BRUCE</b> is the Senior Learning Officer at Elder Research, Inc., author of several best-selling texts on data science, and Founder of the Institute for Statistics Education at Statistics.com, an Elder Research Company.</p>
<p><b>A PRACTICAL GUIDE TO IDENTIFYING AND REDUCING BIAS AND UNFAIRNESS IN DATA SCIENCE</b></p><p>Rapid advancements in data science are causing increasing alarm around the world as governments, companies, other organizations, and individuals put new technologies to uses that were unimaginable just a decade ago. Medicine, finance, criminal justice, law enforcement, communication, marketing and other functions are all being transformed by the implementation of techniques and methods made possible by progressively more obscure manipulations of larger and larger data sets. Almost every day, new stories of AI gone awry appear. What can be done to avoid these issues?</p><p><i>Responsible Data Science</i> is an insightful and practical exploration of the ethical issues that arise when the newest AI technologies are applied to the largest and most sensitive data sets on the planet. The book walks you through how to implement and audit cutting-edge AI models in ways that minimize the risks of unanticipated harms. It combines detailed technical analysis with perceptive social observations to offer data scientists a real-world perspective on their field.</p><p>The inability to explain how an artificial intelligence model uses inputs can jeopardize the willingness of regulators to even consider whether these technologies comply with existing and future regulatory and legal requirements. In this book you’ll learn how to improve the interpretability of AI models, and audit them to reduce bias and unfairness, thereby inspiring greater confidence in the minds of customers, employees, regulators, legislators and other stakeholders.</p><p>Perfect for data science practitioners, statisticians, software engineers, and technically aware managers and solutions architects, <i>Responsible Data Science</i> will also earn a place in the libraries of regulators, lawyers, and policy makers whose decisions will determine how and when data solutions are implemented.</p><p><b>This groundbreaking book also covers:</b></p><ul><b><li>The various types of ethical challenges confronting modern day data scientists</li><li>How the adoption of “black box” models can aggravate issues of model transparency, bias, and fairness</li><li>How moral concepts like fairness translate (or fail to translate) into a modeling context</li><li>How model-agnostic methods can be used to make models more interpretable, identify issues of bias, and mitigate the bias discovered</li></b></ul>

Diese Produkte könnten Sie auch interessieren:

MDX Solutions
MDX Solutions
von: George Spofford, Sivakumar Harinath, Christopher Webb, Dylan Hai Huang, Francesco Civardi
PDF ebook
53,99 €
Concept Data Analysis
Concept Data Analysis
von: Claudio Carpineto, Giovanni Romano
PDF ebook
107,99 €
Handbook of Virtual Humans
Handbook of Virtual Humans
von: Nadia Magnenat-Thalmann, Daniel Thalmann
PDF ebook
150,99 €