Details

Data Mining Techniques


Data Mining Techniques

For Marketing, Sales, and Customer Relationship Management
3. Aufl.

von: Gordon S. Linoff, Michael J. A. Berry

39,99 €

Verlag: Wiley
Format: EPUB
Veröffentl.: 23.03.2011
ISBN/EAN: 9781118087459
Sprache: englisch
Anzahl Seiten: 896

DRM-geschütztes eBook, Sie benötigen z.B. Adobe Digital Editions und eine Adobe ID zum Lesen.

Beschreibungen

The leading introductory book on data mining, fully updated and revised! <p>When Berry and Linoff wrote the first edition of <i>Data Mining Techniques</i> in the late 1990s, data mining was just starting to move out of the lab and into the office and has since grown to become an indispensable tool of modern business. This new edition—more than 50% new and revised— is a significant update from the previous one, and shows you how to harness the newest data mining methods and techniques to solve common business problems. The duo of unparalleled authors share invaluable advice for improving response rates to direct marketing campaigns, identifying new customer segments, and estimating credit risk. In addition, they cover more advanced topics such as preparing data for analysis and creating the necessary infrastructure for data mining at your company. </p> <ul> <li>Features significant updates since the previous edition and updates you on best practices for using data mining methods and techniques for solving common business problems</li> <li>Covers a new data mining technique in every chapter along with clear, concise explanations on how to apply each technique immediately</li> <li>Touches on core data mining techniques, including decision trees, neural networks, collaborative filtering, association rules, link analysis, survival analysis, and more</li> <li>Provides best practices for performing data mining using simple tools such as Excel</li> </ul> <p><i>Data Mining Techniques, Third Edition</i> covers a new data mining technique with each successive chapter and then demonstrates how you can apply that technique for improved marketing, sales, and customer support to get immediate results.</p>
<p>Introduction xxxvii</p> <p><b>Chapter 1 What Is Data Mining and Why Do It? 1</b></p> <p>What Is Data Mining? 2</p> <p>Data Mining Is a Business Process 2</p> <p>Large Amounts of Data 3</p> <p>Meaningful Patterns and Rules 3</p> <p>Data Mining and Customer Relationship Management 4</p> <p>Why Now? 6</p> <p>Data Is Being Produced 6</p> <p>Data Is Being Warehoused 6</p> <p>Computing Power Is Affordable 7</p> <p>Interest in Customer Relationship Management Is Strong 7</p> <p>Commercial Data Mining Software Products Have Become Available 8</p> <p>Skills for the Data Miner 9</p> <p>The Virtuous Cycle of Data Mining 9</p> <p>A Case Study in Business Data Mining 11</p> <p>Identifying BofA’s Business Challenge 12</p> <p>Applying Data Mining 12</p> <p>Acting on the Results 13</p> <p>Measuring the Effects of Data Mining 14</p> <p>Steps of the Virtuous Cycle 15</p> <p>Identify Business Opportunities 16</p> <p>Transform Data into Information 17</p> <p>Act on the Information 19</p> <p>Measure the Results 20</p> <p>Data Mining in the Context of the Virtuous Cycle 23</p> <p>Lessons Learned 26</p> <p><b>Chapter 2 Data Mining Applications in Marketing and Customer Relationship Management 27</b></p> <p>Two Customer Lifecycles 27</p> <p>The Customer’s Lifecycle 28</p> <p>The Customer Lifecycle 28</p> <p>Subscription Relationships versus Event-Based Relationships 30</p> <p>Organize Business Processes Around the Customer Lifecycle 32</p> <p>Customer Acquisition 33</p> <p>Customer Activation 36</p> <p>Customer Relationship Management 37</p> <p>Winback 38</p> <p>Data Mining Applications for Customer Acquisition 38</p> <p>Identifying Good Prospects 39</p> <p>Choosing a Communication Channel 39</p> <p>Picking Appropriate Messages 40</p> <p>A Data Mining Example: Choosing the Right Place to Advertise 40</p> <p>Who Fits the Profile? 41</p> <p>Measuring Fitness for Groups of Readers 44</p> <p>Data Mining to Improve Direct Marketing Campaigns 45</p> <p>Response Modeling 46</p> <p>Optimizing Response for a Fixed Budget 47</p> <p>Optimizing Campaign Profitability 49</p> <p>Reaching the People Most Influenced by the Message 53</p> <p>Using Current Customers to Learn About Prospects 54</p> <p>Start Tracking Customers Before They Become “Customers” 55</p> <p>Gather Information from New Customers 55</p> <p>Acquisition-Time Variables Can Predict Future Outcomes 56</p> <p>Data Mining Applications for Customer Relationship Management 56</p> <p>Matching Campaigns to Customers 56</p> <p>Reducing Exposure to Credit Risk 58</p> <p>Determining Customer Value 59</p> <p>Cross-selling, Up-selling, and Making Recommendations 60</p> <p>Retention 60</p> <p>Recognizing Attrition 60</p> <p>Why Attrition Matters 61</p> <p>Different Kinds of Attrition 62</p> <p>Different Kinds of Attrition Model 63</p> <p>Beyond the Customer Lifecycle 64</p> <p>Lessons Learned 65</p> <p><b>Chapter 3 The Data Mining Process 67</b></p> <p>What Can Go Wrong? 68</p> <p>Learning Things That Aren’t True 68</p> <p>Learning Things That Are True, but Not Useful 73</p> <p>Data Mining Styles 74</p> <p>Hypothesis Testing 75</p> <p>Directed Data Mining 81</p> <p>Undirected Data Mining 81</p> <p>Goals, Tasks, and Techniques 82</p> <p>Data Mining Business Goals 82</p> <p>Data Mining Tasks 83</p> <p>Data Mining Techniques 88</p> <p>Formulating Data Mining Problems: From Goals to Tasks to Techniques 88</p> <p>What Techniques for Which Tasks? 95</p> <p>Is There a Target or Targets? 96</p> <p>What Is the Target Data Like? 96</p> <p>What Is the Input Data Like? 96</p> <p>How Important Is Ease of Use? 97</p> <p>How Important Is Model Explicability? 97</p> <p>Lessons Learned 98</p> <p><b>Chapter 4 Statistics 101: What You Should Know About Data 101</b></p> <p>Occam’s Razor 103</p> <p>Skepticism and Simpson’s Paradox 103</p> <p>The Null Hypothesis 104</p> <p>P-Values 105</p> <p>Looking At and Measuring Data 106</p> <p>Categorical Values 106</p> <p>Numeric Variables 117</p> <p>A Couple More Statistical Ideas 120</p> <p>Measuring Response 120</p> <p>Standard Error of a Proportion 121</p> <p>Comparing Results Using Confidence Bounds 123</p> <p>Comparing Results Using Difference of Proportions 124</p> <p>Size of Sample 125</p> <p>What the Confidence Interval Really Means 126</p> <p>Size of Test and Control for an Experiment 127</p> <p>Multiple Comparisons 129</p> <p>The Confidence Level with Multiple Comparisons 129</p> <p>Bonferroni’s Correction 129</p> <p>Chi-Square Test 130</p> <p>Expected Values 130</p> <p>Chi-Square Value 132</p> <p>Comparison of Chi-Square to Difference of Proportions 134</p> <p>An Example: Chi-Square for Regions and Starts 134</p> <p>Case Study: Comparing Two Recommendation Systems with an A/B Test 138</p> <p>First Metric: Participating Sessions 140</p> <p>Data Mining and Statistics 144</p> <p>Lessons Learned 148</p> <p><b>Chapter 5 Descriptions and Prediction: Profiling and Predictive Modeling 151</b></p> <p>Directed Data Mining Models 152</p> <p>Defining the Model Structure and Target 152</p> <p>Incremental Response Modeling 154</p> <p>Model Stability 156</p> <p>Time-Frames in the Model Set 157</p> <p>Directed Data Mining Methodology 159</p> <p>Step 1: Translate the Business Problem into a Data Mining Problem 161</p> <p>How Will Results Be Used? 163</p> <p>How Will Results Be Delivered? 163</p> <p>The Role of Domain Experts and Information Technology 164</p> <p>Step 2: Select Appropriate Data 165</p> <p>What Data Is Available? 166</p> <p>How Much Data Is Enough? 167</p> <p>How Much History Is Required? 167</p> <p>How Many Variables? 168</p> <p>What Must the Data Contain? 168</p> <p>Step 3: Get to Know the Data 169</p> <p>Examine Distributions 169</p> <p>Compare Values with Descriptions 170</p> <p>Validate Assumptions 170</p> <p>Ask Lots of Questions 171</p> <p>Step 4: Create a Model Set 172</p> <p>Assembling Customer Signatures 172</p> <p>Creating a Balanced Sample 172</p> <p>Including Multiple Timeframes 174</p> <p>Creating a Model Set for Prediction 174</p> <p>Creating a Model Set for Profiling 176</p> <p>Partitioning the Model Set 176</p> <p>Step 5: Fix Problems with the Data 177</p> <p>Categorical Variables with Too Many Values 177</p> <p>Numeric Variables with Skewed Distributions and Outliers 178</p> <p>Missing Values 178</p> <p>Values with Meanings That Change over Time 179</p> <p>Inconsistent Data Encoding 179</p> <p>Step 6: Transform Data to Bring Information to the Surface 180</p> <p>Step 7: Build Models 180</p> <p>Step 8: Assess Models 180</p> <p>Assessing Binary Response Models and Classifiers 181</p> <p>Assessing Binary Response Models Using Lift 182</p> <p>Assessing Binary Response Model Scores Using Lift Charts 184</p> <p>Assessing Binary Response Model Scores Using Profitability Models 185</p> <p>Assessing Binary Response Models Using ROC Charts 186</p> <p>Assessing Estimators 188</p> <p>Assessing Estimators Using Score Rankings 189</p> <p>Step 9: Deploy Models 190</p> <p>Practical Issues in Deploying Models 190</p> <p>Optimizing Models for Deployment 191</p> <p>Step 10: Assess Results 191</p> <p>Step 11: Begin Again 193</p> <p>Lessons Learned 193</p> <p><b>Chapter 6 Data Mining Using Classic Statistical Techniques 195</b></p> <p>Similarity Models 196</p> <p>Similarity and Distance 196</p> <p>Example: A Similarity Model for Product Penetration 197</p> <p>Table Lookup Models 203</p> <p>Choosing Dimensions 204</p> <p>Partitioning the Dimensions 205</p> <p>From Training Data to Scores 205</p> <p>Handling Sparse and Missing Data by Removing Dimensions 205</p> <p>RFM: A Widely Used Lookup Model 206</p> <p>RFM Cell Migration 207</p> <p>RFM and the Test-and-Measure Methodology 208</p> <p>RFM and Incremental Response Modeling 209</p> <p>Naïve Bayesian Models 210</p> <p>Some Ideas from Probability 210</p> <p>The Naïve Bayesian Calculation 212</p> <p>Comparison with Table Lookup Models 213</p> <p>Linear Regression 213</p> <p>The Best-fit Line 215</p> <p>Goodness of Fit 217</p> <p>Multiple Regression 220</p> <p>The Equation 220</p> <p>The Range of the Target Variable 221</p> <p>Interpreting Coefficients of Linear Regression Equations 221</p> <p>Capturing Local Effects with Linear Regression 223</p> <p>Additional Considerations with Multiple Regression 224</p> <p>Variable Selection for Multiple Regression 225</p> <p>Logistic Regression 227</p> <p>Modeling Binary Outcomes 227</p> <p>The Logistic Function 229</p> <p>Fixed Effects and Hierarchical Effects 231</p> <p>Hierarchical Effects 232</p> <p>Within and Between Effects 232</p> <p>Fixed Effects 233</p> <p>Lessons Learned 234</p> <p><b>Chapter 7 Decision Trees 237</b></p> <p>What Is a Decision Tree and How Is It Used? 238</p> <p>A Typical Decision Tree 238</p> <p>Using the Tree to Learn About Churn 240</p> <p>Using the Tree to Learn About Data and Select Variables 241</p> <p>Using the Tree to Produce Rankings 243</p> <p>Using the Tree to Estimate Class Probabilities 243</p> <p>Using the Tree to Classify Records 244</p> <p>Using the Tree to Estimate Numeric Values 244</p> <p>Decision Trees Are Local Models 245</p> <p>Growing Decision Trees 247</p> <p>Finding the Initial Split 248</p> <p>Growing the Full Tree 251</p> <p>Finding the Best Split 252</p> <p>Gini (Population Diversity) as a Splitting Criterion 253</p> <p>Entropy Reduction or Information Gain as a Splitting Criterion 254</p> <p>Information Gain Ratio 256</p> <p>Chi-Square Test as a Splitting Criterion 256</p> <p>Incremental Response as a Splitting Criterion 258</p> <p>Reduction in Variance as a Splitting Criterion for Numeric Targets 259</p> <p>F Test 262</p> <p>Pruning 262</p> <p>The CART Pruning Algorithm 263</p> <p>Pessimistic Pruning: The C5.0 Pruning Algorithm 267</p> <p>Stability-Based Pruning 268</p> <p>Extracting Rules from Trees 269</p> <p>Decision Tree Variations 270</p> <p>Multiway Splits 270</p> <p>Splitting on More Than One Field at a Time 271</p> <p>Creating Nonrectangular Boxes 271</p> <p>Assessing the Quality of a Decision Tree 275</p> <p>When Are Decision Trees Appropriate? 276</p> <p>Case Study: Process Control in a Coffee Roasting Plant 277</p> <p>Goals for the Simulator 277</p> <p>Building a Roaster Simulation 278</p> <p>Evaluation of the Roaster Simulation 278</p> <p>Lessons Learned 279</p> <p><b>Chapter 8 Artificial Neural Networks 281</b></p> <p>A Bit of History 282</p> <p>The Biological Model 283</p> <p>The Biological Neuron 285</p> <p>The Biological Input Layer 286</p> <p>The Biological Output Layer 287</p> <p>Neural Networks and Artificial Intelligence 287</p> <p>Artificial Neural Networks 288</p> <p>The Artificial Neuron 288</p> <p>The Multi-Layer Perceptron 291</p> <p>A Network Example 292</p> <p>Network Topologies 293</p> <p>A Sample Application: Real Estate Appraisal 295</p> <p>Training Neural Networks 299</p> <p>How Does a Neural Network Learn Using Back Propagation? 299</p> <p>Pruning a Neural Network 300</p> <p>Radial Basis Function Networks 303</p> <p>Overview of RBF Networks 303</p> <p>Choosing the Locations of the Radial Basis Functions 305</p> <p>Universal Approximators 305</p> <p>Neural Networks in Practice 308</p> <p>Choosing the Training Set 309</p> <p>Coverage of Values for All Features 309</p> <p>Number of Features 310</p> <p>Size of Training Set 310</p> <p>Number and Range of Outputs 310</p> <p>Rules of Thumb for Using MLPs 310</p> <p>Preparing the Data 311</p> <p>Interpreting the Output from a Neural Network 313</p> <p>Neural Networks for Time Series 315</p> <p>Time Series Modeling 315</p> <p>A Neural Network Time Series Example 316</p> <p>Can Neural Network Models Be Explained? 317</p> <p>Sensitivity Analysis 318</p> <p>Using Rules to Describe the Scores 318</p> <p>Lessons Learned 319</p> <p><b>Chapter 9 Nearest Neighbor Approaches: Memory-Based Reasoning and Collaborative Filtering 321</b></p> <p>Memory-Based Reasoning 322</p> <p>Look-Alike Models 323</p> <p>Example: Using MBR to Estimate Rents in Tuxedo, New York 324</p> <p>Challenges of MBR 327</p> <p>Choosing a Balanced Set of Historical Records 328</p> <p>Representing the Training Data 328</p> <p>Determining the Distance Function, Combination Function, and Number of Neighbors 331</p> <p>Case Study: Using MBR for Classifying Anomalies in Mammograms 331</p> <p>The Business Problem: Identifying Abnormal Mammograms 332</p> <p>Applying MBR to the Problem 332</p> <p>The Total Solution 334</p> <p>Measuring Distance and Similarity 335</p> <p>What Is a Distance Function? 335</p> <p>Building a Distance Function One Field at a Time 337</p> <p>Distance Functions for Other Data Types 340</p> <p>When a Distance Metric Already Exists 341</p> <p>The Combination Function: Asking the Neighbors for Advice 342</p> <p>The Simplest Approach: One Neighbor 342</p> <p>The Basic Approach for Categorical Targets: Democracy 342</p> <p>Weighted Voting for Categorical Targets 344</p> <p>Numeric Targets 344</p> <p>Case Study: Shazam — Finding Nearest Neighbors for Audio Files 345</p> <p>Why This Feat Is Challenging 346</p> <p>The Audio Signature 347</p> <p>Measuring Similarity 348</p> <p>Collaborative Filtering: A Nearest-Neighbor Approach to Making Recommendations 351</p> <p>Building Profiles 352 Comparing Profiles 352</p> <p>Making Predictions 353</p> <p>Lessons Learned 354</p> <p><b>Chapter 10 Knowing When to Worry: Using Survival Analysis to Understand Customers 357</b></p> <p>Customer Survival 360</p> <p>What Survival Curves Reveal 360</p> <p>Finding the Average Tenure from a Survival Curve 362</p> <p>Customer Retention Using Survival 364</p> <p>Looking at Survival as Decay 365</p> <p>Hazard Probabilities 367</p> <p>The Basic Idea 368</p> <p>Examples of Hazard Functions 369</p> <p>Censoring 371</p> <p>The Hazard Calculation 372</p> <p>Other Types of Censoring 375</p> <p>From Hazards to Survival 376</p> <p>Retention 376</p> <p>Survival 378</p> <p>Comparison of Retention and Survival 378</p> <p>Proportional Hazards 380</p> <p>Examples of Proportional Hazards 381</p> <p>Stratification: Measuring Initial Effects on Survival 382</p> <p>Cox Proportional Hazards 382</p> <p>Survival Analysis in Practice 385</p> <p>Handling Different Types of Attrition 385</p> <p>When Will a Customer Come Back? 387</p> <p>Understanding Customer Value 389</p> <p>Forecasting 392</p> <p>Hazards Changing over Time 393</p> <p>Lessons Learned 394</p> <p><b>Chapter 11 Genetic Algorithms and Swarm Intelligence 397</b></p> <p>Optimization 398</p> <p>What Is an Optimization Problem? 398</p> <p>An Optimization Problem in Ant World 399</p> <p>E Pluribus Unum 400</p> <p>A Smarter Ant 401</p> <p>Genetic Algorithms 403</p> <p>A Bit of History 404</p> <p>Genetics on Computers 404</p> <p>Representing the Genome 413</p> <p>Schemata: The Building Blocks of Genetic Algorithms 414</p> <p>Beyond the Simple Algorithm 417</p> <p>The Traveling Salesman Problem 418</p> <p>Exhaustive Search 419</p> <p>A Simple Greedy Algorithm 419</p> <p>The Genetic Algorithms Approach 419</p> <p>The Swarm Intelligence Approach 420</p> <p>Case Study: Using Genetic Algorithms for Resource Optimization 421</p> <p>Case Study: Evolving a Solution for Classifying Complaints 423</p> <p>Business Context 424</p> <p>Data 425</p> <p>The Comment Signature 425</p> <p>The Genomes 426</p> <p>The Fitness Function 427</p> <p>The Results 427</p> <p>Lessons Learned 427</p> <p><b>Chapter 12 Tell Me Something New: Pattern Discovery and Data Mining 429</b></p> <p>Undirected Techniques, Undirected Data Mining 431</p> <p>Undirected versus Directed Techniques 431</p> <p>Undirected versus Directed Data Mining 431</p> <p>Case Study: Undirected Data Mining Using Directed Techniques 432</p> <p>What is Undirected Data Mining? 435</p> <p>Data Exploration 435</p> <p>Segmentation and Clustering 436</p> <p>Target Variable Definition, When the Target Is Not Explicit 438</p> <p>Simulation, Forecasting, and Agent-Based Modeling 443</p> <p>Methodology for Undirected Data Mining 455</p> <p>There Is No Methodology 456</p> <p>Things to Keep in Mind 456</p> <p>Lessons Learned 457</p> <p><b>Chapter 13 Finding Islands of Similarity: Automatic Cluster Detection 459</b></p> <p>Searching for Islands of Simplicity 461</p> <p>Customer Segmentation and Clustering 461</p> <p>Similarity Clusters 463</p> <p>Tracking Campaigns by Cluster-Based Segments 464</p> <p>Clustering Reveals an Overlooked Market Segment 466</p> <p>Fitting the Troops 467</p> <p>The K-Means Clustering Algorithm 468</p> <p>Two Steps of the K-Means Algorithm 468</p> <p>Voronoi Diagrams and K-Means Clusters 471</p> <p>Choosing the Cluster Seeds 473<br /><br />Choosing K 473</p> <p>Using K-Means to Detect Outliers 474</p> <p>Semi-Directed Clustering 475</p> <p>Interpreting Clusters 475</p> <p>Characterizing Clusters by Their Centroids 476</p> <p>Characterizing Clusters by What Differentiates Them 477</p> <p>Using Decision Trees to Describe Clusters 478</p> <p>Evaluating Clusters 479</p> <p>Cluster Measurements and Terminology 480</p> <p>Cluster Silhouettes 480</p> <p>Limiting Cluster Diameter for Scoring 483</p> <p>Case Study: Clustering Towns 484</p> <p>Creating Town Signatures 484</p> <p>Creating Clusters 486</p> <p>Determining the Right Number of Clusters 486</p> <p>Evaluating the Clusters 487</p> <p>Using Demographic Clusters to Adjust Zone Boundaries 488</p> <p>Business Success 490</p> <p>Variations on K-Means 490</p> <p>K-Medians, K-Medoids, and K-Modes 490</p> <p>The Soft Side of K-Means 494</p> <p>Data Preparation for Clustering 495</p> <p>Scaling for Consistency 496</p> <p>Use Weights to Encode Outside Information 496</p> <p>Selecting Variables for Clustering 497</p> <p>Lessons Learned 497</p> <p><b>Chapter 14 Alternative Approaches to Cluster Detection 499</b></p> <p>Shortcomings of K-Means 500</p> <p>Reasonableness 500</p> <p>An Intuitive Example 501</p> <p>Fixing the Problem by Changing the Scales 503</p> <p>What This Means in Practice 504</p> <p>Gaussian Mixture Models 505</p> <p>Adding “Gaussians” to K-Means 505</p> <p>Back to Gaussian Mixture Models 508</p> <p>Scoring GMMs 510</p> <p>Applying GMMs 511</p> <p>Divisive Clustering 513</p> <p>A Decision Tree–Like Method for Clustering 513</p> <p>Scoring Divisive Clusters 515</p> <p>Clusters and Trees 515</p> <p>Agglomerative (Hierarchical) Clustering 516</p> <p>Overview of Agglomerative Clustering Methods 516</p> <p>Clustering People by Age: An Example of An Agglomerative Clustering Algorithm 520</p> <p>Scoring Agglomerative Clusters 522</p> <p>Limitations of Agglomerative Clustering 523</p> <p>Agglomerative Clustering in Practice 525</p> <p>Combining Agglomerative Clustering and K-Means 526</p> <p>Self-Organizing Maps 527</p> <p>What Is a Self-Organizing Map? 527</p> <p>Training an SOM 530</p> <p>Scoring an SOM 531</p> <p>The Search Continues for Islands of Simplicity 532</p> <p>Lessons Learned 533</p> <p><b>Chapter 15 Market Basket Analysis and Association Rules 535</b></p> <p>Defining Market Basket Analysis 536</p> <p>Four Levels of Market Basket Data 537</p> <p>The Foundation of Market Basket Analysis: Basic Measures 539</p> <p>Order Characteristics 540</p> <p>Item (Product) Popularity 541</p> <p>Tracking Marketing Interventions 542</p> <p>Case Study: Spanish or English 543</p> <p>The Business Problem 543</p> <p>The Data 544</p> <p>Defining “Hispanicity” Preference 545</p> <p>The Solution 546</p> <p>Association Analysis 547</p> <p>Rules Are Not Always Useful 548</p> <p>Item Sets to Association Rules 551</p> <p>How Good Is an Association Rule? 553</p> <p>Building Association Rules 555</p> <p>Choosing the Right Set of Items 556</p> <p>Anonymous Versus Identified 561</p> <p>Generating Rules from All This Data 561</p> <p>Overcoming Practical Limits 565</p> <p>The Problem of Big Data 567</p> <p>Extending the Ideas 569</p> <p>Different Items on the Right- and Left-Hand Sides 569</p> <p>Using Association Rules to Compare Stores 570</p> <p>Association Rules and Cross-Selling 572</p> <p>A Typical Cross-Sell Model 572</p> <p>A More Confident Approach to Product Propensities 573</p> <p>Results from Using Confidence 574</p> <p>Sequential Pattern Analysis 574</p> <p>Finding the Sequences 575</p> <p>Sequential Association Rules 578</p> <p>Sequential Analysis Using Other Data Mining Techniques 579</p> <p>Lessons Learned 579</p> <p><b>Chapter 16 Link Analysis 581</b></p> <p>Basic Graph Theory 582</p> <p>What Is a Graph? 582</p> <p>Directed Graphs 584</p> <p>Weighted Graphs 585</p> <p>Seven Bridges of Königsberg 585</p> <p>Detecting Cycles in a Graph 588</p> <p>The Traveling Salesman Problem Revisited 589</p> <p>Social Network Analysis 593</p> <p>Six Degrees of Separation 593</p> <p>What Your Friends Say About You 595</p> <p>Finding Childcare Benefits Fraud 596</p> <p>Who Responds to Whom on Dating Sites 597</p> <p>Social Marketing 598</p> <p>Mining Call Graphs 598</p> <p>Case Study: Tracking Down the Leader of the Pack 601</p> <p>The Business Goal 601</p> <p>The Data Processing Challenge 601</p> <p>Finding Social Networks in Call Data 602</p> <p>How the Results Are Used for Marketing 602</p> <p>Estimating Customer Age 603</p> <p>Case Study: Who Is Using Fax Machines from Home? 604</p> <p>Why Finding Fax Machines Is Useful 604</p> <p>How Do Fax Machines Behave? 604</p> <p>A Graph Coloring Algorithm 605</p> <p>“Coloring” the Graph to Identify Fax Machines 606</p> <p>How Google Came to Rule the World 607</p> <p>Hubs and Authorities 608</p> <p>The Details 609</p> <p>Hubs and Authorities in Practice 611</p> <p>Lessons Learned 612</p> <p><b>Chapter 17 Data Warehousing, OLAP, Analytic Sandboxes, and Data Mining 613</b></p> <p>The Architecture of Data 615</p> <p>Transaction Data, the Base Level 616</p> <p>Operational Summary Data 617</p> <p>Decision-Support Summary Data 617</p> <p>Database Schema/Data Models 618</p> <p>Metadata 623</p> <p>Business Rules 623</p> <p>A General Architecture for Data Warehousing 624</p> <p>Source Systems 624</p> <p>Extraction, Transformation, and Load 626</p> <p>Central Repository 627</p> <p>Metadata Repository 630</p> <p>Data Marts 630</p> <p>Operational Feedback 631</p> <p>Users and Desktop Tools 631</p> <p>Analytic Sandboxes 633</p> <p>Why Are Analytic Sandboxes Needed? 634</p> <p>Technology to Support Analytic Sandboxes 636</p> <p>Where Does OLAP Fit In? 639</p> <p>What’s in a Cube? 641</p> <p>Star Schema 646</p> <p>OLAP and Data Mining 648</p> <p>Where Data Mining Fits in with Data Warehousing 650</p> <p>Lots of Data 651</p> <p>Consistent, Clean Data 651</p> <p>Hypothesis Testing and Measurement 652</p> <p>Scalable Hardware and RDBMS Support 653</p> <p>Lessons Learned 653</p> <p><b>Chapter 18 Building Customer Signatures 655</b></p> <p>Finding Customers in Data 656</p> <p>What Is a Customer? 657</p> <p>Accounts? Customers? Households? 658</p> <p>Anonymous Transactions 658</p> <p>Transactions Linked to a Card 659</p> <p>Transactions Linked to a Cookie 659</p> <p>Transactions Linked to an Account 660</p> <p>Transactions Linked to a Customer 661</p> <p>Designing Signatures 661</p> <p>Is a Customer Signature Necessary? 666</p> <p>What Does a Row Represent? 666</p> <p>Will the Signature Be Used for Predictive Modeling? 671</p> <p>Has a Target Been Defined? 672</p> <p>Are There Constraints Imposed by the Particular Data Mining Techniques to be Employed? 672</p> <p>Which Customers Will Be Included? 673</p> <p>What Might Be Interesting to Know About Customers? 673</p> <p>What a Signature Looks Like 674</p> <p>Process for Creating Signatures 677</p> <p>Some Data Is Already at the Right Level of Granularity 678</p> <p>Pivoting a Regular Time Series 679</p> <p>Aggregating Time-Stamped Transactions 680</p> <p>Dealing with Missing Values 685</p> <p>Missing Values in Source Data 685</p> <p>Unknown or Non-Existent? 687</p> <p>What Not to Do 687</p> <p>Things to Consider 689</p> <p>Lessons Learned 691</p> <p><b>Chapter 19 Derived Variables: Making the Data Mean More 693</b></p> <p>Handset Churn Rate as a Predictor of Churn 694</p> <p>Single-Variable Transformations 696</p> <p>Standardizing Numeric Variables 696</p> <p>Turning Numeric Values into Percentiles 697</p> <p>Turning Counts into Rates 698</p> <p>Relative Measures 699</p> <p>Replacing Categorical Variables with Numeric Ones 700</p> <p>Combining Variables 707</p> <p>Classic Combinations 707</p> <p>Combining Highly Correlated Variables 710</p> <p>Rent to Home Value 712</p> <p>Extracting Features from Time Series 718</p> <p>Trend 719</p> <p>Seasonality 721</p> <p>Extracting Features from Geography 722</p> <p>Geocoding 722</p> <p>Mapping 723</p> <p>Using Geography to Create Relative Measures 724</p> <p>Using Past Values of the Target Variable 725</p> <p>Using Model Scores as Inputs 725</p> <p>Handling Sparse Data 726</p> <p>Account Set Patterns 726</p> <p>Binning Sparse Values 727</p> <p>Capturing Customer Behavior from Transactions 727</p> <p>Widening Narrow Data 728</p> <p>Sphere of Influence as a Predictor of Good Customers 728</p> <p>An Example: Ratings to Rater Profile 730</p> <p>Sample Fields from the Rater Signature 730</p> <p>The Rating Signature and Derived Variables 732</p> <p>Lessons Learned 733</p> <p><b>Chapter 20 Too Much of a Good Thing? Techniques for Reducing the Number of Variables 735</b></p> <p>Problems with Too Many Variables 736</p> <p>Risk of Correlation Among Input Variables 736</p> <p>Risk of Overfitting 738</p> <p>The Sparse Data Problem 738</p> <p>Visualizing Sparseness 739</p> <p>Independence 740</p> <p>Exhaustive Feature Selection 743</p> <p>Flavors of Variable Reduction Techniques 744</p> <p>Using the Target 744</p> <p>Original versus New Variables 744</p> <p>Sequential Selection of Features 745</p> <p>The Traditional Forward Selection Methodology 745</p> <p>Forward Selection Using a Validation Set 747</p> <p>Stepwise Selection 748</p> <p>Forward Selection Using Non-Regression Techniques 748</p> <p>Backward Selection 748</p> <p>Undirected Forward Selection 749</p> <p>Other Directed Variable Selection Methods 749</p> <p>Using Decision Trees to Select Variables 750</p> <p>Variable Reduction Using Neural Networks 752</p> <p>Principal Components 753</p> <p>What Are Principal Components? 753</p> <p>Principal Components Example 758</p> <p>Principal Component Analysis 763</p> <p>Factor Analysis 767</p> <p>Variable Clustering 768</p> <p>Example of Variable Clusters 768</p> <p>Using Variable Clusters 770</p> <p>Hierarchical Variable Clustering 770</p> <p>Divisive Variable Clustering 773</p> <p>Lessons Learned 774</p> <p><b>Chapter 21 Listen Carefully to What Your Customers Say: Text Mining 775</b></p> <p>What Is Text Mining? 776</p> <p>Text Mining for Derived Columns 776</p> <p>Beyond Derived Features 777</p> <p>Text Analysis Applications 778</p> <p>Working with Text Data 781</p> <p>Sources of Text 781</p> <p>Language Effects 782</p> <p>Basic Approaches to Representing Documents 783</p> <p>Representing Documents in Practice 784</p> <p>Documents and the Corpus 786</p> <p>Case Study: Ad Hoc Text Mining 786</p> <p>The Boycott 787</p> <p>Business as Usual 787</p> <p>Combining Text Mining and Hypothesis Testing 787</p> <p>The Results 788</p> <p>Classifying News Stories Using MBR 789</p> <p>What Are the Codes? 789</p> <p>Applying MBR 790</p> <p>The Results 793</p> <p>From Text to Numbers 794</p> <p>Starting with a “Bag of Words” 794</p> <p>Term-Document Matrix 796</p> <p>Corpus Effects 797</p> <p>Singular Value Decomposition (SVD) 798</p> <p>Text Mining and Naïve Bayesian Models 800</p> <p>Naïve Bayesian in the Text World 801</p> <p>Identifying Spam Using Naïve Bayesian 801</p> <p>Sentiment Analysis 806</p> <p>DIRECTV: A Case Study in Customer Service 809</p> <p>Background 809</p> <p>Applying Text Mining 811</p> <p>Taking the Technical Approach 814</p> <p>Not an Iterative Process 818</p> <p>Continuing to Benefit 818</p> <p>Lessons Learned 819</p> <p>Index 821</p>
<p>GORDON S. LINOFF and MICHAEL J. A. BERRY are the founders of Data Miners, Inc., a consultancy specializing in data mining. They have jointly authored two of the leading data mining titles in the field, Data Mining Techniques and Mastering Data Mining (both from Wiley). They each have decades of experience applying data mining techniques to business problems in marketing and customer relationship management.</p>
<p>The newest edition of the leading introductory book on data mining, fully updated and revised</p> <p>Who will remain a loyal customer and who won't? Which messages are most effective with which segments? How can customer value be maximized? This book supplies powerful tools for extracting the answers to these and other crucial business questions from the corporate databases where they lie buried. In the years since the first edition of this book, data mining has grown to become an indispensable tool of modern business. In this latest edition, Linoff and Berry have made extensive updates and revisions to every chapter and added several new ones. The book retains the focus of earlier editions—showing marketing analysts, business managers, and data mining specialists how to harness data mining methods and techniques to solve important business problems. While never sacrificing accuracy for the sake of simplicity, Linoff and Berry present even complex topics in clear, concise English with minimal use of technical jargon or mathematical formulas. Technical topics are illustrated with case studies and practical real-world examples drawn from the authors' experiences, and every chapter contains valuable tips for practitioners. Among the techniques newly covered, or covered in greater depth, are linear and logistic regression models, incremental response (uplift) modeling, naïve Bayesian models, table lookup models, similarity models, radial basis function networks, expectation maximization (EM) clustering, and swarm intelligence. New chapters are devoted to data preparation, derived variables, principal components and other variable reduction techniques, and text mining.</p> <p>After establishing the business context with an overview of data mining applications, and introducing aspects of data mining methodology common to all data mining projects, the book covers each important data mining technique in detail.</p> <p>This third edition of Data Mining Techniques covers such topics as:</p> <ul> <li> <p>How to create stable, long-lasting predictive models</p> </li> <li> <p>Data preparation and variable selection</p> </li> <li> <p>Modeling specific targets with directed techniques such as regression, decision trees, neural networks, and memory based reasoning</p> </li> <li> <p>Finding patterns with undirected techniques such as clustering, association rules, and link analysis</p> </li> <li> <p>Modeling business time-to-event problems such as time to next purchase and expected remaining lifetime</p> </li> <li> <p>Mining unstructured text</p> </li> </ul> <p>The companion website provides data that can be used to test out the various data mining techniques in the book.</p>

Diese Produkte könnten Sie auch interessieren:

The CISO Evolution
The CISO Evolution
von: Matthew K. Sharp, Kyriakos Lambros
PDF ebook
33,99 €
Data Mining and Machine Learning Applications
Data Mining and Machine Learning Applications
von: Rohit Raja, Kapil Kumar Nagwanshi, Sandeep Kumar, K. Ramya Laxmi
EPUB ebook
190,99 €