《数据挖掘:实用机器学习工具与技术(英文版.第3版)》
PREFACE
Updated and Revised Content
Second Edition
Third Edition
ACKNOWLEDGMENTS
ABOUT THE AUTHORS
PART I INTRODUCTION TO DATA MINING
CHAPTER 1 What's It All About?
1.1 Data Mining and Machine Learning
Describing Structural Patterns
Machine Learning
Data Mining
1.2 Simple Examples: TheWeather Problem and Others
The Weather Problem
Contact Lenses: An Idealized Problem
Irises: A Classic Numeric Dataset
CPU Performance: Introducing Numeric Prediction.
Labor Negotiations: A More Realistic Example
Soybean Classification: A Classic Machine Learning Success.
1.3 Fielded Applications
Web Mining
Decisions Involving Judgment
Screening Images
Load Forecasting
Diagnosis
Marketing and Sales
Other Applications
1.4 Machine Learning and Statistics
1.5 Generalization as Search
1,6 Data Mining and Ethics
Reidentification
Using Personal Information
Wider Issues
1.7 Further Reading
CHAPTER 2 Input: Concepts, Instances, and Attributes
2.1 What's a Concept?
2.2 What's in an Example?
Relations
Other Example Types
2.3 What's in an Attribute?
2.4 Preparing the Input
Gathering the Data Together
ARFF Format
Sparse Data
Attribute Types
Missing Values
Inaccurate Values
Getting to Know Your Data
2.5 Further Reading
CHAPTER 3 Output: Knowledge Representation
3.1 Tables
3.2 Linear Models
3.3 Trees
3.4 Rules
Classification Rules
Association Rules
Rules with Exceptions
More Expressive Rules
3.5 Instance-Based Representation
3.6 Clusters
3.7 Further Reading
CHAPTER 4 Aig0rithms: The Basic Methods
4.1 Inferring Rudimentary Rules
Missing Values and Numeric Attributes
Discussion
4.2 Statistical Modeling
Missing Values and Numeric Attributes
Naive Bayes for Document Classification
Discussion
4.3 Divide-and-Conquer: Constructing Decision Trees
Calculating Information
Highly-Branching Attributes
4.4 Covering Algorithms: Constructing Rules
Rules versus Trees
A Simple Covering Algorithm
Rules versus Decision Lists
4.5 Mining Association Rules
Item Sets
Association Rules
Generating Rules Efficiently
Discussion
4.8 Linear Models
Numeric Prediction: Linear Regression
Linear Classification: Logistic Regression
Linear Classification Using the Perceptron
Linear Classification Using Winnow
4.7 Instance-Based Learning
Distance Function
Finding Nearest Neighbors Efficiently
Discussion
4.8 Clustering
Iterative Distance-Based Clustering
Faster Distance Calculations
Discussion
4.9 Multi-Instance Learning
Aggregating the Input
Aggregating the Output
Discussion
4.10 Further Reading
4.11 Weka Implementations
CHAPTER 5 Credibility: Evaluating What's Been Learned
5.1 Training and Testing
5.2 Predicting Performance
5.3 Cross-Validation
5.4 Other Estimates
Leave-One-Out Cross-Validation
The Bootstrap
5.5 Comparing Data Mining Schemes
5.B Predicting Probabilities
Quadratic Loss Function
Informational Loss Function
Discussion
5.7 Counting the Cost
Cost-Sensitive Classification
Cost-Sensitive Learning
Lift Charts
ROC Curves
Recall-Precision Curves
Discussion
Cost Curves
5.0 Evaluating Numeric Prediction
5.9 Minimum Description Length Principle
5.10 Applying the MDL Principle to Clustering.
5.11 Further Reading
PART II ADVANCED DATA MINING
CHAPTER 6 Implementations: Real Machine Learning Schemes.
6.1 Decision Trees
Numeric Attributes
Missing Values
Pruning
Estimating Error Rates
Complexity of Decision Tree Induction
From Trees to Rules
C4.5: Choices and Options
Cost-Complexity Pruning
Discussion
6.2 Classification Rules
Criteria for Choosing Tests
Missing Values, Numeric Attributes
Generating Good Rules
Using Global Optimization
Obtaining Rules from Partial Decision Trees
Rules with Exceptions
Discussion
6.3 Association Rules
Building a Frequent-Pattern Tree
Finding Large Item Sets
Discussion
6.4 Extending Linear Models
Maximum-Margin Hyperplane
Nonlinear Class Boundaries
Support Vector Regression
Kernel Ridge Regression
Kernel Perceptron
Multilayer Perceptrons
Radial Basis Function Networks
Stochastic Gradient Descent
Discussion
6.5 Instance-Based Learning
Reducing the Number of Exemplars
Pruning Noisy Exemplars
Weighting Attributes
Generalizing Exemplars
Distance Functions for Generalized
Exemplars
Generalized Distance Functions
Discussion
6.6 Numeric Prediction with Local Linear Models
Model Trees
Building the Tree
Pruning the Tree
Nominal Attributes
Missing Values
Pseudocode for Model Tree Induction
Rules from Model Trees
Locally Weighted Linear Regression
Discussion
6.7 Bayesian Networks
Making Predictions
Learning Bayesian Networks
Specific Algorithms
Data Structures for Fast Learning
Discussion
6.8 Clustering
Choosing the Number of Clusters
Hierarchical Clustering
Example of Hierarchical Clustering
Incremental Clustering
Category Utility
Probability-Based Clustering
The EM Algorithm
Extending the Mixture Model
Bayesian Clustering
Discussion
6.0 Semisupervised Learning
Clustering for Classification
Co4raining
EM and Co-training
Discussion
6.10 Multi-Instance Learning
Converting to Single-Instance Learning
Upgrading Learning Algorithms
Dedicated Multi-Instance Methods
Discussion
6.11 Weka Implementations
CHAPTER 7 Data Transformations
7.1 Attribute Selection
Scheme-Independent Selection
Searching the Attribute Space
Scheme-Specific Selection
7.2 Discretizing Numeric Attributes
Unsupervised Discretization
Entropy-Based Discretization
Other Discretization Methods
Entropy-Based versus Error-Based Discretization
Converting Discrete Attributes to Numeric Attributes
7.3 Projections
Principal Components Analysis
Random Projections
Partial Least-Squares Regression
Text to Attribute Vectors
Time Series
7.4 Sampling
Reservoir Sampling
7.5 Cleansing
Improving Decision Trees
Robust Regression
Detecting Anomalies
One-Class Learning
7.6 Transforming Multiple Classes to Binary Ones
Simple Methods
Error-Correcting Output Codes
Ensembles of Nested Dichotomies
7.7 Calibrating Class Probabilities
7.8 Further Reading
7.9 Weka Implementations
CHAPTER 8 Ensemble Learning
8.1 Combining Multiple Models
8.9 Bagging
Bias-Variance Decomposition
Bagging with Costs
8.3 Randomization
Randomization versus Bagging
Rotation Forests
8.4 Boosting
AdaBoost
The Power of Boosting
8.5 Additive Regression
Numeric Prediction
Additive Logistic Regression
8.6 Interpretable Ensembles
Option Trees
Logistic Model Trees
8.7 Stacking
8.8 Further Reading
8.9 Weka Implementations
Chapter 9 Moving on: Applications and Beyond
9.1 Applying Data Mining
9.2 Learning from Massive Datasets
9.3 Data Stream Learning
9.4 Incorporating Domain Knowledge
9.5 Text Mining
9.6 Web Mining
9.7 Adversarial Situations
0.8 Ubiquitous Data Mining
9.9 Further Reading
PART III THE WEKA DATA MINING WORKBENCH
CHAPTER 10 Introduction to Weka
10.1 What's in Weka?
10.2 How Do You Use It?
10.3 What Else Can You Do?
10.4 How Do You Get It?
CHAPTER 11 The Explorer
11.1 Getting Started
Preparing the Data
Loading the Data into the Explorer
Building a Decision Tree
Examining the Output
Doing It Again
Working with Models
When Things Go Wrong
11.2 Exploring the Explorer
Loading and Filtering Files
Training and Testing Learning Schemes
Do It Yourself: The User Classifier
Using a Metalearner
Clustering and Association Rules
Attribute Selection
Visualization
11.3 Filtering Algorithms
Unsupervised Attribute Filters
Unsupervised Instance Filters
Supervised Filters
11.4 Learning Algorithms
Bayesian Classifiers
Trees
Rules
Functions
Neural Networks
Lazy Classifiers
Multi-Instance Classifiers
Miscellaneous Classifiers
11.5 Metalearning Algorithms
Bagging and Randomization
Boosting
Combining Classifiers
Cost-Sensitive Learning
Optimizing Performance
Retargeting Classifiers for Different Tasks
11.6 Clustering Algorithms
1 1.7 Association-Rule Learners
11.8 Attribute Selection
Attribute Subset Evaluators
Single-Attribute Evaluators
Search Methods
CHAPTER 12 The Knowledge Flow Interface
12.1 Getting Started
12.2 Components
12.3 Configuring and Connecting the Components
12.4 Incremental Learning
CHAPTER 13 The Experimenter
13.1 Getting Started
Running an Experiment
Analyzing the Results
13.2 Simple Setup
13.3 Advanced Setup
13.4 The Analyze Panel
13.5 Distributing Processing over Several Machines
CHAPTER 14 The Command-Line interface
14.1 Getting Started
14.2 The Structure of Weka
Classes, Instances, and Packages
The weka. core Package
The weka. classifiers Package
Other Packages
Javadoc Indexes
14.3 Command-Line Options
Generic Options
Scheme-Specific Options
CHAPTER 15 Embedded Machine Learning
15.1 A Simple Data Mining Application
MessageClassifierO
updateDataO
classifyMessageO
CHAPTER 16 Writing New Learning Schemes
16.1 An Example Classifier
buildClassifierO
makeTreeO
computelnfoGainO
classifylnstanceO
toSourceO
main()
16.2 Conventions for Implementing Classifiers
Capabilities
CHAPTER 17 Tutorial Exercises for the Weka Explorer
17.1 Introduction to the Explorer Interface
Loading a Dataset
The Dataset Editor
Applying a Filter
The Visualize Panel
The Classify Panel
17.2 Nearest-Neighbor Learning and Decision Trees
The Glass Dataset
Attribute Selection
Class Noise and Nearest-Neighbor Learning
Varying the Amount of Training Data
Interactive Decision Tree Construction
17.3 Classification Boundaries
Visualizing 1R
Visualizing Nearest-Neighbor Learning
Visualizing Naive Bayes
Visualizing Decision Trees and Rule Sets
Messing with the Data
17.4 Preprocessing and Parameter Tuning
Discretization
More on Discretization
Automatic Attribute Selection
More on Automatic Attribute Selection
Automatic Parameter Tuning
17.5 Document Classification
Data with String Attributes
Classifying Actual Documents
Exploring the StringToWordVector Filter
17.6 Mining Association Rules
Association-Rule Mining
Mining a Real-World Dataset
Market Basket Analysis
REFERENCES
INDEX