Example: biology

Data Mining Classification: Basic Concepts and Techniques

data Mining Classification: Basic Concepts and TechniquesLecture Notes for Chapter 3 Introduction to data Mining , 2ndEditionbyTan, Steinbach, Karpatne, Kumar2/1/2021 Introduction to data Mining , 2ndEdition1 Classification: DefinitionlGiven a collection of records (training set ) Each record is by characterized by a tuple (x,y), where x is the attribute set and y is the class label x: attribute, predictor, independent variable, input y: class, response, dependent variable, outputlTask: Learn a model that maps each attribute set x into one of the predefined class labels y2/1/2021 Introduction to data Mining , 2ndEdition212 Examples of Classification TaskTa s kAttribute set, xClass label, yCategorizing email messagesFeatures extracted from email message header and contentspam or non-spamIdentifying tumor cellsFeatures extracted from x-rays or MRI scansmalignant or benign cellsCataloging galaxiesFeatures extracted from telescope imagesElliptical, spiral, or irregular-shaped galaxies2/1/2021 Introduction to data Mining , 2ndEdition3 General Approach for Building Classification Model2/1/2021 Introduction to data Mining , 2ndEdition434 Classification Techniques Base Classifiers Decision Tree based Methods Rule-based Methods Nearest-neighbor Na ve Bayes and

2/1/2021 Introduction to Data Mining, 2nd Edition 23 Test Condition for Ordinal Attributes l Multi-way split: – Use as many partitions as distinct values l Binary split: – Divides values into two subsets – Preserve order property among attribute values This grouping violates order property 2/1/2021 Introduction to Data Mining, 2nd Edition ...

Tags:

  Data, Ordinal

Information

Domain:

Source:

Link to this page:

Please notify us if you found a problem with this document:

Other abuse

Transcription of Data Mining Classification: Basic Concepts and Techniques

1 data Mining Classification: Basic Concepts and TechniquesLecture Notes for Chapter 3 Introduction to data Mining , 2ndEditionbyTan, Steinbach, Karpatne, Kumar2/1/2021 Introduction to data Mining , 2ndEdition1 Classification: DefinitionlGiven a collection of records (training set ) Each record is by characterized by a tuple (x,y), where x is the attribute set and y is the class label x: attribute, predictor, independent variable, input y: class, response, dependent variable, outputlTask: Learn a model that maps each attribute set x into one of the predefined class labels y2/1/2021 Introduction to data Mining , 2ndEdition212 Examples of Classification TaskTa s kAttribute set, xClass label, yCategorizing email messagesFeatures extracted from email message header and contentspam or non-spamIdentifying tumor cellsFeatures extracted from x-rays or MRI scansmalignant or benign cellsCataloging galaxiesFeatures extracted from telescope imagesElliptical, spiral, or irregular-shaped galaxies2/1/2021 Introduction to data Mining , 2ndEdition3 General Approach for Building Classification Model2/1/2021 Introduction to data Mining , 2ndEdition434 Classification Techniques Base Classifiers Decision Tree based Methods Rule-based Methods Nearest-neighbor Na ve Bayes and Bayesian Belief Networks Support Vector Machines Neural Networks.

2 Deep Neural Nets Ensemble Classifiers Boosting, Bagging, Random Forests2/1/2021 Introduction to data Mining , 2ndEdition5 Example of a Decision TreeID Home Owner Marital Status Annual Income Defaulted Borrower 1 Yes Single 125K No 2 No Married 100K No 3 No Single 70K No 4 Yes Married 120K No 5 No Divorced 95K Yes 6 No Married 60K No 7 Yes Divorced 220K No 8 No Single 85K Yes 9 No Married 75K No 10 No Single 90K Yes 10 Home OwnerMarStIncomeYESNONONOYe sN oMarriedSingle, Divorced< 80K> 80 KSplitting AttributesTraining DataModel: Decision Tree2/1/2021 Introduction to data Mining , 2ndEdition656 Apply Model to Test DataHome OwnerMarStIncomeYESNONONOYe sN oMarriedSingle, Divorced< 80K> 80 KHome Owner Marital Status Annual Income Defaulted Borrower No Married 80K ? 10 Test DataStart from the root of to data Mining , 2ndEdition7 Apply Model to Test DataMarStIncomeYESNONONOYe sN oMarriedSingle, Divorced< 80K> 80 KHome Owner Marital Status Annual Income Defaulted Borrower No Married 80K ?

3 10 Test DataHome Owner2/1/2021 Introduction to data Mining , 2ndEdition878 Apply Model to Test DataMarStIncomeYESNONONOYe sNoMarriedSingle, Divorced< 80K> 80 KHome Owner Marital Status Annual Income Defaulted Borrower No Married 80K ? 10 Test DataHome Owner2/1/2021 Introduction to data Mining , 2ndEdition9 Apply Model to Test DataMarStIncomeYESNONONOYe sNoMarriedSingle, Divorced< 80K> 80 KHome Owner Marital Status Annual Income Defaulted Borrower No Married 80K ? 10 Test DataHome Owner2/1/2021 Introduction to data Mining , 2ndEdition10910 Apply Model to Test DataMarStIncomeYESNONONOYe sNoMarried Single, Divorced< 80K> 80 KHome Owner Marital Status Annual Income Defaulted Borrower No Married 80K ? 10 Test DataHome Owner2/1/2021 Introduction to data Mining , 2ndEdition11 Apply Model to Test DataMarStIncomeYESNONONOYe sNoMarried Single, Divorced< 80K> 80 KHome Owner Marital Status Annual Income Defaulted Borrower No Married 80K ?

4 10 Test DataAssign Defaulted to No Home Owner2/1/2021 Introduction to data Mining , 2ndEdition121112 Another Example of Decision TreeMarStHome OwnerIncomeYESNONONOYe sNoMarriedSingle, Divorced< 80K> 80 KThere could be more than one tree that fits the same data !ID Home Owner Marital Status Annual Income Defaulted Borrower 1 Yes Single 125K No 2 No Married 100K No 3 No Single 70K No 4 Yes Married 120K No 5 No Divorced 95K Yes 6 No Married 60K No 7 Yes Divorced 220K No 8 No Single 85K Yes 9 No Married 75K No 10 No Single 90K Yes 10 2/1/2021 Introduction to data Mining , 2ndEdition13 Decision Tree Classification TaskApply ModelLearn ModelTid Attrib1 Attrib2 Attrib3 Class 1 Yes Large 125K No 2 No Medium 100K No 3 No Small 70K No 4 Yes Medium 120K No 5 No Large 95K Yes 6 No Medium 60K No 7 Yes Large 220K No 8 No Small 85K Yes 9 No Medium 75K No 10 No Small 90K Yes 10 Tid Attrib1 Attrib2 Attrib3 Class 11 No Small 55K ?

5 12 Yes Medium 80K ? 13 Yes Large 110K ? 14 No Small 95K ? 15 No Large 67K ? 10 Decision Tree2/1/2021 Introduction to data Mining , 2ndEdition141314 Decision Tree Induction Many Algorithms: Hunt s Algorithm (one of the earliest) CART ID3, SLIQ,SPRINT2/1/2021 Introduction to data Mining , 2ndEdition15 General Structure of Hunt s AlgorithmlLet Dtbe the set of training records that reach a node tlGeneral Procedure: If Dtcontains records that belong the same class yt, then t is a leaf node labeled as yt If Dtcontains records that belong to more than one class, use an attribute test to split the data into smaller subsets. Recursively apply the procedure to each Home Owner Marital Status Annual Income Defaulted Borrower 1 Yes Single 125K No 2 No Married 100K No 3 No Single 70K No 4 Yes Married 120K No 5 No Divorced 95K Yes 6 No Married 60K No 7 Yes Divorced 220K No 8 No Single 85K Yes 9 No Married 75K No 10 No Single 90K Yes 10 2/1/2021 Introduction to data Mining , 2ndEdition161516 Hunt s Algorithm(3,0)(4,3)(3,0)(1,3)(3,0)(3,0)( 1,0)(0,3)(3,0)(7,3) ID Home Owner Marital Status Annual Income Defaulted Borrower 1 Yes Single 125K No 2 No Married 100K No 3 No Single 70K No 4 Yes Married 120K No 5 No Divorced 95K Yes 6 No Married 60K No 7 Yes Divorced 220K No 8 No Single 85K Yes 9 No Married 75K No 10 No Single 90K Yes 10 2/1/2021 Introduction to data Mining , 2ndEdition17 Hunt s Algorithm(3,0)(4,3)(3,0)(1,3)(3,0)(3,0)( 1,0)(0,3)(3,0)(7,3)

6 ID Home Owner Marital Status Annual Income Defaulted Borrower 1 Yes Single 125K No 2 No Married 100K No 3 No Single 70K No 4 Yes Married 120K No 5 No Divorced 95K Yes 6 No Married 60K No 7 Yes Divorced 220K No 8 No Single 85K Yes 9 No Married 75K No 10 No Single 90K Yes 10 2/1/2021 Introduction to data Mining , 2ndEdition181718 Hunt s Algorithm(3,0)(4,3)(3,0)(1,3)(3,0)(3,0)( 1,0)(0,3)(3,0)(7,3) ID Home Owner Marital Status Annual Income Defaulted Borrower 1 Yes Single 125K No 2 No Married 100K No 3 No Single 70K No 4 Yes Married 120K No 5 No Divorced 95K Yes 6 No Married 60K No 7 Yes Divorced 220K No 8 No Single 85K Yes 9 No Married 75K No 10 No Single 90K Yes 10 2/1/2021 Introduction to data Mining , 2ndEdition19 Hunt s Algorithm(3,0)(4,3)(3,0)(1,3)(3,0)(3,0)( 1,0)(0,3)(3,0)(7,3) ID Home Owner Marital Status Annual Income Defaulted Borrower 1 Yes Single 125K No 2 No Married 100K No 3 No Single 70K No 4 Yes Married 120K No 5 No Divorced 95K Yes 6 No Married 60K No 7 Yes Divorced 220K No 8 No Single 85K Yes 9 No Married 75K No 10 No Single 90K Yes 10 2/1/2021 Introduction to data Mining , 2ndEdition201920 Design Issues of Decision Tree InductionlHow should training records be split?

7 Method for expressing test condition depending on attribute types Measure for evaluating the goodness of a test conditionlHow should the splitting procedure stop? Stop splitting if all the records belong to the same class or have identical attribute values Early termination 2/1/2021 Introduction to data Mining , 2ndEdition21 Methods for Expressing Test ConditionslDepends on attribute types Binary Nominal ordinal Continuous2/1/2021 Introduction to data Mining , 2ndEdition222122 Test Condition for Nominal Attributes Multi-way split: Use as many partitions as distinct values. Binary split: Divides values into two subsets2/1/2021 Introduction to data Mining , 2ndEdition23 Test Condition for ordinal AttributeslMulti-way split: Use as many partitions as distinct valueslBinary split: Divides values into two subsets Preserve order property among attribute valuesThis grouping violates order property2/1/2021 Introduction to data Mining , 2ndEdition242324 Test Condition for Continuous Attributes2/1/2021 Introduction to data Mining , 2ndEdition25 Splitting Based on Continuous Attributes Different ways of handling Discretizationto form an ordinal categorical attributeRanges can be found by equal interval bucketing, equal frequency bucketing (percentiles), or clustering.

8 Static discretize once at the beginning Dynamic repeat at each node Binary Decision: (A < v) or (A v) consider all possible splits and finds the best cut can be more compute intensive2/1/2021 Introduction to data Mining , 2ndEdition262526 How to determine the Best SplitBefore Splitting: 10 records of class 0,10 records of class 1 Which test condition is the best?2/1/2021 Introduction to data Mining , 2ndEdition27 How to determine the Best SplitlGreedy approach: Nodes with purerclass distribution are preferredlNeed a measure of node impurity:High degree of impurityLow degree of impurity2/1/2021 Introduction to data Mining , 2ndEdition282728 Measures of Node ImpuritylGini IndexlEntropylMisclassification error2/1/2021 Introduction to data Mining , 2ndEdition29 =1 = ( ) =1 max [ ( )]Where is the frequency of class at node t, and is the total number of classes Finding the Best impurity measure (P) before impurity measure (M) after splittinglCompute impurity measure of each child nodelM is the weighted impurity of child the attribute test condition that produces the highest gainGain = P - Mor equivalently, lowest impurity measure after splitting (M) 2/1/2021 Introduction to data Mining , 2ndEdition302930 Finding the Best SplitB?

9 Ye sN oNode N3 Node N4A?Ye sN oNode N1 Node N2 Before Splitting:C0 N10 C1 N11 C0 N20 C1 N21 C0 N30 C1 N31 C0 N40 C1 N41 C0 N00 C1 N01 PM11M12M21M22M1M2 Gain = P M1 vs P M22/1/2021 Introduction to data Mining , 2ndEdition31 Measure of Impurity: GINI Gini Index for a given node Where is the frequency of class at node , and is the total number of classes Maximum of 1 1/ when records are equally distributed among all classes, implying the least beneficial situation for classification Minimum of 0 when all records belong to one class, implying the most beneficial situation for classification Gini index is used in decision tree algorithms such as CART, SLIQ, SPRINT2/1/2021 Introduction to data Mining , 2ndEdition32 =1 3132 Measure of Impurity: GINI Gini Index for a given node t : For 2-class problem (p, 1 p): GINI = 1 p2 (1 p)2= 2p (1-p)C10C26 Gini= to data Mining , 2ndEdition33 =1 Computing Gini Index of a Single NodeC1 0 C2 6 C1 2 C2 4 C1 1 C2 5 P(C1) = 0/6 = 0 P(C2) = 6/6 = 1 Gini = 1 P(C1)2 P(C2)2= 1 0 1 = 0 P(C1) = 1/6 P(C2) = 5/6 Gini = 1 (1/6)2 (5/6)2= (C1) = 2/6 P(C2) = 4/6 Gini = 1 (2/6)2 (4/6)2= to data Mining , 2ndEdition34 =1 3334 Computing Gini Index for a Collection of NodeslWhen a node is split into partitions (children)where, = number of records at child , = number of records at parent node.

10 2/1/2021 Introduction to data Mining , 2ndEdition35 = ( ) Binary Attributes: Computing GINI Index Splits into two partitions (child nodes) Effect of Weighing partitions: Larger and purer partitions are soughtB?Ye sN oNode N1 Node N2 Parent C1 7 C2 5 Gini = N1 N2 C1 5 2 C2 1 4 Gini= Gini(N1) = 1 (5/6)2 (1/6)2= Gini(N2) = 1 (2/6)2 (4/6)2= Gini of N1 N2= 6/12 * + 6/12 * = = to data Mining , 2ndEdition363536 Categorical Attributes: Computing Gini IndexlFor each distinct value, gather counts for each class in the datasetlUse the count matrix to make decisions CarType {Sports, Luxury} {Family} C1 9 1 C2 7 3 Gini CarType {Sports} {Family,Luxury} C1 8 2 C2 0 10 Gini CarType Family Sports Luxury C1 1 8 1 C2 3 0 7 Gini Multi-way splitTwo-way split (find best partition of values)Which of these is the best?


Related search queries