Multi-label classification using rule-based classifier systems
Shabnam Nazmi (PhD candidate) Department of electrical and computer engineering North Carolina A&T state university Advisor: Dr. A. Homaifar
using rule-based classifier systems Shabnam Nazmi (PhD candidate) - - PowerPoint PPT Presentation
Multi-label classification using rule-based classifier systems Shabnam Nazmi (PhD candidate) Department of electrical and computer engineering North Carolina A&T state university Advisor: Dr. A. Homaifar outline Motivation
Shabnam Nazmi (PhD candidate) Department of electrical and computer engineering North Carolina A&T state university Advisor: Dr. A. Homaifar
(LCSs)
Shabnam Nazmi
2
such as classification, estimation and modeling
attribute to more than one class simultaneously
are in advantage
vital to train an accurate machine
adjacent sub-models can be handled using multi-label data with appropriate confidence levels
Shabnam Nazmi
3
Multi-class classification
Shabnam Nazmi
4 Multi-label classification
point, will correctly predict the class to which the new point belongs
Multi-class classification Multi-label classification
Shabnam Nazmi
5
the setting of multi-label classification (MLC) allows an instance to belong to several classes simultaneously.
problems
predefined topics
predicting its functional classes
Multi-class classification Multi-label classification
Shabnam Nazmi
6
πΈ: ππ£ππ’π β πππππ πππ’π π‘ππ’ πΌ: π β π
π, π ππ π
π = π§1, π§2, β¦ , π§π
examples in πΈ
examples in πΈ divided by |π|
πΌπ πΌ, πΈ = 1 |πΈ| |π
πβπ π|
|π|
|πΈ| π=1
ππ π = 1 |πΈ| 1 π |π | |π(π¦)|
|πΈ| π=1
Shabnam Nazmi
7
Shabnam Nazmi
8
labels for each instance
a single label
different label
2 binary label
data sets
Shabnam Nazmi
9
labels for each instance
a single label
random, disjoint or overlapping, subsets
with large number of labels and instances
Shabnam Nazmi
10
generative model is trained according to which, each label generates different words
distributions of its labels using EM
introduction of a new error function similar to ranking loss
the posterior probability of labels assigned to new instances
Shabnam Nazmi
11
partition a dataset with π labels into π
2 double label subsets.
between positive and negative instances
sets using associative rule mining
with this rule set, and recursively learns a new rule set from the remaining examples until no further frequent items are left.
Shabnam Nazmi
12
confidence degree for the predictions of βweakerβ hypotheses
logistic regression, output a value as a probability of a label to be true
step prior to training
Shabnam Nazmi
13
confidence degree for the predictions of βweakerβ hypotheses
logistic regression, output a value as a probability of a label to be true
step prior to training
Encounter confidence levels in training data provided by the expert
Shabnam Nazmi
14
confidence degree for the predictions of βweakerβ hypotheses
logistic regression, output a value as a probability of a label to be true
step prior to training
confidence degree along with its predicted labels for new instances
Encounter confidence levels in training data provided by the expert
Shabnam Nazmi
15
finite set of class labels
π§ β π
πΈ = { π¦1, π1, π·1 , π¦2, π2, π·2 , β¦ π¦π, ππ, π·π }
ππ,π = {1: π§π β π§, 0: π§π β π§|βπ β 1, π , π β [1, π]}
) along with a vector of confidence level (π)of the hypothesis in each
Shabnam Nazmi
16
used to extract knowledge from ML data
learning systems with a fixed rule length
rules
form of βIF condition-THEN actionβ
problems
further modifications to adapt to more general cases of classification problems, namely multi-class and multi-label
Shabnam Nazmi
17
Shabnam Nazmi
19 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
Data set: a set of triples in the form of: (sample, label, confidence level) Training instance: randomly drawn individual from the data set
Shabnam Nazmi
20 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
[P]: population of rules/classifiers Classifier parameters:
π = π₯1, π₯2, β¦ , π₯π
Shabnam Nazmi
21 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
Condition:
composed of {0,1, #}
form of an ordered list of pairs of center and spread (ππ, π‘π)
Shabnam Nazmi
22 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
Action: is an ordered list of 0,1
from a four class data set "0110β
π· = [0 1 0.9 0]
Shabnam Nazmi
23 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
[M]: matching classifiers with provided instance ππ β π‘π < π¦π < ππ + π‘π Covering: creates a matching classifier if [M] is empty
Shabnam Nazmi
24 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
CR: conflict resolution
that gets to classify the instance πΆ = πππβπ½π
generality of the classifier
Shabnam Nazmi
25 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
[A]: classifiers having the same action as the winning classifier [A ]: [M] β [A] Genetic algorithm: randomly picks two classifiers from [A] and creates two off- springs
Shabnam Nazmi
26 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
Genetic algorithm: favors classifiers with higher fitness value and lower confidence estimate error simultaneously
Shabnam Nazmi
27 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
Taxes are deducted from classifiers in both sets ππ = π
π β π· 1
π
π β π π + πΎ π· β π π
recourse sharing scheme ππ = πππβπ½ππ π
ππβπ½ππ πβ[π΅]
π0
Shabnam Nazmi
28 [P] [M] [A] [A ]
Covering CR Genetic algorithm Update rule parameters
Data set
Training instance
Model
Model: the population of trained classifiers (rules) that collectively solve the classification problem, after proper number of training iterations
plotted against training iterations
plotted against training iterations
classifiers that match the instance
average of the confidence estimates of the classifiers that match the instance
Shabnam Nazmi
29
attribute range is (β0.5, 0.5)
Shabnam Nazmi
30
Shabnam Nazmi
31 Method OVO SVM MLP Logistic Regression Random Forest LCS Accuracy 97.33 99.48 98 100 98
οΌStrength-based learning classifier system is employed to design an embedded MLC algorithm οΌClassifier structure is adapted to handle confidence level in labels provided in the training set οΌModel is tested on one real-world data set and two artificial datasets and results are provided ο±Appropriate performance measures for test accuracy needs to be implemented ο±MLC method discussed here will be extended to accuracy- based classifier system (UCS)
Shabnam Nazmi
32
Shabnam Nazmi
33