11/21/2018 1
Scaling up classification rule induction through parallel processing
Presented by Melissa Kremer and Pierre Duez
Background/Motivation
- Large datasets require parallel approaches
- Datasets can be:
Scaling up classification rule induction through parallel - - PDF document
11/21/2018 Scaling up classification rule induction through parallel processing Presented by Melissa Kremer and Pierre Duez Background/Motivation Large datasets require parallel approaches Datasets can be: Extremely large (NASAs
11/21/2018 1
Presented by Melissa Kremer and Pierre Duez
11/21/2018 2
○ Geographically distributed data mining ■ Multiple data sources ■ Collaboration between multiple stakeholders ■ Cost/logistics to collating all data in one location ○ Computationally distributed data mining ■ Also referred to as “parallel data mining” ■ Scaling of data mining by distributing the load to multiple computers ■ Borne of a single, coherent dataset
11/21/2018 3
11/21/2018 4
11/21/2018 5
11/21/2018 6
11/21/2018 7
○ By number of attributes: ■ Keep n top attributes ■ Keep x% of the attributes with highest gain ○ By information gain ■ Keep all attributes whose information again is at least x% of the best attribute ■ Keep all attributes whose information gain is at least x%
11/21/2018 8
How to choose the best attributes? From Han and Kamber (2001):
… just do PCA.
11/21/2018 9
○ SRSWOR (Simple Random Sample WithOut Replacement) ○ SRSWR (Simple Random Sample With Replacement)
○ How to choose sample size? ○ OR How to tell when your sample is sufficiently representative?
11/21/2018 10
○ Start with user‐specified window size ○ Use sample (“window”) to train initial classifier ○ Apply classifier to remaining samples in dataset (until limit of misclassified examples is reached) ○ Add misclassified samples to the window and repeat
○ Does not do well on noisy datasets ○ Multiple classification/training runs ○ Naive stop conditions
Extensions to Windowing (Quinlan, 1993) including:
skewed datasets
11/21/2018 11
○ “Consistent Rule”: rule that did not misclassify a negative example
model accuracy plateaus
about accuracy vs training set size
11/21/2018 12
11/21/2018 13
○ The dataset may be divided into equal sizes, or sizes may reflect speed
○ a learning algorithm is used to learn rules from local dataset ○ Algorithms may or may not communicate training data or information about the training data
○ Use a combining procedure to form a final concept description
11/21/2018 14
○ every rule that is acceptable globally
11/21/2018 15
Conquer” approach:
○ Select attribute (A) to split on ○ Divide instances in the training set into subsets for each value of A ○ Recurse along each leaf until stop condition (pure set, exhausted attributes, no more information gain, etc.)
○ Repeatedly generate rules to cover as much of the outstanding training set as possible ○ Remove samples covered by the rules from the set, and train new rules on the remaining examples
11/21/2018 16
processor has its own memory, bus
distribution characteristics and collectively determine next partition
11/21/2018 17
11/21/2018 18
Advantages
Disadvantages
in higher nodes; cannot be rebalanced if one processor finishes early
○ Begin with synchronous tree construction ○ When communication cost becomes too high, partition tree for later stages
11/21/2018 19
class index> pairs (as well as a <class‐label, node> list for the classes.
○
Sorting for each attribute only has to happen once
projected table (and the class/node table) at a time
the <class‐label, node> table.
○ SLIQ‐R (replicate class list) ○ SLIQ‐D (distributed class list) ○ Scalable Parallelizable Induction of Decision Trees (SPRINT)
11/21/2018 20
SLIQ : SPRINT:
11/21/2018 21
Advantages:
time Disadvantages:
11/21/2018 22
classifiers on new data
○ Voting (majority or weighted) ○ Arbitration (if consensus is not reached) ○ Combining (rule set for classifiers)
11/21/2018 23
Images taken from: https://www.youtube.com/watch?v=N_V2BmjeYLw
11/21/2018 24
○
E.g. instances have the same values for attributes but are assigned to different classes
○
These can be dealt with by discarding the rule if the target class is not the majority class, then deleting the instances that are assigned to the discarded rule’s target class
11/21/2018 25
11/21/2018 26
to induce locally best rule terms.
plus its covering probability on the local Rule Term Partition
Term Partition and writes the name of the KS that induced the best rule term on the "Global Information Partition"
○ IF KS is winning expert { ■ keep locally induced rule term and derive by last induced rule term uncovered ids and write them on the Global Information ■ Partition and delete uncovered list records } ○ ELSE IF KS is not winning KS { ■ delete the locally induced rule term and wait for by best rule term uncovered ids being available on the Global Information Partition, download them ■ delete list records matching the retrieved ids. }
11/21/2018 27
list
Prism algorithm
11/21/2018 28
Strengths:
size of the datasets
Weaknesses:
must also be parallelised. Explored in other papers
11/21/2018 29
that cannot be held in memory
processors & more memory, but slower communication between processors)
construction, ensemble learning
PMCRI
11/21/2018 30
multiprocessor systems: “Nowadays, dual‐core processors are standard technology, but in the near future we will have multi‐core processors in our personal computers.” (emphasis in the original)
that many of these lessons will continue to apply.