Three Data Partitioning Strategies for Building Local Classiers Indrė Žliobaitė TU Eindhoven 2010, September 20

Set up

Ensembles Training set for each member Randomized procedure

Evaluation Competence of each member Assigned region of competence

Deterministic procedure

Ensembles Training set for each member Randomized procedure

Evaluation Competence of each member Assigned region of competence

Deterministic procedure

Set up



Specific types of ensembles, which ●

Partition the data into non intersecting regions



Train one classifier per partition



Use classifier assignment for the final decision

Classifier 4 Classifier 1

Classifier 5 Classifier 2

Classifier 3

Classifier 4 Classifier 1

Classifier 5 Classifier 2

Classifier 3

Set up ● ●

We will explore three data partitioning strategies We will build a meta ensemble consisting of local experts

Set up ● ●



We will explore three data partitioning strategies We will build a meta ensemble consisting of local experts Motivation ●

divide and conquer



use different views to the same learning problem



assess the impact of class labels to partitions



building blocks for handling contexts / concept drift

Partitioning

Three partitioning techniques ●

Cluster the input data



Cluster each class separately



Partition based on a selected feature

Toy data

Clustering all (CLU) Cluster the input data

Clustering all (CLU) Cluster the input data

Build classifiers

Clustering all (CLU) Cluster the input data

Build classifiers

Select the relevant classifier

Clustering within classes Cluster the first class

A B

Clustering within classes Cluster the first class

A B

Cluster the second class

D C

Clustering within classes Cluster the first class

A B

Build the classifiers (pairwise)

A D

B A

Cluster the second class

C

D C

D

B

C

Clustering within classes Build the classifiers (pairwise)

D

A D

Select two closest clusters = the relevant classifier

B A C

B

C

Partitioning based on a feature Slice the data and build classifiers

Partitioning based on a feature Slice the data and build classifiers

Select the relevant classifier

Experiments

Experiments ● ●



CLU, CLU2, FEA and meta ensemble (MMM) Baselines: naive (NAI), random partitoning (RAN) and no partitioning (ALL) Classification datasets from various domains ●

dimensionalities 7-58



sizes 500- 44000



two classes

Intuition ●

Partition makes sense if CLU, CL2, FEA < ALL



Small sample size problem if ALL<


Complex data if NAY< intelligent

Accuracies Credit: EEE>FEA>NAI>ALL>CLU>RAN>CL2 Shut: FEA>EEE>CL2>CLU>RAN>ALL>NAI Marc: EEE>FEA>CLU>CL2>ALL>RAN>NAI Spam: EEE>CLU>FEA>RAN>CL2>ALL>NAI Elec:

EEE>CLU>RAN>FEA>CL2>ALL>NAI

Chess: EEE>CLU>ALL>CL2>RAN>FEA>NAI

0 .3 6

0 .0 6

ALL 0 .3 4

ALL

0 .3 2

C L2

FE A

0 .3

C LU

0 .2 8 0 .2 6

R A N

0 .0 4

's h u t ' d a t a

C LU 0 .0 3

C L2 0 .0 2

M M M 0 .0 1

M M M

0 .2 4 0 .2 2

0 .0 5

R A N te s tin g e r r o r

te s t in g e r r o r

How many partitions?

FEA

'e le c ' d a t a 2

4

6

8

n u m b e r o f p a r t it io n s ( k )

10

0

2

4

6

8

n u m b e r o f p a r titio n s ( k )

10

Summary ●

Better with more partitions, but there is a risk of small taining sample



Mediciore performance individually



Meta ensemble performs well

Outlook ●

Partitioning with constraints

(e.g. Ignoring label informaition) ●

Data partitioning

for contextual and adaptive learning

Thank you

Transition images from www.images.com

Three Data Partitioning Strategies for Building Local ...

Experiments. ○ CLU, CLU2, FEA and meta ensemble (MMM). ○ Baselines: naive (NAI), random partitoning. (RAN) and no partitioning (ALL). ○ Classification datasets from various domains. ○ dimensionalities 7-58. ○ sizes 500- 44000. ○ two classes ...

2MB Sizes 0 Downloads 133 Views

Recommend Documents

Multi-view clustering via spectral partitioning and local ...
(2004), for example, show that exploiting both the textual content of web pages and the anchor text of ..... 1http://www.umiacs.umd.edu/~abhishek/papers.html.

High-Level Data Partitioning for Parallel Computing on ...
Nov 23, 2010 - Comparison of Communications on a Star Topology . ...... (2001b), the future of computing platforms is best described ... of a small number of interconnected heterogeneous computing .... as the computational (number crunching) equivale

Element-local level set method for three-dimensional ...
Jun 23, 2009 - Let Sp be the set of elements that are intact (uncracked) and share the crack front with elements ... where nI is the number of the adjacent cracked elements which share node I with the current element e. Note that ..... Xu XP, Needlem

Building a Sentiment Summarizer for Local ... - Research at Google
Apr 22, 2008 - For example, figure 1 summarizes a restaurant using aspects food ... a wide variety of entities such as hair salons, schools, mu- seums, retailers ...

Building Strong Brands: Three Models for ... - University of Minnesota
Ask About Your Brand,” Harvard Business Review, September, 80 (9), 80-89. Kevin Lane Keller (2001), “Building Customer-Based Brand Equity: A Blueprint for Creating. Strong Brands,” Marketing Management, July/August, 15-19. Kevin Lane Keller and

Optimistic Parallelism Benefits from Data Partitioning
Mar 5, 2008 - may reduce performance. A simple reductio ad absurdum shows .... flow.inc(cap); ..... In our earlier work, we introduced set iterators to express.

[PDF BOOK] Reputation Rules: Strategies for Building ...
Company's Most Valuable Asset READ ONLINE By Daniel. Diermeier. Online PDF Reputation Rules: Strategies for Building Your Company's Most Valuable .... S.C. Johnson &Son Distinguished Professor of International Marketing, Kellogg.

Method for downloading information data in wireless local loop system
Feb 26, 2008 - disadvantage in that terminal subscribers should move to a downloading area to ... data to the plurality of mobile communication terminals through the ... Additional advantages, objects, and features of the inven tion will be set ...