On-line Multi-label Classification

Similar documents
Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

StatPatternRecognition: Status and Plans. Ilya Narsky, Caltech

Music Genre Classification and Variance Comparison on Number of Genres

CS 1674: Intro to Computer Vision. Face Detection. Prof. Adriana Kovashka University of Pittsburgh November 7, 2016

Analysis and Clustering of Musical Compositions using Melody-based Features

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian

arxiv: v1 [cs.ir] 16 Jan 2019

Using Genre Classification to Make Content-based Music Recommendations

Detecting Musical Key with Supervised Learning

Lyrics Classification using Naive Bayes

Audio-Based Video Editing with Two-Channel Microphone

ur-caim: Improved CAIM Discretization for Unbalanced and Balanced Data

Lyric-Based Music Mood Recognition

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections

Noise Flooding for Detecting Audio Adversarial Examples Against Automatic Speech Recognition

Bibliometric evaluation and international benchmarking of the UK s physics research

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Differences Between, Changes Within: Guidelines on When to Create a New Record

CS229 Project Report Polyphonic Piano Transcription

Automatic Music Genre Classification

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

Music Genre Classification

Determining sentiment in citation text and analyzing its impact on the proposed ranking index

Automatic Music Clustering using Audio Attributes

ALF-200k: Towards Extensive Multimodal Analyses of Music Tracks and Playlists

Automatic Piano Music Transcription

Sarcasm Detection in Text: Design Document

Introduction to Knowledge Systems

Finding Sarcasm in Reddit Postings: A Deep Learning Approach

An Effective Filtering Algorithm to Mitigate Transient Decaying DC Offset

Semi-supervised Musical Instrument Recognition

base calling: PHRED...

Neural Network for Music Instrument Identi cation

Enabling editors through machine learning

First Stage of an Automated Content-Based Citation Analysis Study: Detection of Citation Sentences 1

Supervised Learning in Genre Classification

An Efficient Closed Frequent Itemset Miner for the MOA Stream Mining System

Music Information Retrieval with Temporal Features and Timbre

Introduction to Artificial Intelligence. Learning from Oberservations

MUSI-6201 Computational Music Analysis

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

Exploring Melodic Features for the Classification and Retrieval of Traditional Music in the Context of Cultural Source

Auto classification and simulation of mask defects using SEM and CAD images

Introduction to Artificial Intelligence. Learning from Oberservations

Chapter 1 Midterm Review

Automatic Laughter Detection

Singer Traits Identification using Deep Neural Network

Advanced Techniques for Spurious Measurements with R&S FSW-K50 White Paper

Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction

Neural Network Predicating Movie Box Office Performance

Computational Laughing: Automatic Recognition of Humorous One-liners

Multilabel Subject-Based Classification of Poetry

A CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS

Melody classification using patterns

A FEATURE SELECTION APPROACH FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

arxiv: v1 [cs.cl] 23 Aug 2018

Working Paper Series of the German Data Forum (RatSWD)

Multiple classifiers for different features in timbre estimation

New Results on QAM-Based 1000BASE-T Transceiver

LEARNING SPECTRAL FILTERS FOR SINGLE- AND MULTI-LABEL CLASSIFICATION OF MUSICAL INSTRUMENTS. Patrick Joseph Donnelly

Understanding Financial Statements (10th Edition) Free Ebooks PDF

How comprehensive is the PubMed Central Open Access full-text database?

Specifying Features for Classical and Non-Classical Melody Evaluation

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Week 14 Music Understanding and Classification

Topic 10. Multi-pitch Analysis

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

Efficient Testing of Variant-Rich Software Systems

TOWARDS IMPROVING ONSET DETECTION ACCURACY IN NON- PERCUSSIVE SOUNDS USING MULTIMODAL FUSION

Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio

SIGNAL + CONTEXT = BETTER CLASSIFICATION

Web of Science Unlock the full potential of research discovery

Iterative Direct DPD White Paper

An Introduction to Deep Image Aesthetics

Manuel Richey. Hossein Saiedian*

Outline. Why do we classify? Audio Classification

Cross-Dataset Validation of Feature Sets in Musical Instrument Classification

Categorization of ICMR Using Feature Extraction Strategy And MIR With Ensemble Learning

A discretization algorithm based on Class-Attribute Contingency Coefficient

Riccardo Farinelli. Charge Centroid Feasibility

Single-switch Scanning Example. Learning Objectives. Enhancing Efficiency for People who Use Switch Scanning. Overview. Part 1. Single-switch Scanning

Research & Development. White Paper WHP 232. A Large Scale Experiment for Mood-based Classification of TV Programmes BRITISH BROADCASTING CORPORATION

Automatic Rhythmic Notation from Single Voice Audio Sources

A Large Scale Experiment for Mood-Based Classification of TV Programmes

LAUGHTER serves as an expressive social signal in human

Evaluation Tools. Journal Impact Factor. Journal Ranking. Citations. H-index. Library Service Section Elyachar Central Library.

VLSI Chip Design Project TSEK06

YOU ARE WHAT YOU LIKE INFORMATION LEAKAGE THROUGH USERS INTERESTS

An Impact Analysis of Features in a Classification Approach to Irony Detection in Product Reviews

Indexing local features and instance recognition

Bibliometric measures for research evaluation

Using Generic Summarization to Improve Music Information Retrieval Tasks

Singer Recognition and Modeling Singer Error

ADE Assembler Flow for Rapid Design of High-Speed Low-Power Circuits

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

Multi-modal Analysis of Music: A large-scale Evaluation

Joint Image and Text Representation for Aesthetics Analysis

HDMI Extender via 4 Single-mode fibers 4 x LC Connector Extends HDMI + RS232 full-duplex up to 1500 meters

Transcription:

On-line Multi-label Classification A Problem Transformation Approach Jesse Read Supervisors: Bernhard Pfahringer, Geoff Holmes Hamilton, New Zealand

Outline Multi label Classification Problem Transformation Binary Method Combination Method Pruned Sets Method () Results On line Applications Summary

Multi label Classification Single label Classification Set of instances, set of labels Assign one label to each instance e.g. Shares plunge on financial fears, Economy

Multi label Classification Single label Classification Set of instances, set of labels Assign one label to each instance e.g. Shares plunge on financial fears, Economy Multi label Classification Set of instances, set of labels Assign a subset of labels to each instance e.g. Germany agrees bank rescue, {Economy,Germany}

Applications Text Classification: News articles; Encyclopedia articles; Academic papers; Web directories; E mail; Newsgroups Images, Video, Music: Scene classification; Genre classification Other: Medical classification; Bioinformatics N.B. Not the same as tagging / keywords.

Multi label Issues Relationships between labels e.g. consider: {US, Iraq} vs {Iraq, Antarctica} Extra dimension Imbalances exaggerated Extra complexity Evaluation methods Evaluate by label? by example? How to do Multi label Classification?

Problem Transformation 1.Transform multi label data into single label data 2.Use one or more single label classifiers 3.Transform classifications back into multi label representation Can employ any single label classifier Naive Bayes, SVMs, Decision Trees, etc,... e.g. Binary Method, Combination Method,.. (overview by (Tsoumakas & Katakis, 2005))

Algorithm Transformation 1.Adapts a single label algorithm to make multilabel classifications 2.Runs directly on multi label data Specific to a particular type of classifier Does some form of Problem Transformation internally e.g. To AdaBoost (Schapire & Singer, 2000), Decision Trees (Blockheel et al. 2008), knn (Zhang & Zhou. 2005), NB (McCallum. 1999),...

Outline Multi label Classification Problem Transformation Binary Method Combination Method Pruned Sets Method () Results On line Applications Summary

Binary Method One binary classifier for each label A label is either relevant or!relevant

Binary Method One binary classifier for each label A label is either relevant or!relevant Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c}

Binary Method One binary classifier for each label A label is either relevant or!relevant Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} SL Train L' = {A,!A} d0,a d1,!a d2,a d3,!a SL Train L' = {B,!B} d0,!b d1,!b d2,!b d3,b SL Train L' = {C,!C} d0,!c d1,c d2,!c d3,c SL Train L' = {D,!D} d0,d d1,d d2,!d d3,!d

Binary Method One binary classifier for each label A label is either relevant or!relevant Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} SL Train L' = {A,!A} d0,a d1,!a d2,a d3,!a SL Train L' = {B,!B} d0,!b d1,!b d2,!b d3,b SL Train L' = {C,!C} d0,!c d1,c d2,!c d3,c SL Train L' = {D,!D} d0,d d1,d d2,!d d3,!d Single label Test: dx,? dx,? dx,? dx,?

Binary Method One binary classifier for each label A label is either relevant or!relevant Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} SL Train L' = {A,!A} d0,a d1,!a d2,a d3,!a SL Train L' = {B,!B} d0,!b d1,!b d2,!b d3,b SL Train L' = {C,!C} d0,!c d1,c d2,!c d3,c SL Train L' = {D,!D} d0,d d1,d d2,!d d3,!d Single label Test: dx,!a dx,!b dx,c dx,d

Binary Method One binary classifier for each label A label is either relevant or!relevant Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} SL Train L' = {A,!A} d0,a d1,!a d2,a d3,!a SL Train L' = {B,!B} d0,!b d1,!b d2,!b d3,b SL Train L' = {C,!C} d0,!c d1,c d2,!c d3,c SL Train L' = {D,!D} d0,d d1,d d2,!d d3,!d Single label Test: dx,!a dx,!b dx,c dx,d Multi label Test L = {A,B,C,D} dx,???

Binary Method One binary classifier for each label A label is either relevant or!relevant Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} SL Train L' = {A,!A} d0,a d1,!a d2,a d3,!a SL Train L' = {B,!B} d0,!b d1,!b d2,!b d3,b SL Train L' = {C,!C} d0,!c d1,c d2,!c d3,c SL Train L' = {D,!D} d0,d d1,d d2,!d d3,!d Single label Test: dx,!a dx,!b dx,c dx,d Multi label Test L = {A,B,C,D} dx,{c,d}

Binary Method One binary classifier for each label A label is either relevant or!relevant Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} SL Train L' = {A,!A} d0,a d1,!a d2,a d3,!a SL Train L' = {B,!B} d0,!b d1,!b d2,!b d3,b SL Train L' = {C,!C} d0,!c d1,c d2,!c d3,c SL Train L' = {D,!D} d0,d d1,d d2,!d d3,!d Single label Test: Assumes label independence dx,!a dx,!b dx,c dx,d Multi label Test L = {A,B,C,D} dx,{c,d} Often unbalanced by many negative examples

Combination Method One decision involves multiple labels Each subset becomes a single label

Combination Method One decision involves multiple labels Each subset becomes a single label Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c}

Combination Method One decision involves multiple labels Each subset becomes a single label Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} Single label Train L' = {A,AD,BC,CD} d0,ad d1,cd d2,a d3,bc

Combination Method One decision involves multiple labels Each subset becomes a single label Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} Single label Train L' = {A,AD,BC,CD} d0,ad d1,cd d2,a d3,bc Single label Test L' = {A,AD,BC,CD} dx,???

Combination Method One decision involves multiple labels Each subset becomes a single label Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} Single label Train L' = {A,AD,BC,CD} d0,ad d1,cd d2,a d3,bc Single label Test L' = {A,AD,BC,CD} dx,cd

Combination Method One decision involves multiple labels Each subset becomes a single label Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} Single label Train L' = {A,AD,BC,CD} d0,ad d1,cd d2,a d3,bc Single label Test L' = {A,AD,BC,CD} dx,cd Multi label Test L = {A,B,C,D} dx,{c,d}

Combination Method One decision involves multiple labels Each subset becomes a single label Multi label Train L = {A,B,C,D} d0,{a,d} d1,{c,d} d2,{a} d3,{b,c} Single label Train L' = {A,AD,BC,CD} d0,ad d1,cd d2,a d3,bc Single label Test L' = {A,AD,BC,CD} dx,cd Multi label Test L = {A,B,C,D} dx,{c,d} May generate too many single labels Can only predict combinations seen in the training set

A Pruned Sets Method () Binary Method Assumes label independence Combination Method Takes into account combinations Can't adapt to new combinations High complexity (~ distinct label sets) Pruned Sets Method Use pruning to focus on core combinations

A Pruned Sets Method () Concept: Prune away and break apart infrequent label sets Form new examples with more frequent label sets

A Pruned Sets Method () E.g. 12 examples, 6 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d06,{animation,comedy,family,musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d12,{adult,animation}

A Pruned Sets Method () 1.Count label sets E.g. 12 examples, 6 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d06,{animation,comedy,family,musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d12,{adult,animation} {Animation,Comedy} 3 {Animation,Family} 2 {Adult} 3 {Animation,Comedy,Family,Musical} 1 {Musical} 2 {Adult,Animation} 1

A Pruned Sets Method () 1.Count label sets 2.Prune infrequent sets (e.g. count < 2) E.g. 12 examples, 6 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d12,{adult,animation} d06,{animation,comedy,family,musical} {Animation,Comedy} 3 {Animation,Family} 2 {Adult} 3 {Animation,Comedy,Family,Musical} 1 {Musical} 2 {Adult,Animation} 1 Information loss!

A Pruned Sets Method () 1.Count label sets 2.Prune infrequent sets (e.g. count < 2) 3.Break up infrequent sets into frequent sets (e.g. count >= 2) {Animation,Comedy} 3 {Animation,Family} 2 {Adult} 3 {Animation,Comedy,Family,Musical} 1 {Musical} 2 {Adult,Animation} 1 E.g. 12 examples, 6 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d12,{adult,animation} d12,{adult} d06,{animation,comedy,family,musical} d06,{animation,comedy} d06,{animation,family} d06,{musical}

A Pruned Sets Method () 1.Count label sets 2.Prune infrequent sets (e.g. count < 2) 3.Break up infrequent sets into frequent sets (e.g. count >= 2) 4.Decide which subsets to reintroduce (!) Too many (esp. small) sets will: 'dillute' the dataset with single labels vastly increase the training set size i.e. frequent item sets not desireable {Animation,Comedy} 3 {Animation,Family} 2 {Adult} 3 {Animation,Comedy,Family,Musical} 1 {Musical} 2 {Adult,Animation} 1 E.g. 12 examples, 6 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d12,{adult,animation} d12,{adult} d06,{animation,comedy,family,musical} d06,{animation,comedy} d06,{animation,family} d06,{musical}

A Pruned Sets Method () 1.Count label sets 2.Prune infrequent sets (e.g. count < 2) 3.Break up infrequent sets into frequent sets (e.g. count >= 2) 4.Decide which subsets to reintroduce Strategies: A. Keep the top n subsets (ranked by number of labels and count) or B. Keep all subsets of size greater than n {Animation,Comedy} 3 {Animation,Family} 2 {Adult} 3 {Animation,Comedy,Family,Musical} 1 {Musical} 2 {Adult,Animation} 1 E.g. 12 examples, 6 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d12,{adult,animation} d12,{adult} d06,{animation,comedy,family,musical} d06,{animation,comedy} d06,{animation,family} d06,{musical}

A Pruned Sets Method () 1.Count label sets 2.Prune infrequent sets (e.g. count < 2) 3.Break up infrequent sets into frequent sets (e.g. count >= 2) 4.Decide which subsets to reintroduce 5.Add new instances {Animation,Comedy} 3 {Animation,Family} 2 {Adult} 3 {Animation,Comedy,Family,Musical} 1 {Musical} 2 {Adult,Animation} 1 E.g. 12 examples, 6 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d12,{adult,animation} d12,{adult} d06,{animation,comedy,family,musical} d06,{animation,comedy} d06,{animation,family} d06,{musical}

A Pruned Sets Method () 1.Count label sets 2.Prune infrequent sets (e.g. count < 2) 3.Break up infrequent sets into frequent sets (e.g. count >= 2) 4.Decide which subsets to reintroduce 5.Add new instances 6.Use Combination Method transformation E.g. 15 examples, 4 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d06,{animation,comedy} d06,{animation,family} d12,{adult} {Animation,Comedy} 4 {Animation,Family} 3 {Adult} 4 {Musical} 2

A Pruned Sets Method () 1.Count label sets 2.Prune infrequent sets (e.g. count < 2) 3.Break up infrequent sets into frequent sets (e.g. count >= 2) 4.Decide which subsets to reintroduce 5.Add new instances 6.Use Combination Method transformation Accounts for label relationships Reduced complexity Cannot form new combinations (e.g. {Animation,Family,Musical}) E.g. 15 examples, 4 combinations d01,{animation,family} d02,{musical} d03,{animation,comedy } d04,{animation,comedy} d05,{musical} d07,{adult} d08,{adult} d09,{animation,comedy} d10,{animation,family} d11,{adult} d06,{animation,comedy} d06,{animation,family} d12,{adult} {Animation,Comedy} 4 {Animation,Family} 3 {Adult} 4 {Musical} 2

Ensembles of Pruned Sets (E.) Creating new label set classifications 1. Train an Ensemble of e.g. Bagging (introduces variation!)

Ensembles of Pruned Sets (E.) Creating new label set classifications 1. Train an Ensemble of e.g. Bagging (introduces variation!) 2. Get preditions {Musical} {Animation,Family} {Animation, Comedy} {Animation, Family} {Musical} {Musical}

Ensembles of Pruned Sets (E.) Creating new label set classifications 1. Train an Ensemble of e.g. Bagging (introduces variation!) 2. Get preditions 3. Calculate a score {Musical} {Animation,Family} Musical: 3 (0.33) Animation: 3 (0.33) Family: 2 (0.22) Comedy: 1 (0.11) {Animation, Comedy} {Animation, Family} {Musical} {Musical}

Ensembles of Pruned Sets (E.) Creating new label set classifications 1. Train an Ensemble of e.g. Bagging (introduces variation!) 2. Get preditions 3. Calculate a score 4. Form a classification set {Musical} dx,{animation, Family, Musical} {Animation,Family} Musical: 3 (0.33) Animation: 3 (0.33) Family: 2 (0.22) Comedy: 1 (0.11) Threshold = 0.15 {Animation, Comedy} {Animation, Family} {Musical} {Musical}

Ensembles of Pruned Sets (E.) Creating new label set classifications 1. Train an Ensemble of e.g. Bagging (introduces variation!) 2. Get preditions 3. Calculate a score 4. Form a classification set {Musical} dx,{animation, Family, Musical} Can form new combinations Musical: 3 (0.33) Animation: 3 (0.33) Family: 2 (0.22) Comedy: 1 (0.11) Threshold = 0.15 {Animation,Family} {Animation, Comedy} {Animation, Family} {Musical} {Musical}

Results F1 Measure D.SET size / #lbls / avg.lbls BM [CM] E. RAK. Scene 2407 6 1.1 0.671 0.729 0.730 0.752 0.735 Medical 978 45 1.3 0.791 0.767 0.766 0.764 0.784 Yeast 2417 14 4.2 0.630 0.633 0.643 0.665 0.664 Enron 1702 53 3.4 0.504 0.502 0.520 0.543 0.543 Reuters 6000 103 1.5 0.421 0.482 0.496 0.499 0.418 Combination Method (CM) improves Binary Method (BM) Puned Sets method () improves Combination Method (CM) Except Medical: maybe label relationships not as important E. is best overall. RAKEL ~ E similar What about complexity? J. Read, B. Pfahringer, G. Homes. To Appear ICDM 08.

Complexity Build Time RAKEL may not be able to find ideal parameter value 'Worst case' scenarios are similar, but different in practice J. Read, B. Pfahringer, G. Homes. To Appear ICDM 08.

Complexity Memory Use Reuters Dataset transformation: ~2,500 instances E transformation: ~25,000 instances (for 10 iterations) RAKEL transformation: 3,090,000 instances (for 10 iterations) Number of instances generated during the Problem Transformation procedure for most complex parameter setting J. Read, B. Pfahringer, G. Homes. To Appear ICDM 08.

Outline Multi label Classification Problem Transformation Binary Method Combination Method Pruned Sets Method () Results On line Applications Summary

On line Multi label Classification Many multi label data sources are on line: New instances incoming Data can be time ordered Possibly large collections Concept drift An on line multi label algorithm should be: Adaptive Efficient

On line Multi label Classification

Multi label Concept Drift Measuring concept drift Observing indiv. labels? Complicated (may be 1000's of labels) May need domain knowledge Counting distinct label sets? Doesn't tell us much Transformation? Focus on core combinations

Multi label Concept Drift 20NG; News; Enron (On line data) Slow; medium; rapid concept drift YEAST Randomised SCENE Ordered Train/Test Split MEDICAL??? 1. transformation on first 50 instances 2.Measure the % coverage 3.Measure on the next 50, and etc..

Preliminary Results 'On line' Binary Method vs E. Model(s) built on 100 instances Thresholds updated every instance Model(s) rebuilt every 25 instances Enron Dataset Subsets Accuracy

Summary Multi label Classification Problem Transformation Binary Method (BM), Combination Method (CM) Pruned Sets () and Ensembles of (E.) Focus on core label relationships via pruning Outperforms standard and state of the art methods Multi label Classification in an On line Context Naive methods (eg. BM) can perform better than E in an on line context (future work!)

Questions?