Classifier ensembles

Download Report

Transcript Classifier ensembles

Stochastic Text Models for Music Categorization
Carlos Pérez-Sancho, José M. Iñesta, David Rizo
Pattern Recognition and Artificial Intelligence group
Department of Software and Computing Systems
University of Alicante, Spain
Outline
► Introduction
► Music
encoding
 Melody
 Harmony
► Experiments
 Plain classification
 Classifier ensembles
 Hierarchical classification
► Conclusions
SSPR 2008
2
Introduction
► Premise:
music content can be used to model
musical style
► We
use language modeling techniques to
classify symbolic digital scores
► For
that, digital scores need to be encoded into
sequences of symbols
SSPR 2008
3
Music encoding
► Two
different sources of information
Melody
Harmony
SSPR 2008
4
Melody encoding
►
►
►
Polyphonic sequences are reduced to monophonic using skyline
Pitch intervals and duration ratios are computed for each pair of
consecutive notes
Numeric values are encoded into ASCII symbols
Bf AZ aF bZ
BZ
dZ bZ DZ
input string
(2,×½) → Bf
(1,×1) → AZ
(-1,×2) → aF
(-2,×1) → bZ
(2,×1) → BZ
( interval , duration ratio )
(-4,×1) → dZ
(-2,×1) → bZ
(4,×1) → DZ
SSPR 2008
5
Harmony encoding
Key: E flat
►
►
VIm
V
I
input string
Chords are encoded as degrees relative to the tonality
for transposition invariance
Only chord changes are encoded
SSPR 2008
6
Experiments
► Dataset:
music from 3 genres and 9 sub-genres
(around 60 hours of music)
► Classification
techniques
 Naïve Bayes
 Language modeling (n-grams), classifying by lowest
perplexity
SSPR 2008
7
Experimental setup
► First
step: plain classification
 80% of the dataset used
 10-fold cross validation
► Second
step: hierarchical classification using
classifier ensembles
 Weights of the ensembles adjusted using previous

results
Remaining 20% of dataset used for validation
SSPR 2008
8
Classification results
Chords (harmony)
Melody
2-grams 3-grams 4-grams N.B. 2-grams 3-grams 4-grams N.B.
►
Best classification rates in the 3-classes problem were obtained
using harmonic information
►
When classifying sub-genres, melody usually performs better
►
Naïve Bayes performs better most of the times
►
No significant differences for different context sizes in n-grams.
SSPR 2008
9
Confussion matrix
×100 %
►
►
Misclassifications occur more frequently within broad domains
Try to prevent intra-domain errors by using a hierarchical classifier
SSPR 2008
10
Hierarchical classification
SSPR 2008
11
Hierarchical classification
►
Harmony (chord progressions) is used at the first level
►
Melody is used at the second level
►
Instead of using single classifiers, an ensemble of
classifiers is used at each level to increase robustness
SSPR 2008
12
Classifier ensembles
►
►
Decisions are made by weighted majority vote
Two weighting schemes
 Linear best-worst weighting vote
 Quadratic best-words weighting vote
weight
# errors
# errors
SSPR 2008
13
Hierarchical classification results
►
With the remaining 20% of the dataset
Single classifiers
2-grams
3-grams
4-grams
N.Bayes
3 classes
Using harmony
87.8
88.4
88.4
87.8
9 classes
Using melody
54.7
48.8
47.7
58.7
Classifier ensembles
Best single
3 classes
88.4
(melody 2-grams)
Linear
Quadratic
90.1
90.1
Hierarchical classification
9 classes
60.5
62.8
63.4
1st level (3 classes): harmony 2-grams
2nd level (3x3 classes):
academic – melody NB
jazz – melody NB
popular – melody 4-grams
SSPR 2008
14
Conclusions
►
Harmony and melody are suitable features for music
genre classification
►
Harmony is better for classifying broad musical domains,
while melody is better for distinguishing sub-genres
►
Misclassifications occur more frequently within broad
domains
►
Hierarchical classification and classifier ensembles
outperformed the best single classifiers
SSPR 2008
15
Stochastic Text Models for Music Categorization
Carlos Pérez-Sancho, José M. Iñesta, David Rizo
Pattern Recognition and Artificial Intelligence group
Department of Software and Computing Systems
University of Alicante, Spain