SlideShare une entreprise Scribd logo
1  sur  79
Télécharger pour lire hors ligne
From Music Information Retrieval
to Music Emotion Recognition
Center for Informatics and Systems of the University of Coimbra (CISUC)
Portugal
Rui Pedro Paiva
March 2012
2 © Rui Pedro Paiva, “From MIR to MER”, 2012
Contents
 Music Information Retrieval
 Music Emotion Recognition
 References
3 © Rui Pedro Paiva, “From MIR to MER”, 2012
MUSIC INFORMATION RETRIEVAL
- MIR: What and Why?
- Applications
- Short MIR Tale
- Representations
- Techniques
4 © Rui Pedro Paiva, “From MIR to MER”, 2012
MIR: What and Why?
 MIR
• Music Information Retrieval: interdisciplinary research field
devoted to the study of information extraction mechanisms from
musical pieces, retrieval methodologies, as well as all the processes
involved in those tasks in different music representation media.
 Why MIR?
• MIR emerges from the necessity to manage huge collections of
digital music for “preservation, access, research and other uses”
[Futrelle and Downie, 2003]
5 © Rui Pedro Paiva, “From MIR to MER”, 2012
MIR: What and Why?
 Music and Man
• Music expresses “that which cannot be put into words and that
which cannot remain silent” (Victor Hugo)
• “We associate music with the most unique moments of our
lives and music is part of our individual and social imaginary”
[Paiva, 2006]
- “By listening to music, emotions and memories, thoughts and reactions,
are awakened” [Paiva, 2006]
• “Life has a soundtrack” [Gomes, 2005] (“Festivais de Verão”,
Jornal “Público”)
• “The history of a people is found in its songs” (George Jellinek)
6 © Rui Pedro Paiva, “From MIR to MER”, 2012
MIR: What and Why?
 Music and World economy
• Music industry runs, only in the USA an amount of money in
the order of several billion US dollars per year.
• Explosion of the Electronic Music Industry (EMD)
- Widespread access to the Internet
- Bandwidth increasing in domestic and mobile accesses
- Compact audio formats with near CD quality (mp3, wma)
- Portable music devices (iPod, mp3 readers)
- Peer-to-peer networks (Napster, Kazaa, eMule)
- Online music stores (iTunes, Calabash Music, Sapo Music) 
resolution is the song, not the CD
- Music identification platforms (Shazam, 411-Song, Gracenote MusicID /
TrackID)
- Music recommendation systems (MusicSurfer)
7 © Rui Pedro Paiva, “From MIR to MER”, 2012
MIR: What and Why?
 Music and World economy (cont.)
• By 2005, Apple iTunes was selling  1.25 million songs each
day [TechWhack, 2005]
- Until January 2009, over 6 billion songs had been sold in total
[TechCrunch, 2009]
• By 2007, music shows in Portugal sold 30 M€ in tickets [RTP,
2009]
• Number and dimension of digital music archives continuously
growing
- Database size (these days, over 2 million songs)
- Genres covered
• Challenges to music providers and music librarians
- Organization, maintenance, labeling, user interaction
- Any large music database is only really useful if users can find what
they are looking for in an efficient manner!
8 © Rui Pedro Paiva, “From MIR to MER”, 2012
MIR: What and Why?
 Database Organization and Music Retrieval
• Presently, databases are manually annotated  search and
retrieval is mostly textual (artist, title, album, genre)
- Service providers
– Difficulties regarding manual song labeling: subjective and time-consuming,
- Customers
– Difficulties in performing “content-based” queries
• “Music’s preeminent functions are social and psychological”, and so “the most useful
retrieval indexes are those that facilitate searching in conformity with such social and
psychological functions. Typically, such indexes will focus on stylistic, mood, and
similarity information” [Huron, 2000].
•  Music Information Retrieval (MIR) emerges from the
necessity to manage huge collections of digital music for
“preservation, access, research and other uses” [Futrelle and
Downie, 2003].
9 © Rui Pedro Paiva, “From MIR to MER”, 2012
MIR: What and Why?
 Database Organization and Music Retrieval
Client
Internet Ordering of
Results and
Summarization
Comparison
of Signatures
Database Search
Database
Classification
Raw Audio
Recordings
Segmentation
and
Summarization
Feature
Extraction
Signature
Feature Extraction
Query Signature
Visualization
1. Tom Jobim - Corcovado
2. Dido - Thank You
3. João Gilberto - Meia Luz
4. .....
5. .....
Jazz Classical Latin
Swing Baroque Mambo Salsa
.
Interactive Edition
Query Creation
© Rui Pedro Paiva, 2002
10 © Rui Pedro Paiva, “From MIR to MER”, 2012
Applications
 Platforms for EMD
• Similarity-based retrieval tools
- Query-by-example [Welsh et al.,
1999]
– Music identification (Shazam,
trackID, Tunatic) [Wang, 2003;
Haitsma and Kalker, 2002;
TMN, 2009]
– Query-by-mood [Cardoso et al.,
2011]
– Music recommendation [Celma &
Lamere, 2007]
– Islands of music [Pampalk, 2001]
• Metaphor of geographic maps:
similar genres close together
- Automatic playlist generation
[Pauws and Wijdeven, 2005;
Alghoniemy and Tewfik, 2000] © Elias Pampalk, 2001
11 © Rui Pedro Paiva, “From MIR to MER”, 2012
 Platforms for EMD
• Similarity-based retrieval tools
- Query-by-melody
(query-by-humming, QBS) [Parker, 2005; Ghias et al., 1995]
- Plagiarism detection [Paiva et al., 2006]
- Music web crawlers [Huron, 2000]
 Music education and training
• Automatic music transcription [Ryynänen, 2008; Kashino et al.,
1995]
- Music composition, analysis, performance evaluation, plagiarism
detection
 Digital music libraries
• For research issues involving music retrieval, training (learning
activities, evaluation, etc.)  Variations [Dunn, 2000]
Applications
12 © Rui Pedro Paiva, “From MIR to MER”, 2012
Applications
 Audio software
• Intelligent audio (music)
editors  automatic indexing
[Tzanetakis, 2002]
 Multimedia databases and
operating systems
[Burad, 2006]
 Video indexing and searching
• Segmentation based on audio (music) content  detection of
scene transitions [Pfeiffer, 1996]
© George Tzanetakis, 2002
13 © Rui Pedro Paiva, “From MIR to MER”, 2012
Applications
 Advertisement and cinema
• Tools for mood-based retrieval [Cardoso et al. 2011]
 Sports
• Music to induce a certain cardiac frequency [Matesic and
Cromartie, 2002]
 … and so forth…
14 © Rui Pedro Paiva, “From MIR to MER”, 2012
Short MIR Tale
 Precursors of computer-based MIR: incipit and theme indexes, e.g.,
Harold Barlow and Sam Morganstern’s dictionary of musical
themes [Barlow and Morganstern, 1948]
15 © Rui Pedro Paiva, “From MIR to MER”, 2012
Short MIR Tale
 1966: potential of applying automatic information retrieval
techniques to music was recognized [Kassler, 1966]
 1970s and 1980s: automatic music transcription systems
 1990s: surge of interest, mostly in topics such as query-by-
humming (impulse from research on digital libraries)
 2000: 1st International Symposium on Music Information Retrieval
(ISMIR)
• Now “International Society for Music Information Retrieval Conference”
16 © Rui Pedro Paiva, “From MIR to MER”, 2012
Short MIR Tale
 2000-present: active multidisciplinary research field
© Joe Futrelle and J. Stephen Downie, 2003
17 © Rui Pedro Paiva, “From MIR to MER”, 2012
Short MIR Tale
 2000-present: active multidisciplinary research field
• Content analysis and similarity assessment and retrieval in
audio song databases
- Metrics of similarity, music identification, music recommendation, audio
fingerprinting, music classification and feature extraction, tempo and
melody detection, music summarization
• Databases systems, indexing, query languages
• Knowledge representation, metadata
• Theme extraction, harmonic and motive analysis, tonality
• Emotion
• Perception and cognition
• Social, legal, ethical and business issues
• User interface design
18 © Rui Pedro Paiva, “From MIR to MER”, 2012
Representations
© Donald Byrd and Tim Crawford, 2002
19 © Rui Pedro Paiva, “From MIR to MER”, 2012
Representations
 Audio (music content analysis)
• Object
- Audio recordings, streaming audio
• Topics
- Automatic transcription, QBE, classification, recommendation,
identification, …
 Symbolic
• Object
- Scores or event-based representations (e.g., MIDI)
• Topics
- Melodic matching, theme extraction, harmonic analysis, …
20 © Rui Pedro Paiva, “From MIR to MER”, 2012
Representations
 Visual
• Object
- Printed music
• Topics
- Optical music recognition
 Metadata
• Object
- Any kind
• Topics
- Digital libraries, music ontologies (semantic web)
21 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Idea
• Extract semantic information from low-level data
• Feature extraction
- Physical: F0, intensity, centroid, uniformity, rolloff, flux
- Perceptual: pitch, loudness, timbre, beat
- Musicological: notes, melodies, measures, motives, themes
- Higher-level (semantic) features: emotion, genre, instruments, artist
• Content and Representation
- Audio: which instruments, notes, artist
- Symbolic: themes, motives
- Visual: notation, basic units
- Metadata: genre, artist, MPEG-7 descriptors
22 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Physical features
• Time domain
- Waveform analysis: energy contour, amplitude-based segmentation,
auto-correlation, peak detection
0 0.005 0.01 0.015 0.02 0.025 0.03 0.035 0.04 0.045 0.05
-1
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
0.8
1
time (s)
Intensity
0 0.005 0.01 0.015 0.02 0.025 0.03 0.035 0.04 0.045 0.05
-800
-600
-400
-200
0
200
400
600
800
Time (s)
R(k)
      000000 2;200;1,5sin
5
3sin
3
sin)( fHzfAt
A
t
A
tAtx  
23 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Physical features (cont.)
• Time domain
- Auditory model-based F0 detectors, beat detectors (energy-based)
0.21 0.22 0.23 0.24 0.25
-0.4
-0.2
0
0.2
0.4
Time (s)
x(t)
a) Sound waveform
Time (s)
Frequencychannel
b) Cochleagram frame
0.21 0.22 0.23 0.24 0.25
20
40
60
80
Time Lag (ms)
Frequncychannel
c) Correlogram frame
0 5 10 15 20
20
40
60
80
0 5 10 15 20 25
0
0.1
0.2
0.3
0.4
Time Lag (ms)
Pitchsalience
d) Summary correlogram
24 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Physical features
• Frequency domain
- DFT, STFT, spectrogram
- F0 detectors
Time
Frequency
0 0.5 1 1.5 2
0
0.5
1
1.5
2
x 10
4
25 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Physical features (cont.)
• Spectral features
- MFCCs, centroid, rolloff, flux, harmonicity, high-frequency content, …
• Sub-band features
© MIR Toolbox, 2008
26 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Physical features (cont.)
• Spectral features + sub-band features (e.g., audio
fingerprinting)
© Haitsma & Kalker, 2002
27 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Perceptual features
• Pitch
- Frequency
- Intensity
- Context
- Ear physiology (age)
© MIR Toolbox, 2008
© http://www.ai.rug.nl/~tjeerd/CPSP/docs/cochleaModel.html
28 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Perceptual features (cont.)
• Loudness
- Intensity
- Frequency
- Context
- Ear physiology (age)
Fletcher-Munson equal loudness contours
29 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Perceptual features (cont.)
• Timbre
- No physical correlate
- “what something sounds like”:
– Spectral content at steady-sate
• Centroid, rolloff, relative amplitudes of harmonic components, inharmonicty…
– Signal’s temporal envelope
• Attack transient
– Temporal behavior of the harmonics
• Melodic contour
- UDUEEUUD
• Rhythm contour
- FSSFEEFS
• Beat
30 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Musicological features
• Notes from audio
• Notes from optical music recognition
0 100 200 300 400 500
4800
4900
5000
5100
5200
5300
5400
Time (msec)
f[k](cents)
a) Eliades Ochoa's excerpt
0 200 400 600 800
6100
6200
6300
6400
6500
6600
6700
6800
Time (msec)f[k](cents)
b) Female opera excerpt
© Rui Pedro Paiva, 2006
31 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Musicological features (cont.)
• Melody
© Rui Pedro Paiva, 2006
32 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Musicological features (cont.)
• Themes
© Colin Meek and William Birmingham, 2001
33 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Higher-level features  top-down information flow
© Xavier Serra, 2005
34 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Higher-level features (cont.)
• Bridge the semantic gap
• Memory, context, expectations
- Repetitions, sonic environment, modeling the individual, musicological
knowledge
• Emotion: valence (happy/anxious) and arousal (calm/energetic)
- Classification approaches resorting to low-level features
35 © Rui Pedro Paiva, “From MIR to MER”, 2012
MUSIC EMOTION RECOGNITION
(MER)
- MER: What and Why?
- Applications
- Short MER Tale
- Emotion Models
- Techniques
- MER@CISUC
- Current Limitations and Open Problems
36 © Rui Pedro Paiva, “From MIR to MER”, 2012
MER: What and Why?
 MER
• Music Emotion Recognition: branch of MIR devoted to the
identification of emotions/moods in musical pieces
• Emotion vs mood
- MIR researchers  use both terms interchangeably
- Psychologists  clear distinction [Sloboda and Juslin, 2001]
– Emotion = a short experience in response to an object (e.g., music)
– Mood = longer experience without specific object connection
• Categories of emotions [Gabrielsson, 2002)]
- Expressed emotion: emotion the performer tries to communicate to the
listeners
- Perceived emotion: emotion the listener perceives as being expressed
in a song (which may be different than the emotion the performer tried
to communicate)  scope of MIR researchers
- Felt (evoked) emotion: emotion felt by the listener, in response to the
song and performance
37 © Rui Pedro Paiva, “From MIR to MER”, 2012
MER: What and Why?
 Why MER?
• “Music’s preeminent functions are social and psychological”,
and so “the most useful retrieval indexes are those that
facilitate searching in conformity with such social and
psychological functions. Typically, such indexes will focus on
stylistic, mood, and similarity information” [Huron, 2000].
- Studies on music information behaviour  music mood is an important
criterion for music retrieval and organization [Juslin and Laukka, 2004]
• “In the academia, more and more multimedia systems that involve
emotion analysis of music signals have been developed, such as
Moodtrack, LyQ, MusicSense, Mood Cloud, Moody and i.MTV, just
to name a few. In the industry, many music companies, such as
AMG, Gracenote, MoodLogic, Musicovery, Syntonetic, and
Sourcetoneuse emotion as a cue for music retrieval.” [Yang and
Chen, 2012]
38 © Rui Pedro Paiva, “From MIR to MER”, 2012
MER: What and Why?
 Difficulties [Yang and Chen, 2012]
• Emotion perception is by nature subjective [Yang and Chen,
2012]
- People can perceive different emotions for the same song
•  Performance evaluation of an MER systems is difficult [Yang
and Chen, 2012]
- Common agreement on the recognition result is hard to obtain
• Still not fully understood how music and emotion are related
- Despite several studies on music psychology
39 © Rui Pedro Paiva, “From MIR to MER”, 2012
Applications
 Platforms for EMD
• Mood-based retrieval tools
- Music recommendation, automatic playlist generation, classification, …
 Game development
 Cinema
 Advertising
 Health
• Sports
• Stress management
40 © Rui Pedro Paiva, “From MIR to MER”, 2012
Short MER Tale
 19th century: initial research on the relations between
music and emotion [Gabrielsson and Lindström, 2001]
 20th century: more active study of music and emotion
• Relationship between emotions and musical attributes
- Mode, harmony, tempo, rhythm, dynamics [Meyers, 2007]
• Mood models
- E.g., Kate Hevner [Hevner, 1935], James Russell [Russell, 1980]
 1980-1990s: Initial computational models  mostly
emotion synthesis (e.g., [Juslin, 1997])
• Relationships between emotion and music composition and
music expressivity
41 © Rui Pedro Paiva, “From MIR to MER”, 2012
Short MER Tale
 1980-1990s: only a few works on emotion analysis
• Mostly in the symbolic/MIDI domain [Lu et al., 2006]
 2003: first work on mood detection in audio musical
signals (by Yazhong Feng [Feng et al., 2003])
• 4 mood categories (happiness, sadness, anger and fear)
• 2 musical attributes: tempo and articulation
• Neural network classifier trained using 200 musical pieces
• Test corpus of 23 pieces, with average precision and recall of
67 and 66%
 2004: first multi-modal approach combining audio and
text (lyrics) [Yang and Lee, 2004]
42 © Rui Pedro Paiva, “From MIR to MER”, 2012
Short MER Tale
 2006: first study on MEVD (Music Emotion Variation
Detection [Lu et al., 2006]
• Temporal variable: emotions may change throughout a song
 2007: first mood analysis model based on a
dimensional approach [Yang et al., 2007]
• Thayer 2D mood model  continuous emotion as function of
arousal and valence
43 © Rui Pedro Paiva, “From MIR to MER”, 2012
Short MER Tale
 2007: first MIREX (MIR Evaluation eXchange) track
devoted to mood classification [MIREX, 2012]
• Before:
- Each approach used a different (and limited) set of features, mood
taxonomies, number of classes and test sets.
- Some approaches constrained the analysis to a particular musical style
(e.g., classical music)
 Present: active research in the field
44 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Two main conceptualizations of emotion
• Categorical models
- Emotions as categories: limited number of discrete emotions
(adjectives)
• Dimensional models
- Emotions as continuous values, depending on 2 or 3 axes
45 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Categorical Models
• Main ideas
- “people experience emotions as categories that are distinct from each
other” [Yang and Chen, 2012]
- Existence of basic emotions
– Limited number of universal and primary emotion classes (e.g.,
happiness, sadness, anger, fear, disgust, surprise) from which all other
“secondary” emotion classes can be derived
46 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Categorical Models
• Examples
- Hevners’s 8 clusters of affective terms (1935)
- Regrouped into 10 adjective groups by Farnsworth [Farnsworth, 1954]
and into 9 adjective groups by Schubert [Schubert 2003].
© Yang and Chen, 2012
47 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Categorical Models
• Examples
- Tellegen-Watson-Clark
model (1999)
© Tellegen, Watson and
Clark, 1999
48 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Categorical Models
• Limitations
- Limited number of adjectives
- Larger number may ne impractical for psychological studies
- Adjectives may be ambiguous
49 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Dimensional Models
• Main ideas
- Emotions experience as a continuous
– Each emotion is a locations in a multi-dimensional plane, based on a
reduced number of axes (2D or 3D)
– Argument: correspond to internal human representations of emotions
- 3 main dimensions of emotion [Yang and chen, 2012]
– valence (or pleasantness; positive and negative affective states),
– arousal (or activation; energy and stimulation level)
– potency (or dominance; a sense of control or freedom to act)
- 2D used in practice
– Valence and arousal regarded as the “core processes” of affect [Yang and
Chen, 2002]
– Simpler to visualize emotions
50 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Dimensional Models
• Examples
- James Russell’s
circumplex model
[Russell, 1980]
© Kim et al., 2010
51 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Dimensional Models
• Examples
- Robert Thayer’s model [Thayer, 1989]
© Meyers, 2007
52 © Rui Pedro Paiva, “From MIR to MER”, 2012
Emotion Models
 Dimensional Models
• Limitations
- Obscures important aspects of emotion
– Anger and fear are placed close in the valence-arousal plane
– Very different in terms of their implications
–  potency (dominant–submissive) as the third dimension
53 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Idea
• Extract musical features correlated to emotion from the
information source under analysis (audio, MIDI, lyrics, …)
• Temporal variable
- Emotions may change throughout a song  MEVD
54 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Relevant musical attributes
•  Studies by music psychologists, e.g., [Gabrielsson and
Lindström, 2001]
- Dynamics, articulation, timbre, pitch, interval, melody, harmony, tonality
and rhythm [Friberg, 2008], mode, harmony, tempo, rhythm, dynamics,
musical form [Meyers, 2007]
– Modes: major modes related to happiness or solemnity, minor modes
associated with sadness or anger [Meyers, 2007].
– Harmonies: simple, consonant, harmonies are usually happy, pleasant or
relaxed; complex, dissonant, harmonies relate to emotions such as
excitement, tension or sadness, as they create instability in a musical piece
[Meyers, 2007].
- Few relevant audio features proposed so far [Friberg, 2008]
– Difficult to extract from audio signals  but easier to extract from
symbolic representations (some features are score-based in nature)
55 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Relevant musical attributes
• Timing: Tempo, tempo variation, duration contrast
• Dynamics: overall level, crescendo/decrescendo, accents
• Articulation: overall (staccato/legato), variability
• Timbre: Spectral richness, onset velocity, harmonic richness
• Pitch (high/low)
• Interval (small/large)
• Melody: range (small/large), direction (up/down)
56 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Relevant musical attributes
• Harmony (consonant/complex-dissonant)
• Tonality (chromatic-atonal/key-oriented)
• Rhythm (regular-smooth/firm/flowing-fluent/irregular-rough)
• Mode (major/minor)
• Loudness (high/low)
• Musical form (complexity, repetition, new ideas, disruption)
57 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Overall approach: classical pattern recognition
approach
• 1) Selection of a mood model
- Categorical or dimensional?
- How many categories / variables? Multi-label or probabilistic
classification?
• 2) Acquisition of training and testing data
- Necessary to perform manual data annotation? How many annotators,
what profiles, how many songs, song balance, etc.?
• 3) Feature extraction and selection
- Which features? How to extract them from the information source?
- How to evaluate the impact of different features in each class?
58 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Overall approach: classical pattern recognition
approach
• 4) Selection of a representation model
- Black-box, gray-box, white-box?
– Black-box paradigm dominates. Gray-box should be researched.
- SVM, neural networks, GMM, KNN, neural-fuzzy classifiers, rule-base
systems?
– SVMs: most accurate results currently
- NN: how many hidden layers, etc? SVM: which parameters?...
– SVM parameter optimization: grid-search
• 5) Model training
- How to organize the training and testing experience? Cross validation?
How many folds?
– Tipically, 10-fold cross validation
59 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Overall approach: classical pattern recognition
approach
• 6) Model validation
- How to quantitatively evaluate the results? RMSE, SSE, correlation, R2
statistics? Precision, recall and F-measure? Confusion matrix?
– Regression: RMSE and R2 are tipically used
– Classification: Precision, recall and F-measure are standard
- What kinds of errors are more prevalent? Why? Where is the root of the
problem: classifier/regressor, extracted features, method employed for
extraction, feature selection approach?
– Currently, the bottleneck seems to be the lack of semantic richness of
current low-level features and the innacuracy of particular feature
extraction algorithms (e.g., tempo, etc.)
- What classes/variables show low accuracy? Why?
– Valence: important features might be missing
- Same questions for individual songs.
60 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Acquisition of training and testing data [Kim et al., 2010]
• Subjective tests
• Annotation games
© Kim et al., 2010
61 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 Feature Extraction
• Music platforms
- Audio: Marsyas, MIR Toolbox, PsySound, …
- MIDI: jMIR, …
- Lyrics: ConceptNet, …
• Audio: acoustic features developed in other contexts
- Music genre classification (e.g., spectral shape features, Mel-frequency
cepstral coefficients (MFCC), etc.) and speech recognition
– Low-level audio descriptors (LLDs): spectral shape features such as
centroid, spread, bandwidth, skewness, kurtosis, slope, decrease, rolloff,
flux contrast, MFCCs, etc.
• Necessary to
- Evaluate the relevance of LLDs to mood detection
- Program actual mood-related features
62 © Rui Pedro Paiva, “From MIR to MER”, 2012
Techniques
 MEVD
• Segment-based classification [Panda and Paiva, 2011]
- Divide audio signal into small segments
- Classify each of them as before
• Ad-hoc strategies (e.g., [Lu et al., 2006])
- Segmentation based on feature variations
– Thresholds used and difficult to tune
63 © Rui Pedro Paiva, “From MIR to MER”, 2012
MER@CISUC
 MOODetector project
• Search
64 © Rui Pedro Paiva, “From MIR to MER”, 2012
MER@CISUC
 MOODetector project
• Generate playlist
65 © Rui Pedro Paiva, “From MIR to MER”, 2012
MER@CISUC
 MOODetector project
• Mood Tracking
66 © Rui Pedro Paiva, “From MIR to MER”, 2012
Limitations and Open Problems
 Current Limitations
• Agreement on a usable mood taxonomy
- MIREX mood: only 5 moods, with semantic and acoustic overlap [Yang
and Chen, 2012]
• Lack of sizeable real-world datasets
- Dimensional approaches
– Yang only uses 194 songs
- Categorical approaches
– MIREX mood validates using 600 songs [MIREX, 2012]
• Accuracy of current systems is too low for most real-world
applications
- MIREX best algorithm ~ 65% accuracy in a 5-class mood problem
[MIREXresults, 2010]
67 © Rui Pedro Paiva, “From MIR to MER”, 2012
Limitations and Open Problems
 Open problems
• Semantic gap
- Novel, semantically-relevant features necessary, able to capture the
relevant musical attributes
– Most important limitation, according to [Friberg, 2008]
- Multi-modal approaches: combination of different information sources
(audio, midi, lyrics)
– Use MIDI resulting from automatic music transcription
• Data sets
- Real-world sizeable datasets for categorical, dimensional and MEVD
approaches
• MEVD
- Other techniques, e.g., self-similarity techniques [Foote, 1999]
- Quality datasets
68 © Rui Pedro Paiva, “From MIR to MER”, 2012
Limitations and Open Problems
 Open problems
• Multi-label classification (e.g., [Sanden and Zhang, 2011])
- Same song belonging to more than one mood category
• Extraction of knowledge from computational models
- E.g., neural-fuzzy approaches [Paiva and Dourado, 2004]
69 © Rui Pedro Paiva, “From MIR to MER”, 2012
REFERENCES
- Suggested Initial Literature
- Cited References
70 © Rui Pedro Paiva, “From MIR to MER”, 2012
Suggested Initial Literature
 Yang Y.-H. and Chen H.-H. (2011). "Music Emotion Recognition," CRC
Press, Boca Raton, Florida, USA, February 2011, 261 pages.
 Friberg A. (2008). “Digital Audio Emotions - an Overview of Computer Analysis
and Synthesis of Emotional Expression in Music”. In Proc. International
Conference on Digital Audio Effects - DAFx-08.
 Kim Y. E., Schmidt E. M., Migneco R., Morton B. G., Richardson P., Scott J.,
Speck J. A. and Turnbull D. (2010). “Music emotion recognition: a state of the art
review”. International Society for Music Information Retrieval Conference –
ISMIR’2010.
 Lu L., Liu D. and Zhang H.-J. (2006). “Automatic Mood Detection and Tracking
of Music Audio Signals”. IEEE Transactions on Audio, Speech and Language
Processing, Vol. 14, No. 1, pp. 5-18.
 Yang Y.-H., Lin Y.-C., Su Y.-F. and Chen H. H. (2008). “A Regression Approach
to Music Emotion Recognition”. IEEE Transactions on Audio, Speech and
Language Processing, Vol. 16, No. 2, pp. 448-457.
 Yang Y.-H. and Chen H.-H. (2012). "Machine recognition of music emotion: a
review“, To Appear in ACM Trans. Intelligent Systems and Technology.
71 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 Alghoniemy M. and Tewfik A. (2000). “Personalized Music Distribution”, In Proc.
IEEE International Conference on Acoustics, Speech and Signal Processing –
ICASSP’2000.
 Barlow H. and Morganstern S. (1948). A Dictionary of Musical Themes, New
York, Crown.
 Burad A. (2006). “Multimedia Databases”, Seminar Report, Computer Science
and Engineering, Indian Institute of Technology.
 Byrd D. (2002). “The History of ISMIR - A Short Happy Tale”, D-Lib Magazine,
Vol. 8, No. 1, URL: http://www.dlib.org/dlib/november02/11inbrief.html#BYRD.
 Byrd D. and Crawford T. (2002). “Problems of Music Information Retrieval in the
Real World”, Information Systems and Management, Vol. 38, pp. 249-272
 Cardoso L., Panda R. and Paiva R. P. (2011). “MOODetector: A Prototype
Software Tool for Mood-based Playlist Generation”. Simpósio de Informática –
INForum 2011.
 Celma O. and Lamere P. (2007). “Music Recommendation Tutorial”,
International Society for Music Information Retrieval Conference – ISMIR’2007,
tutorial.
 Dunn J. W. (2000). “Beyond VARIATIONS: Creating a Digital Music Library”, In
International Society for Music Information Retrieval Conference – ISMIR’2000.
72 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 Farnsworth P. R. (1954). “A study of the Hevner adjective list”, J. Aesthetics and
Art Criticism 13, 97–103.
 Feng Y., Zhuang Y. and Pan Y. (2003). “Music information retrieval by detecting
mood via computational media aesthetics”. In Proc. International Conference on
Web Intelligence - IEEE/WIC.
 Friberg A. (2008). “Digital Audio Emotions - an Overview of Computer Analysis
and Synthesis of Emotional Expression in Music”. In Proc. International
Conference on Digital Audio Effects - DAFx-08.
 Foote J. (1999). “Visualizing music and audio using self-similarity”. In Proc. ACM
Multimedia, pp. 77–84, Orlando, Florida, USA.
 Futrelle J. and Downie J. S. (2003). “Interdisciplinary Research Issues in Music
Information Retrieval: ISMIR 2000-2002”, Journal of New Music Research, Vol.
32, No. 2, pp. 121-131.
 Gabrielsson A. (2002). Emotion perceived and emotion felt: Same or different?
Musicae Scientiae,123–147. special issue.
 Gabrielsson A. and Lindström E. (2001). “The influence of musical structure on
emotional expression”. In P. N. Juslin, & J. A. Sloboda (Eds.), Music and
emotion: Theory and Research New York: Oxford University Press, 2001, pp.
223-248.
73 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 Ghias A., Logan J., Chamberlin D. and Smith B. C. (1995). “Query by Humming:
Musical Information Retrieval in an Audio Database”, In Proc. ACM Multimedia
Conference – Multimedia’95.
 Gomes A. R. (2005). “Festivais de Verão: Para Além da Música, Uma
Experiência”, Revista XIS – Jornal Público, August 6, 2005, pp. 18-19 (in
Portuguese).
 Haitsma J. and Kalker T. (2002). “A highly robust audio-fingerprinting system”,
International Society for Music Information Retrieval Conference – ISMIR’2002.
 Hevner K. (1935). “Expression in music: A discussion of experimental studies
and theories”, Psychological Review 48, 2, 186–204.
 Huron D. (2000). “Perceptual and Cognitive Applications in Music Information
Retrieval”, International Society for Music Information Retrieval Conference –
ISMIR’2000.
 Juslin P. N. (1997). “Perceived emotional expression in synthesized
performances of a short melody: capturing the listener’s judgment policy,” Music.
Sci., Vol. 1, No. 2, pp. 225–256.
 Juslin P. N. and Laukka J. (2004). “Expression, Perception, and Induction of
Musical Emotions: A Review and a Questionnaire Study of Everyday Listening”.
Journal of New Music Research, Vol. 33, No. 3, pp. 217-38, 2004.
74 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 Kashino K., Nakadai K., Kinoshita T. and Tanaka H. (1995). “Organization of
Hierarchical Perceptual Sounds: Music Scene Analysis with Autonomous
Processing Modules and a Quantitative Information Integration Mechanism”, In
Proc. International Joint Conference on Artificial Intelligence – IJCAI’95, pp.
158-164.
 Kassler M. (1966). “Toward Musical Information Retrieval”, Perspectives of New
Music, Vol. 4, No. 2, pp. 59-67.
 Kim Y. E., Schmidt E. M., Migneco R., Morton B. G., Richardson P., Scott J.,
Speck J. A. and Turnbull D. (2010). “Music emotion recognition: a state of the art
review”. International Society for Music Information Retrieval Conference –
ISMIR’2010.
 Lu L., Liu D. and Zhang H.-J. (2006). “Automatic Mood Detection and Tracking
of Music Audio Signals”. IEEE Transactions on Audio, Speech and Language
Processing, Vol. 14, No. 1, pp. 5-18.
 Matesic B. C. and Cromartie F. (2002). “Effects music has on lap pace, heart
rate, and perceived exertion rate during a 20-minute self-paced run”, The Sports
Journal, Vol. 5. No. 1.
 Meyers O. C. (2007). “A Mood-Based Classification and Exploration System”.
MSc Thesis, School of Architecture and Planning, Massachusetts Institute of
Technology.
75 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 MIREX (2012). “MIREX Home”, URL: http://www.music-
ir.org/mirex/wiki/MIREX_HOME, available by March 1, 2012.
 MIREXresults (2010). “MIREX 2010 Results”. URL: http://www.music-
ir.org/mirex/wiki/2010:MIREX2010_Results, available by March 1, 2012.
 Orio N. (2006). “Music Retrieval: A Tutorial and Review”, Foundations and
Trends in Information Retrieval, Vol. 1, No. 1.
 Paiva R. P. (2002). “Content-based Classification and Retrieval of Music:
Overview and Research Trends”. Technical Report, Centre of Informatics and
Systems of the University of Coimbra, Portugal.
 Paiva R. P. and Dourado A. (2004). “Interpretability and Learning in
Neuro-Fuzzy Systems”. Fuzzy Sets and Systems, Vol. 147 (1), pp. 17-38,
Elsevier.
 Paiva R. P. (2006). “Melody Detection in Polyphonic Audio”, PhD Thesis,
University of Coimbra, Department of Informatics Engineering.
 Pampalk E. (2001). Islands of Music: Analysis, Organization and Visualization of
Music Archives, MSc Thesis, Department of Software Technology and
Interactive Systems, Vienna University of Technology, Austria.
 Panda R. and Paiva R. P. (2011). “Using Support Vector Machines for
Automatic Mood Tracking in Audio Music”. Proceedings of the 130th Audio
Engineering Society Convention – AES 130, London, UK.
76 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 Parker C. (2005). “Applications of Binary Classification and Adaptive Boosting to
the Query-by-Humming Problem”, International Society for Music Information
Retrieval Conference – ISMIR’2005.
 Pauws S. and Wijdeven S. (2005). “User Evaluation of a New Interactive Playlist
Generation Concept”, International Society for Music Information Retrieval
Conference – ISMIR’2005.
 Pfeiffer S., Fischer S. and Effelsberg W. (1996). “Automatic Audio Content
Analysis”, In Proc. ACM Multimedia Conference – Multimedia’96, pp. 21-30.
 Ryynänen M. (2008). “Automatic Transcription of Pitch Content in Music and
Selected Applications”. PhD Thesis, Tampere University of Technology, Finland.
 RTP (2009). “Cultura: Concertos de música ligeira geraram 30 milhões em
2007”. URL: http://ww1.rtp.pt/noticias/?article=385096&visual=26&tema=5/.
Published in January 29, 2009, available by February 2, 2009.
 Russell J. A. (1980). “A circumspect model of affect”, Journal of Psychology and
Social Psychology, Vol. 39, No. 6, p. 1161.
 Sanden C and Zhang J. Z. (2011). “An Empirical Study of Multi-Label
Classifiers for Music Tag Annotation”, International Society for Music Information
Retrieval Conference – ISMIR’2003.
77 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 Schubert E. (2003). “Update of the Hevner adjective checklist”, Perceptual and
Motor Skills 96, 1117–1122.
 Serra X. (2005). “Bridging the Music Semantic Gap”, talk given at the
Information Day on “Audio-Visual Search Technologies” organized by the
Commission services of DG Information Society of the EC in Brussels.
 Sloboda J. A. and Juslin P. N. (2001). “Psychological perspectives on music and
emotion”. In Music and Emotion: Theory and Research, P. N. Juslin and J. A.
Sloboda, Eds. Oxford, University Press, New York.
 Thayer R. E. (1989). “The Biopsychology of Mood and Arousal”, New York,
Oxford University Press.
 TechCrunch (2009). “iTunes Sells 6 Billion Songs, And Other Fun Stats From
The Philnote”. URL: http://www.techcrunch.com/2009/01/06/itunes-sells-6-
billion-songs-and-other-fun-stats-from-the-philnote/, published in January 6,
2009, available by February 2, 2009.
 TechWhack (2005). “Apple iTunes Touches an Impressive 250 Million Sales
Figure”. URL: http://news.techwhack.com/690/apple-itunes-250-million/,
published in January 25, 2005, available by February 2, 2009.
78 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 TMN (2009). “TMN lança inovador serviço de identificação, pesquisa e
aquisição de música através do telemóvel”. URL:
http://www.tmn.pt/portal/site/tmn/menuitem.db67f528a6dbaa5ac8a71c10a51056
a0/?vgnextoid=d072fbb324238110VgnVCM1000005401650aRCRD&vgnextcha
nnel=15b5fa2eb8b37110VgnVCM1000005401650aRCRD&vgnextfmt=default2,
published in February 18, 2008, available by February 2, 2009.
 Tzanetakis G. (2002). Manipulation, Analysis and Retrieval Systems for Audio
Signals, PhD Thesis, Department of Computer Science, Princeton University,
USA.
 Wang A. (2003). “An Industrial-Strength Audio Search Algorithm”, International
Society for Music Information Retrieval Conference – ISMIR’2003, presentation.
 Welsh M., Borisov N., Hill J., von Behren R. and Woo A. (1999). Querying Large
Collections of Music for Similarity, Technical Report, Computer Science Division,
University of California, Berkeley, USA.
 Yang D. and Lee W. (2004). “Disambiguating music emotion using software
agents”, International Society for Music Information Retrieval Conference –
ISMIR’2004
 Yang Y.-H. and Chen H.-H. (2012). "Machine recognition of music emotion: a
review“, To Appear in ACM Trans. Intelligent Systems and Technology.
79 © Rui Pedro Paiva, “From MIR to MER”, 2012
Cited References
 Yang Y.-H., Lin Y.-C., Su Y.-F and Chen H.-H. (2007), "Music emotion
classification: A regression approach“, IEEE Int. Conf. Multimedia and Expo –
ICME’2007.

Contenu connexe

Tendances

Emotion recognition using image processing in deep learning
Emotion recognition using image     processing in deep learningEmotion recognition using image     processing in deep learning
Emotion recognition using image processing in deep learningvishnuv43
 
SPATIAL FILTERING IN IMAGE PROCESSING
SPATIAL FILTERING IN IMAGE PROCESSINGSPATIAL FILTERING IN IMAGE PROCESSING
SPATIAL FILTERING IN IMAGE PROCESSINGmuthu181188
 
Age and Gender Detection-converted.pdf
Age and Gender Detection-converted.pdfAge and Gender Detection-converted.pdf
Age and Gender Detection-converted.pdfMohammedMuzammil83
 
Crime Analysis & Prediction System
Crime Analysis & Prediction SystemCrime Analysis & Prediction System
Crime Analysis & Prediction SystemBigDataCloud
 
synopsis report on BIOMETRIC ONLINE VOTING SYSTEM
synopsis report on BIOMETRIC ONLINE VOTING SYSTEMsynopsis report on BIOMETRIC ONLINE VOTING SYSTEM
synopsis report on BIOMETRIC ONLINE VOTING SYSTEMNitesh Dubey
 
Gas Management System by Vikash Mainanwal
Gas Management System by Vikash MainanwalGas Management System by Vikash Mainanwal
Gas Management System by Vikash MainanwalVIKASH MAINANWAL
 
Prediction of rainfall using image processing
Prediction of rainfall using image processingPrediction of rainfall using image processing
Prediction of rainfall using image processingVineesh Kumar
 
Interpixel redundancy
Interpixel redundancyInterpixel redundancy
Interpixel redundancyNaveen Kumar
 
Image restoration and enhancement #2
Image restoration and enhancement #2 Image restoration and enhancement #2
Image restoration and enhancement #2 Gera Paulos
 
FACE RECOGNITION USING NEURAL NETWORK
FACE RECOGNITION USING NEURAL NETWORKFACE RECOGNITION USING NEURAL NETWORK
FACE RECOGNITION USING NEURAL NETWORKcodebangla
 
image basics and image compression
image basics and image compressionimage basics and image compression
image basics and image compressionmurugan hari
 
Output analysis for simulation models / Elimination of initial Bias
Output analysis for simulation models / Elimination of initial BiasOutput analysis for simulation models / Elimination of initial Bias
Output analysis for simulation models / Elimination of initial BiasTilakpoudel2
 
IT6005 digital image processing question bank
IT6005   digital image processing question bankIT6005   digital image processing question bank
IT6005 digital image processing question bankGayathri Krishnamoorthy
 
Crime Data Analysis, Visualization and Prediction using Data Mining
Crime Data Analysis, Visualization and Prediction using Data MiningCrime Data Analysis, Visualization and Prediction using Data Mining
Crime Data Analysis, Visualization and Prediction using Data MiningAnavadya Shibu
 

Tendances (20)

Emotion recognition using image processing in deep learning
Emotion recognition using image     processing in deep learningEmotion recognition using image     processing in deep learning
Emotion recognition using image processing in deep learning
 
SPATIAL FILTERING IN IMAGE PROCESSING
SPATIAL FILTERING IN IMAGE PROCESSINGSPATIAL FILTERING IN IMAGE PROCESSING
SPATIAL FILTERING IN IMAGE PROCESSING
 
Age and Gender Detection-converted.pdf
Age and Gender Detection-converted.pdfAge and Gender Detection-converted.pdf
Age and Gender Detection-converted.pdf
 
Forensics Analysis and Validation
Forensics Analysis and Validation  Forensics Analysis and Validation
Forensics Analysis and Validation
 
Crime Analysis & Prediction System
Crime Analysis & Prediction SystemCrime Analysis & Prediction System
Crime Analysis & Prediction System
 
synopsis report on BIOMETRIC ONLINE VOTING SYSTEM
synopsis report on BIOMETRIC ONLINE VOTING SYSTEMsynopsis report on BIOMETRIC ONLINE VOTING SYSTEM
synopsis report on BIOMETRIC ONLINE VOTING SYSTEM
 
Gas Management System by Vikash Mainanwal
Gas Management System by Vikash MainanwalGas Management System by Vikash Mainanwal
Gas Management System by Vikash Mainanwal
 
Audio mining
Audio miningAudio mining
Audio mining
 
Prediction of rainfall using image processing
Prediction of rainfall using image processingPrediction of rainfall using image processing
Prediction of rainfall using image processing
 
Interpixel redundancy
Interpixel redundancyInterpixel redundancy
Interpixel redundancy
 
Image restoration and enhancement #2
Image restoration and enhancement #2 Image restoration and enhancement #2
Image restoration and enhancement #2
 
FACE RECOGNITION USING NEURAL NETWORK
FACE RECOGNITION USING NEURAL NETWORKFACE RECOGNITION USING NEURAL NETWORK
FACE RECOGNITION USING NEURAL NETWORK
 
movie ticket pricing prediction
movie ticket pricing prediction movie ticket pricing prediction
movie ticket pricing prediction
 
image basics and image compression
image basics and image compressionimage basics and image compression
image basics and image compression
 
E voting
E voting E voting
E voting
 
Output analysis for simulation models / Elimination of initial Bias
Output analysis for simulation models / Elimination of initial BiasOutput analysis for simulation models / Elimination of initial Bias
Output analysis for simulation models / Elimination of initial Bias
 
IT6005 digital image processing question bank
IT6005   digital image processing question bankIT6005   digital image processing question bank
IT6005 digital image processing question bank
 
Noise Models
Noise ModelsNoise Models
Noise Models
 
Crime Data Analysis, Visualization and Prediction using Data Mining
Crime Data Analysis, Visualization and Prediction using Data MiningCrime Data Analysis, Visualization and Prediction using Data Mining
Crime Data Analysis, Visualization and Prediction using Data Mining
 
PPT.pptx
PPT.pptxPPT.pptx
PPT.pptx
 

En vedette

MOODetector: Automatic Music Emotion Recognition
MOODetector: Automatic Music Emotion RecognitionMOODetector: Automatic Music Emotion Recognition
MOODetector: Automatic Music Emotion RecognitionRui Pedro Paiva
 
Emotion based music player
Emotion based music playerEmotion based music player
Emotion based music playerNizam Muhammed
 
MusicMood - Machine Learning in Automatic Music Mood Prediction Based on Song...
MusicMood - Machine Learning in Automatic Music Mood Prediction Based on Song...MusicMood - Machine Learning in Automatic Music Mood Prediction Based on Song...
MusicMood - Machine Learning in Automatic Music Mood Prediction Based on Song...Sebastian Raschka
 
Interdisciplinary Perspectives on Emotion, Music and Technology
Interdisciplinary Perspectives on Emotion, Music and TechnologyInterdisciplinary Perspectives on Emotion, Music and Technology
Interdisciplinary Perspectives on Emotion, Music and TechnologyScottBeveridge
 
A framework for emotion mining from text in online social networks(final)
A framework for emotion mining from text in online social networks(final)A framework for emotion mining from text in online social networks(final)
A framework for emotion mining from text in online social networks(final)es712
 
Music Mood Detection (Lyrics based Approach)
Music Mood Detection (Lyrics based Approach)Music Mood Detection (Lyrics based Approach)
Music Mood Detection (Lyrics based Approach)Akhil Panchal
 
Emotion Detection from Text
Emotion Detection from TextEmotion Detection from Text
Emotion Detection from TextIJERD Editor
 
Emotion Detection in text
Emotion Detection in text Emotion Detection in text
Emotion Detection in text kashif kashif
 
Emotion based music player
Emotion based music playerEmotion based music player
Emotion based music playerNizam Muhammed
 
Emotion detection from text using data mining and text mining
Emotion detection from text using data mining and text miningEmotion detection from text using data mining and text mining
Emotion detection from text using data mining and text miningSakthi Dasans
 

En vedette (12)

MOODetector: Automatic Music Emotion Recognition
MOODetector: Automatic Music Emotion RecognitionMOODetector: Automatic Music Emotion Recognition
MOODetector: Automatic Music Emotion Recognition
 
Emotion based music player
Emotion based music playerEmotion based music player
Emotion based music player
 
MusicMood - Machine Learning in Automatic Music Mood Prediction Based on Song...
MusicMood - Machine Learning in Automatic Music Mood Prediction Based on Song...MusicMood - Machine Learning in Automatic Music Mood Prediction Based on Song...
MusicMood - Machine Learning in Automatic Music Mood Prediction Based on Song...
 
Interdisciplinary Perspectives on Emotion, Music and Technology
Interdisciplinary Perspectives on Emotion, Music and TechnologyInterdisciplinary Perspectives on Emotion, Music and Technology
Interdisciplinary Perspectives on Emotion, Music and Technology
 
Music Information Retrieval
Music Information RetrievalMusic Information Retrieval
Music Information Retrieval
 
A framework for emotion mining from text in online social networks(final)
A framework for emotion mining from text in online social networks(final)A framework for emotion mining from text in online social networks(final)
A framework for emotion mining from text in online social networks(final)
 
Music Mood Detection (Lyrics based Approach)
Music Mood Detection (Lyrics based Approach)Music Mood Detection (Lyrics based Approach)
Music Mood Detection (Lyrics based Approach)
 
Emotion mining in text
Emotion mining in textEmotion mining in text
Emotion mining in text
 
Emotion Detection from Text
Emotion Detection from TextEmotion Detection from Text
Emotion Detection from Text
 
Emotion Detection in text
Emotion Detection in text Emotion Detection in text
Emotion Detection in text
 
Emotion based music player
Emotion based music playerEmotion based music player
Emotion based music player
 
Emotion detection from text using data mining and text mining
Emotion detection from text using data mining and text miningEmotion detection from text using data mining and text mining
Emotion detection from text using data mining and text mining
 

Similaire à From MIR to MER: Exploring Music Information Retrieval and Music Emotion Recognition

Music Information Retrieval: Overview and Current Trends 2008
Music Information Retrieval: Overview and Current Trends 2008Music Information Retrieval: Overview and Current Trends 2008
Music Information Retrieval: Overview and Current Trends 2008Rui Pedro Paiva
 
The convergence of "hard" and "soft"in music technology, Rolf Inge Godøy, UiO
The convergence of "hard" and "soft"in music technology, Rolf Inge Godøy, UiOThe convergence of "hard" and "soft"in music technology, Rolf Inge Godøy, UiO
The convergence of "hard" and "soft"in music technology, Rolf Inge Godøy, UiOThe Research Council of Norway, IKTPLUSS
 
Introduction to Fast by Professor Mark Sandler
Introduction to Fast by  Professor Mark SandlerIntroduction to Fast by  Professor Mark Sandler
Introduction to Fast by Professor Mark SandlerFASTIMPACT
 
Music Objects to Social Machines
Music Objects to Social MachinesMusic Objects to Social Machines
Music Objects to Social MachinesDavid De Roure
 
machine learning x music
machine learning x musicmachine learning x music
machine learning x musicYi-Hsuan Yang
 
Music data analysis big data presentation
Music data analysis big data presentationMusic data analysis big data presentation
Music data analysis big data presentationShubhanshu Gupta
 
Music Information Retrieval: A Literature Review
Music Information Retrieval: A Literature ReviewMusic Information Retrieval: A Literature Review
Music Information Retrieval: A Literature ReviewJane Frazier
 
Denktank 2010
Denktank 2010Denktank 2010
Denktank 2010ocor203
 
Essentials of Digital Music
Essentials of Digital MusicEssentials of Digital Music
Essentials of Digital Musicijtsrd
 
Brief ismir intro + Spotify usecase @ Webinar Radio 2.0 MediaDATA Ismir2018
Brief ismir intro + Spotify usecase @ Webinar Radio 2.0 MediaDATA Ismir2018Brief ismir intro + Spotify usecase @ Webinar Radio 2.0 MediaDATA Ismir2018
Brief ismir intro + Spotify usecase @ Webinar Radio 2.0 MediaDATA Ismir2018ACTUONDA
 
J. S. Downie, D. De Roure, K. Page.Towards Web-Scale Analysis of Musical Stru...
J. S. Downie, D. De Roure, K. Page.Towards Web-Scale Analysis of Musical Stru...J. S. Downie, D. De Roure, K. Page.Towards Web-Scale Analysis of Musical Stru...
J. S. Downie, D. De Roure, K. Page.Towards Web-Scale Analysis of Musical Stru...MusicNet
 
Towards Web-Scale Analysis of Musical Structure
Towards Web-Scale Analysis of Musical Structure Towards Web-Scale Analysis of Musical Structure
Towards Web-Scale Analysis of Musical Structure David De Roure
 
Towards a musical Semantic Web
Towards a musical Semantic WebTowards a musical Semantic Web
Towards a musical Semantic WebYves Raimond
 
Caroline Ardrey (University of Birmingham)
Caroline Ardrey (University of Birmingham)Caroline Ardrey (University of Birmingham)
Caroline Ardrey (University of Birmingham)Renata Brandão
 
1. paul dwyer internet radio emma
1. paul dwyer internet radio emma1. paul dwyer internet radio emma
1. paul dwyer internet radio emmaPaul Dwyer
 
Lorenzo Porcaro PhD Defense
Lorenzo Porcaro PhD Defense Lorenzo Porcaro PhD Defense
Lorenzo Porcaro PhD Defense Lorenzo Porcaro
 

Similaire à From MIR to MER: Exploring Music Information Retrieval and Music Emotion Recognition (20)

Music Information Retrieval: Overview and Current Trends 2008
Music Information Retrieval: Overview and Current Trends 2008Music Information Retrieval: Overview and Current Trends 2008
Music Information Retrieval: Overview and Current Trends 2008
 
The convergence of "hard" and "soft"in music technology, Rolf Inge Godøy, UiO
The convergence of "hard" and "soft"in music technology, Rolf Inge Godøy, UiOThe convergence of "hard" and "soft"in music technology, Rolf Inge Godøy, UiO
The convergence of "hard" and "soft"in music technology, Rolf Inge Godøy, UiO
 
Introduction to Fast by Professor Mark Sandler
Introduction to Fast by  Professor Mark SandlerIntroduction to Fast by  Professor Mark Sandler
Introduction to Fast by Professor Mark Sandler
 
Music Objects to Social Machines
Music Objects to Social MachinesMusic Objects to Social Machines
Music Objects to Social Machines
 
machine learning x music
machine learning x musicmachine learning x music
machine learning x music
 
Music data analysis big data presentation
Music data analysis big data presentationMusic data analysis big data presentation
Music data analysis big data presentation
 
楊奕軒/音樂資料檢索
楊奕軒/音樂資料檢索楊奕軒/音樂資料檢索
楊奕軒/音樂資料檢索
 
Music Information Retrieval: A Literature Review
Music Information Retrieval: A Literature ReviewMusic Information Retrieval: A Literature Review
Music Information Retrieval: A Literature Review
 
Ism2011
Ism2011Ism2011
Ism2011
 
Denktank 2010
Denktank 2010Denktank 2010
Denktank 2010
 
Essentials of Digital Music
Essentials of Digital MusicEssentials of Digital Music
Essentials of Digital Music
 
Brief ismir intro + Spotify usecase @ Webinar Radio 2.0 MediaDATA Ismir2018
Brief ismir intro + Spotify usecase @ Webinar Radio 2.0 MediaDATA Ismir2018Brief ismir intro + Spotify usecase @ Webinar Radio 2.0 MediaDATA Ismir2018
Brief ismir intro + Spotify usecase @ Webinar Radio 2.0 MediaDATA Ismir2018
 
J. S. Downie, D. De Roure, K. Page.Towards Web-Scale Analysis of Musical Stru...
J. S. Downie, D. De Roure, K. Page.Towards Web-Scale Analysis of Musical Stru...J. S. Downie, D. De Roure, K. Page.Towards Web-Scale Analysis of Musical Stru...
J. S. Downie, D. De Roure, K. Page.Towards Web-Scale Analysis of Musical Stru...
 
Towards Web-Scale Analysis of Musical Structure
Towards Web-Scale Analysis of Musical Structure Towards Web-Scale Analysis of Musical Structure
Towards Web-Scale Analysis of Musical Structure
 
Big Data in the Music Industries, Dagfinn Bach, Bach Technology
Big Data in the Music Industries, Dagfinn Bach, Bach TechnologyBig Data in the Music Industries, Dagfinn Bach, Bach Technology
Big Data in the Music Industries, Dagfinn Bach, Bach Technology
 
Towards a musical Semantic Web
Towards a musical Semantic WebTowards a musical Semantic Web
Towards a musical Semantic Web
 
Bowman_resume_LIS
Bowman_resume_LISBowman_resume_LIS
Bowman_resume_LIS
 
Caroline Ardrey (University of Birmingham)
Caroline Ardrey (University of Birmingham)Caroline Ardrey (University of Birmingham)
Caroline Ardrey (University of Birmingham)
 
1. paul dwyer internet radio emma
1. paul dwyer internet radio emma1. paul dwyer internet radio emma
1. paul dwyer internet radio emma
 
Lorenzo Porcaro PhD Defense
Lorenzo Porcaro PhD Defense Lorenzo Porcaro PhD Defense
Lorenzo Porcaro PhD Defense
 

Dernier

Yaroslav Rozhankivskyy: Три складові і три передумови максимальної продуктивн...
Yaroslav Rozhankivskyy: Три складові і три передумови максимальної продуктивн...Yaroslav Rozhankivskyy: Три складові і три передумови максимальної продуктивн...
Yaroslav Rozhankivskyy: Три складові і три передумови максимальної продуктивн...Lviv Startup Club
 
Unlocking the Secrets of Affiliate Marketing.pdf
Unlocking the Secrets of Affiliate Marketing.pdfUnlocking the Secrets of Affiliate Marketing.pdf
Unlocking the Secrets of Affiliate Marketing.pdfOnline Income Engine
 
Call Girls Pune Just Call 9907093804 Top Class Call Girl Service Available
Call Girls Pune Just Call 9907093804 Top Class Call Girl Service AvailableCall Girls Pune Just Call 9907093804 Top Class Call Girl Service Available
Call Girls Pune Just Call 9907093804 Top Class Call Girl Service AvailableDipal Arora
 
Mysore Call Girls 8617370543 WhatsApp Number 24x7 Best Services
Mysore Call Girls 8617370543 WhatsApp Number 24x7 Best ServicesMysore Call Girls 8617370543 WhatsApp Number 24x7 Best Services
Mysore Call Girls 8617370543 WhatsApp Number 24x7 Best ServicesDipal Arora
 
It will be International Nurses' Day on 12 May
It will be International Nurses' Day on 12 MayIt will be International Nurses' Day on 12 May
It will be International Nurses' Day on 12 MayNZSG
 
Event mailer assignment progress report .pdf
Event mailer assignment progress report .pdfEvent mailer assignment progress report .pdf
Event mailer assignment progress report .pdftbatkhuu1
 
RSA Conference Exhibitor List 2024 - Exhibitors Data
RSA Conference Exhibitor List 2024 - Exhibitors DataRSA Conference Exhibitor List 2024 - Exhibitors Data
RSA Conference Exhibitor List 2024 - Exhibitors DataExhibitors Data
 
Call Girls Navi Mumbai Just Call 9907093804 Top Class Call Girl Service Avail...
Call Girls Navi Mumbai Just Call 9907093804 Top Class Call Girl Service Avail...Call Girls Navi Mumbai Just Call 9907093804 Top Class Call Girl Service Avail...
Call Girls Navi Mumbai Just Call 9907093804 Top Class Call Girl Service Avail...Dipal Arora
 
M.C Lodges -- Guest House in Jhang.
M.C Lodges --  Guest House in Jhang.M.C Lodges --  Guest House in Jhang.
M.C Lodges -- Guest House in Jhang.Aaiza Hassan
 
The Path to Product Excellence: Avoiding Common Pitfalls and Enhancing Commun...
The Path to Product Excellence: Avoiding Common Pitfalls and Enhancing Commun...The Path to Product Excellence: Avoiding Common Pitfalls and Enhancing Commun...
The Path to Product Excellence: Avoiding Common Pitfalls and Enhancing Commun...Aggregage
 
Monthly Social Media Update April 2024 pptx.pptx
Monthly Social Media Update April 2024 pptx.pptxMonthly Social Media Update April 2024 pptx.pptx
Monthly Social Media Update April 2024 pptx.pptxAndy Lambert
 
A305_A2_file_Batkhuu progress report.pdf
A305_A2_file_Batkhuu progress report.pdfA305_A2_file_Batkhuu progress report.pdf
A305_A2_file_Batkhuu progress report.pdftbatkhuu1
 
7.pdf This presentation captures many uses and the significance of the number...
7.pdf This presentation captures many uses and the significance of the number...7.pdf This presentation captures many uses and the significance of the number...
7.pdf This presentation captures many uses and the significance of the number...Paul Menig
 
Value Proposition canvas- Customer needs and pains
Value Proposition canvas- Customer needs and painsValue Proposition canvas- Customer needs and pains
Value Proposition canvas- Customer needs and painsP&CO
 
Best VIP Call Girls Noida Sector 40 Call Me: 8448380779
Best VIP Call Girls Noida Sector 40 Call Me: 8448380779Best VIP Call Girls Noida Sector 40 Call Me: 8448380779
Best VIP Call Girls Noida Sector 40 Call Me: 8448380779Delhi Call girls
 
Regression analysis: Simple Linear Regression Multiple Linear Regression
Regression analysis:  Simple Linear Regression Multiple Linear RegressionRegression analysis:  Simple Linear Regression Multiple Linear Regression
Regression analysis: Simple Linear Regression Multiple Linear RegressionRavindra Nath Shukla
 
Call Girls Jp Nagar Just Call 👗 7737669865 👗 Top Class Call Girl Service Bang...
Call Girls Jp Nagar Just Call 👗 7737669865 👗 Top Class Call Girl Service Bang...Call Girls Jp Nagar Just Call 👗 7737669865 👗 Top Class Call Girl Service Bang...
Call Girls Jp Nagar Just Call 👗 7737669865 👗 Top Class Call Girl Service Bang...amitlee9823
 
Understanding the Pakistan Budgeting Process: Basics and Key Insights
Understanding the Pakistan Budgeting Process: Basics and Key InsightsUnderstanding the Pakistan Budgeting Process: Basics and Key Insights
Understanding the Pakistan Budgeting Process: Basics and Key Insightsseri bangash
 
Boost the utilization of your HCL environment by reevaluating use cases and f...
Boost the utilization of your HCL environment by reevaluating use cases and f...Boost the utilization of your HCL environment by reevaluating use cases and f...
Boost the utilization of your HCL environment by reevaluating use cases and f...Roland Driesen
 

Dernier (20)

Yaroslav Rozhankivskyy: Три складові і три передумови максимальної продуктивн...
Yaroslav Rozhankivskyy: Три складові і три передумови максимальної продуктивн...Yaroslav Rozhankivskyy: Три складові і три передумови максимальної продуктивн...
Yaroslav Rozhankivskyy: Три складові і три передумови максимальної продуктивн...
 
Unlocking the Secrets of Affiliate Marketing.pdf
Unlocking the Secrets of Affiliate Marketing.pdfUnlocking the Secrets of Affiliate Marketing.pdf
Unlocking the Secrets of Affiliate Marketing.pdf
 
Call Girls Pune Just Call 9907093804 Top Class Call Girl Service Available
Call Girls Pune Just Call 9907093804 Top Class Call Girl Service AvailableCall Girls Pune Just Call 9907093804 Top Class Call Girl Service Available
Call Girls Pune Just Call 9907093804 Top Class Call Girl Service Available
 
Mysore Call Girls 8617370543 WhatsApp Number 24x7 Best Services
Mysore Call Girls 8617370543 WhatsApp Number 24x7 Best ServicesMysore Call Girls 8617370543 WhatsApp Number 24x7 Best Services
Mysore Call Girls 8617370543 WhatsApp Number 24x7 Best Services
 
It will be International Nurses' Day on 12 May
It will be International Nurses' Day on 12 MayIt will be International Nurses' Day on 12 May
It will be International Nurses' Day on 12 May
 
Event mailer assignment progress report .pdf
Event mailer assignment progress report .pdfEvent mailer assignment progress report .pdf
Event mailer assignment progress report .pdf
 
VVVIP Call Girls In Greater Kailash ➡️ Delhi ➡️ 9999965857 🚀 No Advance 24HRS...
VVVIP Call Girls In Greater Kailash ➡️ Delhi ➡️ 9999965857 🚀 No Advance 24HRS...VVVIP Call Girls In Greater Kailash ➡️ Delhi ➡️ 9999965857 🚀 No Advance 24HRS...
VVVIP Call Girls In Greater Kailash ➡️ Delhi ➡️ 9999965857 🚀 No Advance 24HRS...
 
RSA Conference Exhibitor List 2024 - Exhibitors Data
RSA Conference Exhibitor List 2024 - Exhibitors DataRSA Conference Exhibitor List 2024 - Exhibitors Data
RSA Conference Exhibitor List 2024 - Exhibitors Data
 
Call Girls Navi Mumbai Just Call 9907093804 Top Class Call Girl Service Avail...
Call Girls Navi Mumbai Just Call 9907093804 Top Class Call Girl Service Avail...Call Girls Navi Mumbai Just Call 9907093804 Top Class Call Girl Service Avail...
Call Girls Navi Mumbai Just Call 9907093804 Top Class Call Girl Service Avail...
 
M.C Lodges -- Guest House in Jhang.
M.C Lodges --  Guest House in Jhang.M.C Lodges --  Guest House in Jhang.
M.C Lodges -- Guest House in Jhang.
 
The Path to Product Excellence: Avoiding Common Pitfalls and Enhancing Commun...
The Path to Product Excellence: Avoiding Common Pitfalls and Enhancing Commun...The Path to Product Excellence: Avoiding Common Pitfalls and Enhancing Commun...
The Path to Product Excellence: Avoiding Common Pitfalls and Enhancing Commun...
 
Monthly Social Media Update April 2024 pptx.pptx
Monthly Social Media Update April 2024 pptx.pptxMonthly Social Media Update April 2024 pptx.pptx
Monthly Social Media Update April 2024 pptx.pptx
 
A305_A2_file_Batkhuu progress report.pdf
A305_A2_file_Batkhuu progress report.pdfA305_A2_file_Batkhuu progress report.pdf
A305_A2_file_Batkhuu progress report.pdf
 
7.pdf This presentation captures many uses and the significance of the number...
7.pdf This presentation captures many uses and the significance of the number...7.pdf This presentation captures many uses and the significance of the number...
7.pdf This presentation captures many uses and the significance of the number...
 
Value Proposition canvas- Customer needs and pains
Value Proposition canvas- Customer needs and painsValue Proposition canvas- Customer needs and pains
Value Proposition canvas- Customer needs and pains
 
Best VIP Call Girls Noida Sector 40 Call Me: 8448380779
Best VIP Call Girls Noida Sector 40 Call Me: 8448380779Best VIP Call Girls Noida Sector 40 Call Me: 8448380779
Best VIP Call Girls Noida Sector 40 Call Me: 8448380779
 
Regression analysis: Simple Linear Regression Multiple Linear Regression
Regression analysis:  Simple Linear Regression Multiple Linear RegressionRegression analysis:  Simple Linear Regression Multiple Linear Regression
Regression analysis: Simple Linear Regression Multiple Linear Regression
 
Call Girls Jp Nagar Just Call 👗 7737669865 👗 Top Class Call Girl Service Bang...
Call Girls Jp Nagar Just Call 👗 7737669865 👗 Top Class Call Girl Service Bang...Call Girls Jp Nagar Just Call 👗 7737669865 👗 Top Class Call Girl Service Bang...
Call Girls Jp Nagar Just Call 👗 7737669865 👗 Top Class Call Girl Service Bang...
 
Understanding the Pakistan Budgeting Process: Basics and Key Insights
Understanding the Pakistan Budgeting Process: Basics and Key InsightsUnderstanding the Pakistan Budgeting Process: Basics and Key Insights
Understanding the Pakistan Budgeting Process: Basics and Key Insights
 
Boost the utilization of your HCL environment by reevaluating use cases and f...
Boost the utilization of your HCL environment by reevaluating use cases and f...Boost the utilization of your HCL environment by reevaluating use cases and f...
Boost the utilization of your HCL environment by reevaluating use cases and f...
 

From MIR to MER: Exploring Music Information Retrieval and Music Emotion Recognition

  • 1. From Music Information Retrieval to Music Emotion Recognition Center for Informatics and Systems of the University of Coimbra (CISUC) Portugal Rui Pedro Paiva March 2012
  • 2. 2 © Rui Pedro Paiva, “From MIR to MER”, 2012 Contents  Music Information Retrieval  Music Emotion Recognition  References
  • 3. 3 © Rui Pedro Paiva, “From MIR to MER”, 2012 MUSIC INFORMATION RETRIEVAL - MIR: What and Why? - Applications - Short MIR Tale - Representations - Techniques
  • 4. 4 © Rui Pedro Paiva, “From MIR to MER”, 2012 MIR: What and Why?  MIR • Music Information Retrieval: interdisciplinary research field devoted to the study of information extraction mechanisms from musical pieces, retrieval methodologies, as well as all the processes involved in those tasks in different music representation media.  Why MIR? • MIR emerges from the necessity to manage huge collections of digital music for “preservation, access, research and other uses” [Futrelle and Downie, 2003]
  • 5. 5 © Rui Pedro Paiva, “From MIR to MER”, 2012 MIR: What and Why?  Music and Man • Music expresses “that which cannot be put into words and that which cannot remain silent” (Victor Hugo) • “We associate music with the most unique moments of our lives and music is part of our individual and social imaginary” [Paiva, 2006] - “By listening to music, emotions and memories, thoughts and reactions, are awakened” [Paiva, 2006] • “Life has a soundtrack” [Gomes, 2005] (“Festivais de Verão”, Jornal “Público”) • “The history of a people is found in its songs” (George Jellinek)
  • 6. 6 © Rui Pedro Paiva, “From MIR to MER”, 2012 MIR: What and Why?  Music and World economy • Music industry runs, only in the USA an amount of money in the order of several billion US dollars per year. • Explosion of the Electronic Music Industry (EMD) - Widespread access to the Internet - Bandwidth increasing in domestic and mobile accesses - Compact audio formats with near CD quality (mp3, wma) - Portable music devices (iPod, mp3 readers) - Peer-to-peer networks (Napster, Kazaa, eMule) - Online music stores (iTunes, Calabash Music, Sapo Music)  resolution is the song, not the CD - Music identification platforms (Shazam, 411-Song, Gracenote MusicID / TrackID) - Music recommendation systems (MusicSurfer)
  • 7. 7 © Rui Pedro Paiva, “From MIR to MER”, 2012 MIR: What and Why?  Music and World economy (cont.) • By 2005, Apple iTunes was selling  1.25 million songs each day [TechWhack, 2005] - Until January 2009, over 6 billion songs had been sold in total [TechCrunch, 2009] • By 2007, music shows in Portugal sold 30 M€ in tickets [RTP, 2009] • Number and dimension of digital music archives continuously growing - Database size (these days, over 2 million songs) - Genres covered • Challenges to music providers and music librarians - Organization, maintenance, labeling, user interaction - Any large music database is only really useful if users can find what they are looking for in an efficient manner!
  • 8. 8 © Rui Pedro Paiva, “From MIR to MER”, 2012 MIR: What and Why?  Database Organization and Music Retrieval • Presently, databases are manually annotated  search and retrieval is mostly textual (artist, title, album, genre) - Service providers – Difficulties regarding manual song labeling: subjective and time-consuming, - Customers – Difficulties in performing “content-based” queries • “Music’s preeminent functions are social and psychological”, and so “the most useful retrieval indexes are those that facilitate searching in conformity with such social and psychological functions. Typically, such indexes will focus on stylistic, mood, and similarity information” [Huron, 2000]. •  Music Information Retrieval (MIR) emerges from the necessity to manage huge collections of digital music for “preservation, access, research and other uses” [Futrelle and Downie, 2003].
  • 9. 9 © Rui Pedro Paiva, “From MIR to MER”, 2012 MIR: What and Why?  Database Organization and Music Retrieval Client Internet Ordering of Results and Summarization Comparison of Signatures Database Search Database Classification Raw Audio Recordings Segmentation and Summarization Feature Extraction Signature Feature Extraction Query Signature Visualization 1. Tom Jobim - Corcovado 2. Dido - Thank You 3. João Gilberto - Meia Luz 4. ..... 5. ..... Jazz Classical Latin Swing Baroque Mambo Salsa . Interactive Edition Query Creation © Rui Pedro Paiva, 2002
  • 10. 10 © Rui Pedro Paiva, “From MIR to MER”, 2012 Applications  Platforms for EMD • Similarity-based retrieval tools - Query-by-example [Welsh et al., 1999] – Music identification (Shazam, trackID, Tunatic) [Wang, 2003; Haitsma and Kalker, 2002; TMN, 2009] – Query-by-mood [Cardoso et al., 2011] – Music recommendation [Celma & Lamere, 2007] – Islands of music [Pampalk, 2001] • Metaphor of geographic maps: similar genres close together - Automatic playlist generation [Pauws and Wijdeven, 2005; Alghoniemy and Tewfik, 2000] © Elias Pampalk, 2001
  • 11. 11 © Rui Pedro Paiva, “From MIR to MER”, 2012  Platforms for EMD • Similarity-based retrieval tools - Query-by-melody (query-by-humming, QBS) [Parker, 2005; Ghias et al., 1995] - Plagiarism detection [Paiva et al., 2006] - Music web crawlers [Huron, 2000]  Music education and training • Automatic music transcription [Ryynänen, 2008; Kashino et al., 1995] - Music composition, analysis, performance evaluation, plagiarism detection  Digital music libraries • For research issues involving music retrieval, training (learning activities, evaluation, etc.)  Variations [Dunn, 2000] Applications
  • 12. 12 © Rui Pedro Paiva, “From MIR to MER”, 2012 Applications  Audio software • Intelligent audio (music) editors  automatic indexing [Tzanetakis, 2002]  Multimedia databases and operating systems [Burad, 2006]  Video indexing and searching • Segmentation based on audio (music) content  detection of scene transitions [Pfeiffer, 1996] © George Tzanetakis, 2002
  • 13. 13 © Rui Pedro Paiva, “From MIR to MER”, 2012 Applications  Advertisement and cinema • Tools for mood-based retrieval [Cardoso et al. 2011]  Sports • Music to induce a certain cardiac frequency [Matesic and Cromartie, 2002]  … and so forth…
  • 14. 14 © Rui Pedro Paiva, “From MIR to MER”, 2012 Short MIR Tale  Precursors of computer-based MIR: incipit and theme indexes, e.g., Harold Barlow and Sam Morganstern’s dictionary of musical themes [Barlow and Morganstern, 1948]
  • 15. 15 © Rui Pedro Paiva, “From MIR to MER”, 2012 Short MIR Tale  1966: potential of applying automatic information retrieval techniques to music was recognized [Kassler, 1966]  1970s and 1980s: automatic music transcription systems  1990s: surge of interest, mostly in topics such as query-by- humming (impulse from research on digital libraries)  2000: 1st International Symposium on Music Information Retrieval (ISMIR) • Now “International Society for Music Information Retrieval Conference”
  • 16. 16 © Rui Pedro Paiva, “From MIR to MER”, 2012 Short MIR Tale  2000-present: active multidisciplinary research field © Joe Futrelle and J. Stephen Downie, 2003
  • 17. 17 © Rui Pedro Paiva, “From MIR to MER”, 2012 Short MIR Tale  2000-present: active multidisciplinary research field • Content analysis and similarity assessment and retrieval in audio song databases - Metrics of similarity, music identification, music recommendation, audio fingerprinting, music classification and feature extraction, tempo and melody detection, music summarization • Databases systems, indexing, query languages • Knowledge representation, metadata • Theme extraction, harmonic and motive analysis, tonality • Emotion • Perception and cognition • Social, legal, ethical and business issues • User interface design
  • 18. 18 © Rui Pedro Paiva, “From MIR to MER”, 2012 Representations © Donald Byrd and Tim Crawford, 2002
  • 19. 19 © Rui Pedro Paiva, “From MIR to MER”, 2012 Representations  Audio (music content analysis) • Object - Audio recordings, streaming audio • Topics - Automatic transcription, QBE, classification, recommendation, identification, …  Symbolic • Object - Scores or event-based representations (e.g., MIDI) • Topics - Melodic matching, theme extraction, harmonic analysis, …
  • 20. 20 © Rui Pedro Paiva, “From MIR to MER”, 2012 Representations  Visual • Object - Printed music • Topics - Optical music recognition  Metadata • Object - Any kind • Topics - Digital libraries, music ontologies (semantic web)
  • 21. 21 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Idea • Extract semantic information from low-level data • Feature extraction - Physical: F0, intensity, centroid, uniformity, rolloff, flux - Perceptual: pitch, loudness, timbre, beat - Musicological: notes, melodies, measures, motives, themes - Higher-level (semantic) features: emotion, genre, instruments, artist • Content and Representation - Audio: which instruments, notes, artist - Symbolic: themes, motives - Visual: notation, basic units - Metadata: genre, artist, MPEG-7 descriptors
  • 22. 22 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Physical features • Time domain - Waveform analysis: energy contour, amplitude-based segmentation, auto-correlation, peak detection 0 0.005 0.01 0.015 0.02 0.025 0.03 0.035 0.04 0.045 0.05 -1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1 time (s) Intensity 0 0.005 0.01 0.015 0.02 0.025 0.03 0.035 0.04 0.045 0.05 -800 -600 -400 -200 0 200 400 600 800 Time (s) R(k)       000000 2;200;1,5sin 5 3sin 3 sin)( fHzfAt A t A tAtx  
  • 23. 23 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Physical features (cont.) • Time domain - Auditory model-based F0 detectors, beat detectors (energy-based) 0.21 0.22 0.23 0.24 0.25 -0.4 -0.2 0 0.2 0.4 Time (s) x(t) a) Sound waveform Time (s) Frequencychannel b) Cochleagram frame 0.21 0.22 0.23 0.24 0.25 20 40 60 80 Time Lag (ms) Frequncychannel c) Correlogram frame 0 5 10 15 20 20 40 60 80 0 5 10 15 20 25 0 0.1 0.2 0.3 0.4 Time Lag (ms) Pitchsalience d) Summary correlogram
  • 24. 24 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Physical features • Frequency domain - DFT, STFT, spectrogram - F0 detectors Time Frequency 0 0.5 1 1.5 2 0 0.5 1 1.5 2 x 10 4
  • 25. 25 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Physical features (cont.) • Spectral features - MFCCs, centroid, rolloff, flux, harmonicity, high-frequency content, … • Sub-band features © MIR Toolbox, 2008
  • 26. 26 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Physical features (cont.) • Spectral features + sub-band features (e.g., audio fingerprinting) © Haitsma & Kalker, 2002
  • 27. 27 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Perceptual features • Pitch - Frequency - Intensity - Context - Ear physiology (age) © MIR Toolbox, 2008 © http://www.ai.rug.nl/~tjeerd/CPSP/docs/cochleaModel.html
  • 28. 28 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Perceptual features (cont.) • Loudness - Intensity - Frequency - Context - Ear physiology (age) Fletcher-Munson equal loudness contours
  • 29. 29 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Perceptual features (cont.) • Timbre - No physical correlate - “what something sounds like”: – Spectral content at steady-sate • Centroid, rolloff, relative amplitudes of harmonic components, inharmonicty… – Signal’s temporal envelope • Attack transient – Temporal behavior of the harmonics • Melodic contour - UDUEEUUD • Rhythm contour - FSSFEEFS • Beat
  • 30. 30 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Musicological features • Notes from audio • Notes from optical music recognition 0 100 200 300 400 500 4800 4900 5000 5100 5200 5300 5400 Time (msec) f[k](cents) a) Eliades Ochoa's excerpt 0 200 400 600 800 6100 6200 6300 6400 6500 6600 6700 6800 Time (msec)f[k](cents) b) Female opera excerpt © Rui Pedro Paiva, 2006
  • 31. 31 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Musicological features (cont.) • Melody © Rui Pedro Paiva, 2006
  • 32. 32 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Musicological features (cont.) • Themes © Colin Meek and William Birmingham, 2001
  • 33. 33 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Higher-level features  top-down information flow © Xavier Serra, 2005
  • 34. 34 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Higher-level features (cont.) • Bridge the semantic gap • Memory, context, expectations - Repetitions, sonic environment, modeling the individual, musicological knowledge • Emotion: valence (happy/anxious) and arousal (calm/energetic) - Classification approaches resorting to low-level features
  • 35. 35 © Rui Pedro Paiva, “From MIR to MER”, 2012 MUSIC EMOTION RECOGNITION (MER) - MER: What and Why? - Applications - Short MER Tale - Emotion Models - Techniques - MER@CISUC - Current Limitations and Open Problems
  • 36. 36 © Rui Pedro Paiva, “From MIR to MER”, 2012 MER: What and Why?  MER • Music Emotion Recognition: branch of MIR devoted to the identification of emotions/moods in musical pieces • Emotion vs mood - MIR researchers  use both terms interchangeably - Psychologists  clear distinction [Sloboda and Juslin, 2001] – Emotion = a short experience in response to an object (e.g., music) – Mood = longer experience without specific object connection • Categories of emotions [Gabrielsson, 2002)] - Expressed emotion: emotion the performer tries to communicate to the listeners - Perceived emotion: emotion the listener perceives as being expressed in a song (which may be different than the emotion the performer tried to communicate)  scope of MIR researchers - Felt (evoked) emotion: emotion felt by the listener, in response to the song and performance
  • 37. 37 © Rui Pedro Paiva, “From MIR to MER”, 2012 MER: What and Why?  Why MER? • “Music’s preeminent functions are social and psychological”, and so “the most useful retrieval indexes are those that facilitate searching in conformity with such social and psychological functions. Typically, such indexes will focus on stylistic, mood, and similarity information” [Huron, 2000]. - Studies on music information behaviour  music mood is an important criterion for music retrieval and organization [Juslin and Laukka, 2004] • “In the academia, more and more multimedia systems that involve emotion analysis of music signals have been developed, such as Moodtrack, LyQ, MusicSense, Mood Cloud, Moody and i.MTV, just to name a few. In the industry, many music companies, such as AMG, Gracenote, MoodLogic, Musicovery, Syntonetic, and Sourcetoneuse emotion as a cue for music retrieval.” [Yang and Chen, 2012]
  • 38. 38 © Rui Pedro Paiva, “From MIR to MER”, 2012 MER: What and Why?  Difficulties [Yang and Chen, 2012] • Emotion perception is by nature subjective [Yang and Chen, 2012] - People can perceive different emotions for the same song •  Performance evaluation of an MER systems is difficult [Yang and Chen, 2012] - Common agreement on the recognition result is hard to obtain • Still not fully understood how music and emotion are related - Despite several studies on music psychology
  • 39. 39 © Rui Pedro Paiva, “From MIR to MER”, 2012 Applications  Platforms for EMD • Mood-based retrieval tools - Music recommendation, automatic playlist generation, classification, …  Game development  Cinema  Advertising  Health • Sports • Stress management
  • 40. 40 © Rui Pedro Paiva, “From MIR to MER”, 2012 Short MER Tale  19th century: initial research on the relations between music and emotion [Gabrielsson and Lindström, 2001]  20th century: more active study of music and emotion • Relationship between emotions and musical attributes - Mode, harmony, tempo, rhythm, dynamics [Meyers, 2007] • Mood models - E.g., Kate Hevner [Hevner, 1935], James Russell [Russell, 1980]  1980-1990s: Initial computational models  mostly emotion synthesis (e.g., [Juslin, 1997]) • Relationships between emotion and music composition and music expressivity
  • 41. 41 © Rui Pedro Paiva, “From MIR to MER”, 2012 Short MER Tale  1980-1990s: only a few works on emotion analysis • Mostly in the symbolic/MIDI domain [Lu et al., 2006]  2003: first work on mood detection in audio musical signals (by Yazhong Feng [Feng et al., 2003]) • 4 mood categories (happiness, sadness, anger and fear) • 2 musical attributes: tempo and articulation • Neural network classifier trained using 200 musical pieces • Test corpus of 23 pieces, with average precision and recall of 67 and 66%  2004: first multi-modal approach combining audio and text (lyrics) [Yang and Lee, 2004]
  • 42. 42 © Rui Pedro Paiva, “From MIR to MER”, 2012 Short MER Tale  2006: first study on MEVD (Music Emotion Variation Detection [Lu et al., 2006] • Temporal variable: emotions may change throughout a song  2007: first mood analysis model based on a dimensional approach [Yang et al., 2007] • Thayer 2D mood model  continuous emotion as function of arousal and valence
  • 43. 43 © Rui Pedro Paiva, “From MIR to MER”, 2012 Short MER Tale  2007: first MIREX (MIR Evaluation eXchange) track devoted to mood classification [MIREX, 2012] • Before: - Each approach used a different (and limited) set of features, mood taxonomies, number of classes and test sets. - Some approaches constrained the analysis to a particular musical style (e.g., classical music)  Present: active research in the field
  • 44. 44 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Two main conceptualizations of emotion • Categorical models - Emotions as categories: limited number of discrete emotions (adjectives) • Dimensional models - Emotions as continuous values, depending on 2 or 3 axes
  • 45. 45 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Categorical Models • Main ideas - “people experience emotions as categories that are distinct from each other” [Yang and Chen, 2012] - Existence of basic emotions – Limited number of universal and primary emotion classes (e.g., happiness, sadness, anger, fear, disgust, surprise) from which all other “secondary” emotion classes can be derived
  • 46. 46 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Categorical Models • Examples - Hevners’s 8 clusters of affective terms (1935) - Regrouped into 10 adjective groups by Farnsworth [Farnsworth, 1954] and into 9 adjective groups by Schubert [Schubert 2003]. © Yang and Chen, 2012
  • 47. 47 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Categorical Models • Examples - Tellegen-Watson-Clark model (1999) © Tellegen, Watson and Clark, 1999
  • 48. 48 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Categorical Models • Limitations - Limited number of adjectives - Larger number may ne impractical for psychological studies - Adjectives may be ambiguous
  • 49. 49 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Dimensional Models • Main ideas - Emotions experience as a continuous – Each emotion is a locations in a multi-dimensional plane, based on a reduced number of axes (2D or 3D) – Argument: correspond to internal human representations of emotions - 3 main dimensions of emotion [Yang and chen, 2012] – valence (or pleasantness; positive and negative affective states), – arousal (or activation; energy and stimulation level) – potency (or dominance; a sense of control or freedom to act) - 2D used in practice – Valence and arousal regarded as the “core processes” of affect [Yang and Chen, 2002] – Simpler to visualize emotions
  • 50. 50 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Dimensional Models • Examples - James Russell’s circumplex model [Russell, 1980] © Kim et al., 2010
  • 51. 51 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Dimensional Models • Examples - Robert Thayer’s model [Thayer, 1989] © Meyers, 2007
  • 52. 52 © Rui Pedro Paiva, “From MIR to MER”, 2012 Emotion Models  Dimensional Models • Limitations - Obscures important aspects of emotion – Anger and fear are placed close in the valence-arousal plane – Very different in terms of their implications –  potency (dominant–submissive) as the third dimension
  • 53. 53 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Idea • Extract musical features correlated to emotion from the information source under analysis (audio, MIDI, lyrics, …) • Temporal variable - Emotions may change throughout a song  MEVD
  • 54. 54 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Relevant musical attributes •  Studies by music psychologists, e.g., [Gabrielsson and Lindström, 2001] - Dynamics, articulation, timbre, pitch, interval, melody, harmony, tonality and rhythm [Friberg, 2008], mode, harmony, tempo, rhythm, dynamics, musical form [Meyers, 2007] – Modes: major modes related to happiness or solemnity, minor modes associated with sadness or anger [Meyers, 2007]. – Harmonies: simple, consonant, harmonies are usually happy, pleasant or relaxed; complex, dissonant, harmonies relate to emotions such as excitement, tension or sadness, as they create instability in a musical piece [Meyers, 2007]. - Few relevant audio features proposed so far [Friberg, 2008] – Difficult to extract from audio signals  but easier to extract from symbolic representations (some features are score-based in nature)
  • 55. 55 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Relevant musical attributes • Timing: Tempo, tempo variation, duration contrast • Dynamics: overall level, crescendo/decrescendo, accents • Articulation: overall (staccato/legato), variability • Timbre: Spectral richness, onset velocity, harmonic richness • Pitch (high/low) • Interval (small/large) • Melody: range (small/large), direction (up/down)
  • 56. 56 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Relevant musical attributes • Harmony (consonant/complex-dissonant) • Tonality (chromatic-atonal/key-oriented) • Rhythm (regular-smooth/firm/flowing-fluent/irregular-rough) • Mode (major/minor) • Loudness (high/low) • Musical form (complexity, repetition, new ideas, disruption)
  • 57. 57 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Overall approach: classical pattern recognition approach • 1) Selection of a mood model - Categorical or dimensional? - How many categories / variables? Multi-label or probabilistic classification? • 2) Acquisition of training and testing data - Necessary to perform manual data annotation? How many annotators, what profiles, how many songs, song balance, etc.? • 3) Feature extraction and selection - Which features? How to extract them from the information source? - How to evaluate the impact of different features in each class?
  • 58. 58 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Overall approach: classical pattern recognition approach • 4) Selection of a representation model - Black-box, gray-box, white-box? – Black-box paradigm dominates. Gray-box should be researched. - SVM, neural networks, GMM, KNN, neural-fuzzy classifiers, rule-base systems? – SVMs: most accurate results currently - NN: how many hidden layers, etc? SVM: which parameters?... – SVM parameter optimization: grid-search • 5) Model training - How to organize the training and testing experience? Cross validation? How many folds? – Tipically, 10-fold cross validation
  • 59. 59 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Overall approach: classical pattern recognition approach • 6) Model validation - How to quantitatively evaluate the results? RMSE, SSE, correlation, R2 statistics? Precision, recall and F-measure? Confusion matrix? – Regression: RMSE and R2 are tipically used – Classification: Precision, recall and F-measure are standard - What kinds of errors are more prevalent? Why? Where is the root of the problem: classifier/regressor, extracted features, method employed for extraction, feature selection approach? – Currently, the bottleneck seems to be the lack of semantic richness of current low-level features and the innacuracy of particular feature extraction algorithms (e.g., tempo, etc.) - What classes/variables show low accuracy? Why? – Valence: important features might be missing - Same questions for individual songs.
  • 60. 60 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Acquisition of training and testing data [Kim et al., 2010] • Subjective tests • Annotation games © Kim et al., 2010
  • 61. 61 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  Feature Extraction • Music platforms - Audio: Marsyas, MIR Toolbox, PsySound, … - MIDI: jMIR, … - Lyrics: ConceptNet, … • Audio: acoustic features developed in other contexts - Music genre classification (e.g., spectral shape features, Mel-frequency cepstral coefficients (MFCC), etc.) and speech recognition – Low-level audio descriptors (LLDs): spectral shape features such as centroid, spread, bandwidth, skewness, kurtosis, slope, decrease, rolloff, flux contrast, MFCCs, etc. • Necessary to - Evaluate the relevance of LLDs to mood detection - Program actual mood-related features
  • 62. 62 © Rui Pedro Paiva, “From MIR to MER”, 2012 Techniques  MEVD • Segment-based classification [Panda and Paiva, 2011] - Divide audio signal into small segments - Classify each of them as before • Ad-hoc strategies (e.g., [Lu et al., 2006]) - Segmentation based on feature variations – Thresholds used and difficult to tune
  • 63. 63 © Rui Pedro Paiva, “From MIR to MER”, 2012 MER@CISUC  MOODetector project • Search
  • 64. 64 © Rui Pedro Paiva, “From MIR to MER”, 2012 MER@CISUC  MOODetector project • Generate playlist
  • 65. 65 © Rui Pedro Paiva, “From MIR to MER”, 2012 MER@CISUC  MOODetector project • Mood Tracking
  • 66. 66 © Rui Pedro Paiva, “From MIR to MER”, 2012 Limitations and Open Problems  Current Limitations • Agreement on a usable mood taxonomy - MIREX mood: only 5 moods, with semantic and acoustic overlap [Yang and Chen, 2012] • Lack of sizeable real-world datasets - Dimensional approaches – Yang only uses 194 songs - Categorical approaches – MIREX mood validates using 600 songs [MIREX, 2012] • Accuracy of current systems is too low for most real-world applications - MIREX best algorithm ~ 65% accuracy in a 5-class mood problem [MIREXresults, 2010]
  • 67. 67 © Rui Pedro Paiva, “From MIR to MER”, 2012 Limitations and Open Problems  Open problems • Semantic gap - Novel, semantically-relevant features necessary, able to capture the relevant musical attributes – Most important limitation, according to [Friberg, 2008] - Multi-modal approaches: combination of different information sources (audio, midi, lyrics) – Use MIDI resulting from automatic music transcription • Data sets - Real-world sizeable datasets for categorical, dimensional and MEVD approaches • MEVD - Other techniques, e.g., self-similarity techniques [Foote, 1999] - Quality datasets
  • 68. 68 © Rui Pedro Paiva, “From MIR to MER”, 2012 Limitations and Open Problems  Open problems • Multi-label classification (e.g., [Sanden and Zhang, 2011]) - Same song belonging to more than one mood category • Extraction of knowledge from computational models - E.g., neural-fuzzy approaches [Paiva and Dourado, 2004]
  • 69. 69 © Rui Pedro Paiva, “From MIR to MER”, 2012 REFERENCES - Suggested Initial Literature - Cited References
  • 70. 70 © Rui Pedro Paiva, “From MIR to MER”, 2012 Suggested Initial Literature  Yang Y.-H. and Chen H.-H. (2011). "Music Emotion Recognition," CRC Press, Boca Raton, Florida, USA, February 2011, 261 pages.  Friberg A. (2008). “Digital Audio Emotions - an Overview of Computer Analysis and Synthesis of Emotional Expression in Music”. In Proc. International Conference on Digital Audio Effects - DAFx-08.  Kim Y. E., Schmidt E. M., Migneco R., Morton B. G., Richardson P., Scott J., Speck J. A. and Turnbull D. (2010). “Music emotion recognition: a state of the art review”. International Society for Music Information Retrieval Conference – ISMIR’2010.  Lu L., Liu D. and Zhang H.-J. (2006). “Automatic Mood Detection and Tracking of Music Audio Signals”. IEEE Transactions on Audio, Speech and Language Processing, Vol. 14, No. 1, pp. 5-18.  Yang Y.-H., Lin Y.-C., Su Y.-F. and Chen H. H. (2008). “A Regression Approach to Music Emotion Recognition”. IEEE Transactions on Audio, Speech and Language Processing, Vol. 16, No. 2, pp. 448-457.  Yang Y.-H. and Chen H.-H. (2012). "Machine recognition of music emotion: a review“, To Appear in ACM Trans. Intelligent Systems and Technology.
  • 71. 71 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  Alghoniemy M. and Tewfik A. (2000). “Personalized Music Distribution”, In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing – ICASSP’2000.  Barlow H. and Morganstern S. (1948). A Dictionary of Musical Themes, New York, Crown.  Burad A. (2006). “Multimedia Databases”, Seminar Report, Computer Science and Engineering, Indian Institute of Technology.  Byrd D. (2002). “The History of ISMIR - A Short Happy Tale”, D-Lib Magazine, Vol. 8, No. 1, URL: http://www.dlib.org/dlib/november02/11inbrief.html#BYRD.  Byrd D. and Crawford T. (2002). “Problems of Music Information Retrieval in the Real World”, Information Systems and Management, Vol. 38, pp. 249-272  Cardoso L., Panda R. and Paiva R. P. (2011). “MOODetector: A Prototype Software Tool for Mood-based Playlist Generation”. Simpósio de Informática – INForum 2011.  Celma O. and Lamere P. (2007). “Music Recommendation Tutorial”, International Society for Music Information Retrieval Conference – ISMIR’2007, tutorial.  Dunn J. W. (2000). “Beyond VARIATIONS: Creating a Digital Music Library”, In International Society for Music Information Retrieval Conference – ISMIR’2000.
  • 72. 72 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  Farnsworth P. R. (1954). “A study of the Hevner adjective list”, J. Aesthetics and Art Criticism 13, 97–103.  Feng Y., Zhuang Y. and Pan Y. (2003). “Music information retrieval by detecting mood via computational media aesthetics”. In Proc. International Conference on Web Intelligence - IEEE/WIC.  Friberg A. (2008). “Digital Audio Emotions - an Overview of Computer Analysis and Synthesis of Emotional Expression in Music”. In Proc. International Conference on Digital Audio Effects - DAFx-08.  Foote J. (1999). “Visualizing music and audio using self-similarity”. In Proc. ACM Multimedia, pp. 77–84, Orlando, Florida, USA.  Futrelle J. and Downie J. S. (2003). “Interdisciplinary Research Issues in Music Information Retrieval: ISMIR 2000-2002”, Journal of New Music Research, Vol. 32, No. 2, pp. 121-131.  Gabrielsson A. (2002). Emotion perceived and emotion felt: Same or different? Musicae Scientiae,123–147. special issue.  Gabrielsson A. and Lindström E. (2001). “The influence of musical structure on emotional expression”. In P. N. Juslin, & J. A. Sloboda (Eds.), Music and emotion: Theory and Research New York: Oxford University Press, 2001, pp. 223-248.
  • 73. 73 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  Ghias A., Logan J., Chamberlin D. and Smith B. C. (1995). “Query by Humming: Musical Information Retrieval in an Audio Database”, In Proc. ACM Multimedia Conference – Multimedia’95.  Gomes A. R. (2005). “Festivais de Verão: Para Além da Música, Uma Experiência”, Revista XIS – Jornal Público, August 6, 2005, pp. 18-19 (in Portuguese).  Haitsma J. and Kalker T. (2002). “A highly robust audio-fingerprinting system”, International Society for Music Information Retrieval Conference – ISMIR’2002.  Hevner K. (1935). “Expression in music: A discussion of experimental studies and theories”, Psychological Review 48, 2, 186–204.  Huron D. (2000). “Perceptual and Cognitive Applications in Music Information Retrieval”, International Society for Music Information Retrieval Conference – ISMIR’2000.  Juslin P. N. (1997). “Perceived emotional expression in synthesized performances of a short melody: capturing the listener’s judgment policy,” Music. Sci., Vol. 1, No. 2, pp. 225–256.  Juslin P. N. and Laukka J. (2004). “Expression, Perception, and Induction of Musical Emotions: A Review and a Questionnaire Study of Everyday Listening”. Journal of New Music Research, Vol. 33, No. 3, pp. 217-38, 2004.
  • 74. 74 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  Kashino K., Nakadai K., Kinoshita T. and Tanaka H. (1995). “Organization of Hierarchical Perceptual Sounds: Music Scene Analysis with Autonomous Processing Modules and a Quantitative Information Integration Mechanism”, In Proc. International Joint Conference on Artificial Intelligence – IJCAI’95, pp. 158-164.  Kassler M. (1966). “Toward Musical Information Retrieval”, Perspectives of New Music, Vol. 4, No. 2, pp. 59-67.  Kim Y. E., Schmidt E. M., Migneco R., Morton B. G., Richardson P., Scott J., Speck J. A. and Turnbull D. (2010). “Music emotion recognition: a state of the art review”. International Society for Music Information Retrieval Conference – ISMIR’2010.  Lu L., Liu D. and Zhang H.-J. (2006). “Automatic Mood Detection and Tracking of Music Audio Signals”. IEEE Transactions on Audio, Speech and Language Processing, Vol. 14, No. 1, pp. 5-18.  Matesic B. C. and Cromartie F. (2002). “Effects music has on lap pace, heart rate, and perceived exertion rate during a 20-minute self-paced run”, The Sports Journal, Vol. 5. No. 1.  Meyers O. C. (2007). “A Mood-Based Classification and Exploration System”. MSc Thesis, School of Architecture and Planning, Massachusetts Institute of Technology.
  • 75. 75 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  MIREX (2012). “MIREX Home”, URL: http://www.music- ir.org/mirex/wiki/MIREX_HOME, available by March 1, 2012.  MIREXresults (2010). “MIREX 2010 Results”. URL: http://www.music- ir.org/mirex/wiki/2010:MIREX2010_Results, available by March 1, 2012.  Orio N. (2006). “Music Retrieval: A Tutorial and Review”, Foundations and Trends in Information Retrieval, Vol. 1, No. 1.  Paiva R. P. (2002). “Content-based Classification and Retrieval of Music: Overview and Research Trends”. Technical Report, Centre of Informatics and Systems of the University of Coimbra, Portugal.  Paiva R. P. and Dourado A. (2004). “Interpretability and Learning in Neuro-Fuzzy Systems”. Fuzzy Sets and Systems, Vol. 147 (1), pp. 17-38, Elsevier.  Paiva R. P. (2006). “Melody Detection in Polyphonic Audio”, PhD Thesis, University of Coimbra, Department of Informatics Engineering.  Pampalk E. (2001). Islands of Music: Analysis, Organization and Visualization of Music Archives, MSc Thesis, Department of Software Technology and Interactive Systems, Vienna University of Technology, Austria.  Panda R. and Paiva R. P. (2011). “Using Support Vector Machines for Automatic Mood Tracking in Audio Music”. Proceedings of the 130th Audio Engineering Society Convention – AES 130, London, UK.
  • 76. 76 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  Parker C. (2005). “Applications of Binary Classification and Adaptive Boosting to the Query-by-Humming Problem”, International Society for Music Information Retrieval Conference – ISMIR’2005.  Pauws S. and Wijdeven S. (2005). “User Evaluation of a New Interactive Playlist Generation Concept”, International Society for Music Information Retrieval Conference – ISMIR’2005.  Pfeiffer S., Fischer S. and Effelsberg W. (1996). “Automatic Audio Content Analysis”, In Proc. ACM Multimedia Conference – Multimedia’96, pp. 21-30.  Ryynänen M. (2008). “Automatic Transcription of Pitch Content in Music and Selected Applications”. PhD Thesis, Tampere University of Technology, Finland.  RTP (2009). “Cultura: Concertos de música ligeira geraram 30 milhões em 2007”. URL: http://ww1.rtp.pt/noticias/?article=385096&visual=26&tema=5/. Published in January 29, 2009, available by February 2, 2009.  Russell J. A. (1980). “A circumspect model of affect”, Journal of Psychology and Social Psychology, Vol. 39, No. 6, p. 1161.  Sanden C and Zhang J. Z. (2011). “An Empirical Study of Multi-Label Classifiers for Music Tag Annotation”, International Society for Music Information Retrieval Conference – ISMIR’2003.
  • 77. 77 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  Schubert E. (2003). “Update of the Hevner adjective checklist”, Perceptual and Motor Skills 96, 1117–1122.  Serra X. (2005). “Bridging the Music Semantic Gap”, talk given at the Information Day on “Audio-Visual Search Technologies” organized by the Commission services of DG Information Society of the EC in Brussels.  Sloboda J. A. and Juslin P. N. (2001). “Psychological perspectives on music and emotion”. In Music and Emotion: Theory and Research, P. N. Juslin and J. A. Sloboda, Eds. Oxford, University Press, New York.  Thayer R. E. (1989). “The Biopsychology of Mood and Arousal”, New York, Oxford University Press.  TechCrunch (2009). “iTunes Sells 6 Billion Songs, And Other Fun Stats From The Philnote”. URL: http://www.techcrunch.com/2009/01/06/itunes-sells-6- billion-songs-and-other-fun-stats-from-the-philnote/, published in January 6, 2009, available by February 2, 2009.  TechWhack (2005). “Apple iTunes Touches an Impressive 250 Million Sales Figure”. URL: http://news.techwhack.com/690/apple-itunes-250-million/, published in January 25, 2005, available by February 2, 2009.
  • 78. 78 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  TMN (2009). “TMN lança inovador serviço de identificação, pesquisa e aquisição de música através do telemóvel”. URL: http://www.tmn.pt/portal/site/tmn/menuitem.db67f528a6dbaa5ac8a71c10a51056 a0/?vgnextoid=d072fbb324238110VgnVCM1000005401650aRCRD&vgnextcha nnel=15b5fa2eb8b37110VgnVCM1000005401650aRCRD&vgnextfmt=default2, published in February 18, 2008, available by February 2, 2009.  Tzanetakis G. (2002). Manipulation, Analysis and Retrieval Systems for Audio Signals, PhD Thesis, Department of Computer Science, Princeton University, USA.  Wang A. (2003). “An Industrial-Strength Audio Search Algorithm”, International Society for Music Information Retrieval Conference – ISMIR’2003, presentation.  Welsh M., Borisov N., Hill J., von Behren R. and Woo A. (1999). Querying Large Collections of Music for Similarity, Technical Report, Computer Science Division, University of California, Berkeley, USA.  Yang D. and Lee W. (2004). “Disambiguating music emotion using software agents”, International Society for Music Information Retrieval Conference – ISMIR’2004  Yang Y.-H. and Chen H.-H. (2012). "Machine recognition of music emotion: a review“, To Appear in ACM Trans. Intelligent Systems and Technology.
  • 79. 79 © Rui Pedro Paiva, “From MIR to MER”, 2012 Cited References  Yang Y.-H., Lin Y.-C., Su Y.-F and Chen H.-H. (2007), "Music emotion classification: A regression approach“, IEEE Int. Conf. Multimedia and Expo – ICME’2007.