A Two-Step Approach for Classifying Music Genre on the Strength of AHP Weighted Musical Features
Abstract
:1. Introduction
2. The Design Concepts
2.1. Entropy Analysis Method
2.2. Exponential Distribution
2.3. Analytic Hierarchy Process
2.4. Machine Learning
2.5. Design Highlights
3. The Proposed TSMGC Approach
3.1. The Operation of the TSMGC
3.2. Musical Content Retrieval
3.3. Musical Content Analysis
- (1)
- Pitch. The pitch of each musical note can be generated as a numerical value according to the defined numerical value of a musical alphabet (as shown in Table 2). Any two notes that have the same pitch but different musical alphabets are seen as enharmonic equivalents. For instance, the notes as alphabet values C and B♯ are enharmonic notes, thus the transformed numerical values of these two notes are both equal to 1. Based on this rule, a total of 78 pitches can be referred in a music composition. The value of each feature depends on its frequency. Furthermore, since the highest and lowest pitch values are also considered, 80 values in terms of pitch-oriented features can be collected.
- (2)
- Musical Interval. An n-gram segmentation approach is applied to transform a note into a gram. Assuming the number of musical composition (N) equals 2, Equation (8) is used to measure the value of the musical interval between two successive notes. According to Wikipedia, the traditional musical theory has defined basic musical intervals, but the influence by semitone and enharmonic notes should also be considered. The use of a semitone may form an additional musical interval [29]; on the other hand, enharmonic notes may conduct an equivalent interval. For instance, the interval from C to D♯ is an augmented second (A2), while that from C to E♭ is a minor third (m3). Since D♯ and E♭ are enharmonic notes, the musical intervals of A2 and m3 are equivalent. Furthermore, the transformation from a compound interval into a simple interval can significantly simplify the identification of musical interval. Based on the mentioned rules, the numerical value of every musical interval can thus be calculated. For example, on a semitone basis, the interval from C (i.e., the pitch value is 1) to F♯ (i.e., the pitch value is 7) is an augmented fourth (A4), the numerical value of this interval is set to 6 (i.e., 7 minus 1 leaves 6). Table 2 shows the numerical value of the defined musical intervals. The value of each musical interval feature is the number of times it appears in the musical composition. As a result, 12 musical interval-oriented features are determined.
- (3)
- Chord. The set of pitches from a musical composition can be used to determine its tonality type by comparison with tonality characteristics. Each music has a different set of chords according to its tonality; there are seven basic chords in a tonality. Take the 12 Variations on “Ah, vous dirai-je, Maman” by Mozart as an example; since it contains no rising-falling tone (as the symbols shown in Figure 2), the tonality of this musical composition is perceived as a major scale based on C. In the first measure, only chord I is included. The second measure contains chords IV and I, the third measure contains chords vii and I, and the fourth measure contains chords V and I. The number of times each chord appears in a musical composition is recorded as the value of the respective chord-oriented features. In this part, 7 chord-oriented features can be generated.
- (4)
- Rhythm. According to the Oxford English Dictionary II, a rhythm generally means a "movement marked by the regulated succession of strong and weak elements, or of opposite or different conditions." In this study, 3 rhythm-related features are defined to record the lowest note value (i.e., the shortest duration), the highest note value (i.e., the longest duration), and the average note value of the musical composition.
- (5)
- Entropy of the pitch. The ApEn and SampEn methods are adopted in this analysis. First, the pitches of a musical composition are transformed into a time series. The data set of this time series is the core material for entropy computation. If 3 is chosen as the threshold and 2 as the number of dimensions, then and can be calculated as the pitch entropies.
3.4. Feature Extraction
3.5. Two-Step Classification
4. Experiments and Analyses
4.1. Tools for Experiment Implementation
4.2. Samples and the Classes
4.3. The Evaluation Factor
4.4. The Experiment Execution
4.5. Results
4.6. Discussions
5. Conclusions and Future Work
Author Contributions
Funding
Conflicts of Interest
References
- Lee, J.H.; Downie, J.S. Survey of music information needs, uses, and seeking behaviours: Preliminary findings. In Proceedings of the 5th International Symposium on Music Information Retrieval, Barcelona, Spain, 10–14 October 2004; pp. 1–4. [Google Scholar]
- Lo, C.C.; Kuo, T.H.; Kung, H.Y.; Chen, C.H.; Lin, C.H. Personalization music therapy service recommendation system using information retrieval technology. J. Inf. Manag. 2014, 21, 1–24. [Google Scholar]
- Corrêa, D.C.; Rodrigues, F.A. A survey on symbolic data-based music genre classification. Expert Syst. Appl. 2016, 60, 190–210. [Google Scholar] [CrossRef]
- Conklin, D. Multiple viewpoint systems for music classification. J. New Music Res. 2014, 42, 19–26. [Google Scholar] [CrossRef]
- Zhong, J.; Cheng, Y.F. Research on music mood classification integrating audio and lyrics. Comput. Eng. 2012, 38, 144–146. [Google Scholar]
- Hu, X.; Downie, J.S.; Ehmann, A.F. Lyric text mining in music mood classification. In Proceedings of the 10th International Society for Music Information Retrieval Conference, Kobe, Japan, 26–30 October 2009. [Google Scholar]
- Zeng, Z.; Pantic, M.; Roisman, G.I.; Huang, T.S. A survey of affect recognition methods: Audio, visual, and spontaneous expressions. IEEE Trans. Pattern Anal. Ma-Chine Intell. 2009, 31, 39–58. [Google Scholar] [CrossRef]
- Kermanidis, K.L.; Karydis, I.; Koursoumis, A.; Talvis, K. Combining language modeling and LSA on Greek song “words” for mood classification. Int. J. Artif. Intell. Tools 2014, 23, 17. [Google Scholar] [CrossRef]
- Silla, C.N., Jr.; Freitas, A.A. Novel top-down approaches for hierarchical classification and their application to automatic music genre classification. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, San Antonio, TX, USA, 11–14 October 2009; pp. 3499–3504. [Google Scholar]
- Scaringella, N.; Zoia, G.; Mlynek, D. Automatic genre classification of music content—A survey. IEEE Signal Process. Mag. 2006, 23, 133–141. [Google Scholar] [CrossRef]
- Fu, Z.; Lu, G.; Ting, K.M.; Zhang, D. A survey of audio-based music classification and annotation. IEEE Trans. Mult. 2011, 13, 303–319. [Google Scholar] [CrossRef]
- Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification; John Wiley & Sons Inc.: New York, NY, USA, 2001. [Google Scholar]
- Da Fontoura Costa, L.; César, R.M., Jr. Shape Analysis and Classification; CRC Press: Boca Raton, FL, USA, 2001. [Google Scholar]
- Lykartsis, A.; Wu, C.W.; Lerch, A. Beat histogram features from NMF-based novelty functions for music classification. In Proceedings of the International Conference on Music Information Retrieval, Malaga, Spain, 26–30 October 2015. [Google Scholar]
- Lin, C.R.; Liu, N.H.; Wu, Y.H.; Chen, A.L.P. Music classi-fication using significant repeating patterns. Lect. Notes Comput. Sci. 2004, 2973, 506–518. [Google Scholar]
- Chew, E.; Volk, A.; Lee, C.Y. Dance music classification using inner metric analysis. Oper. Res./Comput. Sci. Interfaces Ser. 2005, 29, 355–370. [Google Scholar]
- Thomas, L.S. Music helps heal mind, body, and spirit. Nurs. Crit. Care 2014, 9, 28–31. [Google Scholar] [CrossRef] [Green Version]
- Drossinou-Korea, M.; Fragkouli, A. Emotional readiness and music therapeutic activities. J. Res. Spec. Educ. Needs 2016, 16, 440–444. [Google Scholar] [CrossRef]
- Karydis, I. Symbolic music genre classification based on note pitch and duration. In Advances in Databases and Information Systems (ADBIS 2006); Lecture Notes on Computer Science; Manolopoulos, Y., Pokorný, J., Sellis, T.K., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4152, pp. 329–338. [Google Scholar]
- Saaty, T.L. A scaling method for priorities in hierarchical structures. J. Math. Psychol. 1977, 15, 234–281. [Google Scholar] [CrossRef]
- Pincus, S.M.; Gladstone, I.M.; Ehrenkranz, R.A. A regularity statistic for medical data analysis. J. Clin. Monit. Comput. 1991, 7, 335–345. [Google Scholar] [CrossRef]
- Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed]
- Bishop, C. Pattern Recognition and Machine Learning; Springer: Berlin, Germany, 2006; ISBN 0-387-31073-8. [Google Scholar]
- Saaty, T.L. How to make a decision: The analytic hierarchy process. Eur. J. Oper. Res. 1990, 48, 9–26. [Google Scholar] [CrossRef]
- Amancio, D.R.; Comin, C.H.; Casanova, D.; Travieso, G.; Bruno, O.M.; Rodrigues, F.A.; Costa, L.D.F. A systematic comparison of supervised classifiers. PLoS ONE 2014, 9, e94137. [Google Scholar] [CrossRef]
- Mandel, M.I.; Ellis, D.P.W. Song-level features and support vector machines for music classification. In Proceedings of the 6th International Conference on Music Information Retrieval: ISMIR-05, London, UK, 11–15 September 2005; pp. 594–599. [Google Scholar]
- Lee, H.; Largman, Y.; Pham, P.; Ng, A.Y. Unsupervised feature learning for audio classification using convolutional deep belief networks. Adv. Neural Inf. Process. Syst. 2009, 22, 1096–1104. [Google Scholar]
- Brown, A.R. Making Music with Java; Lulu Press, Inc.: Morrisville, NC, USA, 2009. [Google Scholar]
- Aruffo, C.; Goldstone, R.L.; Earn, D.J.D. Absolute judgment of musical interval width. Music Percept. Interdiscip. J. 2014, 32, 186–200. [Google Scholar] [CrossRef]
- Han, J.; Kamber, M.; Pei, J. Data Mining: Concepts and Techniques, 3rd ed.; Morgan Kaufmann Publishers: Chusetts, MA, USA, 2011. [Google Scholar]
- Chen, C.H.; Lin, H.F.; Chang, H.C.; Ho, P.H.; Lo, C.C. An analytical framework of a deployment strategy for cloud computing services: A case study of academic websites. Math. Probl. Eng. 2013, 2013, 14. [Google Scholar] [CrossRef]
- Shan, M.K.; Kuo, F.F. Music style mining and classification by melody. IEICE Trans. Inf. Syst. 2003, E86-D, 655–659. [Google Scholar]
Parameter | Definition |
---|---|
N | The number of musical composition |
n | The total number of classes |
nm | The number of main classes |
nsi | The number of sub-classes belonging to the ith main class, where |
m | The number of features |
nns | The number of musical notes in the sth musical composition |
The pitch value of the ith musical note of the sth musical composition, where | |
The musical interval between the ith musical note and the (i+1)th musical note of the sth musical composition, where | |
The set of pitch of the sth musical composition | |
The set of musical intervals of the sth musical composition | |
The set of chords of the sth musical composition | |
The set of rhythms of the sth musical composition | |
The set of pitch entropy in the sth musical composition | |
The value of the ith feature of the sth musical composition, where | |
The weight of the ith feature of the sth musical composition, where |
Musical Alphabet | Enharmonic Note | Numerical Value |
---|---|---|
C | B♯ | 1 |
C♯ | D♭ | 2 |
D | - | 3 |
D♯ | E♭ | 4 |
E | F♭ | 5 |
F | E♯ | 6 |
F♯ | G♭ | 7 |
G | - | 8 |
G♯ | A♭ | 9 |
A | - | 10 |
A♯ | B♭ | 11 |
B | C♭ | 12 |
Main-Class | Sub-Class | Sample Size of Sub-Class | Sample Size of Main-Class | Sample Size in Total |
---|---|---|---|---|
Classical | Medieval music | 11 | 51 | 141 |
Baroque music | 10 | |||
Classical era music | 10 | |||
Romantic music | 11 | |||
Modern music | 9 | |||
Popular | Pop 1960s music | 12 | 48 | |
Pop 1970s music | 10 | |||
Pop 1980s music | 7 | |||
Pop 1990s music | 8 | |||
Pop 2000s music | 11 | |||
Rock | 1960s music | 6 | 42 | |
Classic rock music | 10 | |||
Hard rock music | 8 | |||
Psychedelic rock music | 10 | |||
2000s music | 8 |
CI | SC | SN | LN | MM | FW | FE | TS | AC |
1 | Main | 141 | 3 | kNN | N | Y | N | 77.30% |
2 | N | 75.89% | ||||||
3 | Y | Y | 80.14% | |||||
4 | N | 85.82% | ||||||
5 | ANN | N | Y | 85.11% | ||||
6 | N | 78.72% | ||||||
7 | Y | Y | 87.23% | |||||
8 | N | 84.40% | ||||||
9 | Classical | 51 | 5 | kNN | N | Y | 35.29% | |
10 | N | 45.10% | ||||||
11 | Y | Y | 35.29% | |||||
12 | N | 27.45% | ||||||
13 | ANN | N | Y | 52.94% | ||||
14 | N | 23.53% | ||||||
15 | Y | Y | 78.43% | |||||
16 | N | 49.02% | ||||||
17 | Popular | 48 | kNN | N | Y | 14.58% | ||
18 | N | 14.58% | ||||||
19 | Y | Y | 14.58% | |||||
20 | N | 22.92% | ||||||
21 | ANN | N | Y | 35.42% | ||||
22 | N | 54.17% | ||||||
23 | Y | Y | 35.42% | |||||
24 | N | 68.75% | ||||||
25 | Rock | 42 | kNN | N | Y | 30.95% | ||
26 | N | 21.43% | ||||||
27 | Y | Y | 30.95% | |||||
28 | N | 26.19% | ||||||
29 | ANN | N | Y | 38.1% | ||||
30 | N | 42.86% | ||||||
31 | Y | Y | 42.86% | |||||
32 | N | 40.48% | ||||||
33 | All-mixed | 141 | 15 | kNN | N | Y | 14.18% | |
34 | N | 14.89% | ||||||
35 | Y | Y | 12.77% | |||||
36 | N | 10.64% | ||||||
37 | ANN | N | Y | 17.73% | ||||
38 | N | 13.48% | ||||||
39 | Y | Y | 19.15% | |||||
40 | N | 23.40% | ||||||
41 | kNN | N | Y | Y | 24.30% | |||
42 | N | 28.48% | ||||||
43 | Y | Y | 21.73% | |||||
44 | N | 27.91% | ||||||
45 | ANN | N | Y | 43.83% | ||||
46 | N | 35.10% | ||||||
47 | Y | Y | 57.19% | |||||
48 | N | 44.52% |
Feature Extraction Method | Number of Feature Values | Main-Class | Subclass of Classical Music | Subclass of Popular Music | Subclass of Rock Music |
---|---|---|---|---|---|
Musical interval only | 12 | 59.57% | 21.57% | 22.92% | 19.50% |
Pitch only | 50 | 74.47% | 25.49% | 20.83% | 19.05% |
RMH | 104 | 85.11% | 52.94% | 35.42% | 38.10% |
RMH and AHP | 104 | 87.23% | 78.43% | 35.42% | 42.86% |
Feature Extraction Method | One-Step Classification | Two-Step Classification |
---|---|---|
Musical interval only | 7% | 24% |
Pitch only | 15% | 20% |
RMH | 18% | 44% |
RMH and AHP | 19% | 57% |
© 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chen, Y.-T.; Chen, C.-H.; Wu, S.; Lo, C.-C. A Two-Step Approach for Classifying Music Genre on the Strength of AHP Weighted Musical Features. Mathematics 2019, 7, 19. https://doi.org/10.3390/math7010019
Chen Y-T, Chen C-H, Wu S, Lo C-C. A Two-Step Approach for Classifying Music Genre on the Strength of AHP Weighted Musical Features. Mathematics. 2019; 7(1):19. https://doi.org/10.3390/math7010019
Chicago/Turabian StyleChen, Yu-Tso, Chi-Hua Chen, Szu Wu, and Chi-Chun Lo. 2019. "A Two-Step Approach for Classifying Music Genre on the Strength of AHP Weighted Musical Features" Mathematics 7, no. 1: 19. https://doi.org/10.3390/math7010019
APA StyleChen, Y.-T., Chen, C.-H., Wu, S., & Lo, C.-C. (2019). A Two-Step Approach for Classifying Music Genre on the Strength of AHP Weighted Musical Features. Mathematics, 7(1), 19. https://doi.org/10.3390/math7010019