Identification of Greek Orthodox Church Chants Using Fuzzy Entropy
Abstract
:1. Introduction
1.1. Music Identification
1.2. Related Works
1.3. Motivation
- The measure of entropy has been swapped with that of fuzzy entropy, which has been reported in the literature to be more robust in comparing signals.
- The frequency domain signal is not segmented in frequency bins, but rather its entropy as a whole is computed. This results in a less detailed but much shorter vector entropy measure, as opposed to a matrix entropy feature (entropygram).
- The comparison is performed on the entropy vector rather than on a binary fingerprint.
- Following [31], the correlation coefficient is calculated after applying DTW to the entropy vectors, which significantly improves performance.
2. Characterization of Byzantine Chants Using Fuzzy Entropy
2.1. The Dataset
2.2. Signal Preprocessing
- For dual-channel signals, an averaged single-channel audio signal is computed.
- Following [17,32], a pre-emphasis filter is applied to the signal, which can emphasize human voice patterns. For a signal , where k denotes its samples, the filter’s function is given by
2.3. Computation of Fuzzy Entropy
- Construct the vectors
- The distance between two vectors and is computed, which is the maximum absolute difference of the following two vectors:
- A fuzzy membership function is applied to measure the similarity degree between two vectors and . Here, the exponential function is considered, given by
- The following function is computed.
- The previous steps are repeated for vectors , in order to compute
- The fuzzy entropy is finally defined as
2.4. Extraction of Fuzzy Entropy Measure
- The time series is broken down into overlapping segments of duration . The overlap between consecutive segments is taken as 50%.
- For each segmented part , a Hanning window is applied, and the signal is then transformed into the frequency domain, using an FFT transform. The resulting signal is denoted as , and it is a complex vector.
- The fuzzy entropy of magnitude is calculated for each segment. Since is symmetric, only its first half is considered for this. The resulting entropy values are added to a vector. This results in the fuzzy entropy vector of the recording, which represents the changes between the FuzzEn values across all segments of the track.
- Finally, the entropy vector is normalized so that its values lie in the interval . Normalization of a vector x is performed as
2.5. Track Comparison
2.5.1. Comparison of Whole Tracks
- Let and be two different recordings in the dataset, and and their corresponding fuzzy entropy vectors. Let , , and assume, without loss of generality, that , so the second recording is longer.
- For the longer vector , segments of length , that is, , are considered. Here, i denotes the iteration step and z denotes the sliding window jump. Its default value is , but higher values can be considered to improve the execution speed. Here, we choose . This corresponds to a 1.05 s jump. The notation denotes the elements from the position i to the position .
- A DTW is applied on all pairs of and to stretch the two vectors into the new vectors and , whose Euclidean distance is the smallest. Note that, in general, the stretched signals have a length longer than , so some entries may be repeated.
- The vector is compared with each segment . For this, the correlation coefficients are computed. The maximum value among these is taken as the highest similarity index between the two tracks and .
2.5.2. Comparison of Segments
- Let and be two different recordings in the dataset, and and their corresponding fuzzy entropy vectors. Let and .
- For both vectors and , segments of 100 values are considered, that is, , , . Here, i and j denote the iteration step and z denotes the jump of the sliding window. Here, we choose .
- A DTW is applied to all pairs of and to stretch the two vectors into the new vectors , and , whose Euclidean distance is the smallest. Note that, in general, the stretched signals have lengths longer than 100, so some entries may be repeated.
- Each entropy vector is compared to each . For this, the correlation coefficients are calculated. The maximum value among these is taken as the highest similarity index between the two tracks and .
3. Comparison Results
- For a track , the similarity measures (maximum correlation coefficient) that were computed between all songs are sorted in descending order.
- If the top three most similar entries include a performance of the same chant, the identification is considered successful. Self-matches are, of course, excluded.
Execution Time
4. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Wang, A. An industrial strength audio search algorithm. In Proceedings of the Ismir, Washington, DC, USA, 26–30 October 2003; Volume 2003, pp. 7–13. [Google Scholar]
- Son, W.; Cho, H.T.; Yoon, K.; Lee, S.P. Sub-fingerprint masking for a robust audio fingerprinting system in a real-noise environment for portable consumer devices. IEEE Trans. Consum. Electron. 2010, 56, 156–160. [Google Scholar] [CrossRef]
- Borkar, N.; Patre, S.; Khalsa, R.S.; Kawale, R.; Chakurkar, P. Music plagiarism detection using audio fingerprinting and segment matching. In Proceedings of the IEEE 2021 Smart Technologies, Communication and Robotics (STCR), Sathyamangalam, India, 9–10 October 2021; pp. 1–4. [Google Scholar]
- Karasavvidis, K.; Kampelopoulos, D.; Moysis, L.; Boursianis, A.D.; Nikolaidis, S.; Sarigiannidis, P.; Goudos, S.K. Recognition of Greek Orthodox Hymns Using Audio Fingerprint Techniques. In Proceedings of the IEEE 2023 8th South-East Europe Design Automation, Computer Engineering, Computer Networks and Social Media Conference (SEEDA-CECNSM), Piraeus, Greece,, 10–12 November 2023; pp. 1–6. [Google Scholar]
- Kampelopoulos, D.; Moysis, L.; Karasavvidis, K.; Boursianis, A.D.; Goudos, S.K.; Nikolaidis, S. Byzantine Hymn Recognition with Audio Fingerprints Resistant to Noise, Tempo and Scale Changes. In Proceedings of the IEEE 2024 13th International Conference on Modern Circuits and Systems Technologies (MOCAST), Sofia, Bulgaria, 26–28 June 2024; pp. 1–4. [Google Scholar]
- Su, X.; Nongpong, K. Audio Fingerprinting Based Music Recommendation Algorithm. In Proceedings of the 2023 6th International Conference on Algorithms, Computing and Artificial Intelligence, Sanya, China, 22–24 December 2023; pp. 52–56. [Google Scholar]
- Kishor, K.; Venkatesh, S.; Koolagudi, S.G. Audio fingerprinting system to detect and match audio recordings. In Proceedings of the International Conference on Pattern Recognition and Machine Intelligence, Kolkata, India, 12–15 December 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 683–690. [Google Scholar]
- Htun, M.T.; Oo, T.T. Broadcast Monitoring System using MFCC-based Audio Fingerprinting. In Proceedings of the 2023 IEEE Conference on Computer Applications (ICCA), Cairo, Egypt, 28–30 November 2023; pp. 243–247. [Google Scholar]
- Kritopoulou, P.; Stergiaki, A.; Kokkinidis, K. Optimizing human computer interaction for byzantine music learning: Comparing HMMs with RDFs. In Proceedings of the IEEE 2020 9th International Conference on Modern Circuits and Systems Technologies (MOCAST), Bremen, Germany, 7–9 September 2020; pp. 1–4. [Google Scholar]
- Fang, J.T.; Day, C.T.; Chang, P.C. Deep feature learning for cover song identification. Multimed. Tools Appl. 2017, 76, 23225–23238. [Google Scholar] [CrossRef]
- Jin, Y.; Cai, W.; Chen, L.; Zhang, Y.; Doherty, G.; Jiang, T. Exploring the Design of Generative AI in Supporting Music-based Reminiscence for Older Adults. In Proceedings of the CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 11–16 May 2024; pp. 1–17. [Google Scholar]
- Tsallis, C. Entropy. Encyclopedia 2022, 2, 264–300. [Google Scholar] [CrossRef]
- Camarena-Ibarrola, A.; Chávez, E. Identifying music by performances using an entropy based audio-fingerprint. In Proceedings of the Mexican International Conference on Artificial Intelligence (MICAI), Apizaco, Mexico, 13–17 November 2006. [Google Scholar]
- Yin, C.; Li, W.; Luo, Y.; Tseng, L.C. Robust online music identification using spectral entropy in the compressed domain. In Proceedings of the 2014 IEEE Wireless Communications and Networking Conference Workshops (WCNCW), Istanbul, Turkey, 6–9 April 2014; pp. 128–133. [Google Scholar]
- Li, W.; Liu, Y.; Xue, X. Robust audio identification for MP3 popular music. In Proceedings of the 33rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Geneva, Switzerland, 18–23 July 2010; pp. 627–634. [Google Scholar]
- Camarena-Ibarrola, A.; Chávez, E. Robust Audio-Fingerprinting With Spectral Entropy Signatures; Universidad Michoacana de San Nicolás de Hidalgo: Morelia, Mexico, 2007. [Google Scholar]
- Camarena-Ibarrola, A.; Figueroa, K.; Tejeda-Villela, H. Entropy per chroma for Cover song identification. In Proceedings of the 2016 IEEE International Autumn Meeting on Power, Electronics and Computing (ROPEC), Ixtapa, Mexico, 9–11 November 2016; pp. 1–6. [Google Scholar]
- Ibarrola, A.C.; Chávez, E. A robust entropy-based audio-fingerprint. In Proceedings of the 2006 IEEE International Conference on Multimedia and Expo, Toronto, ON, Canada, 9–12 July 2006; pp. 1729–1732. [Google Scholar]
- Camarena-Ibarrola, A.; Chávez, E.; Tellez, E.S. Robust radio broadcast monitoring using a multi-band spectral entropy signature. In Proceedings of the Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications: 14th Iberoamerican Conference on Pattern Recognition, CIARP 2009, Guadalajara, Jalisco, Mexico, 15–18 November 2009; Proceedings 14. Springer: Berlin/Heidelberg, Germany, 2009; pp. 587–594. [Google Scholar]
- Chen, W.; Wang, Z.; Xie, H.; Yu, W. Characterization of surface EMG signal based on fuzzy entropy. IEEE Trans. Neural Syst. Rehabil. Eng. 2007, 15, 266–272. [Google Scholar] [CrossRef] [PubMed]
- Zadeh, L. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef]
- Zimmermann, H.J. Fuzzy Set Theory—And Its Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
- Azami, H.; Li, P.; Arnold, S.E.; Escudero, J.; Humeau-Heurtier, A. Fuzzy entropy metrics for the analysis of biomedical signals: Assessment and comparison. IEEE Access 2019, 7, 104833–104847. [Google Scholar] [CrossRef]
- Xiang, J.; Li, C.; Li, H.; Cao, R.; Wang, B.; Han, X.; Chen, J. The detection of epileptic seizure signals based on fuzzy entropy. J. Neurosci. Methods 2015, 243, 18–25. [Google Scholar] [CrossRef] [PubMed]
- Li, P.; Karmakar, C.; Yearwood, J.; Venkatesh, S.; Palaniswami, M.; Liu, C. Detection of epileptic seizure based on entropy analysis of short-term EEG. PLoS ONE 2018, 13, e0193691. [Google Scholar] [CrossRef]
- Chen, W.; Zhuang, J.; Yu, W.; Wang, Z. Measuring complexity using FuzzyEn, ApEn, and SampEn. Med Eng. Phys. 2009, 31, 61–68. [Google Scholar] [CrossRef] [PubMed]
- Xie, H.B.; Chen, W.T.; He, W.X.; Liu, H. Complexity analysis of the biomedical signal using fuzzy entropy measurement. Appl. Soft Comput. 2011, 11, 2871–2879. [Google Scholar] [CrossRef]
- Dong, C.; Rajagopal, K.; He, S.; Jafari, S.; Sun, K. Chaotification of Sine-series maps based on the internal perturbation model. Results Phys. 2021, 31, 105010. [Google Scholar] [CrossRef]
- Johny Elton, R.; Vasuki, P.; Mohanalin, J. Voice activity detection using fuzzy entropy and support vector machine. Entropy 2016, 18, 298. [Google Scholar] [CrossRef]
- Baldini, G.; Amerini, I. An evaluation of entropy measures for microphone identification. Entropy 2020, 22, 1235. [Google Scholar] [CrossRef] [PubMed]
- Pandria, N.; Kugiumtzis, D. Testing the correlation of time series using dynamic time warping. In Proceedings of the 27th Panhellenic Conference of Statistics, Thessaloniki, Greece, 13–15 June 2014. [Google Scholar]
- Ramírez-Hernández, J.I.; Manzo-Martínez, A.; Gaxiola, F.; González-Gurrola, L.C.; Álvarez-Oliva, V.C.; López-Santillán, R. A Comparison Between MFCC and MSE Features for Text-Independent Speaker Recognition Using Machine Learning Algorithms. In Fuzzy Logic and Neural Networks for Hybrid Intelligent System Design; Springer: Berlin/Heidelberg, Germany, 2023; pp. 123–140. [Google Scholar]
- Laney, R.; Samuels, R.; Capulet, E. Cross entropy as a measure of musical contrast. In Proceedings of the Mathematics and Computation in Music: 5th International Conference, MCM 2015, London, UK, 22–25 June 2015; Proceedings 5. Springer: Berlin/Heidelberg, Germany, 2015; pp. 193–198. [Google Scholar]
- Rohila, A.; Sharma, A. Phase entropy: A new complexity measure for heart rate variability. Physiol. Meas. 2019, 40, 105006. [Google Scholar] [CrossRef] [PubMed]
- Zhang, T.; Chen, W.; Li, M. Fuzzy distribution entropy and its application in automated seizure detection technique. Biomed. Signal Process. Control 2018, 39, 360–377. [Google Scholar] [CrossRef]
- Seetharaman, P.; Rafii, Z. Cover song identification with 2D Fourier transform sequences. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 616–620. [Google Scholar]
FuzzEn | En | ||||
---|---|---|---|---|---|
Method | Metric | Top 3 | Top 1 | Top 3 | Top 1 |
no-DTW | - | 56.82% | 51.25% | 63.23% | 54.03% |
DTW | Euclidean | 95.54% | 90.52% | 91.92% | 87.18% |
DTW | Absolute | 95.54% | 90.52% | 91.92% | 87.18% |
DTW | Squared | 96.10% | 93.03% | 92.20% | 86.90% |
DTW | Symmkl | 93.59% | 88.30% | 90.52% | 86.07% |
DTW-lim100 | Euclidean | 95.82% | 90.25% | 91.64% | 87.18% |
DTW-lim100 | Absolute | 95.82% | 90.25% | 91.64% | 87.18% |
DTW-lim100 | Squared | 96.37% | 92.75% | 91.92% | 87.18% |
DTW-lim100 | Symmkl | 94.98% | 89.97% | 91.64% | 87.46% |
DTW-lim80 | Euclidean | 95.82% | 90.25% | 91.36% | 86.62% |
DTW-lim80 | Absolute | 95.82% | 90.25% | 91.36% | 86.62% |
DTW-lim80 | Squared | 96.37% | 92.47% | 91.36% | 86.90% |
DTW-lim80 | Symmkl | 94.98% | 89.69% | 91.36% | 87.18% |
DTW-lim60 | Euclidean | 94.98% | 89.13% | 91.36% | 84.95% |
DTW-lim60 | Absolute | 94.98% | 89.13% | 91.36% | 84.95% |
DTW-lim60 | Squared | 95.54% | 91.08% | 90.52% | 84.95% |
DTW-lim60 | Symmkl | 95.26% | 89.97% | 92.20% | 85.23% |
DTW-lim40 | Euclidean | 93.03% | 86.62% | 91.08% | 84.67% |
DTW-lim40 | Absolute | 93.03% | 86.62% | 91.08% | 84.67% |
DTW-lim40 | Squared | 93.87% | 88.02% | 90.25% | 83.00% |
DTW-lim40 | Symmkl | 94.70% | 87.46% | 89.97% | 82.17% |
DTW-lim20 | Euclidean | 86.62% | 80.77% | 85.79% | 75.48% |
DTW-lim20 | Absolute | 86.62% | 80.77% | 85.79% | 75.48% |
DTW-lim20 | Squared | 86.90% | 81.05% | 84.12% | 73.81% |
DTW-lim20 | Symmkl | 88.85% | 81.89% | 84.67% | 74.09% |
FuzzEn | En | ||||
---|---|---|---|---|---|
Method | Metric | Top 3 | Top 1 | Top 3 | Top 1 |
no-DTW | - | 85.51% | 73.53% | 80.50% | 69.08% |
DTW | Euclidean | 92.47% | 85.23% | 88.57% | 77.99% |
DTW | Absolute | 92.47% | 85.23% | 88.57% | 77.99% |
DTW | Squared | 93.59% | 86.07% | 87.46% | 79.10% |
DTW | Symmkl | 93.59% | 86.35% | 87.74% | 79.66% |
DTW-lim40 | Euclidean | 92.47% | 85.23% | 88.57% | 78.83% |
DTW-lim40 | Absolute | 92.47% | 85.23% | 88.57% | 78.83% |
DTW-lim40 | Squared | 93.59% | 86.35% | 87.74% | 79.94% |
DTW-lim40 | Symmkl | 93.59% | 86.07% | 88.30% | 81.05% |
DTW-lim20 | Euclidean | 93.31% | 86.07% | 90.25% | 80.77% |
DTW-lim20 | Absolute | 93.31% | 86.07% | 90.25% | 80.77% |
DTW-lim20 | Squared | 93.59% | 86.35% | 89.41% | 81.33% |
DTW-lim20 | Symmkl | 94.42% | 86.07% | 88.57% | 83.00% |
DTW-lim10 | Euclidean | 96.37% | 89.97% | 92.75% | 85.79% |
DTW-lim10 | Absolute | 96.37% | 89.97% | 92.75% | 85.79% |
DTW-lim10 | Squared | 96.10% | 89.69% | 93.31% | 86.35% |
DTW-lim10 | Symmkl | 95.26% | 89.69% | 93.03% | 87.46% |
DTW-lim05 | Euclidean | 94.70% | 90.25% | 92.20% | 86.62% |
DTW-lim05 | Absolute | 94.70% | 90.25% | 92.20% | 86.62% |
DTW-lim05 | Squared | 94.98% | 89.69% | 91.92% | 87.74% |
DTW-lim05 | Symmkl | 94.70% | 88.57% | 93.03% | 87.46% |
DTW-lim02 | Euclidean | 92.75% | 86.35% | 92.20% | 81.33% |
DTW-lim02 | Absolute | 92.75% | 86.35% | 92.20% | 81.33% |
DTW-lim02 | Squared | 91.92% | 86.35% | 90.80% | 81.05% |
DTW-lim02 | Symmkl | 91.92% | 86.35% | 90.80% | 81.89% |
Method | Metric | Whole Track | Method | Metric | 5 s Interval |
---|---|---|---|---|---|
no-DTW | - | 15 | no-DTW | - | 120 |
DTW | Euclidean | 642 | DTW | Euclidean | 651 |
DTW | Absolute | 669 | DTW | Absolute | 650 |
DTW | Squared | 666 | DTW | Squared | 647 |
DTW | Symmkl | 2095 | DTW | Symmkl | 1372 |
DTW-lim100 | Euclidean | 380 | DTW-lim40 | Euclidean | 590 |
DTW-lim100 | Absolute | 393 | DTW-lim40 | Absolute | 603 |
DTW-lim100 | Squared | 395 | DTW-lim40 | Squared | 608 |
DTW-lim100 | Symmkl | 1003 | DTW-lim40 | Symmkl | 1150 |
DTW-lim80 | Euclidean | 354 | DTW-lim20 | Euclidean | 552 |
DTW-lim80 | Absolute | 364 | DTW-lim20 | Absolute | 557 |
DTW-lim80 | Squared | 380 | DTW-lim20 | Squared | 557 |
DTW-lim80 | Symmkl | 917 | DTW-lim20 | Symmkl | 883 |
DTW-lim60 | Euclidean | 326 | DTW-lim10 | Euclidean | 546 |
DTW-lim60 | Absolute | 335 | DTW-lim10 | Absolute | 513 |
DTW-lim60 | Squared | 351 | DTW-lim10 | Squared | 522 |
DTW-lim60 | Symmkl | 780 | DTW-lim10 | Symmkl | 706 |
DTW-lim40 | Euclidean | 299 | DTW-lim05 | Euclidean | 507 |
DTW-lim40 | Absolute | 329 | DTW-lim05 | Absolute | 497 |
DTW-lim40 | Squared | 318 | DTW-lim05 | Squared | 511 |
DTW-lim40 | Symmkl | 585 | DTW-lim05 | Symmkl | 611 |
DTW-lim20 | Euclidean | 262 | DTW-lim02 | Euclidean | 496 |
DTW-lim20 | Absolute | 262 | DTW-lim02 | Absolute | 478 |
DTW-lim20 | Squared | 279 | DTW-lim02 | Squared | 472 |
DTW-lim20 | Symmkl | 430 | DTW-lim02 | Symmkl | 542 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Moysis, L.; Karasavvidis, K.; Kampelopoulos, D.; Boursianis, A.D.; Sotiroudis, S.; Nikolaidis, S.; Volos, C.; Sarigiannidis, P.; Matin, M.A.; Goudos, S.K. Identification of Greek Orthodox Church Chants Using Fuzzy Entropy. Computers 2025, 14, 39. https://doi.org/10.3390/computers14020039
Moysis L, Karasavvidis K, Kampelopoulos D, Boursianis AD, Sotiroudis S, Nikolaidis S, Volos C, Sarigiannidis P, Matin MA, Goudos SK. Identification of Greek Orthodox Church Chants Using Fuzzy Entropy. Computers. 2025; 14(2):39. https://doi.org/10.3390/computers14020039
Chicago/Turabian StyleMoysis, Lazaros, Konstantinos Karasavvidis, Dimitris Kampelopoulos, Achilles D. Boursianis, Sotirios Sotiroudis, Spiridon Nikolaidis, Christos Volos, Panagiotis Sarigiannidis, Mohammad Abdul Matin, and Sotirios K. Goudos. 2025. "Identification of Greek Orthodox Church Chants Using Fuzzy Entropy" Computers 14, no. 2: 39. https://doi.org/10.3390/computers14020039
APA StyleMoysis, L., Karasavvidis, K., Kampelopoulos, D., Boursianis, A. D., Sotiroudis, S., Nikolaidis, S., Volos, C., Sarigiannidis, P., Matin, M. A., & Goudos, S. K. (2025). Identification of Greek Orthodox Church Chants Using Fuzzy Entropy. Computers, 14(2), 39. https://doi.org/10.3390/computers14020039