A Novel Emotion-Aware Hybrid Music Recommendation Method Using Deep Neural Network
Abstract
:1. Introduction
- The proposed models represent music emotions and users’ music emotion preferences under certain emotion states in a continuous form of valence-arousal scores allowing for complex emotion expressions in music recommendations;
- Trained models with a deep neural network (emoDNN) that enable the rapid processing of both music emotions and user emotion states even on time points when no user activity data exists;
- Incorporated event-related information to allow the hybrid approach to generate music recommendations considering the influences posed by events on emotions.
2. Related Works
3. Model Construction
- The data layer. This layer offers 5 kinds of data for further processing. The user portrait data contains information inherited from conventional recommendation systems, such as the user profile from the registration system. The social media data contains the users’ activity data from social media, which is used as implicit user feedback in our recommendation method. The music acoustic data contains the audio signal data of the music. The music metadata contains descriptive data of the music, such as the genre, artist, lyrics, etc. The event data contains public opinion data on certain events.
- The model layer. In the model layer, the music emotion representation model and the emotion state representation model generate the music emotion representation and emotion state representation using data from the data layer. The models are trained to predict music emotions for the music and the music emotion preferences for users with a deep neural network (emoDNN). The hybrid recommendation model combining the content-based and collaborative filtering recommendation approaches uses the data generated by the trained models to make recommendations.
- The application layer. In this layer, the proposed method generates music recommendation lists for users.
3.1. The Music Emotion Representation Model
- Low-level audio features, which extracted from the audio data;
- Music descriptive metadata, such as genre, year, artist, lyrics, etc.
3.1.1. Low-Level Audio Features Extraction
- Pitch: Pitch extraction calculates the distances between the peaks of a given segment of the music audio signal. Let denote the audio segment, k denotes the pitch period of a peak, and denotes the window length of the segment, and the pitch feature can be obtained using Equation (1):
- ZCR: The Zero-Crossing Rate describes the rate of sign changes of the signal during a signal frame. It counts the times the signal changes across positive and negative values. The definition of ZCR is shown in Equation (2),
- LE: This feature estimates the energy of the amplitude of the audio signal. The calculation can be formulated as Equation (4):
- TEO: This feature links to the energy of the audio signal as well, but from a nonlinear perspective. The TEO of a signal segment can be calculated using Equation (5):
- MFCC: The MFCC is derived from a mel-scale frequency filter-bank. The calculation of MFCC can be obtained by first segmenting audio signals into frames and then applying a Hamming Window (HW) defined by Equation (6) to each frame:
3.1.2. Music Metadata Exploitation
- Normalization: The min-max normalization is used to convert the 6 elements in to values within the range of . Let denote the score of the ith lyrics after normalization, and can be described by Equation (9):
- Calculate valence and arousal: The emotions of are treated as positive while are treated as negative. The valence and arousal values of can be obtained by Equation (10):
3.1.3. The Deep Neural Network for the Music Emotion Representation Model
3.2. The Emotion State Representation Model
- The environment module: It utilizes the information that affects user emotion in the environment. The information creates a context that can be identified from social networks and public opinion. This module influences the emotion evaluation of the user and contributes to the exogenous factors of the emotion state;
- The experience module is controlled by the characters of the user and contributes to the endogenous factors of the emotion state;
- The evaluation module synthesis the output of the exogenous and endogenous factors and generate the emotion state at arbitrary time points by evaluation;
- The emotion module identifies the music emotion preference user the emotion state generated by the evaluation module.
3.2.1. The Exogenous Factors of the Emotion State
- Events: Events influence user emotions by raising public opinions. Thus, events can be identified by analyzing the sentiments of the raised public opinions;
- Posts: Users react to certain kinds of emotions on social networks by posting messages. Sentiment analysis on posts within a certain time range can utilize posts to describe user emotion;
- Comments: Comments within a certain time range reflect user emotional responses. The sentiments of these comments help distinguish user emotion states.
3.2.2. The Endogenous Factors of the Emotion State
3.2.3. The Emotion State Representation
3.2.4. The Music Emotion Preference Identification Using Deep Neural Network
3.3. The Recommendation Process
- If the user’s music emotion preference information exists in the system, the music emotion representation is calculated using the music emotion representation model. The emoDNN for the music emotion representation model can be trained, and its parameters can be saved for future use. The emotion representation for the music can also be stored to accelerate future recommendation cycles.
- If the user’s music emotion preference information does not exist in the system, the music emotion preference for the user’s current emotion state is calculated using the emotion state representation model. The parameters for the trained emotion state representation model emoDNN and the calculated user music emotion preference can be stored to accelerate future recommendation cycles.
- Content-based method is used to get a list of music of similar music emotion to the music emotion preference at the given time . Add the music emotion vector to the music feature vector used by existing content-based music recommenders [84]. Calculate the similarities of the music to the emotion preferred music and rank the similarities to generate a list of music.
- Collaborative filtering method is used to get the music with the music emotion preferred by similar music emotion preference users at the given time . Add the music emotion preference vector of the user’s feature vector. Calculate the similarities between and other users. Rank the users according to the similarities. The preferred music lists of the users in the similar user list are ranked by the music emotion preference of the users. Get Top-K music items from the lists of the users.
- Using the generated music list as the recommendation list.
4. Experiments
4.1. Datasets
- Data from the myPersonality [27] dataset. This dataset comes with information about 4 million Facebook users and 22 million posts over their timeline. By filtering through this dataset, we extracted 109 users who post music links and 509 music posted in the form of music links. This dataset provides not only the user tags but also music information and implicit user feedback information to aid the training of the music emotion representation model and the emotion state representation model.
- Data acquired from music social platform. The platform is a Chinese music streaming service (NetEase Cloud Music, https://music.163.com (accessed on 12 July 2021)) with an embedded community for users to share their lives with music. We scraped the music chart with a web crawler (https://github.com/MiChongGET/CloudMusicApi (accessed on 12 July 2021) and got 500 high-ranking songs. The metadata and audio file of the 500 songs and the 509 ones in the myPersonality dataset were acquired. The music in the myPersonality dataset contained almost English songs and was relatively old, while the 500 high-ranking songs were mostly Chinese songs. Therefore, the two datasets have 38 songs in common. We also searched through the community and acquired 200 users related to the 971 songs and 116,079 user activities, including user likes, posts, reposts, and comments.
- Data acquired from the social network (Weibo, https://weibo.com (accessed on 13 July 2021)). We acquired 105,376 event-related data from Weibo. The data were posted within the time window of the user activities of the music social platform data. We use this data to insert event-related public opinion information to emoMR.
4.2. Model Training
4.3. Baseline Algorithms and Metrics
4.4. Results
4.4.1. Performance Comparison in the Top-10 Recommendation Task
4.4.2. Performance Comparison in the Top-N Recommendation Tasks
4.4.3. Performance Comparison in Recommending Top-10 Items during Event-Related Time Ranges
4.5. Discussions
5. Conclusions
- The proposed method predicts music emotions in a continuous form, enabling a more precise and flexible representation of music emotions than traditional discrete representations. By modeling the music emotion representation into a vector based on the emotion valence-arousal model using low-level audio features and music metadata, we build a deep neural network (emoDNN) to predict the emotion representations for the music items. Compared to the discrete representations of music emotions in other studies, our model predicts the music emotion in a continuous form of valence and arousal scores.
- The proposed method predicts users’ music emotion preference whenever it needs while traditional methods can only generate predictions at time points when user feedback data are present. By modeling the users’ emotion states using an artificial emotion generation model with endogenous factors generated by the human circadian rhythm model and exogenous factors consisted of events and implicit user feedback, we use the emoDNN to predict the users’ music emotion preferences under the emotion states represented by the model. Benefiting from the continuity of the human circadian rhythm, the model is able to predict continuously across time regardless of the absence of user feedback data at a given time point.
- The proposed method can utilize event information to refine the emotion generation, which provides a more accurate emotion state description than traditional artificial emotion generation models. With the introduction of events in the exogenous factors of the user emotion state representation model, emoMR is able to express the influence of events (especially major social events) on user emotions.
- The proposed method employs a hybrid approach of combining the advantages of both content-based and collaborative filtering approaches to generate the recommendations. Theoretically, our findings will contribute to the theory of emotion recognition by reflecting the cognitive theory of emotion [68] and the cognitive appraisal theory of emotion [69] with social signals.
Author Contributions
Funding
Institutional Review Board Statement
Data Availability Statement
Conflicts of Interest
References
- Xu, C. A novel recommendation method based on social network using matrix factorization technique. Inf. Process. Manag. 2018, 54, 463–474. [Google Scholar] [CrossRef]
- Xu, C. A big-data oriented recommendation method based on multi-objective optimization. Knowl. Based Syst. 2019, 177, 11–21. [Google Scholar] [CrossRef]
- Xu, C.; Ding, A.S.; Zhao, K. A novel POI recommendation method based on trust relationship and spatial-temporal factors. Electron. Commer. Res. Appl. 2021, 48, 101060. [Google Scholar] [CrossRef]
- Song, Y.; Dixon, S.; Pearce, M. A survey of music recommendation systems and future perspectives. In Proceedings of the 9th International Symposium on Computer Music Modeling and Retrieval, London, UK, 19–22 June 2012; Volume 4, pp. 395–410. [Google Scholar]
- Schedl, M.; Knees, P.; McFee, B.; Bogdanov, D.; Kaminskas, M. Music recommender systems. In Recommender Systems Handbook; Springer: Berlin, Germany, 2015; pp. 453–492. [Google Scholar]
- Xie, Y.; Ding, L. A survey of music personalized recommendation system. In Proceedings of the 2018 International Conference on Network, Communication, Computer Engineering (NCCE 2018), Chongqing, China, 26–27 May 2018; Atlantis Press: Paris, Franch, 2018; pp. 852–855. [Google Scholar]
- Cheng, Z.; Shen, J. On effective location-aware music recommendation. ACM Trans. Inf. Syst. (TOIS) 2016, 34, 1–32. [Google Scholar] [CrossRef]
- Wang, D.; Deng, S.; Xu, G. Sequence-based context-aware music recommendation. Inf. Retr. J. 2018, 21, 230–252. [Google Scholar] [CrossRef] [Green Version]
- Krumhansl, C.L. Music: A link between cognition and emotion. Curr. Dir. Psychol. Sci. 2002, 11, 45–50. [Google Scholar] [CrossRef]
- Yang, Y.H.; Chen, H.H. Machine recognition of music emotion: A review. ACM Trans. Intell. Syst. Technol. (TIST) 2012, 3, 1–30. [Google Scholar] [CrossRef]
- Mahsal Khan, S.; Abdul Hamid, N.; Mohd Rashid, S. Music and its familiarity affection on audience decision making. J. Messenger 2019, 11, 70–80. [Google Scholar] [CrossRef]
- Carrington, S.J.; Bailey, A.J. Are there theory of mind regions in the brain? A review of the neuroimaging literature. Hum. Brain Mapp. 2009, 30, 2313–2335. [Google Scholar] [CrossRef]
- Deng, J.J.; Leung, C. Emotion-based music recommendation using audio features and user playlist. In Proceedings of the 2012 6th International Conference on New Trends in Information Science, Service Science and Data Mining (ISSDM2012), Taipei, Taiwan, 23–25 October 2012; pp. 796–801. [Google Scholar]
- Rachman, F.H.; Sarno, R.; Fatichah, C. Music Emotion Classification based on Lyrics-Audio using Corpus based Emotion. Int. J. Electr. Comput. Eng. 2018, 8, 1720. [Google Scholar] [CrossRef]
- Yang, Y.H.; Su, Y.F.; Lin, Y.C.; Chen, H.H. Music emotion recognition: The role of individuality. In Proceedings of the International Workshop on Human-Centered Multimedia, Augsburg, Bavaria, Germany, 28 September 2007; pp. 13–22. [Google Scholar]
- Hu, B.; Shi, C.; Zhao, W.X.; Yang, T. Local and global information fusion for top-n recommendation in heterogeneous information network. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, Torino, Italy, 22–26 October 2018; pp. 1683–1686. [Google Scholar]
- Ayata, D.; Yaslan, Y.; Kamasak, M.E. Emotion based music recommendation system using wearable physiological sensors. IEEE Trans. Consum. Electron. 2018, 64, 196–203. [Google Scholar] [CrossRef]
- Omar, R.; Henley, S.M.; Bartlett, J.W.; Hailstone, J.C.; Gordon, E.; Sauter, D.A.; Frost, C.; Scott, S.K.; Warren, J.D. The structural neuroanatomy of music emotion recognition: Evidence from frontotemporal lobar degeneration. Neuroimage 2011, 56, 1814–1821. [Google Scholar] [CrossRef] [Green Version]
- Malhotra, A.; Totti, L.; Meira, W., Jr.; Kumaraguru, P.; Almeida, V. Studying user footprints in different online social networks. In Proceedings of the 2012 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Istanbul, Turkey, 26–29 August 2012; pp. 1065–1070. [Google Scholar]
- Rivas, A.; González-Briones, A.; Cea-Morán, J.J.; Prat-Pérez, A.; Corchado, J.M. My-Trac: System for Recommendation of Points of Interest on the Basis of Twitter Profiles. Electronics 2021, 10, 1263. [Google Scholar] [CrossRef]
- Huang, X.; Liao, G.; Xiong, N.; Vasilakos, A.V.; Lan, T. A Survey of Context-Aware Recommendation Schemes in Event-Based Social Networks. Electronics 2020, 9, 1583. [Google Scholar] [CrossRef]
- Park, T.; Jeong, O.R. Social network based music recommendation system. J. Internet Comput. Serv. 2015, 16, 133–141. [Google Scholar] [CrossRef]
- Shen, T.; Jia, J.; Li, Y.; Ma, Y.; Bu, Y.; Wang, H.; Chen, B.; Chua, T.S.; Hall, W. Peia: Personality and emotion integrated attentive model for music recommendation on social media platforms. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 206–213. [Google Scholar]
- Ma, Y.; Li, X.; Xu, M.; Jia, J.; Cai, L. Multi-scale context based attention for dynamic music emotion prediction. In Proceedings of the 25th ACM international conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 1443–1450. [Google Scholar]
- Wu, D. Music personalized recommendation system based on hybrid filtration. In Proceedings of the 2019 International Conference on Intelligent Transportation, Big Data & Smart City (ICITBS), Changsha, China, 12–13 January 2019; pp. 430–433. [Google Scholar]
- Abdul, A.; Chen, J.; Liao, H.Y.; Chang, S.H. An emotion-aware personalized music recommendation system using a convolutional neural networks approach. Appl. Sci. 2018, 8, 1103. [Google Scholar] [CrossRef] [Green Version]
- Polignano, M.; Narducci, F.; de Gemmis, M.; Semeraro, G. Towards Emotion-aware Recommender Systems: An Affective Coherence Model based on Emotion-driven Behaviors. Expert Syst. Appl. 2021, 170, 114382. [Google Scholar] [CrossRef]
- Cowen, A.S.; Keltner, D. Self-report captures 27 distinct categories of emotion bridged by continuous gradients. Proc. Natl. Acad. Sci. USA 2017, 114, E7900–E7909. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Gunes, H.; Schuller, B.; Pantic, M.; Cowie, R. Emotion representation, analysis and synthesis in continuous space: A survey. In Proceedings of the Face and Gesture 2011, Santa Barbara, CA, USA, 21–25 March 2011; pp. 827–834. [Google Scholar]
- White, C.; Ratcliff, R.; Vasey, M.; McKoon, G. Dysphoria and memory for emotional material: A diffusion-model analysis. Cogn. Emot. 2009, 23, 181–205. [Google Scholar] [CrossRef] [Green Version]
- Newman, D.B.; Nezlek, J.B. The Influence of Daily Events on Emotion Regulation and Well-Being in Daily Life. Personal. Soc. Psychol. Bull. 2021, 0146167220980882. [Google Scholar] [CrossRef]
- Basu, S.; Jana, N.; Bag, A.; Mahadevappa, M.; Mukherjee, J.; Kumar, S.; Guha, R. Emotion recognition based on physiological signals using valence-arousal model. In Proceedings of the 2015 Third International Conference on Image Information Processing (ICIIP), Waknaghat, India, 21–24 December 2016. [Google Scholar]
- Paul, D.; Kundu, S. A Survey of Music Recommendation Systems with a Proposed Music Recommendation System. In Emerging Technology in Modelling and Graphics; Springer: Berlin, Germany, 2020; pp. 279–285. [Google Scholar]
- Jazi, S.Y.; Kaedi, M.; Fatemi, A. An emotion-aware music recommender system: Bridging the user’s interaction and music recommendation. Multimed. Tools Appl. 2021, 80, 13559–13574. [Google Scholar] [CrossRef]
- Coutinho, E.; Cangelosi, A. Musical emotions: Predicting second-by-second subjective feelings of emotion from low-level psychoacoustic features and physiological measurements. Emotion 2011, 11, 921. [Google Scholar] [CrossRef] [Green Version]
- Barthet, M.; Fazekas, G.; Sandler, M. Music emotion recognition: From content-to context-based models. In International Symposium on Computer Music Modeling and Retrieval; Springer: Berlin, Germany, 2012; pp. 228–252. [Google Scholar]
- Yoon, K.; Lee, J.; Kim, M.U. Music recommendation system using emotion triggering low-level features. IEEE Trans. Consum. Electron. 2012, 58, 612–618. [Google Scholar] [CrossRef]
- Panda, R.; Malheiro, R.; Paiva, R.P. Musical Texture and Expressivity Features for Music Emotion Recognition; ISMIR: Paris, France, 2018; pp. 383–391. [Google Scholar]
- Wang, J.C.; Yang, Y.H.; Jhuo, I.H.; Lin, Y.Y.; Wang, H.M. The acousticvisual emotion Guassians model for automatic generation of music video. In Proceedings of the 20th ACM international conference on Multimedia, Nara, Japan, 29 October–2 November 2012; pp. 1379–1380. [Google Scholar]
- Aljanaki, A. Emotion in Music: Representation and Computational Modeling. Ph.D. Thesis, Utrecht University, Utrecht, The Netherlands, 2016. [Google Scholar]
- An, Y.; Sun, S.; Wang, S. Naive Bayes classifiers for music emotion classification based on lyrics. In Proceedings of the 2017 IEEE/ACIS 16th International Conference on Computer and Information Science (ICIS), Wuhan, China, 24–26 May 2017; pp. 635–638. [Google Scholar]
- Kim, Y.E.; Schmidt, E.M.; Migneco, R.; Morton, B.G.; Richardson, P.; Scott, J.; Speck, J.A.; Turnbull, D. Music emotion recognition: A state of the art review. Proc. Ismir 2010, 86, 937–952. [Google Scholar]
- Nakasone, A.; Prendinger, H.; Ishizuka, M. Emotion recognition from electromyography and skin conductance. In Proceedings of the 5th International Workshop on Biosignal Interpretation, Citeseer, Palmerston North, New Zealand, 13–15 December 2004; pp. 219–222. [Google Scholar]
- Lin, Y.P.; Wang, C.H.; Jung, T.P.; Wu, T.L.; Jeng, S.K.; Duann, J.R.; Chen, J.H. EEG-based emotion recognition in music listening. IEEE Trans. Biomed. Eng. 2010, 57, 1798–1806. [Google Scholar]
- Hsu, J.L.; Zhen, Y.L.; Lin, T.C.; Chiu, Y.S. Affective content analysis of music emotion through EEG. Multimed. Syst. 2018, 24, 195–210. [Google Scholar] [CrossRef]
- Dornbush, S.; Fisher, K.; McKay, K.; Prikhodko, A.; Segall, Z. XPOD-A human activity and emotion aware mobile music player. In Proceedings of the International Conference on Mobile Technology, Applications and Systems, IET, Guangzhou, China, 15–17 November 2005. [Google Scholar]
- Sarda, P.; Halasawade, S.; Padmawar, A.; Aghav, J. Emousic: Emotion and activity-based music player using machine learning. In Advances in Computer Communication and Computational Sciences; Springer: Berlin, Germany, 2019; pp. 179–188. [Google Scholar]
- Rosa, R.L.; Rodriguez, D.Z.; Bressan, G. Music recommendation system based on user’s sentiments extracted from social networks. IEEE Trans. Consum. Electron. 2015, 61, 359–367. [Google Scholar] [CrossRef]
- Deng, S.; Wang, D.; Li, X.; Xu, G. Exploring user emotion in microblogs for music recommendation. Expert Syst. Appl. 2015, 42, 9284–9293. [Google Scholar] [CrossRef]
- Qian, Y.; Zhang, Y.; Ma, X.; Yu, H.; Peng, L. EARS: Emotion-aware recommender system based on hybrid information fusion. Inf. Fusion 2019, 46, 141–146. [Google Scholar] [CrossRef]
- Lu, C.C.; Tseng, V.S. A novel method for personalized music recommendation. Expert Syst. Appl. 2009, 36, 10035–10044. [Google Scholar] [CrossRef]
- Kim, H.H. A semantically enhanced tag-based music recommendation using emotion ontology. In Asian Conference on Intelligent Information and Database Systems; Springer: Berlin, Germany, 2013; pp. 119–128. [Google Scholar]
- Han, B.j.; Rho, S.; Jun, S.; Hwang, E. Music emotion classification and context-based music recommendation. Multimed. Tools Appl. 2010, 47, 433–460. [Google Scholar] [CrossRef]
- Wang, X.; Wang, Y. Improving content-based and hybrid music recommendation using deep learning. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 627–636. [Google Scholar]
- Awan, M.J.; Khan, R.A.; Nobanee, H.; Yasin, A.; Anwar, S.M.; Naseem, U.; Singh, V.P. A Recommendation Engine for Predicting Movie Ratings Using a Big Data Approach. Electronics 2021, 10, 1215. [Google Scholar] [CrossRef]
- Schedl, M. Deep learning in music recommendation systems. Front. Appl. Math. Stat. 2019, 5, 44. [Google Scholar] [CrossRef] [Green Version]
- Nam, J.; Choi, K.; Lee, J.; Chou, S.Y.; Yang, Y.H. Deep learning for audio-based music classification and tagging: Teaching computers to distinguish rock from bach. IEEE Signal Process. Mag. 2018, 36, 41–51. [Google Scholar] [CrossRef]
- Hossain, M.S.; Muhammad, G. Emotion recognition using deep learning approach from audio–visual emotional big data. Inf. Fusion 2019, 49, 69–78. [Google Scholar] [CrossRef]
- Aldao, A.; Nolen-Hoeksema, S. One versus many: Capturing the use of multiple emotion regulation strategies in response to an emotion-eliciting stimulus. Cogn. Emot. 2013, 27, 753–760. [Google Scholar] [CrossRef]
- Nwe, T.L.; Foo, S.W.; Silva, L. Detection of stress and emotion in speech using traditional and FFT based log energy features. In Proceedings of the Fourth International Conference on Information, Communications and Signal Processing, 2003 and the Fourth Pacific Rim Conference on Multimedia, Proceedings of the 2003 Joint, Singapore, 15–18 December 2004. [Google Scholar]
- Hui, G.; Chen, S.; Su, G. Emotion Classification of Infant Voice Based on Features Derived from Teager Energy Operator. In Proceedings of the 2008 Congress on Image and Signal Processing, Sanya, China, 27–30 May 2008. [Google Scholar]
- Jothilakshmi, S.; Ramalingam, V.; Palanivel, S. Unsupervised speaker segmentation with residual phase and MFCC features. Expert Syst. Appl. 2009, 36, 9799–9804. [Google Scholar] [CrossRef]
- Hasan, M.; Rahman, M.S.; Shimamura, T. Windowless-Autocorrelation-Based Cepstrum Method for Pitch Extraction of Noisy Speech. J. Signal Process. 2013, 16, 231–239. [Google Scholar] [CrossRef] [Green Version]
- Corthaut, N.; Govaerts, S.; Verbert, K.; Duval, E. Connecting the Dots: Music Metadata Generation, Schemas and Applications. In Proceedings of the 9th International Conference on Music Information, Philadelphia, PA, USA, 14–18 September 2008; pp. 249–254. [Google Scholar]
- Tallapally, D.; Sreepada, R.S.; Patra, B.K.; Babu, K.S. User preference learning in multi-criteria recommendations using stacked auto encoders. In Proceedings of the 12th ACM Conference on Recommender Systems, Vancouver, BC, Canada, 2 October 2018; pp. 475–479. [Google Scholar]
- Zarzour, H.; Al-Sharif, Z.A.; Jararweh, Y. RecDNNing: A recommender system using deep neural network with user and item embeddings. In Proceedings of the 2019 10th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 11–13 June 2019; pp. 99–103. [Google Scholar]
- Liu, Z.; Guo, S.; Wang, L.; Du, B.; Pang, S. A multi-objective service composition recommendation method for individualized customer: Hybrid MPA-GSO-DNN model. Comput. Ind. Eng. 2019, 128, 122–134. [Google Scholar] [CrossRef]
- Bever, T.G. A cognitive theory of emotion and aesthetics in music. Psychomusicol. J. Res. Music Cogn. 1988, 7, 165. [Google Scholar] [CrossRef]
- Garcia-Prieto, P.; Scherer, K.R. Connecting social identity theory and cognitive appraisal theory of emotions. In Social Identities: Motivational, Emotional, Cultural Influences; Psychology Press: Hove, UK, 2006; pp. 189–207. [Google Scholar]
- Kim, H.R.; Kwon, D.S. Computational model of emotion generation for human–robot interaction based on the cognitive appraisal theory. J. Intell. Robot. Syst. 2010, 60, 263–283. [Google Scholar] [CrossRef] [Green Version]
- Manera, V.; Samson, A.C.; Pehrs, C.; Lee, I.A.; Gross, J.J. The eyes have it: The role of attention in cognitive reappraisal of social stimuli. Emotion 2014, 14, 833. [Google Scholar] [CrossRef] [PubMed]
- Blechert, J.; Sheppes, G.; Di Tella, C.; Williams, H.; Gross, J.J. See what you think: Reappraisal modulates behavioral and neural responses to social stimuli. Psychol. Sci. 2012, 23, 346–353. [Google Scholar] [CrossRef] [Green Version]
- Pribram, K.H.; Melges, F.T. Psychophysiological basis of emotion. Handb. Clin. Neurol. 1969, 3, 316–341. [Google Scholar]
- Takahashi, K. Remarks on emotion recognition from bio-potential signals. In Proceedings of the 2nd International conference on Autonomous Robots and Agents, Citeseer, Palmerston North, New Zealand, 13–15 December 2004; pp. 186–191. [Google Scholar]
- Hofstra, W.A.; de Weerd, A.W. How to assess circadian rhythm in humans: A review of literature. Epilepsy Behav. 2008, 13, 438–444. [Google Scholar] [CrossRef] [PubMed]
- Cardinali, D.P. The human body circadian: How the biologic clock influences sleep and emotion. Neuroendocrinol. Lett. 2000, 21, 9–16. [Google Scholar]
- Panda, S. Circadian physiology of metabolism. Science 2016, 354, 1008–1015. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- olde Scheper, T.; Klinkenberg, D.; Pennartz, C.; Van Pelt, J. A mathematical model for the intracellular circadian rhythm generator. J. Neurosci. 1999, 19, 40–47. [Google Scholar] [CrossRef] [Green Version]
- Refinetti, R.; Cornélissen, G.; Halberg, F. Procedures for numerical analysis of circadian rhythms. Biol. Rhythm Res. 2007, 38, 275–325. [Google Scholar] [CrossRef]
- Mitsutake, G.; Otsuka, K.; Cornelissen, G.; Herold, M.; Günther, R.; Dawes, C.; Burch, J.; Watson, D.; Halberg, F. Circadian and infradian rhythms in mood. Biomed. Pharmacother. 2000, 55, s94–s100. [Google Scholar] [CrossRef]
- Hansen, C.H.; Hansen, R.D. How rock music videos can change what is seen when boy meets girl: Priming stereotypic appraisal of social interactions. Sex Roles 1988, 19, 287–316. [Google Scholar] [CrossRef]
- Turek, F.W.; Penev, P.; Zhang, Y.; Van Reeth, O.; Zee, P. Effects of age on the circadian system. Neurosci. Biobehav. Rev. 1995, 19, 53–58. [Google Scholar] [CrossRef]
- Zheng, E.; Kondo, G.Y.; Zilora, S.; Yu, Q. Tag-aware dynamic music recommendation. Expert Syst. Appl. 2018, 106, 244–251. [Google Scholar] [CrossRef]
- Soleymani, M.; Aljanaki, A.; Wiering, F.; Veltkamp, R.C. Content-based music recommendation using underlying music preference structure. In Proceedings of the 2015 IEEE International Conference on Multimedia and Expo (ICME), Turin, Italy, 29 June–3 July 2015; pp. 1–6. [Google Scholar]
- Metcalf, L.; Casey, W. Chapter 2—Metrics, similarity, and sets. In Cybersecurity and Applied Mathematics; Metcalf, L., Casey, W., Eds.; Syngress: Boston, MA, USA, 2016; pp. 3–22. [Google Scholar]
- Vigliensoni, G.; Fujinaga, I. The Music Listening Histories Dataset. In Proceedings of the 18th International Society for Music Information Retrieval Conference (ISMIR17), Suzhou, China, 23–28 October 2017; pp. 96–102. [Google Scholar]
- Goddard, M. The EU General Data Protection Regulation (GDPR): European regulation that has a global impact. Int. J. Mark. Res. 2017, 59, 703–705. [Google Scholar] [CrossRef]
- Fessahaye, F.; Perez, L.; Zhan, T.; Zhang, R.; Fossier, C.; Markarian, R.; Chiu, C.; Zhan, J.; Gewali, L.; Oh, P. T-recsys: A novel music recommendation system using deep learning. In Proceedings of the 2019 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 11–13 January 2019; pp. 1–6. [Google Scholar]
- Bogdanov, D.; Haro, M.; Fuhrmann, F.; Gómez, E.; Herrera, P. Content-based music recommendation based on user preference examples. Copyr. Inf. 2010, 33, 1–6. [Google Scholar]
- Su, J.H.; Chang, W.Y.; Tseng, V.S. Effective social content-based collaborative filtering for music recommendation. Intell. Data Anal. 2017, 21, S195–S216. [Google Scholar] [CrossRef]
- Xiang, K.; Xu, C.; Wang, J. Understanding the relationship between tourists’ consumption behavior and their consumption substitution willingness under unusual environment. Psychol. Res. Behav. Manag. 2021, 14, 483–500. [Google Scholar] [CrossRef]
Class | Name | Description |
---|---|---|
Musical information | Musical properties of the audio signal, e.g. duration, key | |
Performance | Descriptions of the people that are involved in a musical performance, e.g. artists | |
Descriptor | Describing the musical work, e.g. title | |
Playback rendition | Information useful for rendering of the music file, e.g. relative volume | |
Lyrics | Text of the musical work and related information, e.g. translated lyrics text | |
Instrumentation & arrangement | Information about the used instruments and orchestration, e.g. instrument type |
Parameter | Setting |
---|---|
Training epoch | 300 |
Batch | 20 |
Optimizer | Adam |
Learn rate | 0.05 |
Parameter | Setting |
---|---|
Training epoch | 500 |
Batch | 10 |
Optimizer | Adam |
Learn rate | 0.01 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, S.; Xu, C.; Ding, A.S.; Tang, Z. A Novel Emotion-Aware Hybrid Music Recommendation Method Using Deep Neural Network. Electronics 2021, 10, 1769. https://doi.org/10.3390/electronics10151769
Wang S, Xu C, Ding AS, Tang Z. A Novel Emotion-Aware Hybrid Music Recommendation Method Using Deep Neural Network. Electronics. 2021; 10(15):1769. https://doi.org/10.3390/electronics10151769
Chicago/Turabian StyleWang, Shu, Chonghuan Xu, Austin Shijun Ding, and Zhongyun Tang. 2021. "A Novel Emotion-Aware Hybrid Music Recommendation Method Using Deep Neural Network" Electronics 10, no. 15: 1769. https://doi.org/10.3390/electronics10151769
APA StyleWang, S., Xu, C., Ding, A. S., & Tang, Z. (2021). A Novel Emotion-Aware Hybrid Music Recommendation Method Using Deep Neural Network. Electronics, 10(15), 1769. https://doi.org/10.3390/electronics10151769