Advancing Semantic Classification: A Comprehensive Examination of Machine Learning Techniques in Analyzing Russian-Language Patient Reviews
Abstract
:1. Introduction
- Sentiment;
- Target;
- Causal relationship; etc.
2. Analysing Medical Service Reviews as a Natural Language Text Classification Task
- Poor understanding and knowledge of healthcare on the part of the service consumers, which casts doubt on the accuracy of their assessments of the physician and medical services provided [21,22]. Patients often use indirect indicators unrelated to the quality of medical services as arguments (for example, their interpersonal experience with the physician [23,24]).
- Lack of clear criteria by which to assess a physician/medical service [23].
- In the capture phase, relevant social media content will be extracted from various sources. Data collection can be done by an individual or third-party providers [65].
- In the second phase, relevant data will be selected for predictive modelling of sentiment analysis.
- In the third phase, important key findings of the analysis will be visualised [66].
3. Classification Models for Text Reviews of the Quality of Medical Services in Social Media
- text sentiment: positive or negative;
- target: a review of a medical facility or an physician.
3.1. LSTM Network
- Embedding—the neural network input layer consisting of neurons (2):
- LSTM Layer—recurrent layer of the neural network; includes 32 blocks.
- Dense Layer—output layer consisting of four neurons. Each neuron is responsible for an output class. The activation function is “Softmax”.
3.2. A Recurrent Neural Network
- Embedding—input layer of the neural network.
- GRU—recurrent layer of the neural network; includes 16 blocks.
- Dense—output layer consisting of four neurons. The activation function is “Softmax”.
3.3. A Convolutional Neural Network
- Embedding—input layer of the neural network.
- Conv1D—convolutional layer required for deep learning. This layer improves the accuracy of text message classification by 5–7%. The activation function is “ReLU”.
- MaxPooling1D—layer which performs dimensionality reduction of generated feature maps. The maximum pooling is equal to 2.
- Dense—first output layer consisting of 128 neurons. The activation function is “ReLU”.
- Dense—final output layer consisting of four neurons. The activation function is “Softmax”.
3.4. Using Linguistic Algorithms
4. Software Implementation of a Text Classification System
- Text tokenisation.
- Removing spelling errors.
- Lemmatisation.
- Removing stop words.
- Tensorflow 2.14.0, an open-source machine learning software library developed by Google for solving neural network construction and training problems.
- Keras 2.15.0, a deep-learning library that is a high-level API written in Python 3.10 and capable of running on top of TensorFlow.
- Numpy 1.23.5, a Python library for working with multidimensional arrays.
- Pandas 2.1.2, a Python library that provides special data structures and operations for manipulating numerical tables and time series.
5. Experimental Results of Text Review Classification
5.1. Using Dataset
- city—city where the review was posted;
- text—feedback text;
- author_name—name of the feedback author;
- date—feedback date;
- day—feedback day;
- month—feedback month;
- year—feedback year;
- doctor_or_clinic—a binary variable (the review is of a physician OR a clinic);
- spec—medical specialty (for feedback on physicians);
- gender—feedback author’s gender;
- id—feedback identification number.
5.2. Experimental Results on Classifying Text Reviews by Sentiment
- Positive review of a physician;
- Positive review of a clinic;
- Negative review of a physician;
- Negative review of a clinic.
5.3. A Text Feedback Classification Experiment Using Various Machine Learning Models
- Some reviews were of both a clinic and a physician without mentioning the latter’s name. This prevented the named entity recognition tool from assigning the reviews to the mixed class. This problem could be solved by parsing the sentences further with identifying a semantically significant object unspecified by a full name.
- Some reviews expressed contrasting opinions about the clinic, related to different aspects of its operation. The opinions often differed on the organisational support versus the level of medical services provided by the clinic.
6. Conclusions
- The neural network classifiers achieve high accuracy in classifying the Russian-language reviews from social media by sentiment (positive or negative) and target (clinic or physician) using various architectures of the LSTM, CNN, or GRU networks, with the GRU-based architecture being the best (val_accuracy = 0.9271).
- The named entity recognition method improves the classification performance for each of the neural network classifiers when applied to the segmented text reviews.
- To further improve the classification accuracy, semantic segmentation of the reviews by target and sentiment is required, as well as a separate analysis of the resulting fragments.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Rajabi, M.; Hasanzadeh, R.P. A Modified adaptive hysteresis smoothing approach for image denoising based on spatial domain redundancy. Sens. Imaging 2021, 22, 1–25. [Google Scholar] [CrossRef]
- Rajabi, M.; Golshan, H.; Hasanzadeh, R.P. Non-local adaptive hysteresis despeckling approach for medical ultrasound images. Biomed. Signal Process. Control. 2023, 85, 105042. [Google Scholar] [CrossRef]
- Borji, A.; Seifi, A.; Hejazi, T.H. An efficient method for detection of Alzheimer’s disease using high-dimensional PET scan images. Intell. Decis. Technol. 2023, 17, 1–21. [Google Scholar] [CrossRef]
- Karimzadeh, M.; Vakanski, A.; Xian, M.; Zhang, B. Post-Hoc Explainability of BI-RADS Descriptors in a Multi-Task Framework for Breast Cancer Detection and Segmentation. In Proceedings of the 2023 IEEE 33rd International Workshop on Machine Learning for Signal Processing (MLSP), Rome, Italy, 17–20 September 2023; IEEE: New York, NY, USA; pp. 1–6. [Google Scholar]
- Rezaei, T.; Khouzani, P.J.; Khouzani, S.J.; Fard, A.M.; Rashidi, S.; Ghazalgoo, A.; Khodashenas, M. Integrating Artificial Intelligence into Telemedicine: Revolutionizing Healthcare Delivery. Kindle 2023, 3, 1–161. [Google Scholar]
- Litvin, S.W.; Goldsmith, R.E.; Pan, B. Electronic word-of-mouth in hospitality and tourism management. Tour. Manag. 2008, 29, 458–468. [Google Scholar] [CrossRef]
- Ismagilova, E.; Dwivedi, Y.K.; Slade, E.; Williams, M.D. Electronic word of mouth (eWOM) in the marketing context: A state of the art analysis and future directions. In Electronic Word of Mouth (Ewom) in the Marketing Context; Springer: Cham, Switzerland, 2017. [Google Scholar]
- Cantallops, A.S.; Salvi, F. New consumer behavior: A review of research on eWOM and hotels. Int. J. Hosp. Manag. 2014, 36, 41–51. [Google Scholar] [CrossRef]
- Mulgund, P.; Sharman, R.; Anand, P.; Shekhar, S.; Karadi, P. Data quality issues with physician-rating websites: Systematic review. J. Med. Internet Res. 2020, 22, e15916. [Google Scholar] [CrossRef]
- Ghimire, B.; Shanaev, S.; Lin, Z. Effects of official versus online review ratings. Ann. Tour. Res. 2022, 92, 103247. [Google Scholar] [CrossRef]
- Xu, Y.; Xu, X. Rating deviation and manipulated reviews on the Internet—A multi-method study. Inf. Manag. 2023, 2023, 103829. [Google Scholar] [CrossRef]
- Hu, N.; Bose, I.; Koh, N.S.; Liu, L. Manipulation of online reviews: An analysis of ratings, readability, and sentiments. Decis. Support Syst. 2012, 52, 674–684. [Google Scholar] [CrossRef]
- Luca, M.; Zervas, G. Fake it till you make it: Reputation, competition, and Yelp review fraud. Manag. Sci. 2016, 62, 3412–3427. [Google Scholar] [CrossRef]
- Namatherdhala, B.; Mazher, N.; Sriram, G.K. Artificial Intelligence in Product Management: Systematic review. Int. Res. J. Mod. Eng. Technol. Sci. 2022, 4, 2914–2917. [Google Scholar]
- Jabeur, S.B.; Ballouk, H.; Arfi, W.B.; Sahut, J.M. Artificial intelligence applications in fake review detection: Bibliometric analysis and future avenues for research. J. Bus. Res. 2023, 158, 113631. [Google Scholar] [CrossRef]
- Emmert, M.; McLennan, S. One decade of online patient feedback: Longitudinal analysis of data from a German physician rating website. J. Med. Internet Res. 2021, 23, e24229. [Google Scholar] [CrossRef] [PubMed]
- Kleefstra, S.M.; Zandbelt, L.C.; Borghans, I.; de Haes, H.J.; Kool, R.B. Investigating the potential contribution of patient rating sites to hospital supervision: Exploratory results from an interview study in The Netherlands. J. Med. Internet Res. 2016, 18, e201. [Google Scholar] [CrossRef]
- Bardach, N.S.; Asteria-Peñaloza, R.; Boscardin, W.J.; Dudley, R.A. The relationship between commercial website ratings and traditional hospital performance measures in the USA. BMJ Qual. Saf. 2013, 22, 194–202. [Google Scholar] [CrossRef]
- Van de Belt, T.H.; Engelen, L.J.; Berben, S.A.; Teerenstra, S.; Samsom, M.; Schoonhoven, L. Internet and social media for health-related information and communication in health care: Preferences of the Dutch general population. J. Med. Internet Res. 2013, 15, e220. [Google Scholar] [CrossRef]
- Hao, H.; Zhang, K.; Wang, W.; Gao, G. A tale of two countries: International comparison of online doctor reviews between China and the United States. Int. J. Med. Inform. 2017, 99, 37–44. [Google Scholar] [CrossRef] [PubMed]
- Bidmon, S.; Elshiewy, O.; Terlutter, R.; Boztug, Y. What patients value in physicians: Analyzing drivers of patient satisfaction using physician-rating website data. J. Med. Internet Res. 2020, 22, e13830. [Google Scholar] [CrossRef]
- Ellimoottil, C.; Leichtle, S.W.; Wright, C.J.; Fakhro, A.; Arrington, A.K.; Chirichella, T.J.; Ward, W.H. Online physician reviews: The good, the bad and the ugly. Bull. Am. Coll. Surg. 2013, 98, 34–39. [Google Scholar]
- Bidmon, S.; Terlutter, R.; Röttl, J. What explains usage of mobile physician-rating apps? Results from a web-based questionnaire. J. Med. Internet Res. 2014, 16, e3122. [Google Scholar] [CrossRef]
- Lieber, R. The Web is Awash in Reviews, but Not for Doctors. Here’s Why. New York Times, 9 March 2012. [Google Scholar]
- Daskivich, T.J.; Houman, J.; Fuller, G.; Black, J.T.; Kim, H.L.; Spiegel, B. Online physician ratings fail to predict actual performance on measures of quality, value, and peer review. J. Am. Med. Inform. Assoc. 2018, 25, 401–407. [Google Scholar] [CrossRef]
- Gray, B.M.; Vandergrift, J.L.; Gao, G.G.; McCullough, J.S.; Lipner, R.S. Website ratings of physicians and their quality of care. JAMA Intern. Med. 2015, 175, 291–293. [Google Scholar] [CrossRef]
- Skrzypecki, J.; Przybek, J. Physician review portals do not favor highly cited US ophthalmologists. In Seminars in Ophthalmology; Taylor & Francis: Abingdon, UK, 2018; Volume 33, pp. 547–551. [Google Scholar]
- Widmer, R.J.; Maurer, M.J.; Nayar, V.R.; Aase, L.A.; Wald, J.T.; Kotsenas, A.L.; Timimi, F.K.; Harper, C.M.; Pruthi, S. Online physician reviews do not reflect patient satisfaction survey responses. In Mayo Clinic Proceedings; Elsevier: Amsterdam, The Netherlands, 2018; Volume 93, pp. 453–457. [Google Scholar]
- Saifee, D.H.; Bardhan, I.; Zheng, Z. Do Online Reviews of Physicians Reflect Healthcare Outcomes? In Proceedings of the Smart Health: International Conference, ICSH 2017, Hong Kong, China, 26–27 June 2017; Springer International Publishing: New York, NY, USA, 2017; pp. 161–168. [Google Scholar]
- Trehan, S.K.; Nguyen, J.T.; Marx, R.; Cross, M.B.; Pan, T.J.; Daluiski, A.; Lyman, S. Online patient ratings are not correlated with total knee replacement surgeon–specific outcomes. HSS J. 2018, 14, 177–180. [Google Scholar] [CrossRef]
- Doyle, C.; Lennox, L.; Bell, D. A systematic review of evidence on the links between patient experience and clinical safety and effectiveness. BMJ Open 2013, 3, e001570. [Google Scholar] [CrossRef] [PubMed]
- Okike, K.; Uhr, N.R.; Shin, S.Y.; Xie, K.C.; Kim, C.Y.; Funahashi, T.T.; Kanter, M.H. A comparison of online physician ratings and internal patient-submitted ratings from a large healthcare system. J. Gen. Intern. Med. 2019, 34, 2575–2579. [Google Scholar] [CrossRef] [PubMed]
- Rotman, L.E.; Alford, E.N.; Shank, C.D.; Dalgo, C.; Stetler, W.R. Is there an association between physician review websites and press ganey survey results in a neurosurgical outpatient clinic? World Neurosurg. 2019, 132, 891–899. [Google Scholar] [CrossRef]
- Lantzy, S.; Anderson, D. Can consumers use online reviews to avoid unsuitable doctors? Evidence from RateMDs. com and the Federation of State Medical Boards. Decis. Sci. 2020, 51, 962–984. [Google Scholar] [CrossRef]
- Gilbert, K.; Hawkins, C.M.; Hughes, D.R.; Patel, K.; Gogia, N.; Sekhar, A.; Duszak, R., Jr. Physician rating websites: Do radiologists have an online presence? J. Am. Coll. Radiol. 2015, 12, 867–871. [Google Scholar] [CrossRef] [PubMed]
- Okike, K.; Peter-Bibb, T.K.; Xie, K.C.; Okike, O.N. Association between physician online rating and quality of care. J. Med. Internet Res. 2016, 18, e324. [Google Scholar] [CrossRef]
- Imbergamo, C.; Brzezinski, A.; Patankar, A.; Weintraub, M.; Mazzaferro, N.; Kayiaros, S. Negative online ratings of joint replacement surgeons: An analysis of 6402 reviews. Arthroplast. Today 2021, 9, 106–111. [Google Scholar] [CrossRef]
- Mostaghimi, A.; Crotty, B.H.; Landon, B.E. The availability and nature of physician information on the internet. J. Gen. Intern. Med. 2010, 25, 1152–1156. [Google Scholar] [CrossRef]
- Lagu, T.; Hannon, N.S.; Rothberg, M.B.; Lindenauer, P.K. Patients’ evaluations of health care providers in the era of social networking: An analysis of physician-rating websites. J. Gen. Intern. Med. 2010, 25, 942–946. [Google Scholar] [CrossRef]
- López, A.; Detz, A.; Ratanawongsa, N.; Sarkar, U. What patients say about their doctors online: A qualitative content analysis. J. Gen. Intern. Med. 2012, 27, 685–692. [Google Scholar] [CrossRef] [PubMed]
- Shah, A.M.; Yan, X.; Qayyum, A.; Naqvi, R.A.; Shah, S.J. Mining topic and sentiment dynamics in physician rating websites during the early wave of the COVID-19 pandemic: Machine learning approach. Int. J. Med. Inform. 2021, 149, 104434. [Google Scholar] [CrossRef] [PubMed]
- Shah, A.M.; Yan, X.; Tariq, S.; Ali, M. What patients like or dislike in physicians: Analyzing drivers of patient satisfaction and dissatisfaction using a digital topic modeling approach. Inf. Process. Manag. 2021, 58, 102516. [Google Scholar] [CrossRef]
- Lagu, T.; Metayer, K.; Moran, M.; Ortiz, L.; Priya, A.; Goff, S.L.; Lindenauer, P.K. Website characteristics and physician reviews on commercial physician-rating websites. JAMA 2017, 317, 766–768. [Google Scholar] [CrossRef] [PubMed]
- Chen, Y.; Xie, J. Online consumer review: Word-of-mouth as a new element of marketing communication mix. Manag. Sci. 2008, 54, 477–491. [Google Scholar] [CrossRef]
- Pavlou, P.A.; Dimoka, A. The nature and role of feedback text comments in online marketplaces: Implications for trust building, price premiums, and seller differentiation. Inf. Syst. Res. 2006, 17, 392–414. [Google Scholar] [CrossRef]
- Terlutter, R.; Bidmon, S.; Röttl, J. Who uses physician-rating websites? Differences in sociodemographic variables, psychographic variables, and health status of users and nonusers of physician-rating websites. J. Med. Internet Res. 2014, 16, e97. [Google Scholar] [CrossRef] [PubMed]
- Emmert, M.; Meier, F. An analysis of online evaluations on a physician rating website: Evidence from a German public reporting instrument. J. Med. Internet Res. 2013, 15, e2655. [Google Scholar] [CrossRef] [PubMed]
- Nwachukwu, B.U.; Adjei, J.; Trehan, S.K.; Chang, B.; Amoo-Achampong, K.; Nguyen, J.T.; Taylor, S.A.; McCormick, F.; Ranawat, A.S. Rating a sports medicine surgeon’s “quality” in the modern era: An analysis of popular physician online rating websites. HSS J. 2016, 12, 272–277. [Google Scholar] [CrossRef] [PubMed]
- Obele, C.C.; Duszak, R., Jr.; Hawkins, C.M.; Rosenkrantz, A.B. What patients think about their interventional radiologists: Assessment using a leading physician ratings website. J. Am. Coll. Radiol. 2017, 14, 609–614. [Google Scholar] [CrossRef]
- Emmert, M.; Meier, F.; Pisch, F.; Sander, U. Physician choice making and characteristics associated with using physician-rating websites: Cross-sectional study. J. Med. Internet Res. 2013, 15, e2702. [Google Scholar] [CrossRef] [PubMed]
- Gao, G.G.; McCullough, J.S.; Agarwal, R.; Jha, A.K. A changing landscape of physician quality reporting: Analysis of patients’ online ratings of their physicians over a 5-year period. J. Med. Internet Res. 2012, 14, e38. [Google Scholar] [CrossRef]
- Rahim, A.I.A.; Ibrahim, M.I.; Musa, K.I.; Chua, S.L.; Yaacob, N.M. Patient satisfaction and hospital quality of care evaluation in malaysia using servqual and facebook. Healthcare 2021, 9, 1369. [Google Scholar] [CrossRef]
- Galizzi, M.M.; Miraldo, M.; Stavropoulou, C.; Desai, M.; Jayatunga, W.; Joshi, M.; Parikh, S. Who is more likely to use doctor-rating websites, and why? A cross-sectional study in London. BMJ Open 2012, 2, e001493. [Google Scholar] [CrossRef]
- Hanauer, D.A.; Zheng, K.; Singer, D.C.; Gebremariam, A.; Davis, M.M. Public awareness, perception, and use of online physician rating sites. JAMA 2014, 311, 734–735. [Google Scholar] [CrossRef] [PubMed]
- McLennan, S.; Strech, D.; Meyer, A.; Kahrass, H. Public awareness and use of German physician ratings websites: Cross-sectional survey of four North German cities. J. Med. Internet Res. 2017, 19, e387. [Google Scholar] [CrossRef]
- Lin, Y.; Hong, Y.A.; Henson, B.S.; Stevenson, R.D.; Hong, S.; Lyu, T.; Liang, C. Assessing patient experience and healthcare quality of dental care using patient online reviews in the United States: Mixed methods study. J. Med. Internet Res. 2020, 22, e18652. [Google Scholar] [CrossRef]
- Emmert, M.; Meier, F.; Heider, A.K.; Dürr, C.; Sander, U. What do patients say about their physicians? An analysis of 3000 narrative comments posted on a German physician rating website. Health Policy 2014, 118, 66–73. [Google Scholar] [CrossRef]
- Greaves, F.; Ramirez-Cano, D.; Millett, C.; Darzi, A.; Donaldson, L. Harnessing the cloud of patient experience: Using social media to detect poor quality healthcare. BMJ Qual. Saf. 2013, 22, 251–255. [Google Scholar] [CrossRef]
- Hao, H.; Zhang, K. The voice of Chinese health consumers: A text mining approach to web-based physician reviews. J. Med. Internet Res. 2016, 18, e108. [Google Scholar] [CrossRef] [PubMed]
- Shah, A.M.; Yan, X.; Shah, S.A.A.; Mamirkulova, G. Mining patient opinion to evaluate the service quality in healthcare: A deep-learning approach. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 2925–2942. [Google Scholar] [CrossRef]
- Wallace, B.C.; Paul, M.J.; Sarkar, U.; Trikalinos, T.A.; Dredze, M. A large-scale quantitative analysis of latent factors and sentiment in online doctor reviews. J. Am. Med. Inform. Assoc. 2014, 21, 1098–1103. [Google Scholar] [CrossRef] [PubMed]
- Ranard, B.L.; Werner, R.M.; Antanavicius, T.; Schwartz, H.A.; Smith, R.J.; Meisel, Z.F.; Asch, D.A.; Ungar, L.H.; Merchant, R.M. Yelp reviews of hospital care can supplement and inform traditional surveys of the patient experience of care. Health Aff. 2016, 35, 697–705. [Google Scholar] [CrossRef] [PubMed]
- Hao, H. The development of online doctor reviews in China: An analysis of the largest online doctor review website in China. J. Med. Internet Res. 2015, 17, e134. [Google Scholar] [CrossRef] [PubMed]
- Jiang, S.; Street, R.L. Pathway linking internet health information seeking to better health: A moderated mediation study. Health Commun. 2017, 32, 1024–1031. [Google Scholar] [CrossRef] [PubMed]
- Hotho, A.; Nürnberger, A.; Paaß, G. A Brief Survey of Text Mining. LDV Forum—GLDV. J. Comput. Linguist. Lang. Technol. 2005, 20, 19–62. [Google Scholar] [CrossRef]
- Păvăloaia, V.; Teodor, E.; Fotache, D.; Danileț, M. Opinion Mining on Social Media Data: Sentiment Analysis of User Preferences. Sustainability 2019, 11, 4459. [Google Scholar] [CrossRef]
- Bespalov, D.; Bing, B.; Yanjun, Q.; Shokoufandeh, A. Sentiment classification based on supervised latent n-gram analysis. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management (CIKM’11), Glasgow, Scotland, 24–28 October 2011; Association for Computing Machinery: New York, NY, USA; pp. 375–382. [Google Scholar]
- Irfan, R.; King, C.K.; Grages, D.; Ewen, S.; Khan, S.U.; Madani, S.A.; Kolodziej, J.; Wang, L.; Chen, D.; Rayes, A.; et al. A Survey on Text Mining in Social Networks. Camb. J. Knowl. Eng. Rev. 2015, 30, 157–170. [Google Scholar] [CrossRef]
- Patel, P.; Mistry, K. A Review: Text Classification on Social Media Data. IOSR J. Comput. Eng. 2015, 17, 80–84. [Google Scholar]
- Lee, K.; Palsetia, D.; Narayanan, R.; Patwary, M.d.M.A.; Agrawal, A.; Choudhary, A.S. Twitter Trending Topic Classification. In Proceeding of the 2011 IEEE 11th International Conference on Data Mining Workshops, ICDW’11, Vancouver, BC, Canada, 11 December 2011; pp. 251–258. [Google Scholar]
- Kateb, F.; Kalita, J. Classifying Short Text in Social Media: Twitter as Case Study. Int. J. Comput. Appl. 2015, 111, 1–12. [Google Scholar] [CrossRef]
- Chirawichitichai, N.; Sanguansat, P.; Meesad, P. A Comparative Study on Feature Weight in Thai Document Categorization Framework. In Proceedings of the 10th International Conference on Innovative Internet Community Services (I2CS), IICS, Bangkok, Thailand, 3–5 June 2010; pp. 257–266. [Google Scholar]
- Theeramunkong, T.; Lertnattee, V. Multi-Dimension Text Classification, SIIT, Thammasat University. 2005. Available online: http://www.aclweb.org/anthology/C02–1155 (accessed on 25 October 2023).
- Viriyayudhakorn, K.; Kunifuji, S.; Ogawa, M. A Comparison of Four Association Engines in Divergent Thinking Support Systems on Wikipedia, Knowledge, Information, and Creativity Support Systems, KICSS2010; Springer: Berlin/Heidelberg, Germany, 2011; pp. 226–237. [Google Scholar]
- Sornlertlamvanich, V.; Pacharawongsakda, E.; Charoenporn, T. Understanding Social Movement by Tracking the Keyword in Social Media, in MAPLEX2015, Yamagata, Japan, February 2015. Available online: https://www.researchgate.net/publication/289035345_Understanding_Social_Movement_by_Tracking_the_Keyword_in_Social_Media (accessed on 22 December 2023).
- Konstantinov, A.; Moshkin, V.; Yarushkina, N. Approach to the Use of Language Models BERT and Word2vec in Sentiment Analysis of Social Network Texts. In Recent Research in Control Engineering and Decision Making; Dolinina, O., Bessmertny, I., Brovko, A., Kreinovich, V., Pechenkin, V., Lvov, A., Zhmud, V., Eds.; ICIT 2020. Studies in Systems, Decision and Control; Springer: Cham, Switzerland, 2020; Volume 337, pp. 462–473. [Google Scholar] [CrossRef]
- Kalabikhina, I.; Zubova, E.; Loukachevitch, N.; Kolotusha, A.; Kazbekova, Z.; Banin, E.; Klimenko, G. Identifying Reproductive Behavior Arguments in Social Media Content Users’ Opinions through Natural Language Processing Techniques. Popul. Econ. 2023, 7, 40–59. [Google Scholar] [CrossRef]
Target | Positive | Negative | |
---|---|---|---|
Sentiment | |||
About a clinic | 54% | 20% | |
About a physician | 21% | 5% |
# | Feedback Text | Feedback Data | Sentiment Class | Target Class |
---|---|---|---|---|
1 | “The doctor was really rude, she had no manners with the patients, she didn’t care about your poor health, all she wanted was to get home early. I never want to see that doctor again. She’s rubbish, I wouldn’t recommend her to anyone.” | Ekaterina, 13 April 2023, Moscow | Negative | physician |
2 | “I had to get an MRI scan of my abdomen. They kept me waiting. They gave me the scan results straight away; I’ll show them to my doctor. It was easy for me to get to the clinic. Their manners were not very good. I won’t be going back there.” | Kamil, 17 April 2023, Moscow | Negative | clinic |
3 | “All those good reviews are written by their staff marketers, they try to stop the bad ones, they don’t let any real complaints get through. The clinic is very pricey, they just want to make money, no one cares about your health there.” | Anonymous, 10 April 2023, Moscow | Negative | clinic |
4 | “What they do in this clinic is rip you off because they make you do checkups and tests that you don’t need. I found out when I was going through all this stuff, and then I wondered why I had to do it all.” | Arina, 2 March 2023, Moscow | Negative | clinic |
5 | “Rubbish doctor. My problem is really bad skin dryness and rashes because of that. ######## just said, “you just moisturise it” and that was it. She didn’t tell me how to moisturise my skin or what to use for moisturiser. I had to push her asking for advice on care and what to do next. She didn’t give me anything except some cream, and that only after I asked her”. | Anonymous, 11 May 2023, Moscow, Russia | Negative | physician |
6 | “My husband had a bad tooth under the crown, the dentist said he had to redo his whole jaw and put all new crowns again, like he had to sort everything out to fit the new crowns after the tooth was fixed. In the end we trusted the dentist and redid my husband’s whole jaw. The bridge didn’t last a month, it kept coming out. In the end we had to do it all over again with another dentist at another clinic. He was awful, he only rips you off. I don’t recommend this dentist to anyone.” | Tatyana, 13 April 2023, Moscow | Negative | physician |
7 | “In 2020, I was going to a doctor at the clinic #######.ru for 3 months for the pain in my left breast. He gave me some cream and told me to go on a diet, but I was getting worse. I went to see another doctor; it turned out it was breast cancer. Nearly killed me…” | Maya, 27 March 2023, Moscow | Negative | physician |
8 | “####### nearly left my child with one leg. A healthy 10-month-old child had to have two surgeries after what this “doctor” had given him. It’s over now, but the “nice” memory of this woman will stay with me forever.” | Elizaveta, 16 March 2023, Moscow | Negative | physician |
# | Target Class/Sentiment Class | Positive | Negative | Total |
---|---|---|---|---|
1 | Clinic | 11,178 | 4121 | 15,299 |
2 | Physician | 9775 | 2025 | 11,800 |
3 | Mixed | 20,374 | 10,773 | 31,147 |
4 | Total | 41,327 | 16,919 | 58,246 |
LSTM | GRU | CNN | SVM | BERT | |
---|---|---|---|---|---|
Accuracy | 0.9369 | 0.9309 | 0.9772 | 0.8441 | 0.8942 |
Val_accuracy | 0.9253 | 0.9271 | 0.9112 | 0.8289 | 0.8711 |
Loss | 0.1859 | 0.2039 | 0.0785 | 0.3769 | 0.1729 |
Val_loss | 0.2248 | 0.2253 | 0.3101 | 0.3867 | 0.2266 |
F1 | 0.8045 | 0.7840 | 0.7461 | 0.6819 | 0.7936 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kalabikhina, I.; Moshkin, V.; Kolotusha, A.; Kashin, M.; Klimenko, G.; Kazbekova, Z. Advancing Semantic Classification: A Comprehensive Examination of Machine Learning Techniques in Analyzing Russian-Language Patient Reviews. Mathematics 2024, 12, 566. https://doi.org/10.3390/math12040566
Kalabikhina I, Moshkin V, Kolotusha A, Kashin M, Klimenko G, Kazbekova Z. Advancing Semantic Classification: A Comprehensive Examination of Machine Learning Techniques in Analyzing Russian-Language Patient Reviews. Mathematics. 2024; 12(4):566. https://doi.org/10.3390/math12040566
Chicago/Turabian StyleKalabikhina, Irina, Vadim Moshkin, Anton Kolotusha, Maksim Kashin, German Klimenko, and Zarina Kazbekova. 2024. "Advancing Semantic Classification: A Comprehensive Examination of Machine Learning Techniques in Analyzing Russian-Language Patient Reviews" Mathematics 12, no. 4: 566. https://doi.org/10.3390/math12040566
APA StyleKalabikhina, I., Moshkin, V., Kolotusha, A., Kashin, M., Klimenko, G., & Kazbekova, Z. (2024). Advancing Semantic Classification: A Comprehensive Examination of Machine Learning Techniques in Analyzing Russian-Language Patient Reviews. Mathematics, 12(4), 566. https://doi.org/10.3390/math12040566