1. Introduction
The rapid development of social media has not only provided people with a convenient platform for communication and information sharing, but it has also become a new channel for the spread of hate speech [
1,
2,
3]. This type of hate speech, including racial discrimination, sexism, and religious hatred, has caused great damage to the network environment and may even threaten social stability [
4,
5,
6]. However, the current methods for detecting hate speech still need to be improved. Traditional methods mainly rely on basic ML (machine learning) or DL (deep learning) algorithms, which have great deficiencies in processing complex language structures and contextual information [
7,
8,
9]. Among them, neural networks and deep learning require big data and computing resources, which are difficult to implement. Moreover, the model is as complex as a “black box”, and the decision-making process is difficult to explain. Although the hybrid model combines the advantages of a variety of algorithms, it is difficult to adjust parameters and easy to overfit, and its ability to process high-dimensional sparse data is weak. Although the hybrid model incorporates the advantages of multiple algorithms, the parameter tuning is complex and easy to overfit. In the high-dimensional sparse text feature space, its performance is unstable, and the false positives rate is high when identifying implied hate speech. The SVM algorithm maps data through kernel functions, maximizes the classification interval, reduces overfitting, and better recognizes implied hate speech. Therefore, SVM has an advantage in dealing with such issues. In particular, when hate speech is expressed in metaphors, sarcasm, etc., traditional methods are often difficult to identify, resulting in high false positive and false negative rates [
10,
11]. This situation limits the effective management of hate speech and also affects the regulation effect of network public opinion. To solve this problem, this article proposes a hate speech detection method combined with the SVM algorithm. The SVM model can efficiently find the optimal classification hyperplane from a high-dimensional feature space, which is very suitable for text classification tasks [
12,
13,
14]. When discussing the detection of hate speech on social media, SVM showed superior performance, especially in processing high-dimensional sparse data, feature selection, and dimension reduction. Compared with hybrid models, SVM simplifies model complexity, reduces the risk of overfitting, and is efficient, accurate, and easy to apply across platforms. Therefore, this paper uses SVM as the core algorithm to improve the accuracy and efficiency of hate speech detection. This research mainly targets text-based social platforms such as Twitter, Facebook, and Weibo. The user-generated content of these platforms is rich and provides sufficient training data for the detection of hate speech. Although the influence of video- and image-dominated platforms, such as TikTok and Instagram, has gradually increased, the collection and analysis of text data is more challenging on these platforms. By verifying the effectiveness of the text analysis method, this research has laid the foundation for expanding the scope of research to multiple modal platforms in the future. The purpose of this study is to enhance the accuracy of hate speech detection by deeply mining text context information and constructing more precise feature representations. This research aims to improve the monitoring capabilities for hate speech on social media while providing a foundation for exploring AI-driven strategies for managing online public opinion, which holds significant theoretical and practical implications.
To achieve this goal, the study integrates the SVM algorithm with advanced text feature extraction techniques, such as Word2Vec and TF-IDF, to optimize data preprocessing and effectively recognize complex expressions like sarcasm and metaphor. By leveraging the SVM classifier, the study seeks to achieve efficient and accurate detection of implicit hate speech. Furthermore, the research evaluates the application of this approach in network public opinion management, demonstrating its practical utility. Experimental results reveal that the proposed SVM-based method outperforms other techniques across multiple performance metrics and exhibits strong real-time processing capabilities. Additionally, the integration of sentiment dictionaries and the BERT model further enhances the recognition accuracy of complex hate speech, offering robust support for effective online public opinion management. The core innovation point of this research lies in the optimization of the algorithm. By dynamically adjusting the RBF nuclear parameter γ, the classification ambiguity caused by multiple synonyms of the text was successfully solved. At the same time, the generalization ability of the SVM model has been verified on the multilingual data of multiple social media platforms, such as Twitter, Facebook, and Weibo, providing a lightweight and efficient solution for social media governance. The specific contributions of this research are as follows: first, a text feature mapping method based on RBF kernel is proposed, which effectively solves the problem of nonlinear classification. For the first time, the performance of SVM was verified on a cross-platform multilingual dataset, thus filling the gap in the current research on the application of lightweight models.
5. Conclusions
This study leverages the SVM algorithm and optimized text feature extraction techniques, including Word2Vec word vector embedding and TF-IDF weighting, to enhance the detection of implicit hate speech on social media. By integrating a sentiment dictionary and the BERT model, the system achieves superior recognition accuracy for complex expressions such as sarcasm and metaphor. Experimental results demonstrate that the SVM model outperforms other baseline methods in terms of performance metrics and computational efficiency, making it suitable for real-time applications. The model’s ability to accurately distinguish between hate speech and non-hate speech provides a robust tool for fostering a healthier online environment. Furthermore, its integration with advanced language models like BERT offers significant potential for improving network public opinion management.
Despite these achievements, this study has certain limitations that warrant further exploration. First, while the model has been validated on multiple datasets, its cross-platform and multilingual applicability remains untested. Future research should focus on evaluating and enhancing the model’s adaptability across diverse platforms and languages to ensure broader usability. Second, the model’s generalization ability is limited when encountering obscure or novel expressions, which are common in dynamic social media environments. Addressing this challenge will require incorporating more adaptive learning mechanisms, such as continual learning or transfer learning, to handle emerging linguistic trends effectively.
Additionally, the rapidly evolving nature of social media data necessitates regular updates to the model to maintain its relevance and accuracy. Future work could explore automated retraining pipelines using real-time data streams to ensure the model adapts to new patterns of hate speech. Privacy protection and ethical considerations also present critical challenges. Researchers must develop efficient data collection and analysis methods that comply with privacy regulations while minimizing bias in model predictions.
Finally, the practical application of the model needs to be validated through field tests in real-world scenarios. Such evaluations would provide valuable insights into its effectiveness under varying conditions and help refine its deployment strategies. In conclusion, while this study establishes a strong foundation for hate speech detection, future research should prioritize cross-platform scalability, multilingual support, dynamic adaptability, ethical compliance, and real-world validation to further enhance its impact and utility.
This study has a significant effect on Twitter, Facebook, and Weibo, but the data sources are limited and may affect the universality of the model. For example, the visual content of TikTok and Instagram may contain unique hate speech, and this model is only optimized for text. Future improvement directions include the following: multimodal data integration: integrating image recognition and text analysis to enhance the recognition of hate speech on video platforms; cross-platform migration learning: using the pre-training model for cross-platform fine-tuning to improve the adaptability to different platform language styles; data collection expansion: explore the API cooperation or public datasets of TikTok and Instagram to supplement the training data of the video/image-led platform. These measures will gradually improve cross-platform research and break through current limitations. In addition, although the validity of the SVM model on text data has been verified, we still need to further expand the recording database, and its sample size is currently only 50,000. The existing audio data are mainly concentrated in English and Swahili and rely on a pre-trained CLIP model for feature extraction. However, the recognition effect of this method in dialects or complex accent environments is not ideal. In order to improve this situation, there are plans to introduce migration learning technology to strengthen the joint training of audio and text and include more recording samples of high-risk languages such as Spanish and Arabic.