Next Article in Journal
Characterizing Microheterogeneity in Liquid Mixtures via Local Density Fluctuations
Previous Article in Journal
Multi-Modal Latent Diffusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on a Framework for Chinese Argot Recognition and Interpretation by Integrating Improved MECT Models

1
School of Information and Network Security, People’s Public Security University of China, Beijing 102206, China
2
Key Laboratory of Security Prevention Technology and Risk Assessment of the Ministry of Public Security, Beijing 100038, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(4), 321; https://doi.org/10.3390/e26040321
Submission received: 25 January 2024 / Revised: 24 March 2024 / Accepted: 30 March 2024 / Published: 6 April 2024

Abstract

:
In underground industries, practitioners frequently employ argots to communicate discreetly and evade surveillance by investigative agencies. Proposing an innovative approach using word vectors and large language models, we aim to decipher and understand the myriad of argots in these industries, providing crucial technical support for law enforcement to detect and combat illicit activities. Specifically, positional differences in semantic space distinguish argots, and pre-trained language models’ corpora are crucial for interpreting them. Expanding on these concepts, the article assesses the semantic coherence of word vectors in the semantic space based on the concept of information entropy. Simultaneously, we devised a labeled argot dataset, MNGG, and developed an argot recognition framework named CSRMECT, along with an argot interpretation framework called LLMResolve. These frameworks leverage the MECT model, the large language model, prompt engineering, and the DBSCAN clustering algorithm. Experimental results demonstrate that the CSRMECT framework outperforms the current optimal model by 10% in terms of the F1 value for argot recognition on the MNGG dataset, while the LLMResolve framework achieves a 4% higher accuracy in interpretation compared to the current optimal model.The related experiments undertaken also indicate a potential correlation between vector information entropy and model performance.

1. Introduction

The 51st Statistical Report on the Development of China’s Internet [1] shows that, in 2022, a total of 845 million netizens participated in the purchase and sale of compliant items on online platforms, accounting for 79.2% of the total number of netizens. However, including the dark web and underground forums, illicit transactions persist, posing significant challenges to cyber and social security stability in cyberspace [2]. According to statistics, in its 2.5 years online, the dark web Tor site “Silk Road” amassed 150,000 users and transactions totaling $1.2 billion [3]. The online trading products of the dark web often include illegal and irregular items, such as drugs, electronic fraud materials, hacking tools, and smuggled goods [4]. To avoid scrutiny, industry practitioners often conceal sensitive content within argots mixed with normal content, enhancing transaction concealment. Figure 1, illustrates some hidden words and their explanations in the field of drug trading. In the first example, “宵夜” (midnight snack) refers to “毒品” (drugs), and “猪肉” (pork) refers to “冰毒” (methamphetamine).
The research on argot recognition and interpretation in combat units has a long history. Early law enforcement agencies used manual construction of an argot knowledge base to interpret known argots. For example, the US Drug Enforcement Agency (EDA) intelligence department developed a set of drug codeword libraries to decipher the collected evidence and data containing codewords [5]. Ouyang et al. collected and summarized relevant drug code libraries based on the language characteristics and ethnic customs of the Guangxi region [6]. Ouyang et al. also used railway property infringement criminals as their target for obtaining secret language, and summarized a set of railway property infringement secret language libraries [7]. After summarizing a substantial number of hidden language samples, previous research findings, and the results of our preliminary experiments, we have identified the following characteristics of Chinese argots:
  • Chinese argots and the words they refer to (referred to as pronouns) are normal vocabulary, rather than special characters similar to Morse codes [8].
  • There are inherent connections between Chinese argots and their pronouns, including their shape, pronunciation, and meaning, which are relatively loose and often unknown to outsiders [9].
  • Most pronouns are nouns or verbs, but the part of speech of Chinese argots and pronouns may not be the same, and verbs and adjectives are often used to refer to nouns or verbs.
  • If the lexicon of argot words is concealed, and multiple alternative words are provided for that position, the concealment capability of argot words and the entropy of the set of alternative words for filling in that position are positively correlated.
As the aforementioned algorithm only recognizes a partial set of argot features, the various algorithms mentioned earlier are now inadequate for the current scenario of argot recognition and interpretation, particularly in the task of argot recognition. Therefore, this paper integrates the concept of semantic space, drawing inspiration from the manifold assumption in deep learning. Additionally, it combines the notion of vector information entropy to assess the rationality of word vectors within the semantic space [10].
Specifically, this paper makes the following contributions in the domain of argot recognition and interpretation:
  • We constructed a Chinese long text corpus MNGG dataset using an open-source cant dataset [8] to support research on Chinese argots recognition.
  • A Chinese argots recognition model CSRMECT was proposed based on the MNGG dataset, MECT4CNER, and DBSCAN clustering algorithms
  • Based on the MNGG dataset, the large language model, and prompt engineering, a Chinese argot interpretation framework LLMResolve was constructed to carry out Chinese argot interpretation work.
  • We built a framework for Chinese argot detection, combined with CSRMECT and LLMResolve, to construct a comprehensive cold start Chinese argot recognition and interpretation workflow covering all fields.

2. Related Work

In the early stage of NLP research in academia, the field most related to Chinese argot recognition was called Chinese morphs decoding and resolving. This field focuses on researching the bypass mechanism of sensitive word detection algorithms. Specifically, in order to avoid detection by detection algorithms, users often replace a sensitive word with another word. The replaced word is generally called a reference, and the word used to replace the reference word is called a morph [11].
In the field of Chinese morph interpretation, Huang et al. [12] conducted groundbreaking research, first proposed the concept of morphs, and constructed a morph dataset through Weibo. They also designed various algorithms to interpret morphs. Zhang et al. [13] used Huang et al.’s variant definition algorithm for morph interpretation. Then Zhang et al. [14] constructed a deep neural network-based interpretation algorithm and first proposed the concept of resolve candidate words. Sha et al. [15] proposed a framework based on word embedding for morph resolution. You et al. [16] proposed a variant interpretation method based on an autoencoder combined with contextual information, and the model performance exceeded that for all the aforementioned indicators for morph interpretation. In the field of Chinese morph extraction, Zhang et al. [13] designed various morph generation algorithms by analyzing the construction logic of morphs. They attempted to use these algorithms to generate morphs and used SVM-based detection algorithms for morph extraction, achieving good detection bypass effects. Afterwards, Zhang et al. [14] proposed a morph recognition algorithm based on SVM- and graph-based semi-supervised learning approaches. The morph recognition algorithm achieved an F1 value of 83% on the Weibo dataset designed by Huang et al. [12].
However, there are significant differences between the fields of Chinese argot recognition and interpretation, as well as Chinese morph recognition and interpretation. From the perspective of the research subject, pronouns in the field of variant recognition only include sensitive nouns, such as public figures’ names, well-known place names, and well-known event names. In contrast, the scope of pronouns in argot recognition is broader. Therefore, compared to variant recognition tasks, both argot recognition and interpretation tasks become more complex and challenging. Based on these differences, Xu et al. [8] constructed a dataset of Chinese cant word-pronoun pairs for cant recognition tasks, providing evaluation support for future argot recognition tasks.
Compared with the field of Chinese argot recognition, there has been relatively more progress in the field of English argot recognition. Due to the dynamic and rapidly evolving nature of cybercrime, argot vocabulary undergoes continuous changes, with additions and deletions occurring. Additionally, each criminal group may establish its own industry-specific argot (e.g., drug traffickers) [17]. Consequently, there has been a shift towards machine learning methods for argot recognition, gradually replacing traditional manual construction of argot knowledge bases. In 2015, Dhuliawala et al. [18] proposed an English slang dictionary called SlangNet, aiming to complement WordNet for use in natural language processing (NLP) applications. The research team also evaluated the resource using the Lesk algorithm and the Extended Lesk algorithm. Furthermore, this work showed how to leverage online crowdsourcing resources to build high-quality language resources. In 2016, Wu et al. [19] constructed the slang dataset SlangSD for sentiment analysis of social media. Greg Durrett et al. [20] focused on the task of product keyword identification in online cybercrime forums and studied the effects of different research methods on product keyword identification through custom datasets. Later in 2018, Yuan et al. [5] proposed an argot recognition framework, Cantreader, incorporating improved word2vec and Hypernym identification, achieving commendable results in identifying English argot words across various forums on the dark web. In the 2020 study, Wilson et al. [21] used the Urban Dictionary dataset to train a set of word vectors and evaluated them in multiple slang-related tasks. The set of word vectors achieved significant improvements in specific tasks. Aravinda et al. [22], integrating language models and knowledge graphs, introduced a framework for detecting English slang in social media in 2022. Their approach demonstrated good performance in downstream experimental tasks, such as emotion detection, hate speech detection, and crime detection.
In summary, the current state of automated argot recognition faces several challenges:
  • Lack of research and datasets specifically focused on argot recognition in the Chinese language domain.
  • Existing studies on argot recognition are often domain-specific, lacking the development of a universally applicable framework for argot recognition across diverse domains.
  • Most existing models rely on extensive prior data for training, hindering the generalization and cold start capabilities of argot recognition algorithms in unfamiliar domains.
Inspired by the manifold assumption [23], this paper posits that the commendable performance of numerous deep learning models based on word embeddings indicates that vectors obtained through word embeddings carry specific semantic meanings within their high-dimensional space. The lower the entropy of the set of vectors obtained through word embeddings, the more distinct the semantic meanings conveyed by the word embeddings, indicating a more effective performance of word embeddings.
Therefore, based on this assumption, to address the aforementioned issues, this paper has preliminarily established the Chinese argot dataset MNGG. Subsequently, the CSRMECT argot recognition model and the LLMResolve argot interpretation framework are proposed. Both the model and the framework are designed with a cold start approach, leveraging extensive knowledge embedded in pretrained texts to achieve generalization in unfamiliar domains, eliminating the need for domain-specific datasets for training.

3. Chinese Argot Recognition Based on CSRMECT Model

3.1. Entrophy Based Semantic Space

By combining the concept of information entropy from information theory with the semantic meanings of word vectors in word embeddings, this paper proposes a semantic space based on information entropy. This space is utilized to assess the semantic coherence and richness of word vectors.

3.1.1. Vector Embedding and Semantic Space

In a series of papers around the year 2000, Joshua Bengio and others [24] employed neural probabilistic language models to enable machines to “learn a distributed representation for words”, thereby achieving the goal of dimensionality reduction in the word space. Subsequently, over the following decades, various well-known word-embedding algorithms emerged, including Word2Vec [25], GloVe [26], and others.
The essence of word embeddings lies in reducing words with rich semantics to vectors of specific dimensions, where each dimension carries a specific meaning, as shown in Figure 2.

3.1.2. Word Vector Semantic Rationality Index Based on Information Entropy

In the task of argot recognition, each sentence is treated as a separate corpus for word embedding. Consequently, a single word may have multiple word vectors. This paper posits that when a vocabulary term is used as an argot, its spatial position in the semantic space should exhibit significant differences compared to its position when used in regular contexts alongside other non-argot words. Moreover, vectors generated by more advanced word-embedding algorithms should exhibit more pronounced spatial distribution characteristics, with vectors of words used in similar contexts converging together.
For example, the Chinese word “打击” has two meanings in English, namely “hit” and “catastrophe”. In an ideal word-embedding vector result, as shown on the left side of Figure 3, the vectors for these meanings should be distinct. By contrast, an undesirable result is depicted on the right side, where the vectors fail to adequately differentiate between the meanings.
It is evident that the higher the entropy of the set of word vectors corresponding to multiple usages of a word, the poorer the embedding performance of the current word-embedding algorithm for that specific usage of the word. To measure the entropy of a set of word vectors, we utilize the following formula based on the definition of information entropy:
  • Retrieve the set of word vectors S i corresponding to the i-th usage from the word-embedding result E w o r d of the word w o r d .
  • Let C = j = 1 n S i n denote the core vector.
  • Calculate the distance d i = V i C for each vector V i in S i to the core vector.
  • Assuming P i is the probability for the i-th vector, use the normalized exponential of the distance as the probability: P i = exp ( β d i ) j exp ( β d j ) , where β is a parameter.
  • Calculate the entropy of the vector set S i as H i = i P i log ( P i ) .
  • The vector information entropy for the word w o r d in its word embedding is given by E n t r o p y ( w o r d ) = i = 1 n H i n .

3.2. Enhanced MECT Model

The MECT model, proposed by Wu et al., is a cross-transformer based on multi-modal embeddings, applied in Chinese named entity recognition tasks [27]. As illustrated in Figure 4, the MECT model consists primarily of multi-modal embedding layers and cross-transformer layers. Previous studies have demonstrated the model’s commendable accuracy in identifying Chinese entities, efficient operational speed, and notable interpretability [28,29]. We will employ the MECT model for the word vector embedding tasks described above.

3.2.1. Multivariate Data Embedding Layer

This layer comprises two main components: lattice embedding and Chinese radical-level embedding. Lattice embedding is a crucial element of the FLAT model [30], encompassing semantic and positional boundary information in the lattice data, comprehensively considering contextual features in sentences. Taking the sentence “Nanjing Yangtze River Bridge” as an example, the input situation for lattice embedding is illustrated in Figure 5, containing the head and tail positions of characters and words.
Chinese characters are based on ideograms, representing their meanings through the shapes of objects. For instance, characters with “艹” or “木” as radical components often represent plants and can effectively recognize raw materials used in the production of drugs, such as “cannabis” and “ephedra”. Characters with “月” as a radical component often represent body parts or organs and can adeptly identify euphemisms in the adult content domain. There are various methods for decomposing Chinese characters, including radical decomposition (CR), head-tail decomposition (HT), and structural decomposition (SC), as illustrated in Table 1.
To extract radical-level features of Chinese characters, an improved CNN network is constructed in this paper. CNN was initially proposed in the LeNet-5 model [31] and was applied in AlexNet in 2012 [32], achieving significant breakthroughs in the field of image recognition. Therefore, this paper selects the information-rich structural composition (SC) as the radical-level feature of Chinese characters and utilizes CNN to extract features of the characters. The specific process of embedding Chinese characters at the radical level is as follows:
  • Decompose Chinese characters into radicals (SC) and input them into a CNN network.
  • Embed radicals at the radical level for convolutional operations in the convolutional layer.
  • Utilize max-pooling and fully connected layers to obtain the final embedding vector for Chinese character radicals.

3.2.2. Cross-Transformer Layer

The MECT model introduces a cross-transformer network [27], as illustrated in Figure 6. This network employs two transformer encoders, which independently process information from lattice embeddings and Chinese radical embeddings. It achieves the enrichment of Chinese character semantic information by incorporating contextual and lexical information.
The inputs Q L ( Q R ) , K L ( K R ) , and V L ( V R ) in the cross-transformer network are obtained through linear transformations using lattice embeddings or Chinese radical embeddings, as defined in Equation (1).
Q L ( R ) , i K L ( R ) , i V L ( R ) , i T = E L ( R ) , i · W L ( R ) , Q E W L ( R ) , V T
In the context, E L , i and E R , i represent the i-th lattice embedding vector and Chinese radical embedding vector, respectively. Here, E denotes the unit vector, and W represents learnable parameters. In the cross-transformer network, the attention calculation formula is given by:
A t t A R , V L = S o f t m a x ( A R ) V L
A t t A L , V R = S o f t m a x ( A L ) V R
A L R , i j = ϵ ( u ) K R L , j + ϵ ( v ) R * L R , i j
Wherein, the lower-left corner’s L denotes the values from the lattice embedding side, and R represents the values from the Chinese radical embedding side. The parameters u and v in Formula (4) represent learnable attention offset parameters. Here, R i j * = R i j · W , where W is a learnable parameter, and ϵ ( x ) = Q L ( R ) , i + x L ( R ) T . The calculation of R i j is as follows:
R i j = R e L U W r p h i h j p t i t j
p s p a n ( 2 k ) = sin s p a n 10000 2 k d m o d e l
p s p a n ( 2 k + 1 ) = c o s s p a n 10000 2 k d m o d e l
Among these, R i j represents the computation of the relative distance between positions i and j, where W r denotes a learnable parameter, and h i and t i , respectively, signify the head and tail positions of the Chinese character at position i. The symbol ⨁ signifies a concatenation operation. In Formulas (6) and (7), the term s p a n corresponds to h i h j or t i t j as defined in [33].

3.2.3. CSRMECT Model

To integrate word vectors considering both the context and Chinese character structure, this paper proposes the CSRMECT model, building upon the MECT model with modifications. Specifically, we enhance the MECT model by removing the final CRF layer and directing the output character vectors from the linear layer into the vector aggregator module for word vector synthesis. The final output is a context-encoded word vector, as illustrated in Figure 7. The vector aggregator module takes character vectors as input and produces word vectors. In this paper, we adopt the default approach for constructing the vector aggregator module, as depicted in the pseudocode in Algorithm 1.
Algorithm 1: Vector Aggregator
Entropy 26 00321 i001
The field of argot recognition has long been plagued by the lack of high-quality annotated datasets. Through the above changes, the CSRMECT model successfully solved this problem. Specifically, the training of the CSRMECT model only requires the use of a normal corpus. The trained CSRMECT model can understand the contextual relationships in the sentence and output a word vector with contextual semantics for each word.

3.3. DBSCAN Clustering Algorithm

Clustering, one of the primary methods for knowledge discovery in large datasets, encompasses various prevalent techniques in the field of semantic clustering, including k-means [34], hierarchical clustering [35], and DBSCAN [36]. The k-means algorithm exhibits limitations in handling non-spherical clusters and is susceptible to the choice of initial cluster centers [37], necessitating the pre-specification of the cluster quantity (K value). Hierarchical clustering, often proceeding in a top-down or bottom-up hierarchical decomposition due to its simplicity, tends to form cluster chains. In contrast, DBSCAN possesses the capability to cluster shapes of arbitrary forms, such as linear, concave, elliptical, without the need for predefined cluster quantities. Additionally, DBSCAN has been proven effective in handling massive databases [36,38,39]. Consequently, we employ the DBSCAN algorithm for clustering, aiming to extract argot vocabulary from extensive sets of word vectors.
The DBSCAN clustering algorithm determines the density of a point by calculating the number of points within a specified radius. Points with densities exceeding a specified threshold are grouped into clusters. Given the high dimensionality and sparsity between word vectors in this research, the Euclidean distance proves inadequate for accurately measuring the vector similarity. Hence, the cosine distance is chosen for DBSCAN clustering, with the formula as follows:
D ( p , q ) = i = 1 n p i × q i i = 1 n p i 2 × i = 1 n q i 2
In the context where q and p represent arbitrary word vectors, and n denotes the dimensionality of the word vectors, with p i and q i indicating the values in the i-th dimension of the word vectors, the workflow of the DBSCAN clustering algorithm is outlined as follows:
  • Randomly select a word vector q as the object, defining its neighborhood as E q , and compute the cosine distance values between it and other word vectors p.
  • If D p , q < ε , categorize the word vector p into E q . If D p , q > ε , ignore the word vector p.
  • Tally the number of word vectors in E q . If c o u n t E q > m i n p t s , designate E q as a cluster and recursively process other word vectors in the same manner. Otherwise, label the word vector as noise data.
Here, ε represents the scanning radius distance, and m i n p t s stands for the minimum number of enclosed points. Both are selectable parameters.

3.4. Chinese Argot Recognition Work

The CSRMECT model, tailored to the structural similarities between Chinese argot vocabulary and reference terms, as well as the contextual disparities with the original meanings of argot vocabulary, involves a two-stage process. This process incorporates an enhanced MECT model for the fusion of contextual and Chinese character structural features in word vector representation and utilizes the DBSCAN clustering algorithm for the discovery of semantically inconsistent argot vocabulary in the semantic space. The specific workflow is illustrated in Figure 8.
  • Firstly, normal corpora and argot corpora (dataset containing argots) are amalgamated into datasets. The CSRMECT model extracts lattice embedding vectors and Chinese radical-level embedding vectors from sentences, followed by a fusion operation. Subsequently, through context encoding, word vector representations for each Chinese vocabulary in the sentence are obtained.
  • All word vectors derived from the processed normal corpus dataset N are mapped to a high-dimensional space. The DBSCAN clustering algorithm is employed to partition various clusters, yielding the core cluster vector set N ( W ) for each vocabulary W in dataset N.
  • For the argot corpus dataset M, all word vectors are similarly mapped to a high-dimensional space, resulting in the high-dimensional word vector M ( W ) for each vocabulary W in dataset M.
  • In the vocabulary list M L i s t of argot corpus M, the label list is computed as L M = [ Label W i | W i M L i s t ] , where Label W i is determined as follows:
    Label ( W i ) = 1 , if q N ( W i ) : D ( M ( W i ) , q ) > ε 0 , if q N ( W i ) : D ( M ( W i ) , q ) ε
  • At this time, in the list L M , the words marked 1 are argots, and vice versa for normal words.
It is worth mentioning that in this code word recognition framework, the MECT model in the above process exists as a natural language deep learning model. This means that when a better model appears in the future, that model will be able to replace the MECT model here, allowing the clustering algorithm to obtain better word vectors. We believe that the rapid development in the field of deep learning will directly promote the progress of the field of argot recognition through this framework.

4. Argot Interpretation Based on Large Language Models

4.1. Large Language Models

Since the emergence of ChatGPT in 2022 [40], a plethora of research has surfaced regarding the integration of large language models with various classical machine learning tasks to enhance their effectiveness. Yang et al. introduced the PICa few-shot prompting method, applying large language models as knowledge bases in the VQA domain [41]. Building upon PICa, Shao et al. proposed the Prophet few-shot prompting method, achieving commendable performance in VQA by leveraging answer heuristics to prompt GPT-3 [42]. OpenAI’s experiments indicate that simply scaling up language models significantly improves their performance in NLP tasks, such as knowledge-based QA and language understanding [40]. Chen et al. constructed the Codex programming assistance tool based on large language models and prompt engineering, addressing 70.2% of programming problems in testing [43]. Sun et al. tested ChatGPT’s retrieval capability with successful outcomes [44].
Through resource-intensive training, large language models embed a substantial amount of prior knowledge from the corpus into their parameters. Consequently, large language models can function as knowledge engines, providing external knowledge to enhance task performance across various machine learning tasks. For specific domains, fine-tuning the model with domain-specific texts significantly enhances its understanding of that domain, thereby improving task performance.

4.2. Argot Interpretation Based on Large Language Models

Building on the aforementioned analysis, this paper employs large language models for argot interpretation. Specifically, to investigate the feasibility of using large language models for argot interpretation, this study leaves argot vocabulary blank in MNGG. Through the prompt engineering, syntactic information and cue words are conveyed to the large language model. The vast prior knowledge acquired during the pretraining of the large language model is utilized for the task of argot interpretation.

5. Experimental Process

5.1. Datasets and Parameters

5.1.1. MNGG Argot Recognition Evaluation Dataset

Building upon the achievements of the dogwhistle dataset in the work by Xu et al. [8], this paper integrates argot corpora to create the MNGG (Mystique Naming Glossary Gathering) dataset. The task of argot recognition is transformed into a sequence labeling task for training and testing. From the Insider and Outsider subtasks of the dogwhistle dataset, the paper extracts pairs of argots and referential terms, resulting in a total of 1684 annotated argot-referential term pairs. Leveraging these argot pairs, the paper utilizes Chinese text data from THUC News [45] as the base text and replaces referential terms in the base text with argot words from the pairs. After this replacement operation is completed, we use the BIO sequence annotation method to mark the argots in the sequence. As shown in Figure 9, this process produces an annotated argot corpus with contextual information.
To facilitate the verification of the training data’s impact on model performance and to enable rapid testing with reduced data, the MNGG dataset also includes a clipped subset. This subset, denoted by the .clip file name suffix, represents a 10% extraction from the complete dataset. The number of corpora in the MNGG dataset is presented in Table 2.

5.1.2. Enhanced Base Corpora Utilizing Wikipedia and Large Language Models

In order to obtain clustering results for normal text, facilitating subsequent labeling of the argot dataset using clustering algorithms, this paper leverages Wikipedia text to establish a base corpus. Additionally, the paper employs prompt engineering to enhance the base corpus with a massive text dataset, effectively addressing the presence of vocabulary beyond the base corpus in the argot dataset. Combined with the CSRMECT model, this paper generates a substantial collection of high-dimensional word vectors based on words present in the normal corpus of the base dataset. An overview of the base corpora is provided in Table 3.
Here, the inclusion rate in argot denotes the ratio of the number of words in the argot corpus that appear in the base corpus to the total number of words in the argot corpus. By utilizing large language models and iteratively invoking prompts as illustrated in Figure 10, this paper cleans and organizes the obtained data to construct the LLM enhancement corpus, enhancing the inclusion rate of argot vocabulary and the occurrence frequency of argot terms.

5.2. Argot Recognition Experiment

Metric Calculation Based on BIO-Format Sequence Labeling

The BIO annotation scheme is a labeling method introduced and utilized in the field of named entity recognition (NER), indicating whether words or tokens in the labeled sequence belong to an entity. It has become a common annotation scheme in the field of natural language processing. Its design aims to distinguish the beginning (B: Beginning), interior (I: Inside), and non-entity (O: Outside) parts of an entity.
Calculating metrics for BIO-annotated sequences involves comparing the similarity between predicted and actual sequences. Specifically, given the predicted and actual sequences, the ranges of labeled entities in both sequences can be statistically determined. Subsequently, P r e c i s i o n , R e c a l l , and the F 1 are computed to evaluate the effectiveness of the predicted sequence. These metrics are calculated using the following formulas:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
Here, T P represents the number of correctly predicted positive instances, F P denotes the number of incorrectly predicted positive instances, and F N stands for the number of positive instances that were not predicted.

5.3. Experimental Results

In the work by Zhang et al. [14], the SVM classifier, enriched with additional feature extraction from text, achieved notable recognition rates in the field of argot recognition. This study replicates the SVM classifier mentioned in the literature, as the classifier is applied and tested on the MNGG dataset. MNGG transforms the argot recognition experiment into a sequence labeling algorithm. Consequently, this paper explores various classical sequence labeling algorithms, annotates the MNGG dataset using the BIO format, and tests the algorithmic performance. The comparative effectiveness of these algorithms with the proposed CSRMECT model is presented in Table 4.

5.4. Argot Interpretation Experiment

To investigate the feasibility of using large language models as knowledge engines for argot interpretation, this study conducted argot interpretation experiments based on the MNGG dataset, which contains a total of 1684 pairs of argots. GPT-3, GPT-4, and prompt engineering were employed in the experiments.
The specific experimental procedure is as follows:
  • Split all texts in the MNGG dataset into sentences. Extract sentences containing only one argot from the split corpus, denoted as the corpus W.
  • For each sentence W i in the corpus, extract the argot w o r d i from it. Randomly select T 1 words from the argot vocabulary of the MNGG dataset, forming a list of prompt words L s t i .
  • Utilizing the prompt engineering approach shown in Figure 11, input both syntactic information and the prompt word list into the large language model, obtaining the judgment corpus P i .
  • Tokenize P i to obtain the tokenized vocabulary set P i ¯ .
  • For similarity measurement, train a word2vec model using the Wiki + LLM base corpus. Let the word vector for the vocabulary X in the word2vec model be X = w o r d 2 v e c ( X ) .
  • For the i-th sentence W i and its judgment corpus P i ¯ , if there exists a vocabulary p i , j P i ¯ satisfying e p s w o r d 2 v e c ( p i , j ) · w o r d 2 v e c ( w o r d i ) | | w o r d 2 v e c ( p i , j ) | | · | | w o r d 2 v e c ( w o r d i ) | | , the i-th argot recognition is considered successful; otherwise, it fails.
  • For all sentences W i in the corpus W, calculate its accuracy a c c = c o u n t ( W s u c c e s s ) c o u n t ( W ) .
The experimental results are presented in Table 5.

6. Analysis and Discussion

6.1. Discussion on Argot Recognition Experiments

The effectiveness of the CSRMECT model is contingent upon prior conditions, such as the size of the base corpora and the model parameter settings. This section provides a discussion of such issues.

6.1.1. Qualitative Analysis of Clustering Results

The CSRMECT model is employed in this study to obtain word vectors. To analyze the semantic richness of word vectors, the t-SNE dimensionality reduction algorithm [49] is employed. Quantitative analysis is performed on selected argot and non-argot words from certain base datasets. The results are illustrated in Figure 12.
In Figure 12, the term “客场” (can be translated into opponent’s field, away, etc.) is employed as an argot in every sentence it appears, whereas the term “自己” (can be translated into self, oneself, etc.) is a regular word. In the clustering results on the left side of the figure, the distance between the red points representing argots and the blue points representing normal words is about 10 to 100. In contrast, the distance between the red point and the blue point in the picture on the right is about 0.5 to 1. It is evident that there exists semantic differentiation in the spatial representation between argot and non-argot words. Through DBSCAN clustering, we can easily identify the vast majority of argots on the left side.

6.1.2. Analysis of Data Augmentation Effects

As the semantic nature of vocabulary reflects in the relative positioning within the word vector space, the usage of argot terms in the base corpus directly impacts the effectiveness of argot recognition. When there is a scarcity of argot terms or their usage is overly uniform, the overall algorithmic process may be compromised. This study employs an augmentation algorithm based on large language models, enhancing the diversity of argot term usage in the base corpus to improve evaluation outcomes. The cumulative information on successive base data augmentation and corresponding algorithmic improvements is presented in Table 6.

6.1.3. Sensitivity Analysis

The DBSCAN clustering algorithm involves two hyperparameters: m i n p t s and ε . Here, m i n p t s represents the minimum number of points within the same cluster, and ε indicates the size of the clustering boundary. To explore the model’s sensitivity to hyperparameters, including ε , and ensure experimental efficiency, word vectors need to be dimensionally reduced and then subjected to the DBSCAN clustering algorithm. Common dimensionality reduction algorithms include PCA [50], t-SNE, and others. In this study, a subset of vocabulary word vectors is selected, and different algorithms are employed to reduce the vectors to two dimensions for visualization, as depicted in Figure 13.
In the figure, the number of clusters before dimensionality reduction for the i-th row data is i. It is observed that the PCA, MDS [51], and lsomap algorithms show similar effects, while the t-SNE and UMAP [52] algorithms exhibit comparable effects and better performance.
Based on the above analysis, the PCA and t-SNE algorithms are selected for experimentation to investigate the model’s sensitivity to hyperparameters ε , dimensionality reduction algorithms, and reduced dimensions, as shown in Figure 14.
The experimental results indicate that around the optimal ε value, the stability of the F1-score is high. As ε deviates from the optimum, the F1-score gradually decreases. Additionally, due to the adoption of dimensionality reduction algorithms, while the model’s operational efficiency significantly improves, there is a slight reduction in model accuracy. Furthermore, compared to data augmentation methods, the choice of dimensionality reduction algorithms and reduced dimensions has a relatively minor impact on the F1-score.
For the vector aggregator module in CSRMECT, this study explores various implementation approaches and conducts sensitivity tests, as illustrated in Figure 15.
Here, “Sum” refers to adding multiple word vectors of a single word to form the word vector, while “Average” refers to summing and averaging multiple word vectors of a single word to obtain the word vector. In comparison to data augmentation methods, the modification of the vector aggregator implementation has a minimal impact on the results. However, it is notable that the model’s performance improves consistently across different parameters when using the Average algorithm. Hence, it cannot be ruled out that a more rational vector aggregator implementation could significantly enhance the model’s effectiveness.
Another noteworthy observation is that when calculating the rationality of word vectors under different vector aggregators using the information entropy mentioned earlier, the set scores obtained with the “Sum” method tend to be higher than those with the “Average” method, as shown in Table 7. This suggests that more rational word vectors indeed contribute to the improvement of model performance.

6.2. Discussion on Argot Vocabulary Interpretation Experiment

To analyze the recognition effectiveness of the LLMResolve framework, and to explore the strengths and limitations of large language models in the field of argot recognition, statistical and qualitative analyses are employed to discuss the experimental results.

Statistical Analysis

Different large language models exhibit varying performance; theoretically, utilizing more advanced models enhances the task of argot interpretation. A comparison is made between GPT-3.5 and GPT-4, as shown in Table 8.
The results indicate superior performance using GPT-4 compared to GPT-3.5. Thus, in LLMResolve, the performance of large language models significantly influences the accuracy of argot resolution. At the same time, we also found that for large language models, the performance of LLMResolve is not good when there is no hint word. However, from a certain perspective, this is also normal because even for humans, achieving this is very difficult.
POS-tagging is conducted using the jieba tool, with codes corresponding to the meanings detailed in Table 9.
Combining the output results of the LLMResolve framework, a statistical analysis is conducted on the common POS tags, yielding recognition quantities and rates for each POS, as illustrated in Figure 16.
The MNGG dataset proposed in this paper, in contrast to the Weibo dataset used in previous studies [12], differs in that the argot vocabulary in MNGG extends beyond a small subset of nouns like personal and location names. It encompasses various parts of speech, including nouns, verbs, adjectives, and adverbs. Analyzing the figure above reveals that the diversity of parts of speech poses certain challenges for argot interpretation work, with occasional recognition failures observed for verbs and adverbs. Simultaneously, as the most frequently occurring nouns (i.e., labeled as n, nr, ns), their recognition rate averages 82.4 % , comparatively lower within the spectrum of parts of speech recognition rates.

6.3. Qualitative Analysis

In conjunction with the foregoing, this paper conducts a qualitative analysis on failed instances of noun disambiguation. Failed samples extracted from the experimental results are presented in Figure 17.
Summarizing from the table, the following reasons for disambiguation failure can be delineated:
  • Insufficient Contextual Information: In the case of Sample 1, the phrase “蠢蠢的钟摆” (an animated pendulum) employs personification in Chinese rhetoric. Without ample contextual cues, both large language models and humans struggle to accurately discern the intended word for this context. For Sample 2, where both “天气” (weather) and “市场” (market) share the characteristic of change, additional context is essential for auxiliary reasoning.
  • High Similarity of Prompt Words: Illustrated by Sample 3, the words “鲜红色” (crimson) and “白色” (white) both represent colors, making it challenging for the model to distinguish their semantic differences within the sentence.
  • Triggering Safety Mechanism in Large Language Models: When sensitive terms appear in the prompt engineering, the safety mechanism of large language models is triggered. Consequently, the model refrains from providing an effective response and instead elaborates on the reason for refusing to answer. This phenomenon is particularly prevalent in the context of drug-related or explicit argots.

7. Conclusions

This study introduces, for the first time, the concept of utilizing semantic conflicts in argot vocabulary for argot recognition. Leveraging the MECT model, we propose the CSRMECT model for argot recognition and employ LLMResolve for argot interpretation. The proposed argot recognition and interpretation models surpass previous research efforts. Extensive experiments in this study provide insightful analyses of the model performance.
In terms of argot recognition, experiments indicate that improving the rationality of word vectorization methods enhances argot recognition. Furthermore, under the same vectorization algorithm, the similarity between argots and surrounding sentences also influences argot recognition effectiveness. Regarding argot interpretation, the outstanding performance of large language models validates their feasibility as knowledge engines for argot interpretation. Additionally, experiments demonstrate that more powerful models offer stronger background knowledge and better argot recognition capabilities.
For the future development of argot recognition models, a primary task is to investigate more rational word vectorization algorithms to expand the semantic space gap between argot and general vocabulary, thereby improving recognition rates. As for argot interpretation tasks, feasible future research directions include fine-tuning large language models using known argot repositories and exploring methods to bypass security mechanisms in large language models to enhance model response rates.

Author Contributions

Conceptualization, M.L. and X.L.; data curation, M.L. and M.H. Formal analysis, M.L. and M.L.; methodology, M.L.; investigation, M.L. and D.Y.; writing—original draft preparation, M.L.; writing—review and editing, X.L. and D.Y.; visualization, M.H. supervision, M.L. and X.L.; project administration, M.L. and X.L.; funding acquisition, D.Y. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the China University of Political Science and Law (CUPSL) Cybersecurity and Law Enforcement Technology Innovation Project (Grant No. 2023SYL07); Funding for subject innovation and talent introduction bases in colleges and universities (Supported by the ‘111 Center’) (B20087); Research on cross-domain multi-source video surveillance network security system as a national key research and development project (NO.2022YFC3301101).

Data Availability Statement

The research data supporting the reported results in this article are available on GitHub at the following link: https://github.com/Andrew82106/MNGG_Dataset (accessed on 24 November 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. China Internet Network Information Center (CNNIC). The 51st “China Internet Development Statistics Report”; Internet World; China Internet Network Information Center: Beijing, China, 2023; Volume 3, p. 3. [Google Scholar]
  2. Wang, X.; Zhao, Z. Research on the Composition and Investigation Methods about the Personnel Flowing of Fraud Crime in Communication Network. J. People’S Public Secur. Univ. China (Soc. Sci. Ed.) 2022, 38, 53–64. [Google Scholar]
  3. Luo, J.; Yang, M.; Ling, Z.; Wu, W.; Gu, X. Anonymous Communication and Darknet: A Survey. J. Comput. Res. Dev. 2019, 56, 103–130. [Google Scholar]
  4. Hu, C.; Liu, B.; Ye, Y.; Li, X. Fine-grained classification of drug trafficking based on Instagram hashtags. Decis. Support Syst. 2023, 165, 113896. [Google Scholar] [CrossRef]
  5. Yuan, K.; Lu, H.; Liao, X.; Wang, X. Reading Thieves’ Cant: Automatically Identifying and Understanding Dark Jargons from Cybercrime Marketplaces. In Proceedings of the 27th USENIX Security Symposium (USENIX Security 18), Baltimore, MD, USA, 15–17 August 2018; pp. 1027–1041. [Google Scholar]
  6. Ouyang, T.; Chen, Z.; Feng, R. An Initial Exploration of Drug Crime Implicit Language in Guangxi Region from the Perspective of Speech Recognition. J. Guangxi Police Coll. 2017, 30, 74–78. [Google Scholar]
  7. Ouyang, T.; Chen, Z.; Feng, R. Examination and Reflection on the Implicit Language of a Financial Crime Case in a Certain Railway. J. Railw. Police Coll. 2016, 26, 44–46. [Google Scholar] [CrossRef]
  8. Xu, C.; Zhou, W.; Ge, T.; Xu, K.; McAuley, J.; Wei, F. Blow the dog whistle: A Chinese dataset for cant understanding with common sense and world knowledge. arXiv 2021, arXiv:2104.02704. [Google Scholar]
  9. Ji, H.; Knight, K. Creative Language Encoding under Censorship. In Proceedings of the First Workshop on Natural Language Processing for Internet Freedom, Santa Fe, NM, USA, 20–26 August 2018; pp. 23–33. [Google Scholar]
  10. Shannon, C.E. A mathematical theory of communication. SIGMOBILE Mob. Comput. Commun. Rev. 2001, 5, 3–55. [Google Scholar] [CrossRef]
  11. Hsiung, P. Alias Detection in Link Data Sets. Ph.D. Thesis, Carnegie Mellon University, The Robotics Institute, Pittsburgh, PA, USA, 2004. [Google Scholar]
  12. Huang, H.; Wen, Z.; Yu, D.; Ji, H.; Sun, Y.; Han, J.; Li, H. Resolving entity morphs in censored data. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Sofia, Bulgaria, 4–9 August 2013; pp. 1083–1093. [Google Scholar]
  13. Zhang, B.; Huang, H.; Pan, X.; Ji, H.; Knight, K.; Wen, Z.; Sun, Y.; Han, J.; Yener, B. Be appropriate and funny: Automatic entity morph encoding. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Baltimore, MD, USA, 22–27 June 2014; pp. 706–711. [Google Scholar]
  14. Zhang, B.; Huang, H.; Pan, X.; Li, S.; Lin, C.Y.; Ji, H.; Knight, K.; Wen, Z.; Sun, Y.; Han, J.; et al. Context-aware entity morph decoding. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Beijing, China, 26–31 July 2015; pp. 586–595. [Google Scholar]
  15. Sha, Y.; Shi, Z.; Li, R.; Liang, Q.; Wang, B. Resolving Entity Morphs based on Character-Word Embedding. Procedia Comput. Sci. 2017, 108, 48–57. [Google Scholar] [CrossRef]
  16. You, J.; Sha, Y.; Liang, Q.; Wang, B. Morph Resolution Based on Autoencoders Combined with Effective Context Information. In Proceedings of the Computational Science—ICCS 2018, Wuxi, China, 11–13 June 2018; Shi, Y., Fu, H., Tian, Y., Krzhizhanovskaya, V.V., Lees, M.H., Dongarra, J., Sloot, P.M.A., Eds.; Springer International Publishing AG: Cham, Switzerland, 2018; pp. 487–498. [Google Scholar]
  17. Fan, Y. Research on the Detection Method of Drug-Related Hidden Codes Under the Background of “Internet +”. Netw. Secur. Technol. Appl. 2023. [Google Scholar]
  18. Dhuliawala, S.; Kanojia, D.; Bhattacharyya, P. SlangNet: A WordNet like resource for English Slang. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), Portorož, Slovenia, 23–28 May 2016; Calzolari, N., Choukri, K., Declerck, T., Goggi, S., Grobelnik, M., Maegaard, B., Mariani, J., Mazo, H., Moreno, A., Odijk, J., et al., Eds.; European Language Resources Association: Portorož, Slovenia, 2016; pp. 4329–4332. [Google Scholar]
  19. Wu, L.; Morstatter, F.; Liu, H. Slangsd: Building and using a sentiment dictionary of slang words for short-text sentiment classification. arXiv 2016, arXiv:1608.05129. [Google Scholar] [CrossRef]
  20. Durrett, G.; Kummerfeld, J.K.; Berg-Kirkpatrick, T.; Portnoff, R.; Afroz, S.; McCoy, D.; Levchenko, K.; Paxson, V. Identifying Products in Online Cybercrime Marketplaces: A Dataset for Fine-grained Domain Adaptation. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 9–11 September 2017. [Google Scholar] [CrossRef]
  21. Wilson, S.; Magdy, W.; McGillivray, B.; Garimella, K.; Tyson, G. Urban Dictionary Embeddings for Slang NLP Applications. In Proceedings of the Twelfth Language Resources and Evaluation Conference, Marseille, France, 11–16 May 2020; Calzolari, N., Béchet, F., Blache, P., Choukri, K., Cieri, C., Declerck, T., Goggi, S., Isahara, H., Maegaard, B., Mariani, J., et al., Eds.; European Language Resources Association: Marseille, France, 2020; pp. 4764–4773. [Google Scholar]
  22. Kolla1, A.; Ilievski1, F.; Sandlin, H.A. A Study of Slang Representation Methods. arXiv 2022, arXiv:2212.05613. [Google Scholar]
  23. Belkin, M.; Niyogi, P.; Sindhwani, V. Manifold Regularization: A Geometric Framework for Learning from Labeled and Unlabeled Examples. J. Mach. Learn. Res. 2006, 7, 2399–2434. [Google Scholar]
  24. Bengio, Y.; Schwenk, H.; Senécal, J.S.; Morin, F.; Gauvain, J.L. Neural Probabilistic Language Models. In Innovations in Machine Learning: Theory and Applications; Holmes, D.E., Jain, L.C., Eds.; Springer Berlin Heidelberg: Berlin/Heidelberg, Germany, 2006; pp. 137–186. [Google Scholar] [CrossRef]
  25. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed Representations of Words and Phrases and their Compositionality. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; Burges, C., Bottou, L., Welling, M., Ghahramani, Z., Weinberger, K., Eds.; Curran Associates, Inc.: Granada, Spain, 2013; Volume 26. [Google Scholar]
  26. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  27. Wu, S.; Song, X.; Feng, Z. MECT: Multi-metadata embedding based cross-transformer for Chinese named entity recognition. arXiv 2021, arXiv:2107.05418. [Google Scholar]
  28. Liu, P.; Guo, Y.; Wang, F.; Li, G. Chinese named entity recognition: The state of the art. Neurocomputing 2022, 473, 37–53. [Google Scholar] [CrossRef]
  29. Jin, Z.; He, X.; Wu, X.; Zhao, X. A hybrid Transformer approach for Chinese NER with features augmentation. Expert Syst. Appl. 2022, 209, 118385. [Google Scholar] [CrossRef]
  30. Li, X.; Yan, H.; Qiu, X.; Huang, X. FLAT: Chinese NER Using Flat-Lattice Transformer. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; Jurafsky, D., Chai, J., Schluter, N., Tetreault, J., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 6836–6842. [Google Scholar] [CrossRef]
  31. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  32. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Pereira, F., Burges, C., Bottou, L., Weinberger, K., Eds.; Curran Associates, Inc.: Granada, Spain, 2012; Volume 25. [Google Scholar]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Granada, Spain, 2017; Volume 30. [Google Scholar]
  34. Zhang, Y.; He, C.; Wang, Z. Research on Enterprise Competitiveness Factor Analysis Combining Semantic Clustering. Data Anal. Knowl. Discov. 2012, 9, 49–55. [Google Scholar]
  35. Li Nan, W.B. Recognition and Visual Analysis of Interdisciplinary Semantic Drift. Data Anal. Knowl. Discov. 2023, 7, 15–24. [Google Scholar]
  36. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the KDD, Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  37. Arbelaitz, O.; Gurrutxaga, I.; Muguerza, J.; Pérez, J.M.; Perona, I. An extensive comparative study of cluster validity indices. Pattern Recognit. 2013, 46, 243–256. [Google Scholar] [CrossRef]
  38. Zhou, A.; Zhou, S.; Cao, J.; Fan, Y.; Hu, Y. Approaches for scaling DBSCAN algorithm to large spatial databases. J. Comput. Sci. Technol. 2000, 15, 509–526. [Google Scholar] [CrossRef]
  39. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. Clustering for mining in large spatial databases. KI 1998, 12, 18–24. [Google Scholar]
  40. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–12 December 2020; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Granada, Spain, 2020; Volume 33, pp. 1877–1901. [Google Scholar]
  41. Yang, Z.; Gan, Z.; Wang, J.; Hu, X.; Lu, Y.; Liu, Z.; Wang, L. An empirical study of gpt-3 for few-shot knowledge-based vqa. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 3081–3089. [Google Scholar]
  42. Shao, Z.; Yu, Z.; Wang, M.; Yu, J. Prompting large language models with answer heuristics for knowledge-based visual question answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 14974–14983. [Google Scholar]
  43. Chen, M.; Tworek, J.; Jun, H.; Yuan, Q.; Pinto, H.P.D.O.; Kaplan, J.; Edwards, H.; Burda, Y.; Joseph, N.; Brockman, G.; et al. Evaluating Large Language Models Trained on Code. arXiv 2021, arXiv:2107.03374. [Google Scholar]
  44. Sun, W.; Yan, L.; Ma, X.; Ren, P.; Yin, D.; Ren, Z. Is ChatGPT Good at Search? Investigating Large Language Models as Re-Ranking Agents. arXiv 2023, arXiv:2304.09542. [Google Scholar]
  45. Sun, M.; Li, J.; Guo, Z.; Zhao, Y.; Zheng, Y. THUCTC: An Efficient Chinese Text Classifier. 2016. Available online: https://github.com/thunlp/THUCTC (accessed on 24 November 2023).
  46. Gui, T.; Zou, Y.; Zhang, Q.; Peng, M.; Fu, J.; Wei, Z.; Huang, X. A Lexicon-Based Graph Neural Network for Chinese NER. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; Inui, K., Jiang, J., Ng, V., Wan, X., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 1040–1050. [Google Scholar] [CrossRef]
  47. Zhang, Y.; Yang, J. Chinese NER Using Lattice LSTM. arXiv 2018, arXiv:1805.02023. [Google Scholar]
  48. Gui, T.; Ma, R.; Zhang, Q.; Zhao, L.; Jiang, Y.G.; Huang, X. CNN-Based Chinese NER with Lexicon Rethinking. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; Volume 2019. [Google Scholar]
  49. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  50. Wold, S.; Esbensen, K.; Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  51. Torgerson, W.S. Multidimensional scaling: I. Theory and method. Psychometrika 1952, 17, 401–419. [Google Scholar] [CrossRef]
  52. McInnes, L.; Healy, J.; Melville, J. UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv 2020, arXiv:1802.03426. [Google Scholar]
Figure 1. Examples of usage of argots.
Figure 1. Examples of usage of argots.
Entropy 26 00321 g001
Figure 2. Word embedding example.
Figure 2. Word embedding example.
Entropy 26 00321 g002
Figure 3. Comparison of word-embedding effects.
Figure 3. Comparison of word-embedding effects.
Entropy 26 00321 g003
Figure 4. MECT workflow.
Figure 4. MECT workflow.
Entropy 26 00321 g004
Figure 5. MECT Lattice Embedding.
Figure 5. MECT Lattice Embedding.
Entropy 26 00321 g005
Figure 6. Cross-transformer layer.
Figure 6. Cross-transformer layer.
Entropy 26 00321 g006
Figure 7. CSRMECT model.
Figure 7. CSRMECT model.
Entropy 26 00321 g007
Figure 8. Specific workflow of argot discovery.
Figure 8. Specific workflow of argot discovery.
Entropy 26 00321 g008
Figure 9. Build MNGG dataset.
Figure 9. Build MNGG dataset.
Entropy 26 00321 g009
Figure 10. Prompt for LLM enhancement corpus.
Figure 10. Prompt for LLM enhancement corpus.
Entropy 26 00321 g010
Figure 11. Inputting syntactic information and prompt word list into the large language model.
Figure 11. Inputting syntactic information and prompt word list into the large language model.
Entropy 26 00321 g011
Figure 12. Dimensionality reduction visualization.
Figure 12. Dimensionality reduction visualization.
Entropy 26 00321 g012
Figure 13. Comparison of dimensionality reduction algorithms.
Figure 13. Comparison of dimensionality reduction algorithms.
Entropy 26 00321 g013
Figure 14. Sensitivity analysis for ε .
Figure 14. Sensitivity analysis for ε .
Entropy 26 00321 g014
Figure 15. Sensitivity analysis of vector aggregator.
Figure 15. Sensitivity analysis of vector aggregator.
Entropy 26 00321 g015
Figure 16. Argot interpretation effect POS analysis.
Figure 16. Argot interpretation effect POS analysis.
Entropy 26 00321 g016
Figure 17. Failed examples of argot interpretation in LLMResolve.
Figure 17. Failed examples of argot interpretation in LLMResolve.
Entropy 26 00321 g017
Table 1. Character structure decomposition table.
Table 1. Character structure decomposition table.
Chinese CharacterCRHTSC
麻 (numb)广 (wide)广林 (forest)广木木 (wood)
蠕 (worms)虫 (insect)虫需 (need)虫雨 (rain) 而 (but)
挂 (hang)扌 (hand)扌圭 (Gui)扌土 (earth) 土
唱 (sing)口 (mouth)口昌 (thriving)口曰 (speak) 曰
Table 2. Overview of the MNGG dataset.
Table 2. Overview of the MNGG dataset.
DatasetNumber of SentencesNumber of Argot VocabularyAverage Argot Vocabulary per Sentence
train.clip.bio56411161.97
test.clip.bio3387052.09
dev.clip.bio2254702.08
train.bio564511,5952.05
test.bio338773152.15
dev.bio225847512.10
Table 3. Overview of base corpora.
Table 3. Overview of base corpora.
Base CorpusNumber of Entries (Sentences)Inclusion Rate in ArgotAverage Occurrence Frequency of Argot Vocabulary
Wiki194,7490.71103.4
LLM159,2770.453619.9
Wiki + LLM354,0260.991685.5
Table 4. Comparison of argot recognition models.
Table 4. Comparison of argot recognition models.
ModelF1PrecisionRecall
SVM [14]0.080.080.08
LGN [46]0.030.780.02
Lattice-LSTM [47]0.150.620.08
LR-CNN [48]0.230.640.14
CSRMECT0.330.350.31
Table 5. Argot interpretation experiment results.
Table 5. Argot interpretation experiment results.
ModelAccuracy
Huang2013 [12]0.364
Zhang2015 [14]0.383
Sha2017(Acc@20) [15]0.870
LLMResolve0.919
LLMResolve (GPT-4+10 Prompt Words)0.824
LLMResolve (GPT-4+3 Prompt Words)0.919
Table 6. Comparison of base data augmentation information and effects.
Table 6. Comparison of base data augmentation information and effects.
Base CorpusArgot LossF1PR
Wiki31450.040.220.02
LLM *19180.270.570.18
LLM **4080.320.360.28
LLM ***1520.330.350.31
LLM * denotes augmentation once, LLM ** denotes augmentation twice, and LLM *** denotes augmentation three times.
Table 7. Evaluation results of using vector information entropy for different vector aggregators.
Table 7. Evaluation results of using vector information entropy for different vector aggregators.
WordAverageSum
我们 (we)2.9439.643
我 (I)2.70710.223
一种 (a kind of)2.8878.224
方法 (method)3.0406.640
Average entropy value in dataset2.9023.455
Table 8. Experimental results of code interpretation under different large language models.
Table 8. Experimental results of code interpretation under different large language models.
ModelTop-kAccuracy
GPT-430.919
GPT-4100.824
SOTA-0.919
GPT-3.530.768
GPT-3.550.741
GPT-3.5100.648
GPT-3.5200.537
GPT-3.5300.454
GPT-3.500.133
SOTA-0.768
Table 9. Chinese POS tags and meanings.
Table 9. Chinese POS tags and meanings.
POS TagMeaningDetailed Meaning
nNounRepresents people, things, places, etc.
vVerbIndicates action, state, or behavioral existence.
dAdverbUsed to modify verbs, adjectives, other adverbs, etc.
aAdjectiveDescribes the qualities or states of things.
vnNoun-VerbSometimes represents a mixture of nouns and verbs, typically used as a noun.
cConjunctionConnects words, such as “and”, “or”, etc.
nrNameRepresents personal names.
tTime WordRepresents words related to time.
nsPlace NameRepresents names of places.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, M.; Li, X.; Hu, M.; Yuan, D. Research on a Framework for Chinese Argot Recognition and Interpretation by Integrating Improved MECT Models. Entropy 2024, 26, 321. https://doi.org/10.3390/e26040321

AMA Style

Li M, Li X, Hu M, Yuan D. Research on a Framework for Chinese Argot Recognition and Interpretation by Integrating Improved MECT Models. Entropy. 2024; 26(4):321. https://doi.org/10.3390/e26040321

Chicago/Turabian Style

Li, Mingfeng, Xin Li, Mianning Hu, and Deyu Yuan. 2024. "Research on a Framework for Chinese Argot Recognition and Interpretation by Integrating Improved MECT Models" Entropy 26, no. 4: 321. https://doi.org/10.3390/e26040321

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop