Next Article in Journal
Experimental Study on the Impact Dynamics of Cylindrical Baffles with a Rigid Barrier against Debris Flows
Next Article in Special Issue
Spoken Language Identification System Using Convolutional Recurrent Neural Network
Previous Article in Journal
Performance Evaluation of Machine Learning Methods for Anomaly Detection in CubeSat Solar Panels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computational Linguistics with Deep-Learning-Based Intent Detection for Natural Language Understanding

1
Department of Applied Linguistics, College of Languages, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
Department of Computer Sciences, College of Computing and Information System, Umm Al-Qura University, Mecca 24382, Saudi Arabia
3
Department of Computer Science, College of Computing and Information Technology, Shaqra University, Shaqra 11961, Saudi Arabia
4
Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 22254, Saudi Arabia
5
Department of Information Systems and Technology, College of Computer Science and Engineering, University of Jeddah, Jeddah 21589, Saudi Arabia
6
Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam Bin Abdulaziz University, Al-Kharj 16278, Saudi Arabia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(17), 8633; https://doi.org/10.3390/app12178633
Submission received: 5 August 2022 / Revised: 23 August 2022 / Accepted: 25 August 2022 / Published: 29 August 2022
(This article belongs to the Special Issue Recent Trends in Natural Language Processing and Its Applications)

Abstract

:
Computational linguistics explores how human language is interpreted automatically and then processed. Research in this area takes the logical and mathematical features of natural language and advances methods and statistical procedures for automated language processing. Slot filling and intent detection are significant modules in task-based dialogue systems. Intent detection is a critical task in any natural language understanding (NLU) system and constitutes the base of a task-based dialogue system. In order to build high-quality, real-time conversational solutions for edge gadgets, there is a demand for deploying intent-detection methods on devices. This mandates an accurate, lightweight, and fast method that effectively operates in a resource-limited environment. Earlier works have explored the usage of several machine-learning (ML) techniques for detecting intent in user queries. In this article, we propose Computational Linguistics with Deep-Learning-Based Intent Detection and Classification (CL-DLBIDC) for natural language understanding. The presented CL-DLBIDC technique receives word embedding as input and learned meaningful features to determine the probable intention of the user query. In addition, the presented CL-DLBIDC technique uses the GloVe approach. In addition, the CL-DLBIDC technique makes use of the deep learning modified neural network (DLMNN) model for intent detection and classification. For the hyperparameter tuning process, the mayfly optimization (MFO) algorithm was used in this study. The experimental analysis of the CL-DLBIDC method took place under a set of simulations, and the results were scrutinized for distinct aspects. The simulation outcomes demonstrate the significant performance of the CL-DLBIDC algorithm over other DL models.

1. Introduction

With the development of the task-based dialogue mechanism, natural language understanding (NLU), as a critical element of the task-based dialogue system, has gained more interest among researchers [1]. An author could capture context data for identifying the intent of a user by utilizing intellectual, interactive gadgets which talk to humans, in various cases, and then derive the semantic constituents from the text that the end-user inputs into semantic slots which have been previously defined [2]. Two modules, such as slot filling and intent detection, could transform the text into a semantic representation that offers the task data to support the dialogue system and aids users in achieving their goals [3]. “Intent detection” refers to the task of categorizing natural language utterances into semantic intent classes that have previously been defined [4].
Finding message intent in natural language utterances becomes a critical task for conversational mechanisms [5]. In applications ranging from natural-language response production to offering intellectual proposals, understanding the primary intent of a context interaction becomes critical. Conventionally, research on intent detection has focused on this task with the supposition that inference and training are executed using cloud infrastructure or well-equipped servers [6,7]. This results in the inappropriateness of prevailing machine-learning (ML) techniques for real-time dialogue mechanisms in low-resource edge devices because of its higher latency and dependence on large, pre-trained methods. Recently, there has been greater commercial and academic interest in guiding artificial intelligence (AI) solutions that could work in a straightforward manner on a device using local data [8,9]. On-device AI methods are capable of supporting intent detection in real-time at a low latency and aid in guarding the privacy of delicate user data, such as smartphone messages.
Slot filling and intent detection have always been implemented as the pipeline modules in research on conventional, task-based dialogue systems [10]. However, pipeline techniques propagate mistakes effortlessly. In contrast with pipeline techniques, joint methods have the benefit of using the dependency between slots and intent. Such works have been primarily classified into two classes: the use of semantic analysis and a joint model that incorporates both classification and feature design into the learning process [11]. However, few studies have used additional information, such as the incidence relation and the shared resources between slot-filling and intent-detection modules. However, the above-mentioned studies contain a typical issue: that is, the high-quality and standardized public datasets are highly inadequate because labeling and collecting datasets require more effort and time [12]. Thus, researchers have tried to include an external knowledge base in the prevailing datasets so as to solve the issue of data scarcity. A knowledge base can be engineered into a knowledge cluster, and thus it becomes structured, easy to use and operate, and provides comprehensive and organized knowledge [13]. Moreover, it is a set of interrelated knowledge pieces that are managed, stored, used, and organized in computer memory by some knowledge-representation mode to solve issues in certain domains. Recent research studies have shown that the impact of NLU is enhanced by the presentation of a knowledge base.
This article proposes Computational Linguistics with Deep-Learning-Based Intent Detection and Classification (CL-DLBIDC) for NLU. The CL-DLBIDC technique receives word-embedding as input and uses learned meaningful features to determine the probable intention of the user query. In addition, the CL-DLBIDC technique uses the GloVe approach. In addition, the CL-DLBIDC technique makes use of the deep learning modified neural network (DLMNN) model for intent detection and classification. For the hyperparameter tuning process, the mayfly optimization (MFO) algorithm has been used in this study. The experimental analysis of the CL-DLBIDC technique took place under a set of simulations, and distinct aspects of the results have been examined.

2. Related Works

Yan et al. [14] have presented a semantic-enhanced Gaussian mixture method (SEG) for unknown intent identification. In particular, the authors devised utterance embeddings that have a Gaussian mixture distribution, and they infused dynamic-class semantic data into Gaussian means that allow for the learning of more class-concentrated embeddings, which can be helpful in facilitating downstream outlier recognition. Coupled with a density-related, outlier-identification method, SEG has attained competitive outcomes on three real, task-based, dialogue datasets in two languages for unknown intent recognition. Gangadharaiah and Narayanaswamy [15] have examined an attention-related neural network (NN) method which executes multi-label categorization to identify many intents and generates labels for intents, as well as slot-labels, at the token-level. The authors displayed the existing performance for slot-label identification and intent detection through a comparison with strong, newly formulated methods.
Dopierre et al. [16] have presented ProtAugment, a meta-learning technique for short text classification (an intent-detection task). ProtAugment is a new extension of the Prototypical Network, which confines overfitting to the bias presented by a few-shots classifier objective at every episode. It depends on various paraphrasing. It is a conditional language method that has become the initial, fine-tuned language for paraphrasing, and diversity was later presented at the decoding phase of every meta-learning episode. Nigam et al. [17] have proposed a method for forecasting slots and the intent of queries for chatbots that will answer career-based queries. The authors considered a multistage technique in which both the process (slot-tagging and intent-classification) imitate one another’s decision-making at various levels. The method segregates the issue into stages, figuring out single issues, and transferring the related outcomes of the present stage to the following stage, thus minimizing the search space for consequent levels, and finally, it performs categorizations and tagging that are more viable after every stage.
In [18], the authors began from the nature of out-of-domain (OOD) intent classifications and predicted its optimizing objective. The authors even presented an effective technique, called k-nearest-neighbor (KNN) contrastive learning. This technique uses KNN for learning discriminatory semantic features that are highly conducive to OOD identification. [19] enhanced intent-detecting performance by using datasets that were gathered in the same field, but which have various intent groups. In the baseline mechanism, the authors used the existing, pre-trained transforming techniques with an intent-identification-classifier layer. This technique contains two fine-tuned stages. In the initial stage, the variables of the classifier and transformer layers were adapted utilizing an in-domain dataset to the targeted set. Then, the author retained the transformer parameters, which were updated, but replaced the classifier with a newer one designed for the targeted task. In the second stage, the network was finely tuned endwise, utilizing the targeted set.
Shen et al. [20] modeled an innovative domain-regularized module (DRM) for reducing the overconfident phenomenon of a vanilla classifier, attaining superior generalization in both cases. In addition, DRM was utilized as a drop-in replacement for the final layer in any NN-related intent classifier, offering a low-cost method for an important enhancement. Qin et al. [21] introduced an Adaptive Graph-Interactive Framework (AGIF) for joint, multiple-intent recognition and slot-filling, where the authors introduced an intent–slot graph communication layer for modeling strong relations among the intents and slots. Such communication layers can be implemented for every token adaptively, which has the benefit of mechanically deriving the related intent data, forming a finely grained intent data compilation for token-level slot estimation.

3. The Proposed Model

This study presents a novel CL-DLBIDC approach to detecting and classifying intents for natural language understanding. The presented CL-DLBIDC technique receives word embedding as input and uses learned meaningful features to determine the probable intention of the user query. In addition, the CL-DLBIDC technique uses the GloVe approach. In addition, the CL-DLBIDC technique makes use of the DLMNN model for intent detection and classification. For the hyperparameter tuning process, the MFO algorithm was used in this study. Figure 1 demonstrates the overall block diagram of the CL-DLBIDC method.

3.1. Word Embedding

Here, the CL-DLBIDC technique uses the GloVe approach. The GloVe (global vector) model exploits global co-occurrence matrixes to generate representations of word vectors [22]. This study focuses on the co-occurrence of words in the corpora. Hence, words which have the same context would be close to one another. Because antonymous words share similar contexts and they close to one another, the output can be semantically wrong. This kind of semantic incorrectness of the vector space must be solved in order to achieve good outcomes. The first embedding is enriched by a somewhat-adapted form of an enrichment method. The resultant vector space will be semantically correct when compared to the first one. Consider that v = { v 1 ,   v 2 ,   v n } represents the group of word vectors, where v i indicates the vector for j t h words in a vocabulary. Assume A ,   S ,   R , and H denote the collection of indices for pairs of antonymous, synonymous, related, and hyponymous or hypernymous words, correspondingly. For instance, if ( i ,   j ) is a component of S , later i t h and j t h words in the vocabulary are synonyms. The first vector space y is adapted through objective function and the set (constraint) to attain a novel vector space V , where V = { v 1 ,   v 2 ,   v n } . The formula of objective functions is given in the following section.

3.1.1. Enrichment Using Synonyms

Equation (1) brings synonymous words close to one another in the vector space by improving the cosine similarities among them.
S E ( V ) = ( i , j ) ε S   τ ( δ cos ( v i ,   v j ) )
S represents the collection of pairs of synonyms in a vocabulary, and it involves synonyms from WordNet and equivalence relations. δ   is fixed to one and τ ( x ) =   max   ( O ,   x ) .

3.1.2. Enrichment Using Antonyms

Equation (2) moves antonymous words farther from one another by decreasing the cosine similarities among them to 0 or less.
A E ( V ) = ( i , j ) ε A   τ ( cos ( v i ,   v j ) )
A represents the collection of pairs of antonyms that involve an exclusion relation from PPDB and antonyms from WordNet.

3.1.3. Enriching with Related Words

Related words have a similar sense; however, they are not synonyms in a lexicon (e.g., “ride” vs. “commute”). Bringing related words close to one another in a vector space would be useful because they have similar senses. The objective function shown in Equation (3) raises cosine similarities among related words to 0 or greater than 0.
R E ( V ) = ( i , j ) ε R   τ ( δ cos ( v i ,   v j ) )  
R denotes the group of related words from The Macmillan Dictionary, and δ was fixed to 0 . The value of δ guarantees that similarities among related words are 0 or greater than 0. This ensures that related words are not as close to one another as synonyms are.

3.1.4. Enriching with Hypernyms and Hyponyms

Hypernymous terms have wider meanings, whereas hyponymous terms have particular meanings. A hypernym–hyponym is a pair of common terms and its particular categories (e.g., “book”–“recipe book”). The cosine similarities between these words must be 0 or greater than 0. The objective function provided in Equation (4) is the same as that is Equation (3), and it is utilized for achieving this.
H E ( V ) = ( i , j ) ε H   τ ( δ cos ( v i ,   v j ) )  
H is the group of hypernym–hyponym pairs in vocabulary V, with a value of δ as 1 and a path length of 1 in WordNet.

3.1.5. Regularization

When a word vector is adapted by employing the above for constraint, it becomes distinct from the prior, neighboring vector. It alters the semantics of embedding, and regularization must be performed to retain the meaning. The converted vector space must be bent to the original vector space to conserve the meaning. The objective function provided in Equation (5) keeps cosine similarities among its neighbors and an initial word vector greater than or equivalent to the cosine similarities among them.
V S   P ( V ,   V ) = ( i , j ) ε A M j ε N ( i )   τ ( cos ( v i ,   v j ) cos ( v i ,   v j ) )  
M denotes the word count in vocabulary, and N ( i ) denotes the group of neighbors of i t h in the first embedding space. N ( i ) involves the word in V whose cosine similarities among i t h words is higher than 0.8.
The last objective function is a linear integration of the abovementioned terms, as shown in Equation (6).
E ( V ,   V ) = S E ( V ) + A E ( V ) + R E ( V ) + H E ( V ) + V S P ( V ,   V )  
The objective function is minimalized by the stochastic gradient descent (SGD) approach for transformed vector V and initial vector V. SGD ran for 20 epochs with a learning rate of 0.1.

3.2. Intent Detection Using the DLMNN Model

At this stage, the CL-DLBIDC technique makes use of the DLMNN model for intent detection and classification. In the DLMNN model, all of the inputs are offered to separate nodes that are common from the input region of classification [23]. The weight signifies a randomly allocated value and can be connected with all of the inputs. The following is the Hidden Layer (HL). The nodes in this HL are termed “hidden nodes”, and they apply the function of summing the product of input values and the weighted vector of every input node which is connected to it.
In DLMNN, the weighted value is optimized by employing the MFO algorithm and DLMNN. An arbitrary weighted value offers a further backpropagation (BP) procedure for achieving the outcome in an optimized manner. The activation function is then executed, and the resultant layer is carried to following layer. This weight is an utmost control on the classification result. The algorithmic steps of DLMNN classification are as follow:
Step 1: Allocate score value of selective features and its corresponding weight as:
E i = { E 1 , E 2 , E 3 E n } ,  
W i = { W 1 ,   W 2 ,   W 3 W n } ,  
Step 2: Multiply the input with the weighted vectors, which are arbitrarily chosen, and then, it can be entirely summed as:
R = i = 1 n E i W i ,  
where R refers to the summed value, E i denotes the input entropy value, and W i stands for the weight value.
Step 3: Estimate the activation function (AF), employing
A F i = f ( i = 1 n E i W i ) ,  
Step 4: Measure the resultant HL as:
Y i = A i +     G i W i ,  
where A i defines the bias value, W i demonstrates the weight between the input and HLs, and G i refers to the values that are changed by application of A F .
Step 5: Re-perform the earlier three steps on all of the layers of the DLMNN. Last, estimate the resultant unit by adding up every input signal weight to attain the resultant layer neuron value,
O u i = A i +     P i W i ,  
where P i signifies the value of the layer which leads to the output 1, W j represents the weight of HL, and O u i demonstrates the output unit.
Step 6: Contrast the network outcome with the target values. The variance among these two values produces the error signal. This value can be demonstrated mathematically as:
E r = T a i O u i ,  
where, E r represents the error signal, T a i indicates the aimed target output, and O u i stands for the classification current output.
Step 7: At this point, the resultant unit is weighted against the target values. Accordingly, the relative error is defined. According to this error, a value δ i has been calculated, and it can be utilized for allocating the error at the resultant layer back to every other unit on network:
δ i = E r [ f ( O u i ) ] ,
Step 8: The weighted correction is measured by employing the BP approach. This relation is:
w c i = β δ i ( E i ) ,  
where w c i defines the weighted correction, β signifies the momentum term, E i implies the input vector, and δ i stands for the error that is distributed from the network.

3.3. Parameter Tuning Using the MFO Algorithm

For the hyperparameter tuning process, the MFO algorithm was used in this study. The major idea of this method is based on the flight and mating behaviors of the mayfly (MF) [24]. It integrates the benefits of the firefly algorithm (FA), particle swarm optimization (PSO), and the genetic algorithm (GA). Initially, the MFO algorithm generates an MF population encompassing males and females. The existing velocity and location of i - t h MFs are two n -dimensional vectors; they are correspondingly represented as v i = ( v i 1 , v i 2 , v i n ) and x i = ( x i 1 , x i 2 , x i n ) . Every MF adapts its location based on its individual best ( p b e s t )   location and the optimal ( g b e s t )   location originated by the entire MF swarm.
Male MFs cluster together, which recommends that the position is upgraded on the basis of social and personal experience. x i t denotes the current location of i - t h male MFs at t iteration. Assume that x i t denotes the existing location of i - t h MFs at t iteration; the location is upgraded by adding velocity v i t , and it is shown below:
x i t + 1 = x i t + v i t .  
Assume that male MFs often implement their mating dance not distant from the water, and it is updated in the following equation:
v i j t + 1 = g v i j t + a 1 e β r p 2 ( p b e s t i j x i j t ) + a 2 e β r g 2 ( g b e s t j x i j t ) ,  
In Equation (17), β refers to the visibility coefficient, a 1 and a 2 , determined by the positive constant that represents the attraction. p b e s t i j indicates the optimal location attained using the i - t h male MFs in j dimension.  r p and r g indicate the Euclidean distance between χ j and p b e s t j , and between χ j and g b e s t , respectively. The gravity coefficient is denoted as g , which is a fixed number among [0, 1], or it is steadily lowered over the iteration, as follows:
g = g   max   g   max   g   min   i t e r   max   × i t e r ,  
In Equation (18), g   max   and g   min   correspondingly embody the minimum and maximum weights. The current and the overall number of iterations are represented as i t e r   and t e r   max   / . The personal optimal location p b e s t i at iteration t + 1 is defined as follows:
p b e s t i = { x i t + 1 , i f f ( x i t + 1 ) < f ( p b e s t i ) s a m e   a s   b e f o r e ,   o t h e r w i s e .  
The finest male MF continues implementing up and down movements at distinct velocities, and this equation is shown below:
v i j t + 1 = v i j t + d × r ,  
In Equation (20), the mating dance coefficient can be represented as d , and r denotes an arbitrary number within [ 1 , 1]. Female MFs do not cluster together, but they move to male MFs. y i t denotes the current location of the i - t h female MF at t iteration. The change of i - t h female MF in location is evaluated by Equation (21):
y i t + 1 = y i t + v i t .  
In the MFO algorithm, female and male MFs with similar individual fitness rankings attract one another, and the female MF’s location changes in terms of the position of the male MF with a similar ranking, as follows:
v i j t + 1 = { g v i j t + a 2 e β r m f 2 ( x i j t y i j t )   ,   i f   f ( y i ) > f ( X i ) g v i j t + f l × r ,   i f   f ( y i ) f ( x i )   ,  
where v i j t and y i j t denote the i - t h female MF’s velocity and location in j dimension at t   iteration. a 2 and β indicate the attraction constant and visibility coefficient, correspondingly. r m f refers to the Euclidean distance between the i - t h female MF and i - t h male MF, f l denotes a random walking coefficient that recommends that a female is not attracted to a male, and r implies an arbitrary number within [ 1 , 1]. Figure 2 illustrates the steps involved in the MFO technique.
The act of mating can be denoted as the crossover operator that offers the global search ability for the MFO algorithm. In the MFO algorithm, female and male MFs with similar individual fitness rankings mate with one another to generate offspring. The mating process of every pair of female and male MFs generates two offspring, and the crossover operator is given below:
{ γ 1 = L × m a l e + ( 1 L ) × f e m a l e γ 2 = L × f e m a l e + ( 1 L ) × m a l e ,  
In Equation (23), L implies an arbitrary number within [0, 1]. Furthermore, female and male denote the parents. First, the offspring velocity is fixed as 0.
The mutation operator gives partial local search ability to MFO algorithm. The selected offspring is upgraded by adding an arbitrary number to the parameter over this method; hence, the offspring is changed to
γ n ' = γ n + σ N n ( 0 , 1 ) ,  
In Equation (24), the standard deviation and uniform distribution are signified as σ and N n , correspondingly. Even though this operator provides the MFO algorithm with robust local search ability, d and f l should be slowly decreased. Both values are upgraded over the t iterations as follows:
d t = d 0 δ t  
f l t = f l 0 δ t ,  
Here,   δ is a fixed value among [0, 1].

4. Experimental Evaluation

The intention detection results of the CL-DLBIDC model were tested using the SNIPS dataset [25]. In this work, we considered 3500 samples with 7 class labels, as depicted in Table 1.
Figure 3 depicts the classification outcomes of the CL-DLBIDC model in terms of the confusion matrix under distinct sizes of training (TR) data and testing (TS) data. The CL-DLBIDC model classified all of the distinct class labels of the intents under all sizes of TR and TS data.
Table 2 exhibits a brief set of data regarding the intent-detection performance of the CL-DLBIDC model on 80% of the TR data and 20% of the TS data. These results depict that the CL-DLBIDC model has attained enhanced values under all classes. For instance, with Class 1, the CL-DLBIDC method has granted a c c u y , r e c a l , s p e c y , F s c o r e , the Matthews Correlation Coefficient (MCC), and G m e a s u r e of 99.04%, 97.48%, 99.29%, 96.63%, 96.07%, and 96.63%, respectively.
Table 3 displays a detailed set of data on the intent-detection performance of the CL-DLBIDC method on 70% of the TR data and 30% of the TS data. The experimental values indicate that the proposed model has shown enhanced performance on the classification of distinct classes on 70:30% of the TR/TS data.
The training accuracy (TRA) and validation accuracy (VLA) obtained by the CL-DLBIDC methodology on the test dataset are shown in Figure 4. The experimental outcome denotes that the CL-DLBIDC approach has attained maximal values of TRA and VLA. Seemingly, the VLA is greater than the TRA.
The training loss (TRL) and validation loss (VLL) obtained by the CL-DLBIDC method on the test dataset are exhibited in Figure 5. The experimental outcome represented that the CL-DLBIDC algorithm has established minimal values of TRL and VLL. Particularly, the VLL is lesser than TRL.
A clear precision-recall inspection of the CL-DLBIDC approach on the test dataset is portrayed in Figure 6. The figure shows that the CL-DLBIDC technique has resulted in enhanced values of precision-recall values under all classes.
A brief ROC analysis of the CL-DLBIDC algorithm on the test dataset is presented in Figure 7. The results indicate that the CL-DLBIDC method has shown an ability to categorize distinct classes in the test dataset.
Table 4 provides the overall classification outcome of the CL-DLBIDC model as compared with recent models, such as Lightweight on-device Intent Detection (LIDSNet), Stack-Propagation–Bidirectional Encoder Representations from Transformers (Stack-PBERT), stack-propagation, Capsule-NLU, slot-gated bidirectional long short-term memory with attention (Attention-SG-BiLSTM), and slot filling and intent detection using bidirectional long short-term memory (SFID-BLSTM) models [26]. Figure 8 showcases the comparative a c c u y examination of the CL-DLBIDC model with existing techniques. The figure indicates that the Attention-SG-BiLSTM model attained a lower a c c u y of 96.87%, whereas the LIDSNet, Capsule-NLU, and SFID-BLSTM models obtained a clearly raised a c c u y of 97.79%, 97.43%, and 97.36%, respectively. Next, the stack-propagation model resulted in a moderately improved a c c u y of 98.29%. Although the stack-PBERT model obtained a reasonable a c c u y of 99.25%, the CL-DLBIDC model obtained a higher a c c u y of 99.51%.
Figure 9 showcases the comparative F 1 s c o r e analysis of the CL-DLBIDC technique with existing techniques. The figure shows that the Attention-SG-BiLSTM approach obtained a lower F 1 s c o r e of 96.99%, whereas the LIDSNet, Capsule-NLU, and SFID-BLSTM methodologies obtained a clearly raised F 1 s c o r e of 95.61%, 96.40%, and 97.21%, respectively. The stack-propagation method resulted in a moderately improved F 1 s c o r e of 96.23%. Although the stack-PBERT approach has obtained a reasonable F 1 s c o r e of 97.98%, the CL-DLBIDC methodology obtained a higher F 1 s c o r e of 98.29%. These results demonstrate the enhanced outcomes of the CL-DLBIDC model in comparison with recent models.

5. Conclusions

This study has presented a new CL-DLBIDC approach to detect and classify intent for natural language understanding. The CL-DLBIDC technique receives word-embedding as an input and uses learned meaningful features to determine the probable intention of the user query. In addition, the CL-DLBIDC technique uses the GloVe approach. In addition, the CL-DLBIDC technique makes use of the DLMNN model for intent detection and classification. For the hyperparameter tuning process, the MFO algorithm was used in this study. The experimental analysis of the CL-DLBIDC algorithm took place under a set of simulations, and different aspects of the results have been scrutinized. The simulation outcomes demonstrate the significant performance of the CL-DLBIDC method over other DL models, with a maximum accuracy of 99.51%. The enhanced performance of the proposed model is due to the hyperparameter tuning process using the MFO algorithm. In the future, hybrid DL models and the feature-selection process can improve the detection efficiency of the proposed model.

Author Contributions

Conceptualization, H.J.A.; Data curation, K.T.; Formal analysis, K.T.; Funding acquisition, M.A.H.; Investigation, H.A.; Methodology, H.J.A., H.A. and M.A.H.; Project administration, M.A.E. and M.A.H.; Resources, M.A.E. and A.Y.; Software, A.Y. and R.A.; Supervision, R.A.; Validation, O.A.; Visualization, O.A.; Writing—original draft, H.J.A.; Writing—review & editing, M.A.H. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R281), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work with Grant Code: (22UQU4331004DSR01).

Institutional Review Board Statement

This article does not contain any studies with human participants performed by any of the authors.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written with the contributions of all authors. All authors have given approval to the final version of the manuscript.

References

  1. Firdaus, M.; Golchha, H.; Ekbal, A.; Bhattacharyya, P. A deep multi-task model for dialogue act classification, intent detection and slot filling. Cogn. Comput. 2021, 13, 626–645. [Google Scholar] [CrossRef]
  2. Fernández-Martínez, F.; Griol, D.; Callejas, Z.; Luna-Jiménez, C. An approach to intent detection and classification based on attentive recurrent neural networks. In Proceedings of the IberSPEECH, Valladolid, Spain, 24–25 March 2021; pp. 46–50. [Google Scholar]
  3. Song, S.; Chen, X.; Wang, C.; Yu, X.; Wang, J.; He, X. A Two-Stage User Intent Detection Model on Complicated Utterances with Multi-task Learning. In Proceedings of the WebConf 2022, Lyon, France, 25–29 April 2022. [Google Scholar]
  4. Al-Wesabi, F.N. Proposing high-smart approach for content authentication and tampering detection of arabic text transmitted via internet. IEICE Trans. Inf. Syst. 2020, E103, 2104–2112. [Google Scholar] [CrossRef]
  5. Weld, H.; Huang, X.; Long, S.; Poon, J.; Han, S.C. A survey of joint intent detection and slot-filling models in natural language understanding. arXiv 2021, arXiv:2101.08091. [Google Scholar] [CrossRef]
  6. Al-Wesabi, F.N. A hybrid intelligent approach for content authentication and tampering detection of arabic text transmitted via internet. Comput. Mater. Contin. 2021, 66, 195–211. [Google Scholar] [CrossRef]
  7. Liu, J.; Li, Y.; Lin, M. Review of intent detection methods in the human-machine dialogue system. J. Phys. Conf. Ser. 2019, 1267, 012059. [Google Scholar] [CrossRef]
  8. Al-Wesabi, F.N. Entropy-based watermarking approach for sensitive tamper detection of arabic text. Comput. Mater. Contin. 2021, 67, 3635–3648. [Google Scholar] [CrossRef]
  9. Hou, Y.; Lai, Y.; Wu, Y.; Che, W.; Liu, T. Few-shot learning for multi-label intent detection. Proc. AAAI Conf. Artif. Intell. 2021, 35, 13036–13044. [Google Scholar]
  10. Liu, H.; Zhao, S.; Zhang, X.; Zhang, F.; Sun, J.; Yu, H.; Zhang, X. A Simple Meta-learning Paradigm for Zero-shot Intent Classification with Mixture Attention Mechanism. arXiv 2022, arXiv:2206.02179. [Google Scholar]
  11. Siddique, A.B.; Jamour, F.; Xu, L.; Hristidis, V. Generalized zero-shot intent detection via commonsense knowledge. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 11–15 July 2021; pp. 1925–1929. [Google Scholar]
  12. Zhang, J.G.; Hashimoto, K.; Wan, Y.; Liu, Y.; Xiong, C.; Yu, P.S. Are pretrained transformers robust in intent classification? a missing ingredient in evaluation of out-of-scope intent detection. arXiv 2021, arXiv:2106.04564. [Google Scholar]
  13. Zhang, H.; Xu, H.; Lin, T.E. Deep Open Intent Classification with Adaptive Decision Boundary. In Proceedings of the AAAI, Virtual, 2–9 February 2021; pp. 14374–14382. [Google Scholar]
  14. Yan, G.; Fan, L.; Li, Q.; Liu, H.; Zhang, X.; Wu, X.M.; Lam, A.Y. Unknown intent detection using Gaussian mixture model with an application to zero-shot intent classification. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, online, 5–10 July 2020; pp. 1050–1060. [Google Scholar]
  15. Gangadharaiah, R.; Narayanaswamy, B. Joint multiple intent detection and slot labeling for goal-oriented dialog. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 564–569. [Google Scholar]
  16. Dopierre, T.; Gravier, C.; Logerais, W. ProtAugment: Intent detection meta-learning through unsupervised diverse paraphrasing. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), online, 1–6 August 2021; pp. 2454–2466. [Google Scholar]
  17. Nigam, A.; Sahare, P.; Pandya, K. Intent detection and slots prompt in a closed-domain chatbot. In Proceedings of the 2019 IEEE 13th International Conference on Semantic Computing (ICSC), Newport Beach, CA, USA, 30 January–1 February 2019; pp. 340–343. [Google Scholar]
  18. Zhou, Y.; Liu, P.; Qiu, X. Knn-contrastive learning for out-of-domain intent classification. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 5129–5141. [Google Scholar]
  19. Büyük, O.; Erden, M.; Arslan, L.M. Leveraging the information in in-domain datasets for transformer-based intent detection. In Proceedings of the 2021 Innovations in Intelligent Systems and Applications Conference (ASYU), Elazig, Turkey, 6–8 October 2021; pp. 1–4. [Google Scholar]
  20. Shen, Y.; Hsu, Y.C.; Ray, A.; Jin, H. Enhancing the generalization for Intent Classification and Out-of-Domain Detection in SLU. arXiv 2021, arXiv:2106.14464. [Google Scholar]
  21. Qin, L.; Xu, X.; Che, W.; Liu, T. AGIF: An adaptive graph-interactive framework for joint multiple intent detection and slot filling. arXiv 2020, arXiv:2004.10087. [Google Scholar]
  22. Sreelakshmi, K.; Rafeeque, P.C.; Sreetha, S.; Gayathri, E.S. Deep bi-directional lstm network for query intent detection. Procedia Comput. Sci. 2018, 143, 939–946. [Google Scholar] [CrossRef]
  23. Muthu, B.; Cb, S.; Kumar, P.M.; Kadry, S.N.; Hsu, C.H.; Sanjuan, O.; Crespo, R.G. A framework for extractive text summarization based on deep learning modified neural network classifier. Trans. Asian Low-Resour. Lang. Inf. Process. 2021, 20, 1–20. [Google Scholar] [CrossRef]
  24. Mo, S.; Ye, Q.; Jiang, K.; Mo, X.; Shen, G. An improved MPPT method for photovoltaic systems based on mayfly optimization algorithm. Energy Rep. 2022, 8, 141–150. [Google Scholar] [CrossRef]
  25. Coucke, A.; Saade, A.; Ball, A.; Bluche, T.; Caulier, A.; Leroy, D.; Doumouro, C.; Gisselbrecht, T.; Caltagirone, F.; Lavril, T.; et al. Snips voice platform: An embedded spoken language understanding system for private-by-design voice interfaces. arXiv 2018, arXiv:1805.10190. [Google Scholar]
  26. Agarwal, V.; Shivnikar, S.D.; Ghosh, S.; Arora, H.; Saini, Y. Lidsnet: A lightweight on-device intent detection model using deep siamese network. In Proceedings of the 2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA), Pasadena, CA, USA, 13–16 December 2022; pp. 1112–1117. [Google Scholar]
Figure 1. Overall block diagram of the CL-DLBIDC approach.
Figure 1. Overall block diagram of the CL-DLBIDC approach.
Applsci 12 08633 g001
Figure 2. Steps involved in MFO.
Figure 2. Steps involved in MFO.
Applsci 12 08633 g002
Figure 3. Confusion matrices of CL-DLBIDC: (a) 80% of TR data, (b) 20% of TS data, (c) 70% of TR data, and (d) 30% of TS data.
Figure 3. Confusion matrices of CL-DLBIDC: (a) 80% of TR data, (b) 20% of TS data, (c) 70% of TR data, and (d) 30% of TS data.
Applsci 12 08633 g003
Figure 4. TRA and VLA analysis of the CL-DLBIDC approach.
Figure 4. TRA and VLA analysis of the CL-DLBIDC approach.
Applsci 12 08633 g004
Figure 5. TRL and VLL analysis of the CL-DLBIDC approach.
Figure 5. TRL and VLL analysis of the CL-DLBIDC approach.
Applsci 12 08633 g005
Figure 6. Precision-recall analysis of the CL-DLBIDC approach.
Figure 6. Precision-recall analysis of the CL-DLBIDC approach.
Applsci 12 08633 g006
Figure 7. ROC analysis of the CL-DLBIDC approach.
Figure 7. ROC analysis of the CL-DLBIDC approach.
Applsci 12 08633 g007
Figure 8. A c c u y analysis of the CL-DLBIDC approach with recent algorithms.
Figure 8. A c c u y analysis of the CL-DLBIDC approach with recent algorithms.
Applsci 12 08633 g008
Figure 9. F 1 s c o r e analysis of the CL-DLBIDC approach with recent algorithms.
Figure 9. F 1 s c o r e analysis of the CL-DLBIDC approach with recent algorithms.
Applsci 12 08633 g009
Table 1. Dataset details.
Table 1. Dataset details.
LabelClassNo. of Samples
1CreativeWork500
2PlayMusic500
3RateBook500
4ScreeningEvent500
5GetWeather500
6AddToPlaylist500
7BookRestaurant500
Total No. of Samples3500
Table 2. Result analysis of the CL-DLBIDC approach with distinct measures under 80:20 of TR/TS data.
Table 2. Result analysis of the CL-DLBIDC approach with distinct measures under 80:20 of TR/TS data.
LabelsAccuracyRecallSpecificityF-ScoreMCCG-Measure
Training Set (80%)
199.0497.4899.2996.6396.0796.63
299.1494.7899.8796.9596.4896.97
399.1199.7599.0096.9196.4496.95
499.3996.9599.7997.8297.4797.82
599.6498.5399.8398.7798.5698.77
699.1496.7799.5497.0196.5197.01
799.5498.2799.7598.3998.1298.39
Average99.2997.5099.5897.5097.0997.51
Testing Set (20%)
198.8698.0698.9996.1995.5496.21
299.71100.0099.6798.9998.8398.99
399.5799.0799.6698.6098.3598.61
499.8699.07100.0099.5399.4599.53
599.7197.83100.0098.9098.7498.91
699.4396.9499.8397.9497.6197.94
799.4396.8499.8397.8797.5597.88
Average99.5198.2699.7198.2998.0198.30
Table 3. Result analysis of CL-DLBIDC approach with distinct measures under 70:30 of the TR/TS data.
Table 3. Result analysis of CL-DLBIDC approach with distinct measures under 70:30 of the TR/TS data.
LabelsAccuracyRecallSpecificityF-ScoreMCCG-Measure
Training Set (70%)
198.5795.8299.0495.1694.3295.16
298.9495.8499.4796.3895.7696.38
398.3794.7298.9694.1793.2294.17
499.2296.8099.6297.2396.7897.23
599.1097.1099.4396.8296.3096.82
699.0296.2099.4896.4895.9196.48
799.1096.6599.5296.9296.3996.92
Average98.9096.1699.3696.1695.5396.17
Testing Set (30%)
199.2498.5899.3497.2096.7797.21
299.0595.6899.5696.3895.8396.38
399.0598.1199.2196.8996.3496.90
499.2495.5199.8997.3996.9797.40
599.2497.4299.5597.4296.9797.42
699.3397.4799.6697.7897.3997.78
799.3397.8999.5697.5497.1697.54
Average99.2197.2499.5497.2396.7897.23
Table 4. Comparative analysis of the CL-DLBIDC approach with recent algorithms.
Table 4. Comparative analysis of the CL-DLBIDC approach with recent algorithms.
MethodsAccuracy (%)F1-Score (%)
CL-DLBIDC99.5198.29
Stack-PBERT99.2597.98
LIDSNet97.7995.61
Stack-Propagation98.2996.23
Capsule-NLU97.4396.40
SFID-BLSTM97.3697.21
Attention-SG-BiLSTM96.8796.99
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alshahrani, H.J.; Tarmissi, K.; Alshahrani, H.; Ahmed Elfaki, M.; Yafoz, A.; Alsini, R.; Alghushairy, O.; Ahmed Hamza, M. Computational Linguistics with Deep-Learning-Based Intent Detection for Natural Language Understanding. Appl. Sci. 2022, 12, 8633. https://doi.org/10.3390/app12178633

AMA Style

Alshahrani HJ, Tarmissi K, Alshahrani H, Ahmed Elfaki M, Yafoz A, Alsini R, Alghushairy O, Ahmed Hamza M. Computational Linguistics with Deep-Learning-Based Intent Detection for Natural Language Understanding. Applied Sciences. 2022; 12(17):8633. https://doi.org/10.3390/app12178633

Chicago/Turabian Style

Alshahrani, Hala J., Khaled Tarmissi, Hussain Alshahrani, Mohamed Ahmed Elfaki, Ayman Yafoz, Raed Alsini, Omar Alghushairy, and Manar Ahmed Hamza. 2022. "Computational Linguistics with Deep-Learning-Based Intent Detection for Natural Language Understanding" Applied Sciences 12, no. 17: 8633. https://doi.org/10.3390/app12178633

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop