Next Article in Journal
A Review of Recent Literature on Systems and Methods for the Control of Thermal Comfort in Buildings
Previous Article in Journal
Contactless Soil Moisture Mapping Using Inexpensive Frequency-Modulated Continuous Wave RADAR for Agricultural Purposes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Supervised Sentiment Analysis in Spanish to Understand the University Narrative of the Colombian Conflict

by
Paula Rendón-Cardona
1,*,†,
Julian Gil-Gonzalez
2,*,†,
Julián Páez-Valdez
3 and
Mauricio Rivera-Henao
1
1
Faculty of Humanities, Social Sciences, and Education, Universidad Católica de Pereira, Pereira 660005, Colombia
2
Engineering Faculty, Universidad Tecnológica de Pereira, Pereira 660003, Colombia
3
School of Management, Universidad del Rosario, Bogotá 110141, Colombia
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2022, 12(11), 5472; https://doi.org/10.3390/app12115472
Submission received: 1 April 2022 / Revised: 12 May 2022 / Accepted: 19 May 2022 / Published: 28 May 2022

Abstract

:
Sentiment analysis is a relevant area in the natural language processing context–(NLP) that allows extracting opinions about different topics such as customer service and political elections. Sentiment analysis is usually carried out through supervised learning approaches and using labeled data. However, obtaining such labels is generally expensive or even infeasible. The above problems can be faced by using models based on self-supervised learning, which aims to deal with various machine learning paradigms in the absence of labels. Accordingly, we propose a self-supervised approach for sentiment analysis in Spanish that comprises a lexicon-based method and a supervised classifier. We test our proposal over three corpora; the first two are labeled datasets, namely, CorpusCine and PaperReviews. Further, we use an unlabeled corpus conformed by news related to the Colombian conflict to understand the university journalistic narrative of the war in Colombia. Obtained results demonstrate that our proposal can deal with sentiment analysis settings in scenarios with unlabeled corpus; in fact, it acquires competitive performance compared with state-of-the-art techniques in partially-labeled datasets.

1. Introduction

Between 1988 and 2012, Colombia faced an internal armed conflict that plunged the country into a social and political crisis that involved more than 80,000 displaced people, 12,000 violent deaths of community leaders, and at least 5000 attacks on civilian property [1]. In 2016 the final peace agreement was signed; beyond the polarization and bias of public opinion, this negotiation process meant for the victims a transition towards truth, justice, reparation, and non-repetition. The post-conflict is a new stage of recomposition and social justice, which gives way to a society more focused on reconciliation and partial overcoming of that violence. It is the transition, a stage understood as “normalization”, where Ávila and Valencia [2] affirm that institutional transformations, social development, and inclusion will be measured as a guarantee for the exercise of rights, reconciliation, and a new culture.
The events unleashed by the war have been reported by large cultural industries and mass media globally, forming an informative agenda that puts the government, the armed forces, the paramilitaries, the guerrillas, and the victims in the daily headlines. It is in this sense that it is necessary to analyze the various media representations of the conflict and the memory of the victims, in particular those of students belonging to the Colombian Network of College Journalism, since they allow evidence of the stories and construction of information and opinion of subjects in professional training, who under their worldviews and narrative logic, permeate the agenda with different perspectives from the hegemonic ones, but in line with what is discussed from other media; This, in the words of Behar and García [3], occurs because students in their journalistic texts are reconstructing the country’s history, complying with the parameters established from the discipline, permeating the agenda of large media.
On the other hand, we recognize a lack of systems that allow extracting information about how the Colombian conflict has been narrated from different media, specifically college journalism. Accordingly, the use of automatic systems for sentiment analysis can be applied to the media narrative aiming to understand the journalism students’ opinions about the different conflict stages (war, peace talks, and post-conflict) [4].
Sentiment analysis is a hot topic within the Natural language processing area, its principal objective is to assess peoples’ opinions, attitudes, and emotions regarding a specific topic [5]. Sentiment analysis has been utilized in several scenarios, including customer reviews [6], political elections [7], movies reviews [8], and conflict analysis [9]. Regarding the works developed in sentiment analysis, we recognize two main techniques, namely traditional methods and deep learning methods [10].
On the one hand, traditional methods include Lexicon-based approaches and machine learning algorithms. The lexicon-based methods depend on resources composed of words and their corresponding polarity values. Hence, it is possible to evaluate polarity in an unsupervised manner (i.e., without labeled data) [11]; nevertheless, this type of method fails in some tasks mainly because a word can set forth different senses depending on the context [12]. Conversely, in a machine learning context, the sentiment analysis tasks are considered as a binary classification problem (positive and negative) or a multiclass-classification setting (positive, negative, and neutral). Several studies have demonstrated that this type of method can lead to a better sentiment analysis compared with lexicon approaches [13]. However, these approaches require labeled data, which in some scenarios could be infeasible [14].
On the other hand, in recent years, deep learning has shown significant progress in the natural language processing–(NLP) context outperforming traditional NLP approaches. Unlike traditional methods, which use handcrafted features, deep learning models automatically create the required features for the learning process [15]. The most popular deep learning architecture includes Convolutional neural networks–(CNN) [16], Long Short-Term Memory–(LSTM) [17], and Attention-based networks [18]. Still, most deep learning approaches demand a lot of data to guarantee a good generalization performance. The above drawback can be minimized by using transfer learning techniques. Transfer learning uses domain-specific data to fine-tune the pre-trained deep learning models; a common practice is to connect fully connected layers to pre-trained language models (based, for example, on words embedding, CNN, LSTM) [19]. Thereby, the training data is used to compute the parameters belonging to such fully connected layers, while the parameters related to the language model remain untouched [20]. In this regard, the number of trainable parameters is reduced considerably, which gives two crucial benefits: the time spent in the training process is reduced, and it is possible to use small datasets without increasing the probability of overfitting [21]. Nevertheless, despite the transfer learning benefits, the accuracy of sentiment relies on the availability of labeled corpora, and the quality of their labels, which in some practical scenarios can be problematic due to labeling can be tedious, complex, expensive, or even unfeasible [22].
Aiming to tackle sentiment analysis scenarios with limited labels or unlabeled datasets, we recognize two common practices. The first consists of collecting labels from multiple annotators throughout crowd-sourcing platforms such as Amazon Mechanical Turk (https://www.mturk.com/, accessed on 31 March 2021). The attractiveness of these platforms consists in that the labeling processes can be divided into multiple workers obtaining labels (possibly noisy) at a low-cost [22]. Such noisy labels are then used to train a sentiment analysis based on learning from crowds techniques [23]. The second alternative to building sentiment classification systems in presence of limited datasets is to use self-supervised learning methods, which intend to generate pseudo-labels that are used to train a supervised learning model [24]. For sentiment analysis, lexicon-based models are usually employed combined with filtering stages in order to obtain a set of accurate pseud-labels for training a supervised learning algorithm [14].
Thereby, in this work, we present a self-supervised approach for sentiment analysis in unlabeled Spanish text. The proposed self-supervised methodology is a hybrid methodology that combines lexicon-based and classification methods aiming to build a sentiment analysis system without using any labeled data. Our approach is an extension of SSentiA [14] aiming to deal with Spanish-written texts. Our method comprises three stages. First, it employs a machine translation to build an English version of Spanish corpora. Second, a lexicon-based algorithm is applied to build a set of pseudo-labels with corresponding confidence scores. Finally, a binary classification model is trained with highly-confident pseudo-labels. Such a classifier is used to measure semantic orientation (positive or negative) of the texts whose pseudo-label was categorized with low confidence. We test our proposal over two labeled corpora, namely, CorpusCine and PaperReviews, aiming to validate the performance of our approach in a controlled scenario. Besides, we employ the developed model to assess the polarity in college media that have covered the Colombian conflict, presenting new knowledge about the peace process based on the visibility of the country’s college media as a critical scenario of interpretation and as a laboratory of quality production. Achieved results show how our method can predict sentiment polarities from unlabeled or partially labeled data.
The remainder is organized as follows. Section 2 exposes the related work and main contributions of the proposal. Section 3 describes the methodology. Section 4 and Section 5 present the experiments and discuss the results. Finally, Section 6 outlines the conclusions and future work.

2. Related Work

Sentiment analysis is a hot topic in the natural language processing context that comprises the analysis of people’s opinions and sentiments from written language [9]. Sentiment analysis has successfully applied to different scenarios such as customer opinions [25], enterprise marketing [26], and public opinion about local and international political conflicts [5]. In the literature, we recognize several works regarding political disputes. For example, authors in [27] perform a sentiment analysis from tweets during the 2016 primary debates aiming to assess the opinions’ polarity about Donald Trump. Alike, in [28] a sentiment analysis is employed to carry out a bias detection in Western news about the Palestinian/Israeli conflict. On the other hand, we recognize additional sentiment analysis approaches to analyze conflicts such as the Iran–USA nuclear deal [29], the USA–China trade war [30,31], and the problem of Syrian refugees [9]. Conversely, concerning the Colombian conflict, we notice a lack of sentiment analysis-based studies to describe such conflict and its stages (Conflict, peace agreement, and post-conflict). We only identified two relevant works. The first is the proposed in [32], which uses a crawled dataset from Twitter to analyze the perception of the post-conflict in Colombia from two perspectives: local and foreign perception. Besides, [33] uses a qualitative approach to determine the semantic orientation of information captured from Twitter, aiming to evaluate Norway’s contribution to the peace process in Colombia.
On the other hand, we broadly identified three methods for measuring semantic orientation, namely, lexicon-based models, supervised learning methods, and deep learning approaches. Lexicon-based methods comprise the use of a dictionary, where each word is labeled as positive or negative. Then, a sentence is tokenized, and it is matched with the words in the dictionary, aiming to compute the sentiment polarity [34]. However, one of the main drawbacks of lexicon-based methods is that applying one or another dictionary could lead to different polarity assessments [35]. In contrast to lexicon-based methods, machine learning-based approaches use manually labeled training data to code relevant patterns related to the sentiment analysis. Supervised learning methods usually outperform dictionary-based systems, and in fact, several relevant sentiment analysis models are built from machine learning algorithms [36]. Unlike previous methods, which should spend considerable time defining handcrafted features and performing feature engineering, deep learning models automatically perform such tasks [15]. Moreover, deep learning has exhibited important advances in several areas, including Computer Vision [37] and NLP, where the application of deep learning to sentiment analysis has become popular [38]. However, one of the main drawbacks of these methods (machine learning and deep learning) for sentiment analysis is the requirement of a labeled dataset, where the actual label (gold standard) is provided for each document in the corpus. In different scenarios, collecting such true labels is a non-trivial task because it can be tedious, expensive, or even unfeasible [39].
Several self-supervised approaches have been presented to deal with sentiment analysis problems with no labels. Such methods are mainly based on the combination of a lexicon and machine learning algorithms [40,41]. Accordingly, we notice that most of the works related to self-supervised learning for sentiment analysis are developed for documents written in the English language. However, the majority of documents related to the Colombian conflict are Spanish-written. Spanish is one of the most spoken languages in the world (more than 500 million speakers), just behind English, Chinese, and Hindi [42]. However, despite the importance of the Spanish language, we identify that the area of sentiment analysis in the Spanish language is not as well-resourced as in the English language [43].
According to the related previously, we propose a self-supervised approach to face sentiment analysis without labels. Our approach is an extension of the work in [14], so-called SSentiA. Similar to SSentiA, our proposal uses a combination of a lexicon and supervised learning-based methods. However, unlike SSentiA, we employ an additional step (Machine-translation) to deal with Spanish-written texts. Besides, our proposal shares some similarities with the work in [44] because we use a set of basis labels to train a sentiment analysis model, which is then employed to assign labels to unlabeled texts. However, while the basis labels in [44] are given by an expert (Weakly Supervised), our approach does not receive any information about the gold standard, and such basis labels are generated in an unsupervised way, which configures a more challenging scenario.
Our proposal is applied over a corpus composed of documents extracted from college media to understand the Colombian conflict’s university narrative. To the best of our knowledge, this is the first attempt to analyze sentiment polarity speech from the university perspective. Previous works [32,33] have performed similar analyses, but they use information extracted from Twitter.

3. A Self-Supervised Learning Approach for Sentiment Analysis

A machine learning-based approach for sentiment analysis comprises computing a function that assigns a binary number to each document or sentence, where 1 represents a negative polarity, whereas 1 stands for a positive document. Such a function f is estimated by using a training dataset with N samples D = x i , y i i = 1 N , where x i is the i-th document, and y i 1 , 1 is the polarity (gold standard) for document x i . However, such a gold standard is not available or scarce in many scenarios. Thereby, it is necessary to employ approaches that allow building sentiment analysis frameworks in this type of scenario.
Accordingly, in this work, we employ a self-supervised approach based on the Self-supervised Sentiment Analyzer for classification from unlabeled data–(SSentiA) proposed by [14]. Such an approach generates pseudo-labels using a lexicon-based method; then, these labels are enhanced using a supervised classification scheme. Below, we describe the steps related to our proposal.

3.1. Machine-Translated Corpus

As we have pointed out previously, the number of resources for Spanish sentiment analysis is limited; specifically, we recognize a lack of dictionaries to apply lexicon-based strategies. Conversely, for the English language, there is a considerable quantity of dictionaries for sentiment analysis, including domain-specific dictionaries [34]. Accordingly, following the ideas in [11], we used a machine translation (precisely Google translate) in order to build an English version X ˜ of the original corpus X = x i i = 1 N . The above allows us to take a profit from English dictionaries.

3.2. Lexicon-Based Label Generation

The second step in our approach comprises the generation of highly accurate pseudo-labels to feed a supervised learning classifier. For specific testing, we employ the method LRSentia [14]. LRSentia is a lexicon-based method that allows assessing the semantic orientation of a review in an unsupervised manner (i.e., without labels).
Given the translated corpus X ˜ = x ˜ i i = 1 N , LRSentiA divide each translated document x ˜ i into sentences. Accordingly, we have a set S i = { s 1 , i , , s R i , i } containing the sentences conforming the document x ˜ i , where R i is the number of sentences in document x ˜ i . Then, a pre-processing algorithm is applied over each sentence in order to exclude words that do not have significant relevance for the sentence polarity (for more details, see [11]). Once we have performed the pre-processing step, the sentiment polarity of each sentence P ( s j , i ) with j { 1 , R i } is computed.
Given the polarity for all the sentences, P ( s 1 , i ) , , P ( s R i , i ) , the overall polarity for document x ˜ i is estimated as P ( S i ) = j = i R i s j , i . Therefore, the semantic orientation for the i-th document is given as
P ( x i ) = 1 ( P o s i t i v e ) , P ( S i ) 0 1 ( N e g a t i v e ) , P ( S i ) < 0 .
Besides to the semantic orientation, LRSentiA provides the confidence score ρ i for the predictions P ( x i ) . The confidence score for the document x i is estimated as follows:
ρ i = P ( S i ) j = 1 J abs ( P s j , i ) ,
where we recall that P ( S i ) = j = i R i s j , i . As the Equation (2) indicates, the confidence score of review i, depends on the difference between the positive ( s j , i > 0 ) and negative ( s j , i < 0 ) polarities; hence, the presence of highly polarized documents points out a high confidence score. In summary, documents with high confidence scores are supposed to be less prone to misclassification and can be used as pseudo-labels to train a supervised learning algorithm [11].
On the other hand, we define the vector ρ R N containing the confidence scores for the documents in the corpus. Based on such vector, the confidence group ζ i for the i-th document is defined as
ζ i = Very - high , ρ i > δ high , δ 0.5 σ ρ < ρ i δ Low , δ σ ρ < ρ i δ 0.5 σ ρ Very - low , 0 < ρ i δ σ ρ Zero , ρ i = 0 ,
where δ is the threshold used to define the confidence categories, which is computed as, δ = μ ρ + σ ρ , being μ ρ , and σ ρ respectively the mean and standard deviation of vector ρ . The categorization shown in Equation (3) is intended to minimize incorrect predictions while maximizing the number of samples to train the supervised learning algorithm. Previous studies [11,14] have shown the presence of a correlation between the confidence scores and the prediction accuracies.

3.3. Machine Learning Classifier

As we have established below, there is a relation between the prediction accuracy and the confidence scores; the higher the confidence scores, the higher the accuracy. Accordingly, let y ^ 1 { 1 , 1 } N 1 a pseudo-labels vector containing the lexicon-based predictions whose confidence scores lies in the categories “Very-high” and “High” (see Equation (3)). We employ such pseudo-labels and their corresponding documents to train a supervised classification model. Then, the trained model is used to predict the semantic orientation y ^ 2 { 1 , 1 } N 2 of documents whose lexicon-based label had “Low”, “Very-low”, and “Zero” confidence. Finally, the overall prediction y ^ { 1 , 1 } N is composed from the concatenation of pseudo-labels y ^ 1 , and the predictions y ^ 2 , being N = N 1 + N 2 . As in [14], each document is parametrized using unigram and bigram-based term frequency–inverse document frequency (tf–idf).

4. Experimental Set-Up

4.1. Datasets

The presented methodology is tested over three corpora. The first two, CorpusCine and PaperReviews, configure a controlled experiment because, for both corpora, we have access to the ground truth (the actual semantic orientations). Conversely, a third corpus related to the Colombian conflict is employed, representing a more challenging scenario because such a dataset has not been labeled.

4.1.1. Labeled Datasets

CorpusCine is a dataset formed by 3878 Spanish-written movie reviews captured from the MuchoCine website (https://muchocine.net/, accessed on 31 March 2021). Each document is rated using an integer tag ranging from 1 (unpleasant movie) to 5 (excellent movie). The rating distribution for the CorpusCine dataset is presented in Table 1. We remark that using such a dataset for sentiment analysis configures a multi-class classification problem; however, as we have clarified in Section 3, our sentiment analysis approach was designed for binary settings (positive and negative documents). Accordingly, we use the procedure in [45] aiming to convert the CorpusCine into a binary problem, as follows. Documents with a rating of one or two are considered a negative review ( 1 ); conversely, manuscripts with a label of four and five are estimated as positives (1). We discard documents with a tag of three because they correspond to neutral or mixed opinions [46]. This dataset is publicly available (http://www.lsi.us.es/~fermin/corpusCine.zip, accessed on 31 March 2021).
On the other hand, we use the PaperReviews corpus, which comprises paper reviews sent to an international conference in Spanish. The corpus has a total of 382 documents labeled using a five-point scale, indicating the opinion about the paper quality. Table 1 shows the rating distribution for the PaperReviews dataset. Alike for CorpusCine, we use a methodology to treat PaperReviews as a binary classification problem. Samples with ratings of one or two are considered negative reviews; similarly, documents with ratings of four or five are categorized as positive reviews. PaperReviews is publicly available (https://archive.ics.uci.edu/ml/datasets/Paper+Reviews, accessed on 31 March 2021).

4.1.2. Colombian Conflict Dataset

The third dataset used in this work comes from the Universidad Católica de Pereira. It comprises a collection of journalistic articles written by young university students in Colombia, which is a product of the project: Analytical Center of University Cultural Productions in the Context of the Conflict (caPAZ), funded by the Ministry of Science, Technology, and Innovation (Minciencias) and the National Center for Historical Memory (CNM) of Colombia (under the code: 1349-872-76354, agreement 872 of 2020). This corpus includes news written by the 24 college media of the Colombian Network of College Journalism from 2001 to 2021. The dataset includes digital, printed, sound, and audiovisual news, for 3049 news items and 3,451,486 words related to the armed conflict, the memory of the victims, and the peace process in Colombia. For the case of this article, we will focus on the news articles that appeared on the web pages of the 24 college media because they gave us greater diversity, greater reliability in the extraction process, and a more significant impact due to their high production. However, we only used the news articles from 8 college media since they were the ones with more than 50 news items. In conclusion, a total of 2373 digital news articles were analyzed, see Table 2. These news items were collected through a web-scraping technique, using three lemmatized keywords (conflicto armado, memoria de las víctimas, and proceso de paz), with the aim of identifying these regular expressions in the logical operators that run through the HTML structure of each Web page. Likewise, a web-crawling technique was used to explore the addressing links of each page and build a relevant relationship graph. The 2373 news items were converted into raw text format .txt. Finally, the documents’ semantic orientations are generated by performing a lexicometry analysis using the Alceste–Reinert method for textual data clustering [47]. Such dataset is publicly available (https://zenodo.org/record/6384840#.Yj5Av-fMJPZ, accessed on 31 March 2021).
The Alceste–Reinert method focuses on the identification of lexical worlds. This method makes it possible, through statistical analysis, to identify the appearance of specific forms in the content of a speech. Reinert’s main thesis [48] asserts that all discourse expresses a system of lexical words, which organize thought and give coherence to what is described by the enunciators. These lexical worlds are identified through top-down hierarchical clustering exercises and expressed in textual dendrograms, namely, a tree-shaped diagram to represent the hierarchy of categories according to the degree of similarity and shared characteristics. For the case of this research, textual dendrograms were generated for each medium analyzed. These eight dendrograms compile the 2373 news items. The labeling process consisted of qualitative analysis through direct observation of each textual dendrogram. Polarity coding was performed by an investigator, who labeled the material as positive and negative. The assignment of the said label was made with each percentage community expressed in the dendrograms. It is essential to clarify that the labeling process was done for each analyzed college media and not for each news item in the corpus.

4.2. Methodology Training and Validation

The proposed self-supervised learning approach combines the lexicon-based method LRSentiA with supervised learning techniques. LRSentiA generates pseudo-labels with their corresponding confidence scores. Then the machine learning approaches capture critical patterns on the data with the highest confidence scores aiming to enhance the polarity representation of complex documents where the LRSentiA’s performance is poor [14]. Accordingly, for the LRSentiA method, we use the opinion lexicon curated by authors in [49], which comprises 4783 negative and 2006 polarity words. On the other hand, for the machine learning stage, each document is parametrized using the well-known unigram and bigram-based tf–idf (term frequency-inverse document frequency) scores. Besides, we test four binary classification algorithms, namely, Logistic Regression (LR), Support vector machines (SVM), Random Forest (RF), and Naive–Bayes (NB).
Aiming to validate the efficacy of our self-supervised approach, we use the procedure in Figure 1. We employ two state-of-the-art lexicons for English-written texts, namely, VADER [50], TextBlob (https://textblob.readthedocs.io/en/dev/, accessed on 31 March 2021). Moreover, we use a publicly available Spanish lexicon (https://www.kaggle.com/rtatman/sentiment-lexicons-for-81-languages, accessed on 31 March 2021). On the other hand, the quality assessment is performed by computing overall accuracy and the F1 score between the predicted labels and the ground truth. The Python codes for our method are publicly available (https://github.com/juliangilg/SSentiA_Sp, accessed on 31 March 2021).

5. Results and Discussion

5.1. Labeled Datasets

As we established in Section 3.1, we use a machine-translation approach to take advantage of the English resources; thus, it is crucial to verify that the documents’ semantic orientation is not affected by the translation process [51,52]. In this regard, we perform an initial experiment aiming to evaluate the machine-translation impact for sentiment analysis. For such an experiment, we use the CorpusCine and PaperReviews that review dataset, each dataset since these corpora contain the actual labels. We build two sentiment analysis systems (based on machine learning algorithms) for each corpus. The first uses the documents in their original language (Spanish), while the second employs an English-translated version of such documents. Each document is parametrized by using unigram and bigram-based term frequency-inverse document frequency. We use a five-fold validation scheme, where the performance is assessed in terms of the overall accuracy. Table 3 shows a comparison of a machine learning-based model for sentiment analysis trained by using both original and translated corpora. From Table 3, we remark that there are no significant differences between the performance (in terms of overall accuracy) of sentiment analysis systems from Spanish and English-translated corpora. Consequently, we experimentally demonstrate that the sentiment polarity is preserved after the English translation in most cases.
Second, we perform a controlled experiment aiming to verify the capability of our proposal to compute the semantic orientation in the absence of labels. We use the CorpusCine and PaperReviews reviews dataset described previously. We remark that these datasets contain the true semantic orientations; nevertheless, such labels are only used to compute performance metrics.
As we pointed out in Section 3.2 our approach uses a lexicon-based algorithm (LRSentiA) to generate pseudo-labels that are then used to train a machine learning model. Besides, LRSentiA computes the confidence score for each pseudo-label Equation (2), where it has been established that there exists a relation between such confidence scores and the prediction accuracy [14]. Accordingly, we perform a first experiment aiming to confirm previous affirmation experimentally. Table 4 shows the prediction accuracy compared with the confidence groups defined in Equation (3).
From Table 4, it is possible to notice that, for both labeled datasets, there exists a significant relationship between the confidence scores and the prediction accuracy, the Very-high confidence group is related to the highest prediction accuracy. In contrast, the lowest accuracy corresponds to the Zero confidence group. We compute the Pearson coefficient between the prediction accuracy and the mean confidence score in each group to support our observations. Hence, we obtain a value of 0.8853 for CorpusCine and 0.8999 PaperReviews, which indicates a strong linear relationship between the analyzed variables confirming our initial qualitative analysis.
As a second experiment, we use lexicon-based algorithms for the CorpusCine dataset. We employ the well-known English dictionaries TextBlob and VADER, applied over an English version of the studied dataset. Further, we use the Spanish dictionary described in Section 4.2. From Table 5, we observe that CorpusCine, in terms of the F1 score, the dictionary TextBlob obtains the best performance. On the other hand, regarding PaperReview, we note that the VADER lexicon outperforms all its competitors. However, we observe that the lexicon-based methods exhibit a considerably low-performance in general terms. The above behavior is explained in the sense that dictionaries-based algorithms obtain their best yielding with highly polarity text; however, they are ineffective in assessing the sentiment orientation of cases with mixed opinions [14].
Finally, we test our hybrid proposal over the labeled datasets. Table 6 shows the performance of our approach compared with different methods (Unsupervised, supervised, and hybrid) to perform sentiment analysis in Spanish corpora. Analyzing the results for CorpusCine in Table 6, we first evidence that our proposal reaches the best performance by using a logistic regression classifier, which is remarkable due to it can only deal with linear-separable data. The above suggests a linear structure may exist in the features extracted from the documents. On the other hand, we note that the approaches with the worst performance are those based on unsupervised learning algorithms, which is not surprising. As it was argued below, such types of methods cannot measure documents with mixed opinions. Now, concerning the supervised learning methods, we note that the F1 score considerably outperforms unsupervised algorithms. Such behavior is due to supervised models extracting relevant patterns related to the sentiment orientation. Then, we remark that the works based on hybrid approaches (i.e., combining supervised and unsupervised techniques) [53,54] achieve the best classification scores (F1 score, Accuracy). However, regarding our hybrid approach, we observe a significantly lower performance; in fact, it only outperforms the unsupervised models. To explain such an outcome, we highlight that unlike the approaches in [53,54], our proposal does not use any information about the gold standard. Conversely, it generates pseudo-labels using the lexicon-based method LRSentiA, which is then used to train a supervised classifier. Therefore, we argue that the performance of our approach can be increased by providing limited labeled data.
Conversely, regarding the results (Table 6), we notice that, similar to previous outcomes, the best F1 scores come from supervised and hybrid approaches. However, we remark a considerable low yielding from our proposal and the deep learning approach introduced in [58]; in fact, both methods are defeated by the lexicon-based algorithms (see Table 5). Concerning the deep learning-based results, we argue that such low performance can be caused by a lack of generalization (overfitting). Now with respect to our hybrid proposal, we explain the undesired results in two regards. First, like the results for the CorpusCine dataset, the training step does not have access to the actual labels; therefore, the behavior of our approach can be enhanced by using partially labeled corpora. Second, we observe that the results from our method are skewed towards positive labels (the precision is greater than recall), which is caused for a discrepancy between the paper is evaluated and the way the review is written [60].
Finally, aiming to evaluate the behavior of our hybrid proposal in scenarios with limited labeled data, we carry out an additional experiment where we vary the number of labels. Let be p [ 0 , 1 ] the ratio of available labeled data. We split the CorpusCine review dataset into two groups. The first, X l is conformed by p N (being N the number of documents in such corpus) samples with their corresponding true labels y l { 0 , 1 } p N ; conversely, the second group X u contains N ( 1 p ) unlabeled documents. Hence, we apply the LRSentiA algorithm over the unlabeled corpus X u to generate pseudo-samples. LRSentiA classifies such labels into five groups (Very-high, High, Low, Very-low, and zero) according to their confidence. Then, we conform the sets X t , and y t by concatenating the samples and labels from the Very-high and High categories together with the labeled data ( X l , y l ). X t and y t is then used to feed a classification algorithm to predict the label for documents belonging to low, very-low, and zero confidence. For each value of p, we use a five-fold validation scheme, where the sets X l , and X u are conformed randomly. The performance is measured in terms of the overall accuracy. Figure 2 shows the mean and standard deviation for the accuracy of our proposal as a function of the ratio of labeled data. For specific testing, we compare a logistic regression classifier and a Support vector machine since they achieved the highest performances in previous experiments (see Table 6). From Figure 2, we remark that in most cases, the performance of our approach increases as there are more labeled data. Besides, we highlight that with the 40 % of labeled data, our proposal achieves an accuracy near 0.8 for CorpusCine data and near 0.7 for PaperReview dataset, which remarkably is comparable with state-of-the-art results.

5.2. Colombian Conflict Dataset

As a final experiment, we apply our proposal to each document in the Colombian conflict dataset aiming to understand the narrative of the war from the journalism students’ perspective. Aiming to validate the obtained results, we use the qualitative labels explained in Section 4.1.2. However, we note that such labels are not given for each document but for each media. Accordingly, to perform a fair comparison, we use the following procedure. Let X m = { x 1 ( m ) , , x N m ( m ) } a set containing N m documents corresponding to the m-th media. We use our hybrid method to obtain the sentiment polarity y m = [ y 1 ( m ) , , y N m ( m ) ] for each document in X m . Accordingly, the polarity of m-th media is computed as:
P ( X m ) = 1 , i = 1 N m y 1 ( m ) 0 0 , In   other   case .
Table 7 shows a comparison between the semantic orientation obtained by using our approach under four configurations (employing four classification algorithms: LR, SVM, RF, and NB) and a set of qualitative labels, which are obtained using the identification of lexical worlds [47] (see Section 4.1.2).
From Table 7, we first notice that according to the qualitative assessment, all of the considered media expresses positive aspects of the Colombian conflict and its stages. Now, regarding the result of our self-supervised approach, we note a similar tendency, given that most of the media are categorized as positives; such outcomes are consistent when we vary the classification scheme (RL, SVM, RF, and NB), obtaining an overall accuracy of 75%. Specifically, we remark that for media Universidad del Valle and Santiago de Cali the sentiment orientation predicted for our approach differs from the qualitative labels; hence, to explain such outcomes, we analyze each media individually. For the Universidad del Valle media, we observe it contains 60 documents, and only 26 were categorized as very-high or high confidence. The above indicates that the supervised learning approach was trained with 26 samples, which indicates that the overfitting causes the results obtained. Conversely, regarding the Santiago de Cali media, we notice that despite the predicted label being negative, the ratio of documents with a positive semantic orientation is close to 0.5 , which indicates the robustness of our approach to recognizing the polarity of documents related to the Colombian conflict.
On the other hand, the results are complemented by the lexicometric approach. College journalism narrated the conflict with a positive stance; this is reflected in the need to understand Colombia’s conflict and propose solutions and scenarios of agreement. From journalism, we could affirm that college journalism focused on presenting future actions and not falling into the traditional narrative that was merely informative.

6. Conclusions

In this paper, we study the problem of identifying sentiment polarity from unlabeled corpora. The presented approach is based on the so-called SSentiA [14], which is built from a hybrid method that combines lexicon-based algorithms and supervised learning models. Unlike SSentiA, our proposal uses a machine translation approach to perform sentiment analysis in Spanish-written documents. Our strategy was first tested using the CineCorpus, and PaperReview datasets, which configure a controlled scenario because, for such corpora, we have access to the actual labels (polarities). According to the results, we observe that the proposal’s performance is considerably lower compared with state-of-the-art supervised learning and hybrid approaches, which use the gold standard. However, we highlight that when our methodology has access to limited labeled data (around 40%), it can perform similarly to methods with the fully labeled dataset. On the other hand, we tested the proposal in a more challenging corpus related to the Colombian conflict. The outcomes were compared with qualitative labels based on the identification of lexical worlds [47], showing an overall accuracy of 75 % . From the journalism perspective, we can conclude that the two approaches used (quantitative and qualitative) had similar results and evidenced the positive polarity of the university journalistic story. The qualitative approach, based on the Alceste–Reinert lexicometric method, allowed us to identify the polarity of the discourse, and also the journalistic themes worked by the media, for this reason, we can conclude affirming that the Colombian university students narrated the conflict positively from 3 topics: the armed conflict, the memory of the victims and the peace process.
In future work, the authors plan to follow the work in [5] in order to combine the developed methodology with topic modeling such as the Latent Dirichlet Allocation (LDA), aiming to extract the relationship between the sentiment polarities and the most relevant topics in the documents to capture more relevant insights about the development of the Colombian conflict from the perspective of journalism students. On the other hand, we observe that our approach is considerably affected in the presence of corpora with few documents and for class imbalance; accordingly, we plan to use data manipulation techniques such as weighting data examples or data augmentation [59,61] aiming to improve the performance of our approach. Finally, in this work, we use a traditional approach (based on machine learning algorithms) mainly due to the small quantity of labeled data, especially for the Colombian conflict dataset. Thus, we plan to distribute the labeling processing to several experts aiming to build a dataset from multiple annotators. Hence, the objective is to use transfer learning with pre-trained models based on deep learning such as LSTM or attention models combined with learning from crowds layers as in [23,62], which can favor the semantic orientation representation.

Author Contributions

Conceptualization, J.G.-G. and P.R.-C.; data curation, J.P.-V.; methodology, P.R.-C., J.G.-G. and M.R.-H.; project administration, M.R.-H. and P.R.-C.; supervision, J.G.-G. and J.P.-V.; resources, P.R.-C. and J.P.-V. All authors have read and agreed to the published version of the manuscript.

Funding

Under grants provided by the “Fondo nacional de financiamiento para la ciencia, la tecnología y la innovación Francisco José de Caldas”, the “Ministerio de ciencia, tecnología e innovación”—(MinCiencias), the “Centro nacional de memoria histórica”, the “Universidad Católica de Pereira”, and the “Universidad del Quindío” under the project: “Analytical Center of University Cultural Productions in the Context of the Conflict (caPAZ).”—code 1349-872-76354, agreement 872 of 2020.

Institutional Review Board Statement

Institutional review is not applicable because this work does not include data involving humans or animals.

Informed Consent Statement

This study does not involve humans; accordingly, informed consent is not required.

Data Availability Statement

Datasets are publicly available at: http://www.lsi.us.es/~fermin/corpusCine.zip (accessed on 31 March 2021), https://archive.ics.uci.edu/ml/datasets/Paper+Reviews (accessed on 31 March 2021), and https://zenodo.org/record/6384840#.Yj5Av-fMJPZ (accessed on 31 March 2021).

Conflicts of Interest

The authors declare that this research was conducted without any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Peralta, E.A.C. !Basta ya! Colombia: Memorias de guerra y dignidad. Anu. Colomb. Hist. Soc. Cult. 2014, 41, 339. [Google Scholar] [CrossRef]
  2. Martínez, A.F.Á.; Agudelo, L.V. Los Retos del Posconflicto: Justicia, Seguridad y Mercados Ilegales; Ediciones B Colombia: Bogota, Colombia, 2016. [Google Scholar]
  3. Cárdenas, H.M. Cómo pueden los nuevos comunicadores hacer periodismo de investigación en la era digital? Compiladora 2018, 65–84. [Google Scholar]
  4. Calderón Rojas, J. Etapas del conflicto armado en Colombia: Hacia el posconflicto. Rev. Estud. Latinoam. 2016, 62, 227–257. [Google Scholar]
  5. Lee, E.; Rustam, F.; Ashraf, I.; Washington, P.B.; Narra, M.; Shafique, R. Inquest of Current Situation in Afghanistan Under Taliban Rule Using Sentiment Analysis and Volume Analysis. IEEE Access 2022, 10, 10333–10348. [Google Scholar] [CrossRef]
  6. Kumar, S.; Yadava, M.; Roy, P.P. Fusion of EEG response and sentiment analysis of products review to predict customer satisfaction. Inf. Fusion 2019, 52, 41–52. [Google Scholar] [CrossRef]
  7. Sandoval-Almazan, R.; Valle-Cruz, D. Facebook impact and sentiment analysis on political campaigns. In Proceedings of the 19th Annual International Conference on Digital Government Research: Governance in the Data Age, Delft, The Netherlands, 30 May–1 June 2018; pp. 1–7. [Google Scholar]
  8. Ali, N.M.; Abd El Hamid, M.M.; Youssif, A. Sentiment analysis for movies reviews dataset using deep learning models. Int. J. Data Min. Knowl. Manag. Process. 2019, 9, 19–27. [Google Scholar] [CrossRef]
  9. Öztürk, N.; Ayvaz, S. Sentiment analysis on Twitter: A text mining approach to the Syrian refugee crisis. Telemat. Inform. 2018, 35, 136–147. [Google Scholar] [CrossRef]
  10. Liu, R.; Shi, Y.; Ji, C.; Jia, M. A survey of sentiment analysis based on transfer learning. IEEE Access 2019, 7, 85401–85412. [Google Scholar] [CrossRef]
  11. Sazzed, S. Improving Sentiment Classification in Low-Resource Bengali Language Utilizing Cross-Lingual Self-supervised Learning. In Proceedings of the International Conference on Applications of Natural Language to Information Systems, Saarbrucken, Germany, 23–25 June 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 218–230. [Google Scholar]
  12. Paredes-Valverde, M.A.; Colomo-Palacios, R.; Salas-Zárate, M.d.P.; Valencia-García, R. Sentiment analysis in Spanish for improvement of products and services: A deep learning approach. Sci. Program. 2017, 2017, 1329281. [Google Scholar] [CrossRef] [Green Version]
  13. Severyn, A.; Moschitti, A. Unitn: Training deep convolutional neural network for twitter sentiment classification. In Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015), Denver, CO, USA, 4–5 June 2015; pp. 464–469. [Google Scholar]
  14. Sazzed, S.; Jayarathna, S. Ssentia: A self-supervised sentiment analyzer for classification from unlabeled data. Mach. Learn. Appl. 2021, 4, 100026. [Google Scholar] [CrossRef]
  15. Yadav, A.; Vishwakarma, D.K. Sentiment analysis using deep learning architectures: A review. Artif. Intell. Rev. 2020, 53, 4335–4385. [Google Scholar] [CrossRef]
  16. Gong, Y.; Liu, L.; Yang, M.; Bourdev, L. Compressing deep convolutional networks using vector quantization. arXiv 2014, arXiv:1412.6115. [Google Scholar]
  17. Day, M.Y.; Lin, Y.D. Deep learning for sentiment analysis on google play consumer review. In Proceedings of the 2017 IEEE International Conference on Information Reuse and Integration (IRI), San Diego, CA, USA, 4–6 August 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 382–388. [Google Scholar]
  18. Yuan, Z.; Wu, S.; Wu, F.; Liu, J.; Huang, Y. Domain attention model for multi-domain sentiment classification. Knowl.-Based Syst. 2018, 155, 1–10. [Google Scholar] [CrossRef]
  19. Howard, J.; Ruder, S. Universal language model fine-tuning for text classification. arXiv 2018, arXiv:1801.06146. [Google Scholar]
  20. Tao, J.; Fang, X. Toward multi-label sentiment analysis: A transfer learning based approach. J. Big Data 2020, 7, 1–26. [Google Scholar] [CrossRef] [Green Version]
  21. Sun, Q.; Liu, Y.; Chua, T.S.; Schiele, B. Meta-transfer learning for few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 403–412. [Google Scholar]
  22. Gil-Gonzalez, J.; Orozco-Gutierrez, A.; Alvarez-Meza, A. Learning from multiple inconsistent and dependent annotators to support classification tasks. Neurocomputing 2021, 423, 236–247. [Google Scholar] [CrossRef]
  23. Rodrigues, F.; Pereira, F. Deep learning from crowds. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  24. Pandey, A.C.; Rajpoot, D.S.; Saraswat, M. Twitter sentiment analysis using hybrid cuckoo search method. Inf. Process. Manag. 2017, 53, 764–779. [Google Scholar] [CrossRef]
  25. Kwon, H.J.; Ban, H.J.; Jun, J.K.; Kim, H.S. Topic modeling and sentiment analysis of online review for airlines. Information 2021, 12, 78. [Google Scholar] [CrossRef]
  26. Kauffmann, E.; Peral, J.; Gil, D.; Ferrández, A.; Sellers, R.; Mora, H. Managing marketing decision-making with sentiment analysis: An evaluation of the main product features using text data mining. Sustainability 2019, 11, 4235. [Google Scholar] [CrossRef] [Green Version]
  27. Abdullah, M.; Hadzikadic, M. Sentiment analysis of twitter data: Emotions revealed regarding donald trump during the 2015–2016 primary debates. In Proceedings of the 2017 IEEE 29th International Conference on Tools with Artificial Intelligence (ICTAI), Boston, MA, USA, 6–8 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 760–764. [Google Scholar]
  28. Al-Sarraj, W.F.; Lubbad, H.M. Bias Detection of Palestinian/Israeli Conflict in Western Media: A Sentiment Analysis Experimental Study. In Proceedings of the 2018 International Conference on Promising Electronic Technologies (ICPET), Deir El-Balah, Palestine, 3–4 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 98–103. [Google Scholar]
  29. Abedin, E.; Jafarzadeh, H.; Akhlaghpour, S. Opinion mining on Twitter: A sentiment analysis of the Iran deal. In Proceedings of the PACIS 2018, Yokohama, Japan, 26–30 June 2018. [Google Scholar]
  30. AlKhatib, M.; El Barachi, M.; AleAhmad, A.; Oroumchian, F.; Shaalan, K. A sentiment reporting framework for major city events: Case study on the China-United States trade war. J. Clean. Prod. 2020, 264, 121426. [Google Scholar] [CrossRef]
  31. Khun, N.H.; Thant, H.A. Visualization of Twitter Sentiment during the Period of US Banned Huawei. In Proceedings of the 2019 International Conference on Advanced Information Technologies (ICAIT), Chikmagalur, India, 24–27 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 274–279. [Google Scholar]
  32. Vera, V.D.G.; Lopera, I.C.P. Sentiment analysis on post conflict in Colombia: A text mining approach. Asian J. Appl. Sci. 2018, 6, 53–59. [Google Scholar]
  33. Fabra-Mata, J.; Mygind, J. Big data in evaluation: Experiences from using Twitter analysis to evaluate Norway’s contribution to the peace process in Colombia. Evaluation 2019, 25, 6–22. [Google Scholar] [CrossRef]
  34. Van Atteveldt, W.; van der Velden, M.A.; Boukes, M. The validity of sentiment analysis: Comparing manual annotation, crowd-coding, dictionary approaches, and machine learning algorithms. Commun. Methods Meas. 2021, 15, 121–140. [Google Scholar] [CrossRef]
  35. Rodríguez-Ibáñez, M.; Gimeno-Blanes, F.J.; Cuenca-Jiménez, P.M.; Soguero-Ruiz, C.; Rojo-Álvarez, J.L. Sentiment Analysis of Political Tweets from the 2019 Spanish Elections. IEEE Access 2021, 9, 101847–101862. [Google Scholar] [CrossRef]
  36. Rosenthal, S.; Farra, N.; Nakov, P. SemEval-2017 task 4: Sentiment analysis in Twitter. In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), Vancouver, BC, Canada, 3–4 August 2017; pp. 502–518. [Google Scholar]
  37. Chai, J.; Zeng, H.; Li, A.; Ngai, E.W. Deep learning in computer vision: A critical review of emerging techniques and application scenarios. Mach. Learn. Appl. 2021, 6, 100134. [Google Scholar] [CrossRef]
  38. Zhang, L.; Wang, S.; Liu, B. Deep learning for sentiment analysis: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1253. [Google Scholar] [CrossRef] [Green Version]
  39. Raykar, V.C.; Yu, S.; Zhao, L.H.; Valadez, G.H.; Florin, C.; Bogoni, L.; Moy, L. Learning from crowds. J. Mach. Learn. Res. 2010, 11, 1297–1322. [Google Scholar]
  40. Zhang, P.; He, Z. A weakly supervised approach to Chinese sentiment classification using partitioned self-training. J. Inf. Sci. 2013, 39, 815–831. [Google Scholar] [CrossRef]
  41. He, Y.; Zhou, D. Self-training from labeled features for sentiment analysis. Inf. Process. Manag. 2011, 47, 606–616. [Google Scholar] [CrossRef] [Green Version]
  42. de Arriba Serra, A.; Oriol Hilari, M.; Franch Gutiérrez, J. Applying sentiment analysis on Spanish tweets using BETO. In Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2021): Co-Located with the Conference of the Spanish Society for Natural Language Processing (SEPLN 2021), XXXVII International Conference of the Spanish Society for Natural Language Processing, Málaga, Spain, 22–24 September 2021; pp. 1–8. [Google Scholar]
  43. Rosá, A.; Chiruzzo, L. Emotion Classification in Spanish: Exploring the Hard Classes. Information 2021, 12, 438. [Google Scholar] [CrossRef]
  44. Kastrati, Z.; Imran, A.S.; Kurti, A. Weakly supervised framework for aspect-based sentiment analysis on students’ reviews of MOOCs. IEEE Access 2020, 8, 106799–106810. [Google Scholar] [CrossRef]
  45. Vilares, D.; Alonso, M.A.; Gómez-Rodríguez, C. A syntactic approach for opinion mining on Spanish reviews. Nat. Lang. Eng. 2015, 21, 139–163. [Google Scholar] [CrossRef] [Green Version]
  46. Mata, F.L.C. Feature-based opinion extraction: A practical, domain-adaptable approach. AI Commun. 2012, 25, 369–371. [Google Scholar]
  47. Reinert, p.M. Classification descendante hierarchique et analvse lexicale par contexte-application au corpus des poesies D’A. Rihbaud. Bull. Sociol. Methodol. 1987, 13, 53–90. [Google Scholar] [CrossRef]
  48. Reinert, P.M. Postures énonciatives et mondes lexicaux stabilisés en analyse statistique de discours. Langage et Société 2007, 3, 189–202. [Google Scholar] [CrossRef]
  49. Hu, M.; Liu, B. Mining and summarizing customer reviews. In Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; pp. 168–177. [Google Scholar]
  50. Hutto, C.; Gilbert, E. Vader: A parsimonious rule-based model for sentiment analysis of social media text. In Proceedings of the International AAAI Conference on Web and Social Media, Ann Arbor, MI, USA, 1–4 June 2014; Volume 8, pp. 216–225. [Google Scholar]
  51. Araújo, M.; Pereira, A.; Benevenuto, F. A comparative study of machine translation for multilingual sentence-level sentiment analysis. Inf. Sci. 2020, 512, 1078–1102. [Google Scholar] [CrossRef] [Green Version]
  52. Ghafoor, A.; Imran, A.S.; Daudpota, S.M.; Kastrati, Z.; Batra, R.; Wani, M.A. The Impact of Translating Resource-Rich Datasets to Low-Resource Languages Through Multi-Lingual Text Processing. IEEE Access 2021, 9, 124478–124490. [Google Scholar] [CrossRef]
  53. del Hoyo, R.; Hupont, I.; Lacueva, F.J.; Abadía, D. Hybrid text affect sensing system for emotional language analysis. In Proceedings of the International Workshop on Affective-Aware Virtual Agents and Social Robots, Boston, MA, USA, 6 November 2009; pp. 1–4. [Google Scholar]
  54. Martín-Valdivia, M.T.; Martínez-Cámara, E.; Perea-Ortega, J.M.; Ureña-López, L.A. Sentiment polarity detection in Spanish reviews combining supervised and unsupervised approaches. Expert Syst. Appl. 2013, 40, 3934–3942. [Google Scholar] [CrossRef]
  55. Molina-González, M.D.; Martínez-Cámara, E.; Martín-Valdivia, M.T.; Perea-Ortega, J.M. Semantic orientation for polarity classification in Spanish reviews. Expert Syst. Appl. 2013, 40, 7250–7257. [Google Scholar] [CrossRef]
  56. Martínez-Cámara, E.; Martín-Valdivia, M.T.; Urena-López, L.A. Opinion classification techniques applied to a spanish corpus. In Proceedings of the International Conference on Application of Natural Language to Information Systems, Alicante, Spain, 28–30 June 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 169–176. [Google Scholar]
  57. Graovac, J.; Mladenović, M.; Tanasijević, I. NgramSPD: Exploring optimal n-gram model for sentiment polarity detection in different languages. Intell. Data Anal. 2019, 23, 279–296. [Google Scholar] [CrossRef]
  58. Zhao, J.; Huang, F.; Lv, J.; Duan, Y.; Qin, Z.; Li, G.; Tian, G. Do rnn and lstm have long memory? In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, Online, 13–18 July 2020; pp. 1365–11375. [Google Scholar]
  59. Acosta, R.S.; Villegas, C.M.; Norambuena, B.K. Heurísticas para Data Augmentation en NLP: Aplicación a Revisiones de Artículos Científicos. RISTI-Rev. Iber. Sist. Tecnol. Inform. 2019, 34, 44–53. [Google Scholar] [CrossRef] [Green Version]
  60. Keith, B.; Fuentes, E.; Meneses, C. A hybrid approach for sentiment analysis applied to paper reviews. In Proceedings of the ACM SIGKDD Conference, Halifax, NS, Canada, 13–17 August 2017; p. 10. [Google Scholar]
  61. Hu, Z.; Tan, B.; Salakhutdinov, R.R.; Mitchell, T.M.; Xing, E.P. Learning data manipulation for augmentation and weighting. In Proceedings of the 33rd Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Advances in Neural Information Processing Systems. Volume 32. [Google Scholar]
  62. Gil-González, J.; Valencia-Duque, A.; Álvarez-Meza, A.; Orozco-Gutiérrez, Á.; García-Moreno, A. Regularized Chained Deep Neural Network Classifier for Multiple Annotators. Appl. Sci. 2021, 11, 5409. [Google Scholar] [CrossRef]
Figure 1. Validation procedure using some state-of-the-art lexicon-based methods.
Figure 1. Validation procedure using some state-of-the-art lexicon-based methods.
Applsci 12 05472 g001
Figure 2. Accuracy of our approach (LR and SVM) as a function of the number of labeled data.
Figure 2. Accuracy of our approach (LR and SVM) as a function of the number of labeled data.
Applsci 12 05472 g002
Table 1. Rating distribution for the labeled corpora.
Table 1. Rating distribution for the labeled corpora.
RatingNumber of Documents
CorpusCinePaperReview
135136
2923136
31253104
489091
546115
Table 2. Number of news per media in the Colombian conflict corpus.
Table 2. Number of news per media in the Colombian conflict corpus.
MediaN° News
Universidad de Ibagué89
Universidad del Valle60
Universidad de la Sabana281
Uniminuto508
Jorge Tadeo Lozano800
Santiago de Cali94
Universidad del Rosario175
Universidad Autónoma de Bucaramanga366
Table 3. Comparison of sentiment analysis performance (in terms of overall accuracy) between Spanish corpora and English-translated corpora. LR: logistic regression, SVM: support vector machines.
Table 3. Comparison of sentiment analysis performance (in terms of overall accuracy) between Spanish corpora and English-translated corpora. LR: logistic regression, SVM: support vector machines.
Original CorporaMachine-Translated
(Spanish)Corpora (English)
LRSVMLRSVM
CorpusCine 82.30 ± 1.01 83.62 ± 1.06 81.70 ± 1.24 83.40 ± 0.51
PaperReviews 78.09 ± 1.01 79.52 ± 2.54 77.61 ± 2.04 76.66 ± 2.07
Table 4. Prediction accuracy of LRSentiA for the defined confidence groups (Equation (3)) in a machine-translated version of CorpusCine and PaperReview datasets.
Table 4. Prediction accuracy of LRSentiA for the defined confidence groups (Equation (3)) in a machine-translated version of CorpusCine and PaperReview datasets.
Confidence GroupConfidence Score ρ Equation (2)Accuracy %# Reviews
CorpusCineVery-high ( 0.33 , 1.00 ] 74.23788
High ( 0.24 , 0.33 ] 68.23318
Low ( 0.14 , 0.24 ] 61.90504
Very-low ( 0.01 , 0.14 ] 58.06856
Zero ( 0.00 , 0.01 ] 47.16159
PaperReviewsVery-high ( 0.69 , 1.00 ] 63.0492
High ( 0.51 , 0.69 ] 48.2729
Low ( 0.34 , 0.51 ] 46.6645
Very-low ( 0.01 , 0.34 ] 51.2580
Zero ( 0.00 , 0.01 ] 28.1532
Table 5. Performance of lexicon-based approaches for CorpusCine and PaperReview datasets and their machine-translated English versions. For the Spanish lexicon, we use the available in https://www.kaggle.com/general/88685 (accessed on 21 March 2021). The best performance is highlighted in bold.
Table 5. Performance of lexicon-based approaches for CorpusCine and PaperReview datasets and their machine-translated English versions. For the Spanish lexicon, we use the available in https://www.kaggle.com/general/88685 (accessed on 21 March 2021). The best performance is highlighted in bold.
Corpus LanguageLexiconPrecisionRecallF1 ScoreAccuracy
CorpusCineEnglish (translated)TextBlob 0 . 6975 0.5828 0 . 6350 0 . 5939
VADER 0.5966 0.5685 0.5822 0.5764
SpanishSpanish lexicon 0.5500 0 . 6084 0.5900 0.5649
PaperReviewsEnglish (translated)TextBlob 0.6645 0.6062 0.6340 0.5271
VADER 0 . 6918 0 . 6628 0 . 6770 0 . 6007
SpanishSpanish lexicon 0.6382 0.6337 0.6359 0.5921
Table 6. Performance of our hybrid approach using four different classification algorithms and compared with state-of-the-art approaches. LR: Logistic Regression, SVM: Support Vector Machine, RF: Random Forest, NB: Naive Bayes.
Table 6. Performance of our hybrid approach using four different classification algorithms and compared with state-of-the-art approaches. LR: Logistic Regression, SVM: Support Vector Machine, RF: Random Forest, NB: Naive Bayes.
SourceApproachLabels?PrecisionRecallF1 ScoreAccuracy
CorpusCine[55]Unsupervised 0.6393 0.6274 0.6333 0.6316
[56]Supervised 0.8721 0.8701 0.8710 0.8708
[53]Supervised--- 0.7713
[57]Supervised 0.8531 0.8959 0.8739 0.8667
[53]Hybrid--- 0.8086
[54]Hybrid 0.8858 0.8857 0.8856 0.8857
Proposal (LR)Hybrid 0.6727 0.6632 0.6679 0.6664
Proposal (SVM)Hybrid 0.6624 0.6509 0.6566 0.6546
Proposal (RF)Hybrid 0.6484 0.5959 0.6210 0.6042
Proposal (NB)Hybrid 0.6384 0.5744 0.6047 0.5840
PaperReviews[58]Supervised 0.4621 0.3596 - 0.3600
[59]Supervised 0.7600 0.6600 0.7100 0.7400
[60]Hybrid 0.7100 0.7200 0.7100 0.7100
Proposal (LR)Hybrid 0.695 0.5407 0.6082 0.4212
Proposal (SVM)Hybrid 0.6942 0.5378 0.6061 0.4176
Proposal (RF)Hybrid 0.6942 0.5378 0.6061 0.4176
Proposal (NB)Hybrid 0.6942 0.5378 0.6061 0.4176
Table 7. Sentiment analysis for the Colombian conflict dataset. We compare the behavior of our proposal with qualitative labels for each media. For our approach, we present the predicted polarity and the ratio of documents whit positive semantic orientation.
Table 7. Sentiment analysis for the Colombian conflict dataset. We compare the behavior of our proposal with qualitative labels for each media. For our approach, we present the predicted polarity and the ratio of documents whit positive semantic orientation.
MediaQualitative PolarityProposal
RLSVMRFNB
Universidad de IbaguéPP( 0.8000 )P(0.7888)P(0.7666)P(0.8333)
Universidad del VallePN( 0.1016 )N(0.1016)N(0.1016)N(0.1016)
Universidad de la SabanaPP( 0.7651 )P(0.8066)P(0.8259)P(0.8701)
UniminutoPP( 0.7199 )P(0.7317)P(0.8579)P(0.8086)
Jorge Tadeo LozanoPP( 0.7101 )P(0.8412)P(0.8941)P(0.8981)
Santiago de CaliPN( 0.4772 )N(0.4512)N(0.4610)N(0.4090)
Universidad del RosarioPP( 0.5614 )P(0.6359)P(0.6052)P(0.6403)
Universidad Autónoma de BucaramangaPP( 0.5390 )P(0.5306)P(0.6170)P(0.7142)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rendón-Cardona, P.; Gil-Gonzalez, J.; Páez-Valdez, J.; Rivera-Henao, M. Self-Supervised Sentiment Analysis in Spanish to Understand the University Narrative of the Colombian Conflict. Appl. Sci. 2022, 12, 5472. https://doi.org/10.3390/app12115472

AMA Style

Rendón-Cardona P, Gil-Gonzalez J, Páez-Valdez J, Rivera-Henao M. Self-Supervised Sentiment Analysis in Spanish to Understand the University Narrative of the Colombian Conflict. Applied Sciences. 2022; 12(11):5472. https://doi.org/10.3390/app12115472

Chicago/Turabian Style

Rendón-Cardona, Paula, Julian Gil-Gonzalez, Julián Páez-Valdez, and Mauricio Rivera-Henao. 2022. "Self-Supervised Sentiment Analysis in Spanish to Understand the University Narrative of the Colombian Conflict" Applied Sciences 12, no. 11: 5472. https://doi.org/10.3390/app12115472

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop