Next Article in Journal
Modelling French and Portuguese Mortality Rates with Stochastic Differential Equation Models: A Comparative Study
Next Article in Special Issue
Transformer-Based Composite Language Models for Text Evaluation and Classification
Previous Article in Journal
On Bilevel Monotone Inclusion and Variational Inequality Problems
Previous Article in Special Issue
A Study on Double-Headed Entities and Relations Prediction Framework for Joint Triple Extraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sentiment Difficulty in Aspect-Based Sentiment Analysis

by
Adrian-Gabriel Chifu
* and
Sébastien Fournier
Aix-Marseille Université, Université de Toulon, CNRS, LIS, 13007 Marseille, France
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(22), 4647; https://doi.org/10.3390/math11224647
Submission received: 14 October 2023 / Revised: 9 November 2023 / Accepted: 10 November 2023 / Published: 14 November 2023

Abstract

:
Subjectivity is a key aspect of natural language understanding, especially in the context of user-generated text and conversational systems based on large language models. Natural language sentences often contain subjective elements, such as opinions and emotions, that make them more nuanced and complex. The level of detail at which the study of the text is performed determines the possible applications of sentiment analysis. The analysis can be done at the document or paragraph level, or, even more granularly, at the aspect level. Many researchers have studied this topic extensively. The field of aspect-based sentiment analysis has numerous data sets and models. In this work, we initiate the discussion around the definition of sentence difficulty in this context of aspect-based sentiment analysis. To assess and quantify the difficulty of the aspect-based sentiment analysis, we conduct an experiment using three data sets: “Laptops”, “Restaurants”, and “MTSC” (Multi-Target-dependent Sentiment Classification), along with 21 learning models from scikit-learn. We also use two textual representations, TF-IDF (Terms frequency-inverse document frequency) and BERT (Bidirectional Encoder Representations from Transformers), to analyze the difficulty faced by these models in performing aspect-based sentiment analysis. Additionally, we compare the models with a fine-tuned version of BERT on the three data sets. We identify the most challenging sentences using a combination of classifiers in order to better understand them. We propose two strategies for defining sentence difficulty. The first strategy is binary and considers sentences as difficult when the classifiers are unable to correctly assign the sentiment polarity. The second strategy uses a six-level difficulty scale based on how many of the top five best-performing classifiers can correctly identify sentiment polarity. These sentences with assigned difficulty classes are then used to create predictive models for early difficulty detection. The purpose of estimating the difficulty of aspect-based sentiment analysis is to enhance performance while minimizing resource usage.

1. Introduction

Sentiment analysis, also known as opinion mining, is a field of natural language processing that aims to automatically identify and extract subjective information from texts. Sentiment analysis has numerous applications, ranging from marketing to politics, and it has become an increasingly popular topic of research in the past decade [1,2]. Sentiment analysis can be used to identify the sentiments of customers towards a particular product, the opinions of voters towards a political candidate, or the emotions of patients towards their medical condition, among other applications.
Despite recent advancements in sentiment analysis, the detection and analysis of sentiments remain challenging due to several factors. One of the most significant challenges in sentiment analysis is the ambiguity of language [3,4,5]. For example, the word “hot” can refer to temperature, attractiveness, or anger, and it may be difficult for algorithms to determine which meaning is intended in a particular text. In the sentence “This new restaurant has some hot dishes”, the sentiment may be positive because the dishes are delicious or negative because they are too spicy. Ambiguity in language is further complicated by the use of slang, idioms, and regional dialects, which can vary widely even within the same language. For instance, the phrase “cool beans” means “excellent” in American English, but it may be meaningless or confusing to non-native speakers.
Another challenge in sentiment analysis is the detection of tone and sarcasm [6,7,8]. Texts often contain tones that can be difficult for algorithms to detect, and sarcasm can be especially challenging. For example, a person might say “great” in a sarcastic tone to indicate the opposite of what the word usually means. In the sentence “I love spending hours in traffic every day”, the sentiment may be negative despite the positive connotation of the word “love” because the text is sarcastic. Detecting the tone of a text is crucial in understanding the sentiment, as the same words can have different meanings depending on the tone in which they are expressed.
Cultural and contextual differences also pose a challenge to sentiment analysis [9]. Sentiments can vary based on culture and context, and what might be considered positive in one culture may not be the same in another. For example, in some cultures, being direct and blunt is considered a positive trait, while in others, it may be seen as negative. Let us analyze the sentence “The government imposed a strict lockdown to prevent the spread of COVID-19”. This sentence expresses a positive sentiment towards the aspect of lockdown, as it implies that the government is taking proactive measures to protect the public health and safety. However, this sentiment may not be shared by people from cultures that value individual freedom and autonomy more than collective welfare and security. For them, the lockdown may be seen as a negative aspect that infringes on their personal rights and choices. Sentiment analysis algorithms must be trained on diverse data sets to overcome such differences.
Data quality and quantity are also crucial factors in the effectiveness of sentiment analysis algorithms. Sentiment analysis algorithms require large amounts of high-quality data to train effectively. However, it can be challenging to gather such data, especially for less common languages or topics. Additionally, data quality issues such as noise, missing data, or bias [10] can affect the accuracy of sentiment analysis. In the sentence “I bought a phone from XYZ company, and it’s terrible”, the sentiment may be negative towards the phone, but it could also be negative towards the company or the customer service. Without additional context or information, it may be difficult for algorithms to determine the sentiment accurately.
One way to perform sentiment analysis is to examine different levels of granularity: the whole document, a single paragraph, a sentence, or even a specific aspect. However, each level of analysis may encounter the challenges that are discussed previously in this introduction. In this work, we will focus on the most fine-grained level, the aspect level. Indeed, apsect-based is the level for which research is currently the most productive, and consequently also generates the production of corpora whose expressiveness of sentiment is more subtle and therefore potentially more difficult to analyze.
Finally, sentiment analysis algorithms may exhibit bias [10] and subjectivity due to the training data used or the biases of the developers. For example, if a sentiment analysis algorithm was trained on texts from a particular political perspective, it may not perform well on texts from other perspectives. Bias and subjectivity can also arise from the choice of sentiment lexicons, which are dictionaries of words and phrases that are labeled.
Recent years have seen advances in language models, particularly the emergence of BERT [11] and GPT [12]. These have enabled algorithms to better capture the semantics of texts, resulting in a marked improvement in performance. This raises the question of whether the challenges mentioned above still exist, and if so, how algorithms manage to overcome them. In this article, we explore how algorithms handle these difficulties and which subjective sentences pose the greatest challenge.
Contrary to the usual focus in current aspect-based sentiment analysis research, our aim does not involve achieving better results or introducing a new classification model. Rather, it is to comprehend why existing models are not working in some cases and why some data sets are “simpler” than others. The aim is therefore to observe the behavior of classification models in the aspect-based sentiment analysis task and to estimate the degree of difficulty of the analyzed sentences. Thus, estimating the difficulty of sentiment analysis could enhance performance while minimizing resource usage.
In order to better understand the aspect-based sentiment analysis difficulty, we raise the following research questions:
  • RQ1: How to define difficulty in aspect-based sentiment analysis?
  • RQ2: Is difficulty data set-dependent?
  • RQ3: What is the impact of text representation on performance?
  • RQ4: What is the impact of classification models on performance?
  • RQ5: Are we able to predict difficulty?
  • RQ6: How to better understand difficult sentences (qualitative analysis on difficult sentences)?
To summarize, in order to answer these six research questions, we propose in our work to:
  • Select 3 data sets whose purpose is to perform aspect-based classification. The data sets have been created at a 6–7 year time distance, and choosing three data sets that span over such a large period of time would reflect the evolution of the field;
  • Select 21 models and two different text representations in order to analyze their respective behavior and performance when faced with aspect-based sentiment analysis;
  • Conduct numerous experiments in order to better understand the challenges faced by the models;
  • Investigate automatic sentence difficulty definition and estimation using learning-based models.
The remainder of this paper is organized as follows. Section 2 provides an overview of aspect-based sentiment analysis, text representation models and the concept of difficulty in Information Retrieval. Section 3 presents three different data sets used in the experiments. Then, Section 4 presents the 2 text representations that have been used and Section 5 presents the different used models and the employed implementations. Section 6 explains the experimental protocol and presents the aspect-based sentiment polarity classification results. Section 7 proposes difficulty definitions and tests if the difficulty may be predicted, while Section 8 discusses the results and answers the research questions. Finally, Section 9 concludes the paper and suggests directions for future work.

2. Related Work

We further divide our related work into themes that influence our analysis, including Sentiment Analysis, Aspect-based Sentiment Analysis, Text Representation, and Query Difficulty in Information Retrieval.

2.1. Sentiment Analysis and Aspect-Based Sentiment Analysis

The goal of sentiment analysis is to identify and categorize emotions and sentiments expressed in written text automatically. Sentiment analysis is a fairly broad area of research and can be defined at several levels—at a document level, paragraph level, or sentence level—but also at a much finer level, at an aspect level, which is the element on which subjectivity is focused. In the literature research works are often classified into three categories: machine learning, deeplearning, and ensemble learning. Among the most effective machine learning techniques for this task are naive Bayes [13,14,15] and SVM [16,17,18]. Algorithms based on deeplearning [19] include RNNs [20,21], LSTMs [22,23,24], and transformers [25,26]. Ensemble-based methods [27] combine multiple classifiers, which can fall into either of the previous categories. In response to the large amount of work on the subject of sentiment analysis, a number of surveys have been carried out recently [27,28,29,30]. However, for several years now, research has been more focused on multimodal, multilingual sentiment analysis and on the finest level of sentiment analysis: the aspect-based level.
The finesse of the analysis at the aspect level means that in the vast majority of cases we can be sure of having a uniqueness of subjectivity. In other words, there are not two opposite degrees of subjectivity for the same aspect. This aspect of sentiment analysis is often divided into two distinct tasks; the first consists of finding and extracting the aspects. The second is to estimate the subjectivity, generally reducing the problem of estimating the degree of subjectivity to a simpler problem of classification. This involves classifying the sentences or nominal phrases containing the aspects as neutral (no subjectivity), positive (the author speaks positively about the aspect), and negative (the author speaks negatively about the aspect). Sometimes the scale of values used is broader (often 5 classes) and sometimes there are other classes, as is the case for the data we are going to use (e.g., conflict). In this document, we will only deal with the second task, i.e., the classification of subjectivity by taking aspects into account. To carry out this task, the methods and models used are relatively similar to those used in sentiment analysis at a coarser level. This research was introduced by the seminal work of [31] and has been developed with the production of numerous models and data sets in various languages. As in the more general context of sentiment analysis, we found similar algorithms but adapted them to the task. Among these algorithms, we can cite the use of SVMs [32] and Naïve Bayes [33]. More recently, the advent of deeplearning has considerably improved model performance. For example, there are models using RNNs, LSTMs [34,35,36], and transformers [37,38]. For further reading, one can also refer to surveys on the subject [39,40,41]. As the most recent work has focused on aspect-based analysis, the recent corpora produced for this task seem to us to have a more subtle expressivity of sentiment. This is why, in order to have a more thorough study of the difficulty in sentiment analysis, we have focused in this work on the aspect-based sentiment analysis, although the conclusions and models we obtain can be applied to sentiment analysis in a general way.

2.2. Text Representation

Text representation for machine learning models has always been a major issue. Initially, the vector representation of documents was only done by taking into account the presence or absence of terms in the document. This representation was then improved by taking into account the frequency of terms in the documents and in the collection [42]. However, such representations take no account of the semantics expressed in the sentences. With the arrival of deeplearning and the emergence of less sparse representations, semantic aspects have been better taken into account. These representations were democratized thanks to the work of [43] and then improved through the work of [44]. The emergence of transformers has also made it possible to obtain new, more efficient representations [11].

2.3. Difficulty in Information Retrieval

As we have already noted, the various existing challenges in aspect-based sentiment analysis make it difficult to classify sentences. Therefore, it is important to detect such difficult sentences to choose a different strategy in order to extract the expressed sentiment. Having multiple strategies in order to classify sentiment based on the detected difficulty helps to reduce the use of resource-intensive algorithms. Thus, we reduce the economic and ecological costs of the models. Difficulty detection is a key area of research, especially in information retrieval. In the 2000s, research on query difficulty began and many predictors were defined based on distribution [45], ambiguity [46], and complexity [47]. In the field of information retrieval, models predicting difficulty are divided according to whether they use pre- or post-retrieval predictors. Models using pre-retrieval predictors include those presented in [48,49,50]. These models use statistics on the occurrence of query terms. Among the models using post-retrieval predictors we find the work of [51,52,53,54,55]. These models will use the results of the information retrieval models to make their predictions. With the advent of deep neural networks, recent work has used such models to predict difficulty. These studies notably include [55,56]. Recently, [57,58] have raised the question of the effectiveness of evaluating the difficulty of Neural Information Retrieval based on PLM (Pre-trained Language Models).
Predicting the difficulty of a sentence in sentiment classification can not only improve the performance of the algorithms by selecting models according to the difficulty encountered, but can also make the systems more resource-efficient.
The notion of difficulty in aspect-based sentiment analysis has not yet been studied. Inspired by the work that has been conducted on information retrieval, particularly inspired by works on post-retrieval predictors, we are conducting experiments that will enable us to gain a better understanding of the notion of difficulty in this specific case. In addition, we have also sought to automate the estimation of difficulty for aspect-based sentiment analysis. In the remaining part of this paper we present the experiments that have been carried out in order to better understand where the difficulties lie in the sentiment classification task based on aspects.

3. Reference Data Sets Used for Corpus Building

In this section, we will introduce the reference data sets. These data sets have been essential because they provided the data for our corpus. Following that, we will explain the process of building the corpus. After that, we will move on to an exploratory analysis.
We used three data sets for our experiments: “Laptops”, “Restaurants”, and “MTSC”. We provide details on the data sets below. In order to perceive the temporal evolution of the difficulty of the task, we selected three corpora, each spaced about 6–7 years apart. We also chose two different objects of study: reviews and political news. The first and second corpora concern Laptop and restaurant reviews, respectively. They were published in 2009 and 2014. The third concerns political news and was published in 2021.
On each of these corpora and two different representations (TF-IDF and BERT), we carried out experiments with 21 learning models in order to discover on which corpus the models had the most difficulty. We consider a model to have difficulty if its performance is below the median of model performance for a given corpus. We also consider difficulty at a more micro level, by looking at the sentences and paragraphs posing the most difficulty for the selected models.

3.1. Laptops

The SemEval Laptop Reviews data set [59] is often associated with the laptops data set for aspect-based sentiment analysis. It was first used in the SemEval-2014 Task 4: Aspect-Based Sentiment Analysis challenge. This data set is used for two subtasks: aspect identification (Subtask 1) and aspect-oriented sentiment polarity classification (Subtask 2).
It contains more than 3000 English sentences from customer reviews of laptops. It focuses on analyzing sentiments at a more granular level, targeting various aspects or attributes of laptops such as performance, battery life, design, and usability. Expert human annotators have labeled the aspect terms of the sentences and their sentiment. A part of this data set was kept as test data by the organizers of the SemEval-2014 competition.
This data set enables researchers to analyze sentiment polarities towards specific aspects of laptops. It provides insights into customers’ preferences and satisfaction levels.
Each review may contain one or multiple aspects. Each aspect is assigned one of four possible labels: “positive”, “negative”, “neutral”, or “conflict”.

3.2. Restaurants

The restaurant data set for aspect-based sentiment analysis consists of more than 3000 English sentences from restaurant reviews initially proposed by Ganu et al. [60]. As it has already been pointed out, aspect-based sentiment analysis is different from traditional sentiment analysis which focuses on overall sentiment of reviews. The original data set includes annotations for coarse aspect categories and overall sentence polarities. It has been modified for SemEval-2014 [59] to include annotations for aspect terms occurring in the sentences (Subtask 1), aspect term polarities (Subtask 2), and aspect category-specific polarities (Subtask 4). Some errors in the original data set have been corrected. Human annotators identified the aspect terms of the sentences and their polarities (Subtasks 1 and 2). Additional restaurant reviews, not present in the original data set, have been annotated in the same manner and kept by the organisers of SemEval-2014 as test data.
This data set covers reviews related to various aspects of restaurants such as food quality, service, ambiance, price, and cleanliness. Each review is labeled with sentiment ratings for each aspect. The polarities that may be identified are “positive”, “negative”, “neutral”, and “conflict”.

3.3. MSTC

NewsMTSC is an aspect-based sentiment analysis data set proposed by the authors of [61]. It focuses on news articles about policy issues and contains over 11,000 labeled sentences sampled from online US news outlets.
Most of the sentences contain several aspects that are mentioned in the data set. The conflict class does not exist, so only three polarity levels may be encountered: “positive”, “negative”, and “neutral”.
Next, we will describe how our corpus was constructed.

3.4. Corpus Preparation

There are several files formats across the considered data sets and each data set contains multiple files. The number of attributes per data set may also vary. We have unified the data into one file per corpus, in csv format.
We have kept one id generated by us, the id from the original data set (for tracing purposes), the sentences, the start position of the aspect, the end position of the aspect, the aspect, and the polarity class. We will make the data sets publicly available upon acceptance.
For the “Laptops” corpus, there are several files in the original data set. These files come in both csv and xml formats. We excluded the files Laptops_Test_Data_PhaseA and Laptops_Test_Data_PhaseB as they do not contain annotations. We used the annotated sentences from Laptop_Train_v2 as training data and the annotated sentences from laptops-trial as test data. We merged all the data into one file named laptops.csv. It has 2407 rows in total.
The “Restaurants” corpus has the same structure as the Laptops corpus. We selected the same corresponding elements, which resulted in our restaurants.csv file with 3789 rows.
The MTSC data from the NewsMTSC corpus is in json format. The files containing the data are train, devtest_mt, and devtest_rw. We did not consider devtest_mt since it is designed to evaluate a model’s classification performance only on sentences with at least two target mentions, which is out of the scope of our research. Thus, we used train as the train data and devtest_rw (w stands for “real-world”) as the test data. The resulting data file, called MTSC.csv, contains 9885 rows in total.
This last data set contains polarity scores that we have encoded as classes. There are three possible scores: 2.0 = “negative”, 4.0 = “neutral”, and 6.0 = “positive”. We have converted the scores to their corresponding strings. The “conflict” class is not present in this corpus. Sentences with multiple aspects have been duplicated the same number of times as the number of aspects. For example, a sentence with three aspects will appear three times in the data set, one time for each aspect.
The statistics of the data sets used for the experiments are shown in Table 1.
In the following, we present our exploratory analysis conducted on the three data sets.

3.5. Exploratory Analysis

We analyzed the three data sets briefly to gain insight into their structure, detect any biases, and form hypotheses.
We first examined the polarity ratios in each data set. Figure 1 shows these ratios. The most balanced data set is “MTSC”, with 37.9 % of the data belonging to the negative class. The other two data sets are more unbalanced, with “Restaurants” being the most unbalanced, having 58.9 % of the data in the positive class and only 2.4 % in the conflict class. The conflict class is the rarest across all three data sets, with 1.9 % in “Laptops” and none in “MTSC”.
First, we look at the polarity distributions for each data set, taking into account the train/test splits.
Figure 2 shows the class occurrences in the “Laptops” data set with respect to the train and test splits. Notably, the conflict class has no occurrences in the test set. Additionally, the data set relatively maintains the positive/negative/neutral ratios in the train set ( 41.0 % , 36.0 % , 19.1 % ) as compared to the test set ( 1.2 % , 0.7 % , 0.2 % ).
Figure 3 illustrates the “Restaurants” data set class occurrences. The remarks for “Laptops” are concurrent with this data set as well, and we underline once more that the positive class is significantly more represented than the others.
Figure 4 shows the class occurrences for the “MTSC” data set. This data set is balanced in both the train and test splits. The conflict class does not appear in this data set.
We have decided to keep the conflict class, even though it is not found in all data sets and there is no occurrence of it in any test data. This is to preserve the original data distributions for each data set as much as possible.
We now focus on analyzing the tokens and sentences from the data sets. The results are summarized in Table 2. It is common to find duplicate sentences in the data set, as one sentence may contain multiple aspects and thus may appear more than once, by design. However, the number of unique aspects is even lower than the number of unique sentences, meaning there may be multiple sentences for one aspect. Additionally, the maximum number of tokens per aspect ranges from 6 to 31 on the Laptops and MTSC data sets, respectively. This indicates that aspects may be lengthy and not necessarily composed of only one or two words.
Following this interesting observation, we continued analyzing the number of tokens from the aspects. Figure 5a depicts the frequency of the aspects by their length in terms of the number of tokens. One may notice that the vast majority of the tokens have one or two aspects. This holds for the three data sets. There is a long-tail distribution starting from aspects with four tokens per aspect and going up to 31 tokens per aspect in the case of MTSC. The maximum number of aspects for the Restaurants data set is 19. This long-tail distribution is illustrated in Figure 5b.
To gain a better understanding of the data sets, we conducted a linguistic analysis of sentiment polarity for each data set. We calculated the average number of tokens, nouns, verbs, named entities, and adjectives per instance. The results are shown in Table 3. One may notice that the sentences from MTSC are usually longer than the sentences from the other data set. However, it has been proven for query difficulty prediction that the query length is not correlated with the Average Precision performance measure [62]. Another interesting observation is that the average number of named entities in the MTSC corpus is significantly higher than for the other two data sets. This may add up to the difficulty of aspect-based polarity classification.
We wanted to investigate if the amount of nouns, verbs, or adjectives differs depending on the sentiment class. The statistics show that the positive classes have fewer nouns and verbs than the negative classes in the three data sets. However, if we normalize the number of tokens, the negative classes have the lowest values, except for “MTSC” where the ratio is almost the same. Moreover, these values are very close and the difference is not significant enough.

4. Text Representations

To assess the effect of text representations on the accuracy of classification, we selected two different text representation models: TF-IDF and BERT.

4.1. TF-IDF

Term Frequency-Inverse Document Frequency (TF-IDF) [63] is a widely used technique in natural language processing. It assigns weights to words in a document based on their frequency in the document and their rarity across all documents in a corpus. TF-IDF captures the importance of words by emphasizing both their local and global significance. The term frequency component measures the occurrence of a word within a document, while the inverse document frequency factor highlights the rarity of a word across the corpus. By multiplying these values, TF-IDF assigns higher weights to terms that are both frequent within a document and unique across the entire corpus. This approach has been successful in various text mining tasks, such as information retrieval, text classification, and recommendation systems.
Thus, TF-IDF can be calculated following this formula: T F I D F i , j = t f i , j · l o g ( N / d f i ) , where t f i , j is the number of occurrences of the term i in the document j, d f i is the number of documents containing the term i, and N is the total number of documents in the data set.
We normalized the texts by removing HTML tags and non-alphabetic characters, transforming it into lowercase, tokenizing it with the nltk tokenizer (https://www.nltk.org/api/nltk.tokenize.html (accessed on 10 October 2023)), removing the stopwords with the nltk stopword list (https://www.nltk.org/api/nltk.corpus.html#module-nltk.corpus (accessed on 10 October 2023)), and stemming the tokens with Porter Stemmer (https://www.nltk.org/_modules/nltk/stem/porter.html (accessed on 10 October 2023)).

4.2. BERT

Bidirectional Encoder Representations from Transformers (BERT) [64] is a cutting-edge natural language processing technique. It uses transformer-based neural networks to generate contextualized word representations, instead of relying on fixed word embeddings. By pre-training on large amounts of text data and then fine-tuning on specific downstream tasks, BERT models can capture intricate semantic relationships between words and sentences. This leads to effective text vectorization, where each word or sentence is mapped to a dense representation in a high-dimensional vector space. BERT text vectorization has revolutionized many NLP tasks and opened up new possibilities in areas like sentiment analysis, question answering, and language translation.
We combined sentences and aspects into a list in the format [sentence, aspect], and then fed it to the tokenizer (https://huggingface.co/transformers/v3.0.2/model_doc/auto.html#autotokenizer (accessed on 10 October 2023)). We used the AutoModel from the transformers module (https://huggingface.co/transformers/v3.0.2/model_doc/auto.html#automodel (accessed on 10 October 2023)) to vectorize the tokens. Both the tokenizer and the model are based on distilbert-base-uncased, a pretrained model. A basic illustration of the sentence processing pipeline based on BERT is depicted in Figure 6. We tried marking the aspect inside the sentence at its corresponding position instead of adding it at the end, but the resulting representation was less effective.

5. Classification Models

This section presents the different models we used for our experiments. We have used the models proposed by the LazyText python module (https://github.com/jdvala/lazytext (accessed on 10 October 2023)) for text classification tasks. This module makes it easy to build and train text classification models. It provides a user-friendly interface and automates tedious tasks. With LazyText, users can preprocess their text data, apply feature extraction techniques such as TF-IDF or word embeddings, and train different classification models in a few lines of code. The module also offers functions to evaluate model performance and make predictions on new data. The LazyText model does not store details like the predicted class labels. To access elements such as class labels and confusion matrices, we created the models with Scikit-learn (https://scikit-learn.org/stable/supervised_learning.html (accessed on 10 October 2023)). Scikit-learn is also used internally by LazyText to train models and make predictions. The DummyClassifier is one of the classifiers. It makes predictions without considering the input features. This classifier serves as a baseline to compare with other complex classifiers. We applied the default strategy for this classifier, which always returns the most frequent class label in the data given to fit.
Next, we present the classification results using BERT and TF-IDF representations.

6. Experiments and Results

In this section, we present the results of our experiments on the three data sets presented while varying the textual representations and models used. Table 4 summarizes the hardware and software environments used for our experiments.

6.1. Classification Results with TF-IDF Representations

We used the scikit-learn vectorizer (https://scikit-learn.org/stable/modules/generated/sklearn.feature_extraction.text.TfidfVectorizer.html (accessed on 10 October 2023)) with default parameters to vectorize the sentences and the aspects separately. Then, we combined the sentence and aspect vectorizations columnwise, by placing the sentence vector first, followed by the aspect vector.
We used 20 supervised classifiers from Scikit-learn and a DummyClassifier as described in Section 5. The DummyClassifier predicts the most frequent class. We report the macro-averaged metric and weighted-averaged metric results, for all the three data sets. The macro-averaged and weighted-averaged measure, respectively, are computed by the classification\_report function from the scikit-learn python module, as follows: “The reported averages include macro average (averaging the unweighted mean per label)” and “weighted average (averaging the support-weighted mean per label)” (https://scikit-learn.org/stable/modules/generated/sklearn.metrics.classification_report.html (accessed on 10 October 2023)).
Table 5 and Table 6 show the classification results of macro and weighted metrics respectively, for the “Laptops” data set. These results are based on the TF-IDF text representations.
Analyzing Table 5, it is evident that CalibratedClassifierCV is the best classifier, achieving results higher than 90% in F1 measure. This is significantly better than DummyClassifier, indicating that the model was able to accurately distinguish between classes without any bias due to their distribution. Table 6 shows that six models achieved F1 scores of over 98%. This suggests that, even with a simple TF-IDF representation that does not capture advanced language semantics, the models can easily classify the corpus. Thus, we can conclude that this corpus is relatively easy to classify.
In Table 7 and Table 8, we present the precision, recall, and F1-score results for the “Restaurants” data set. These results are based on the same text representation, TF-IDF, and are macro-averaged and weighted, respectively.
Table 7 shows that five models achieved an excellent score of 100% in F1 measure. Table 8 also reveals that the same models achieved a score of 100% when using TF-IDF. These results demonstrate that the models can easily distinguish between classes, even though the classes are imbalanced. Thus, the “Restaurants” data set is easier to classify than the “Laptops” data set.
Finally, we summarize the classification results of the “MTSC” data set using TF-IDF text representations in Table 9 and Table 10. Both macro-averaged and weighted results are presented.
Table 9 shows different results from the previous two data sets. BernoulliNB has the highest F1 measure of 61%. Table 10 also reveals that BernoulliNB is the best model with an F1 measure of 62.6%. This indicates that it is more challenging to differentiate classes in the “MTSC” data set than in the “Restaurants” and “Laptops” corpora. Table 1 reveals that sentences in this data set are longer and the text is from newspapers instead of reviews. The sentiment vocabulary is likely to be more subtle and less direct than in the case of the other two data sets. This raises the question of how to incorporate semantics into the models. In this experiment, we used TF-IDF, which does not capture the semantics intrinsically. Therefore, using BERT to represent the text may be a solution or may at least improve the results. This is what we will explore in the next section.

6.2. Classification Results with BERT Representations

As in Section 6.1, we present here the classification results, macro-averaged and weighted, for the three data sets. The difference is that the text representations are based on BERT instead of TF-IDF.
Table 11 and Table 12 present the classification results for the “Laptops” data set, with BERT text representations.
Eleven out of twenty-one models showed an increase (and even a great increase for some) in results when using BERT, compared to TF-IDF. Four models stayed the same, while six models (BernoulliNB, CalibratedClassifierCV, BaggingClassifier, DecisionTreeClassifier, Perceptron, and PassiveAgressiveClassifier) experienced a decrease. These models did not take into account the new representation or the semantic dimension. The best model was MLPClassifier, with an F1 measure of 98%. This result was 7 points higher than the best result previously observed. On the other hand, results presented in Table 12 showed the same maximum as with TF-IDF, but for different models and only for four of them.
Next, Table 13 and Table 14 summarize the classification results with BERT text representations for the “Restaurants” data set.
TF-IDF already gave excellent results. Four of the models that achieved the best score with TF-IDF kept this score when using BERT-based representation. BaggingClassifier improved and got the highest score, while MLPClassifier decreased slightly. Surprisingly, most models saw a decrease in F1 measures. Twelve models dropped, four increased, and five stayed the same. The models that stayed the same were the decision tree-based models and the DummyClassifier, whose performance was unaffected by the textual representation. The high performance of TF-IDF compared to a more complex representation of the text using BERT may explain this drop. The relative simplicity of the sentences in the corpus means a complex representation of the text is not necessary for classification. This experiment suggests a hypothesis that can be tested further: a sentence can be considered difficult if it requires a complex representation incorporating semantics. Can we then construct an indicator of difficulty on this basis? We believe that a selective model can be trained to determine sentence difficulty. When a sentence is deemed difficult, its text can be represented using a complex model such as BERT instead of TF-IDF. This increases the likelihood of accurately classifying the polarity. The benefit of this approach is improved efficiency. Simple sentences can be processed quickly, while complex, time-, and resource-consuming text representation is reserved for difficult sentences.
Finally, the classification results on the “MTSC” data set, based on BERT text representation, are shown in Table 15 and Table 16, macro-averaged, and weighted, respectively.
We can observe from Table 15 that, in the case of “MTSC”, only BernouilliNB and ExtraTreeClassifier have lower performances when using BERT representations compared to TF-IDF. This indicates that the semantics in the textual representation significantly enhance the model’s performance. The longer sentences and more subtle expressions of sentiment in the data set require additional knowledge to better comprehend the sentences to classify. This confirms our earlier hypothesis, namely that in order to better classify sentiment polarity in the case of difficult sentences, a more complex text representation would be better suited. The same trends can be seen in Table 16. The best-performing model is LogisticegressionCV, with an F1 measure of 70.8%.

6.3. Fine-Tuned BERT

Fine-tuned BERT models have become popular in natural language processing for their capacity to improve performance on various text classification tasks. BERT, which is pre-trained on a large corpus of unlabeled text data, provides a strong base for language comprehension. The fine-tuning process involves training the model on domain-specific labeled data to make it suitable for the target task. A basic illustration of the sentence processing pipeline based on fine-tuned BERT is depicted in Figure 7. By changing the model’s parameters, it learns task-specific patterns and increases its predictive accuracy. This fine-tuning approach has been successful in achieving the best results in sentiment analysis [65], named entity recognition [66], and other classification tasks.
We fine-tuned three BERT models, one for each data set, using BertTokenizer and BertForSequenceClassification (https://huggingface.co/docs/transformers/model_doc/bert (accessed on 10 October 2023)) from the transformers python module (https://github.com/huggingface/transformers (accessed on 10 October 2023)), starting from the bert-base-uncased pre-trained model. We tried the distilled model as well, but the results were very low. We trained the models for three epochs with a batch size of 8, using the default parameters (Adam optimizer, padding, truncation, and a learning rate of 10 5 ).
We emphasize that we used the default parameter settings for all models and representations. Our goal is to gain a better understanding of the difficulty in aspect-based sentiment analysis, not to introduce a new model or enhance existing results.
Table 17, Table 18 and Table 19 show the classification report results of the fine-tuned BERT on the “Laptops”, “Restaurants”, and “MTSC” data sets, respectively.
Comparing with previous experiments, fine-tuned BERT is better than TF-IDF for any model in the “Laptops” data set for the same reasons mentioned in the comparison between BERT and TF-IDF. We observe the same behavior in the “Restaurants” data set, where the use of BERT does not improve the results. However, the improvement is greater in the “MTSC” data set. Here, using the fine-tuned BERT model is even better than just using BERT as a textual representation. This improvement shows that the model takes advantage of the semantics embedded in the BERT model and also benefits from the adaptation of BERT to the “MTSC” data set, particularly the adaptation to the numerous named entities present in the “MTSC” data set.

6.4. Ensemble Learning to Improve Performance

Ensemble learning is a machine learning technique that boosts prediction accuracy and robustness. It combines the outputs of multiple models to make collective predictions, leading to a more reliable result. Bagging, boosting, and stacking are popular ensemble methods. This approach leverages the diversity of models, reducing biases and variances, and improving overall model performance. It has been effective in various domains, such as classification, regression, and anomaly detection, resulting in significant advancements [67,68,69].
We used ensemble learning (majority vote) for the three collections with both TF-IDF and BERT text representations. We employed two strategies: one that included all the classification models, and the other that only included the top 5 models with the highest accuracy.

6.4.1. Majority Vote for TF-IDF

Table 20 shows the majority vote classification report for the “Laptops” data set. Table 21 lists the top 5 models based on accuracy. Table 22 displays the classification report for these 5 models. In this scenario, we do not see any advantage in using a combination of classifiers. The top classifiers have a very low error rate, and the error stays the same for all classifiers. Because the outcomes are not diverse enough, the ensemble of classifiers has the same error as the individual classifiers.
Similarly, Table 23, Table 24 and Table 25, referring to the “Restaurants” data set, present the classification report for all the models, the top 5 models, and the classification report for the top 5 models, respectively.
A perfect score is achieved by the top classifiers on the restaurant corpus. The overall accuracy drops when all the classifiers are combined by majority vote, as the best ones are outnumbered by the rest. However, when only the five best classifiers are chosen, the resulting classifiers are flawless.
Finally, Table 26, Table 27 and Table 28 present the classification report for all the models, the top five models, and the classification report for the top five models, respectively, for the “MTSC” data set.
The “MTSC” data set presents more challenges, resulting in diverse outcomes for the models. Therefore, using an ensemble of classifiers is more appropriate compared to the “Restaurants” and “Laptops” data sets. However, using all 21 classifiers leads to a small decrease in results (around 1 point in F1 measure). On the other hand, using the top five models as an ensemble improves accuracy and maintains the F1 measure. The results from these five classifiers do not differ significantly to make a notable impact on the overall outcome.

6.4.2. Majority Vote for BERT

Similarly to the TF-IDF text representations, Table 29, Table 30 and Table 31 display the results for the “Laptops” data set. Table 32, Table 33 and Table 34 show the results for the “Restaurants” data set. Lastly, Table 35, Table 36 and Table 37 present the results for the “MTSC” data set. These results include the classification report for all the models, the top 5 models based on accuracy, and the classification report for the majority vote of the top 5 models corresponding to each data set.
The sentiment analysis task for the “Laptops” data set is relatively simple. This task produces similar results across different models. Therefore, using majority voting with TF-IDF does not improve performance. It does not matter if all the models are used or if only the top five are used.
The majority vote method, using either all the models or the top five models, did not improve the performance of the BERT-based text representation models on the “Restaurants” data set, as it did not on the “Laptops” data set.
When we apply the majority vote to the models on the “MTSC” data set, we can improve performance. This is true whether we use all models or just the top five. It shows that using text representations that are aware of meaning and having diverse model outputs can lead to better results. We can make the task easier by using representations that have semantic information and by combining multiple classifiers.

7. Sentence Difficulty Definition and Prediction

In this section, we first define the difficulty using two strategies. Following that, we attempt to predict difficult sentences automatically using our data sets. We also present two sampling strategies for the classification models. Additionally, we analyze difficult sentences qualitatively and evaluate the outcomes of our difficulty predictions.

7.1. Defining Difficulty

Correctly predicting difficult sentences in the context of aspect-based sentiment analysis can lead to effective selective approaches. We can leave the easy sentences for light models that do not require much computation, and submit the difficult sentences to heavy, complex models. This way, we can achieve a balance between efficiency and effectiveness.
We propose two strategies to define the difficulty classes:
  • Binary classification. For this strategy, we conducted an analysis of the incorrect predictions across all data sets. To do this, we investigated the majority votes produced by the top five classifiers using both TF-IDF and BERT text representations for each data set. We identified the sentences that were incorrectly classified by both text representations as ’difficult’. For instance, a sentence is assigned to the difficult class if it is wrongly classified by the majority votes of both the BERT and TF-IDF models; otherwise, it is assigned to the easy class. In terms of exact numbers, on the test sets, we found one such difficult sentence in the “Laptops” data set, none in the “Restaurants” data set, and 197 in the “MTSC” data set.
  • Fine-grained (multi-class) classification. For this strategy, we established several levels of difficulty, taking into account the number of correct classifications made by each of the top 5 performing models, while considering both text representations. For instance, the most difficult sentences are those of level 0, since no top 5 classifier was able to correctly classify it. On the other hand, the easiest sentences are those of level 5, for which all the top 5 classifiers had the correct polarity. Since “MTSC” was the most difficult data set, we focused this strategy only on this data set. The most represented classes for BERT are level 5, with 693 and level 0, with 217 sentences, respectively. The remaining 236 sentences are relatively evenly distributed among the four remaining classes.
One can easily notice that both strategies yield unbalanced test data, in terms of class membership. Figure 8 depicts this lack of balance. This leads us to consider two types of sampling, the default one, without any class balancing, and the SMOTE sampling [70], an oversampling technique where the synthetic samples are generated for the minority class.
To begin our analysis, we first focus on qualitative aspects, and then we analyse the difficulty prediction performance.

7.2. Qualitative Analysis of Difficult Sentences in the Chosen Data Sets

We examined the sentences that were incorrectly classified, with respect to the binary strategy. In the “Laptops” data set, there is only one such sentence. This sentences is: “But see the macbook pro is different because it may have a huge price tag but it comes with the full software that you would actually need and most of it has free future updates”, its aspect is “price tag”, and the polarity is “negative”. We have tried ChatGPT to predict the polarity of this sentence and it yielded “positive”. The term “huge” usually has a positive connotation. However, when it is used with “price tag”, it becomes negative. This is because “huge” is a specialized vocabulary used to describe prices, and thus has a negative connotation. This example demonstrates that when specialized terms are the same as everyday language, it can make understanding more challenging.
The “Restaurants” data set contains no difficult sentences. However, the “MTSC” data set has 197 difficult sentences. In order to observe the behavior of very large language models on the most difficult sentences, we chose six of them to analyze and predict their sentiment using ChatGPT. The sentences and predictions are summarized in Table 38.
We can observe that two of the six ChatGPT predictions are incorrect. The sentence “His persona is generally adult” is difficult to classify as “positive” even for a human annotator, due to the implicit reference “his”. Similarly, the last sentence about “President Muhamadu Buhari” appears to be quite neutral.

7.3. Difficulty Prediction Results for MTSC

For both binary and fine-grained classification strategies, with or without SMOTE sampling, we used the test data from MTSC and applied 10-fold cross-validation (https://scikit-learn.org/stable/modules/generated/sklearn.model_selection.cross_val_score.html (accessed on 10 October 2023)) to make sure every observation passed as test data. We employed the same 21 classifiers as we did for the previous experiments. As well, both BERT and TF-IDF text representations are considered. To summarize, there are 21 classifiers, two class definition strategies, two text representations, and two sampling strategies. The mean cross-validation scores are depicted in Figure 9 and Figure 10, by the binary classification strategy and by the fine-grained strategy, respectively.
We noticed that the best performing model in all the cases was the DummyClassifier and we first hypothesized that this occurs because it predicts everything towards the majority class. However, even with SMOTE oversampling, the situation does not change in terms of the best performing models.
For the fine-grained strategy, the performance is significantly lower than for the binary strategy, in general. This is justifiable, since it is more difficult to classify with 6 classes than with 2 classes. BERT is better performing with the fine-grained strategy, while the performances with respect to the text representations are close in the case of the binary classification.
The SMOTE oversampling method allows a better evaluation of the quality of the classifiers for the proposed tasks. For instance, we noticed that the KNeighborsClassifier constantly drops in performance when SMOTE is applied.
Decision tree-based models, logistic regressions, and SVC are generally the best-performing models.
Even though the mean accuracy is high across the models for the binary classification strategy, we cannot conclude that this is a good approach for difficulty prediction, mainly due to the score of the DummyClassifier. This suggests that the models are hardly learning to classify difficulty. However, this launches the discussion about the task of predicting difficulty.

8. Discussion

In this section, we address the research questions outlined in the introduction. We do this by examining the results of our experiments and analysis.
  • RQ1: How to define difficulty in aspect-based sentiment analysis? This is not a straightforward question. If we extrapolate to the field of Information Retrieval, many studies on difficulty prediction focus on the correlation between the predicted and actual effectiveness, without requiring an exact definition of query difficulty [71]. We may struggle to assign the right sentiment to a sentence, regardless of the text format, the type of classifier, and so on. We suggest that the definition of “difficulty” could be context dependent and inherently, the quality of an eventual difficulty prediction depends on the criteria of the definition.
  • RQ2: Is difficulty data set dependent? It appears to be. We have observed that on the “Laptops” and “Restaurants” data sets, there are few or no test sentences for which we fail to predict the polarity. However, this is not the case for the third data set, “MTSC”. We believe that this takes place because the first two data sets are more specific to their domain than the third, which is from the wider news domain. We have noticed that the “MTSC” data set contains more named entities and implicit references, which may contribute to the difficulty level. Expressing subjectivity is less overt and more nuanced. We also note that some sentences from “MTSC” are challenging, even for human annotators. Moreover, ”MTSC”, the most recent data set is the most challenging. We hypothesize that this may correlate with the advances in terms of performance of aspect-based sentiment analysis models that require more challenging data sets to accurately quantify effectiveness.
  • RQ3: What is the impact of text representation on performance? When we look at the classification results from Section 6, we can see that BERT, the most advanced text representation, usually performs better than TF-IDF. BERT captures more complex semantics than TF-IDF. Fine-tuning also appears to be beneficial. But we must be careful not to overfit and the fine-tuning process can be time-consuming. In conclusion, the choice of text representation method can affect performance. Choosing the right representation for a task depends on its difficulty. If the task is simple, selecting a complex representation may lead to a decrease in performance and an increase in IT costs.
  • RQ4: What is the impact of classification models on performance? We observed a similar pattern for the classification models as for the text representation in RQ3. The selection of the classifier has a significant impact on the classification performance. Thus, we proposed ensemble learning and a variety of classification methods of different types. For instance, in the “MTSC” data set, we found the fine-tuned BERT model more effective than just BERT, indicating its advantage in leveraging embedded semantics, especially with the data set’s numerous named entities. Moreover, BERT is generally better performing than TF-IDF, and the majority vote yields encouraging results.
  • RQ5: Are we able to predict difficulty? We are far from being able to predict sentence difficulty, as indicated by the results in Section 7.3. Nevertheless, we have several ideas to enhance the prediction accuracy. Our initial attempts have started the discussion on this topic, but we do not claim to propose the best models for predicting sentence difficulty. We hope this work will stimulate the research community to pay more attention to this topic.
  • RQ6: How to better understand difficult sentences (qualitative analysis on difficult sentences)? After looking at the difficult sentences that we identified, we noticed that the difficulty may be raised by several aspects, such as ambiguity, subjectivity, implicit references, and the presence of named entities. Some of the difficult sentences may be considered challenging even for human annotators, and the annotation process may be subject to subjectivity. For some difficult sentences, even advanced LLMs are not able to correctly identify the sentiment polarity.

9. Conclusions and Future Work

The goal of this paper is to better understand sentence difficulty in aspect-based sentiment analysis, and not to introduce new models or enhance current results. To our knowledge, this topic has never been formally discussed before. We conducted thorough experiments on three well-known aspect-based sentiment analysis data sets—“Laptops”, “Restaurants” and “MTSC”—testing more than 20 classification models on two different textual representations: TF-IDF and BERT. In studying performance enhancement, we considered fine-tuned BERT representations and also applied ensemble learning (majority vote).
On the “MTSC” data set, using the fine-tuned BERT model is more effective than just using BERT as a textual representation. This shows that the model utilizes the semantics embedded in the BERT model and benefits from adapting BERT to the “MTSC” data set, particularly to the many named entities present in it. The “MTSC” data set presents further challenges, resulting in diverse outcomes for the models. Using an ensemble of classifiers is more suitable compared to the “Restaurants” and “Laptops” data sets. However, using all 21 classifiers results in a small decrease (around 1 point in F1 measure). However, utilizing the top five models as an ensemble improves accuracy and maintains the F1 measure. The performance results from these top five classifiers do not differ significantly, therefore their collective contribution does not notably influence the overall result.
By applying majority vote to the models on the “MTSC” data set, we can improve performance. This holds whether we use all models or just the top five best performing classifiers. It is clear that using text representations that have a sense of meaning combined with diverse model outputs can produce better results. We can simplify the task by using representations with semantic information and combining multiple classifiers.
Regarding the difficulty of aspect-based sentiment analysis, we identified the sentences as difficult which the classifiers did not judge correctly, implying a binary classification strategy. From this viewpoint, only one sentence was considered difficult in the “Laptops” corpus compared to 197 in the recent “MTSC” corpus, indicating that “MTSC” sentences are more challenging. A different strategy for defining difficulty would be more nuanced, considering six levels of difficulty regarding how many of the top five best performing classifiers can correctly identify sentiment polarity. For instance, level 0 means all five classifiers are wrong, while level 5 means all five were correct. In analyzing sentences identified as difficult, we conclude that defining difficulty in this aspect-based sentiment analysis context is not a straightforward task.
The classification difficulty seems to be reliant on the data set. The text representation and the classification model also influence performance. Implicit references, intricate semantics, ambiguity, and other factors make polarity classification difficult. For example, of the six sentences we analyzed qualitatively, two were very difficult, even for a human. The first contained an implicit reference and the second could have been classified as neutral by a human being. Lastly, we assert that predicting difficulty is not an easy task, but there are signs that it is feasible, at least partially.
As future work, we plan to extend our experiments to other data sets, perform domain adaptation to validate model robustness and verify data set biases, and aim to propose difficulty predictors that are correlated to classification performance, inspired by the works of Query Performance Prediction (QPP) in Information Retrieval. We also intend to propose and analyze various definitions of difficulty classes by adjusting the scale levels differently than our binary proposal, or finer than the 6 level scale, based on the top five majority vote.

Author Contributions

Conceptualization, A.-G.C. and S.F.; methodology, A.-G.C. and S.F.; investigation, S.F.; data curation, A.-G.C.; writing—original draft, A.-G.C. and S.F.; writing—review & editing, A.-G.C. and S.F.; visualization, A.-G.C.; funding acquisition, S.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data will available at https://github.com/adrianchifu/sentimentdifficultyABSA, accessed on 10 October 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. van Atteveldt, W.; van der Velden, M.A.C.G.; Boukes, M. The Validity of Sentiment Analysis: Comparing Manual Annotation, Crowd-Coding, Dictionary Approaches, and Machine Learning Algorithms. Commun. Methods Meas. 2021, 15, 121–140. [Google Scholar] [CrossRef]
  2. Wankhade, M.; Rao, A.C.S.; Kulkarni, C. A survey on sentiment analysis methods, applications, and challenges. Artif. Intell. Rev. 2022, 55, 5731–5780. [Google Scholar] [CrossRef]
  3. Cambria, E.; Schuller, B.; Liu, B.; Wang, H.; Havasi, C. Knowledge-based approaches to concept-level sentiment analysis. IEEE Intell. Syst. 2013, 28, 12–14. [Google Scholar] [CrossRef]
  4. Deng, S.; Sinha, A.P.; Zhao, H. Resolving Ambiguity in Sentiment Classification: The Role of Dependency Features. ACM Trans. Manage. Inf. Syst. 2017, 8, 1–13. [Google Scholar] [CrossRef]
  5. Gref, M.; Matthiesen, N.; Hikkal Venugopala, S.; Satheesh, S.; Vijayananth, A.; Ha, D.B.; Behnke, S.; Köhler, J. A Study on the Ambiguity in Human Annotation of German Oral History Interviews for Perceived Emotion Recognition and Sentiment Analysis; Thirteenth Language Resources and Evaluation Conference; European Language Resources Association: Marseille, France, 2022; pp. 2022–2031.
  6. Maynard, D.G.; Greenwood, M.A. Who cares about sarcastic tweets? investigating the impact of sarcasm on sentiment analysis. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC 2014), Reykjavik, Iceland, 26–31 May 2014. [Google Scholar]
  7. Farias, D.H.; Rosso, P. Irony, sarcasm, and sentiment analysis. In Sentiment Analysis in Social Networks; Elsevier: Amsterdam, The Netherlands, 2017; pp. 113–128. [Google Scholar]
  8. Li, Q.; Zhang, K.; Sun, L.; Xia, R. Detecting Negative Sentiment on Sarcastic Tweets for Sentiment Analysis. In Artificial Neural Networks and Machine Learning: Proceedings of the 2nd International Conference on Artificial Neural Networks, Heraklion, Crete, Greece, 26–29 Septembe 2023; Iliadis, L., Papaleonidas, A., Angelov, P., Jayne, C., Eds.; Springer: Cham, Switzerland, 2023; Volume 14263. [Google Scholar]
  9. Kong, J.; Lou, C. Do cultural orientations moderate the effect of online review features on review helpfulness? A case study of online movie reviews. J. Retail. Consum. Serv. 2023, 73, 103374. [Google Scholar] [CrossRef]
  10. Asyrofi, M.H.; Yang, Z.; Yusuf, I.N.B.; Kang, H.J.; Thung, F.; Lo, D. Biasfinder: Metamorphic test generation to uncover bias for sentiment analysis systems. IEEE Trans. Softw. Eng. 2021, 48, 5087–5101. [Google Scholar] [CrossRef]
  11. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, 2–7 June 2019; Burstein, J., Doran, C., Solorio, T., Eds.; Association for Computational Linguistics: Toronto, ON, Canada, 2019; Volume 1 (Long and Short Papers), pp. 4171–4186. [Google Scholar] [CrossRef]
  12. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. In Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 1877–1901. [Google Scholar]
  13. Villavicencio, C.; Macrohon, J.J.; Inbaraj, X.A.; Jeng, J.H.; Hsieh, J.G. Twitter sentiment analysis towards COVID-19 vaccines in the Philippines using naïve bayes. Information 2021, 12, 204. [Google Scholar] [CrossRef]
  14. Mubarok, M.S.; Adiwijaya.; Aldhi, M.D. Aspect-based sentiment analysis to review products using Naïve Bayes. AIP Conf. Proc. 2017, 1867, 020060. [Google Scholar] [CrossRef]
  15. Goel, A.; Gautam, J.; Kumar, S. Real time sentiment analysis of tweets using Naive Bayes. In Proceedings of the 2016 2nd International Conference on Next Generation Computing Technologies (NGCT), Dehradun, India, 14–16 October 2016; pp. 257–261. [Google Scholar] [CrossRef]
  16. Mittal, P.; Tiwari, K.; Malik, K.; Tyagi, M. Feedback Analysis of Online Teaching Using SVM. In International Conference on Recent Trends in Computing; Mahapatra, R.P., Peddoju, S.K., Roy, S., Parwekar, P., Eds.; Springer Nature Singapore: Singapore, 2023; pp. 119–128. [Google Scholar]
  17. Ahmad, M.; Aftab, S.; Bashir, M.S.; Hameed, N. Sentiment analysis using SVM: A systematic literature review. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 182–188. [Google Scholar] [CrossRef]
  18. Fikri, M.; Sarno, R. A comparative study of sentiment analysis using SVM and SentiWordNet. Indones. J. Electr. Eng. Comput. Sci. 2019, 13, 902–909. [Google Scholar] [CrossRef]
  19. Li, D.; Rzepka, R.; Ptaszynski, M.; Araki, K. HEMOS: A novel deep learning-based fine-grained humor detecting method for sentiment analysis of social media. Inf. Process. Manag. 2020, 57, 102290. [Google Scholar] [CrossRef]
  20. Wang, X.; Jiang, W.; Luo, Z. Combination of convolutional and recurrent neural network for sentiment analysis of short texts. In Proceedings of the COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, Japan, 11–16 December 2016; pp. 2428–2437. [Google Scholar]
  21. Basiri, M.E.; Nemati, S.; Abdar, M.; Cambria, E.; Acharya, U.R. ABCDM: An attention-based bidirectional CNN-RNN deep model for sentiment analysis. Future Gener. Comput. Syst. 2021, 115, 279–294. [Google Scholar] [CrossRef]
  22. Ma, Y.; Peng, H.; Khan, T.; Cambria, E.; Hussain, A. Sentic LSTM: A hybrid network for targeted aspect-based sentiment analysis. Cogn. Comput. 2018, 10, 639–650. [Google Scholar] [CrossRef]
  23. Rehman, A.U.; Malik, A.K.; Raza, B.; Ali, W. A hybrid CNN-LSTM model for improving accuracy of movie reviews sentiment analysis. Multimed. Tools Appl. 2019, 78, 26597–26613. [Google Scholar] [CrossRef]
  24. Ahmed, A.; Yousuf, M.A. Sentiment Analysis on Bangla Text Using Long Short-Term Memory (LSTM) Recurrent Neural Network. In International Conference on Trends in Computational and Cognitive Engineering; Kaiser, M.S., Bandyopadhyay, A., Mahmud, M., Ray, K., Eds.; Springer: Singapore, 2021; pp. 181–192. [Google Scholar]
  25. Hoang, M.; Bihorac, O.A.; Rouces, J. Aspect-based sentiment analysis using bert. In Proceedings of the 22nd Nordic Conference on Computational Linguistics, (NoDaLiDa), Turku, Finland, 30 September–2 October 2019; pp. 187–196. [Google Scholar]
  26. Gao, Z.; Feng, A.; Song, X.; Wu, X. Target-dependent sentiment classification with BERT. IEEE Access 2019, 7, 154290–154299. [Google Scholar] [CrossRef]
  27. Tiwari, D.; Nagpal, B.; Bhati, B.S.; Mishra, A.; Kumar, M. A systematic review of social network sentiment analysis with comparative study of ensemble-based techniques. Artif. Intell. Rev. 2023, 56, 13407–13461. [Google Scholar] [CrossRef] [PubMed]
  28. Liu, R.; Shi, Y.; Ji, C.; Jia, M. A Survey of Sentiment Analysis Based on Transfer Learning. IEEE Access 2019, 7, 85401–85412. [Google Scholar] [CrossRef]
  29. Bordoloi, M.; Biswas, S.K. Sentiment analysis: A survey on design framework, applications and future scopes. Artif. Intell. Rev. 2023, 56, 12505–12560. [Google Scholar] [CrossRef]
  30. Cui, J.; Wang, Z.; Ho, S.B.; Cambria, E. Survey on sentiment analysis: Evolution of research methods and topics. Artif. Intell. Rev. 2023, 56, 8469–8510. [Google Scholar] [CrossRef]
  31. Hu, M.; Liu, B. Mining and Summarizing Customer Reviews. In Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; Association for Computing Machinery: New York, NY, USA, 2004; KDD ’04; pp. 168–177. [Google Scholar] [CrossRef]
  32. Varghese, R.; Jayasree, M. Aspect based Sentiment Analysis using support vector machine classifier. In Proceedings of the 2013 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Mysore, India, 22–25 August 2013; pp. 1581–1586. [Google Scholar] [CrossRef]
  33. Mubarok, M.S.; Adiwijaya, A.; Aldhi, M.D. Aspect-based sentiment analysis to review products using Naïve Bayes. In Proceedings of the International Conference on Mathematics: Pure, Applied and Computation: Empowering Engineering using Mathematics, Surabaya, Indonesia, 1 November 2016. [Google Scholar]
  34. Ma, Y.; Peng, H.; Cambria, E. Targeted aspect-based sentiment analysis via embedding commonsense knowledge into an attentive LSTM. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18), New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  35. Do, H.H.; Prasad, P.W.; Maag, A.; Alsadoon, A. Deep learning for aspect-based sentiment analysis: A comparative review. Expert Syst. Appl. 2019, 118, 272–299. [Google Scholar] [CrossRef]
  36. Liu, H.; Chatterjee, I.; Zhou, M.; Lu, X.S.; Abusorrah, A. Aspect-based sentiment analysis: A survey of deep learning methods. IEEE Trans. Comput. Soc. Syst. 2020, 7, 1358–1375. [Google Scholar] [CrossRef]
  37. Karimi, A.; Rossi, L.; Prati, A. Improving BERT Performance for Aspect-Based Sentiment Analysis. In Proceedings of the International Conference on Natural Language and Speech Processing, Copenhagen, Denmark, 25–26 April 2020. [Google Scholar]
  38. Mutlu, M.M.; Özgür, A. A Dataset and BERT-based Models for Targeted Sentiment Analysis on Turkish Texts. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, Dublin, Ireland, 22–27 May 2022; Association for Computational Linguistics: Dublin, Ireland, 2022; pp. 467–472. [Google Scholar] [CrossRef]
  39. Zhang, W.; Li, X.; Deng, Y.; Bing, L.; Lam, W. A Survey on Aspect-Based Sentiment Analysis: Tasks, Methods, and Challenges. IEEE Trans. Knowl. Data Eng. 2023, 35, 11019–11038. [Google Scholar] [CrossRef]
  40. Brauwers, G.; Frasincar, F. A Survey on Aspect-Based Sentiment Classification. ACM Comput. Surv. 2022, 55, 1–37. [Google Scholar] [CrossRef]
  41. Chauhan, G.S.; Nahta, R.; Meena, Y.K.; Gopalani, D. Aspect based sentiment analysis using deep learning approaches: A survey. Comput. Sci. Rev. 2023, 49, 100576. [Google Scholar] [CrossRef]
  42. Joachims, T. A Probabilistic Analysis of the Rocchio Algorithm with TFIDF for Text Categorization; Technical Report; Carnegie-Mellon Univ Pittsburgh Pa Dept of Computer Science: Pittsburgh, PA, USA, 1996. [Google Scholar]
  43. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. Adv. Neural Inf. Process. Syst. 2013, 26, 3111–3119. [Google Scholar]
  44. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  45. de Loupy, C.; Bellot, P. Evaluation of Document Retrieval Systems and Query Difficulty. In Proceedings of the Second International Conference on Language Resources and Evaluation (LREC 2000) Workshop, Athens, Greece, 31 May–2 June 2000; pp. 32–39. [Google Scholar]
  46. Mothe, J.; Tanguy, L. Linguistic features to predict query difficulty. In Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2005), Salvador de Bahia, Brazil, 15–19 August 2005; pp. 7–10. [Google Scholar]
  47. Goeuriot, L.; Kelly, L.; Leveling, J. An Analysis of Query Difficulty for Information Retrieval in the Medical Domain. In Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval, Gold Coast, Australia, 6–11 July 2014; Association for Computing Machinery: New York, NY, USA, 2014. SIGIR ’14. pp. 1007–1010. [Google Scholar] [CrossRef]
  48. Zhao, Y.; Scholer, F.; Tsegay, Y. Effective Pre-retrieval Query Performance Prediction Using Similarity and Variability Evidence. In Advances in Information Retrieval; Macdonald, C., Ounis, I., Plachouras, V., Ruthven, I., White, R.W., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 52–64. [Google Scholar]
  49. Cronen-Townsend, S.; Zhou, Y.; Croft, W.B. A Language Modeling Framework for Selective Query Expansion; Technical Report, Technical Report IR-338; Center for Intelligent Information Retrieval, University of Massachusetts Amherst: Amherst, MA, USA, 2004. [Google Scholar]
  50. Scholer, F.; Williams, H.E.; Turpin, A. Query association surrogates for Web search. J. Am. Soc. Inf. Sci. Technol. 2004, 55, 637–650. [Google Scholar] [CrossRef]
  51. Carmel, D.; Yom-Tov, E. Estimating the Query Difficulty for Information Retrieval. In Proceedings of the 33rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Geneva, Switzerland, 19–23 July 2010; Association for Computing Machinery: New York, NY, USA, 2010. SIGIR ’10. p. 911. [Google Scholar] [CrossRef]
  52. Cronen-Townsend, S.; Zhou, Y.; Croft, W.B. Predicting Query Performance. In Proceedings of the 25th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Tampere, Finland, 11–15 August 2002; Association for Computing Machinery: New York, NY, USA, 2002. SIGIR ’02. pp. 299–306. [Google Scholar] [CrossRef]
  53. Shtok, A.; Kurland, O.; Carmel, D.; Raiber, F.; Markovits, G. Predicting Query Performance by Query-Drift Estimation. ACM Trans. Inf. Syst. 2012, 30, 1–15. [Google Scholar] [CrossRef]
  54. Zhou, Y.; Croft, W.B. Query Performance Prediction in Web Search Environments. In Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, 23–27 July 2007; Association for Computing Machinery: New York, NY, USA, 2007. SIGIR ’07. pp. 543–550. [Google Scholar] [CrossRef]
  55. Tao, Y.; Wu, S. Query Performance Prediction By Considering Score Magnitude and Variance Together. In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, Shanghai, China, 3–7 November 2014; Association for Computing Machinery: New York, NY, USA, 2014. CIKM ’14. pp. 1891–1894. [Google Scholar] [CrossRef]
  56. Hashemi, H.; Zamani, H.; Croft, W.B. Performance Prediction for Non-Factoid Question Answering. In Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval, Santa Clara, CA, USA, 2–5 October 2019; Association for Computing Machinery: New York, NY, USA, 2019. ICTIR ’19. pp. 55–58. [Google Scholar] [CrossRef]
  57. Faggioli, G.; Formal, T.; Marchesin, S.; Clinchant, S.; Ferro, N.; Piwowarski, B. Query Performance Prediction For Neural IR: Are We There Yet? In Proceedings of the Advances in Information Retrieval: 45th European Conference on Information Retrieval, ECIR 2023, Dublin, Ireland, 2–6 April 2023; Proceedings, Part I. Springer: Berlin/Heidelberg, Germany, 2023; pp. 232–248. [Google Scholar] [CrossRef]
  58. Faggioli, G.; Formal, T.; Lupart, S.; Marchesin, S.; Clinchant, S.; Ferro, N.; Piwowarski, B. Towards Query Performance Prediction for Neural Information Retrieval: Challenges and Opportunities. In Proceedings of the 2023 ACM SIGIR International Conference on Theory of Information Retrieval, Taipei, Taiwan, 23–27 July 2023; Association for Computing Machinery: New York, NY, USA, 2023. ICTIR ’23. pp. 51–63. [Google Scholar] [CrossRef]
  59. Pontiki, M.; Galanis, D.; Pavlopoulos, J.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S. SemEval-2014 task 4: Aspect Based Sentiment Analysis. In Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014), Dublin, Ireland, 23–24 August 2014; pp. 27–35. [Google Scholar]
  60. Ganu, G.; Elhadad, N.; Marian, A. Beyond the stars: Improving rating predictions using review text content. WebDB 2009, 9, 1–6. [Google Scholar]
  61. Hamborg, F.; Donnay, K. NewsMTSC: (Multi-)Target-dependent Sentiment Classification in News Articles. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2021), Online, 19–23 April 2021. [Google Scholar]
  62. He, B.; Ounis, I. Inferring query performance using pre-retrieval predictors. In Proceedings of the String Processing and Information Retrieval: 11th International Conference, SPIRE 2004, Padova, Italy, 5–8 October 2004; Proceedings 11. Springer: Berlin/Heidelberg, Germany, 2004; pp. 43–54. [Google Scholar]
  63. Salton, G.; Wong, A.; Yang, C.S. A vector space model for automatic indexing. Commun. ACM 1975, 18, 613–620. [Google Scholar] [CrossRef]
  64. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  65. Geetha, M.; Renuka, D.K. Improving the performance of aspect based sentiment analysis using fine-tuned Bert Base Uncased model. Int. J. Intell. Netw. 2021, 2, 64–69. [Google Scholar] [CrossRef]
  66. Zhao, X.; Greenberg, J.; An, Y.; Hu, X.T. Fine-Tuning BERT Model for Materials Named Entity Recognition. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data), Orlando, FL, USA, 15–18 December 2021; pp. 3717–3720. [Google Scholar]
  67. Sagi, O.; Rokach, L. Ensemble learning: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
  68. Wang, G.; Sun, J.; Ma, J.; Xu, K.; Gu, J. Sentiment classification: The contribution of ensemble learning. Decis. Support Syst. 2014, 57, 77–93. [Google Scholar] [CrossRef]
  69. Zhang, J.; Li, Z.; Nai, K.; Gu, Y.; Sallam, A. DELR: A double-level ensemble learning method for unsupervised anomaly detection. Knowl.-Based Syst. 2019, 181, 104783. [Google Scholar] [CrossRef]
  70. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  71. Mothe, J.; Laporte, L.; Chifu, A.G. Predicting query difficulty in IR: Impact of difficulty definition. In Proceedings of the 2019 11th International Conference on Knowledge and Systems Engineering (KSE), Da Nang, Vietnam, 24–26 October 2019; pp. 1–6. [Google Scholar]
Figure 1. Polarity distributions for the three data sets.
Figure 1. Polarity distributions for the three data sets.
Mathematics 11 04647 g001
Figure 2. Class occurrences (Train vs. Test) for the “Laptops” data set.
Figure 2. Class occurrences (Train vs. Test) for the “Laptops” data set.
Mathematics 11 04647 g002
Figure 3. Class occurrences (Train vs. Test) for the “Restaurants” data set.
Figure 3. Class occurrences (Train vs. Test) for the “Restaurants” data set.
Mathematics 11 04647 g003
Figure 4. Class occurrences (Train vs. Test) for the MTSC data set.
Figure 4. Class occurrences (Train vs. Test) for the MTSC data set.
Mathematics 11 04647 g004
Figure 5. Number of tokens per aspect, for the three data sets.
Figure 5. Number of tokens per aspect, for the three data sets.
Mathematics 11 04647 g005
Figure 6. Sentence processing pipeline based on pretrained BERT.
Figure 6. Sentence processing pipeline based on pretrained BERT.
Mathematics 11 04647 g006
Figure 7. Sentence processing pipeline based on fine-tuned BERT.
Figure 7. Sentence processing pipeline based on fine-tuned BERT.
Mathematics 11 04647 g007
Figure 8. Class repartition for the binary strategy and for the BERT fine-grained strategy, respectively.
Figure 8. Class repartition for the binary strategy and for the BERT fine-grained strategy, respectively.
Mathematics 11 04647 g008
Figure 9. Mean cross-validation scores by model, by text representation, and by sampling, for difficulty prediction as binary classification.
Figure 9. Mean cross-validation scores by model, by text representation, and by sampling, for difficulty prediction as binary classification.
Mathematics 11 04647 g009
Figure 10. Mean cross-validation scores by model, by text representation, and by sampling, for difficulty prediction as multi-class, fine-grained classification.
Figure 10. Mean cross-validation scores by model, by text representation, and by sampling, for difficulty prediction as multi-class, fine-grained classification.
Mathematics 11 04647 g010
Table 1. Summary of the data set information.
Table 1. Summary of the data set information.
Data SetsTotalTrainTest# of Classes
Laptops24072358494
Restaurants37893693964
MTSC9885873911463
Table 2. Token and sentence statistics for the data sets.
Table 2. Token and sentence statistics for the data sets.
Data Set# of Observations# of Unique Aspects# of Unique SentencesMax # of Tokens per Aspect
Laptops2407104414846
Restaurants37891289202219
MTSC98853525880231
Table 3. Average number of tokens, nouns, verbs, named entities, and adjectives per sentence.
Table 3. Average number of tokens, nouns, verbs, named entities, and adjectives per sentence.
Data Set/ClassTokensNounsVerbsNamed EntitiesAdjectives
LaptopsPositive20.043.721.980.641.94
Negative22.764.042.810.821.42
Neutral25.244.542.831.331.43
Conflict23.843.822.690.842.02
RestaurantsPositive18.813.771.410.542.20
Negative22.504.102.180.501.99
Neutral21.624.142.260.741.48
Conflict22.313.601.760.482.67
MTSCPositive30.125.073.292.951.99
Negative31.585.313.553.271.94
Neutral27.634.133.033.301.40
Conflict0.000.000.000.000.00
Table 4. Hardware and software specifications.
Table 4. Hardware and software specifications.
Hardware (Computing Cluster Node)Specification
CPU40 Intel(R) Xeon(R) Silver 4114 CPU @ 2.20 GHz
RAM354 GB
GPU8 (7 Nvidia A40 and 1 Nvidia GeForce RTX 3090)
Software (Conda virtual environment)Version
python3.11
Python modulesVersion
huggingface_hub0.14.1
lazy-text-predict0.0.11
lazypredict0.2.12
lazytext0.0.2
matplotlib-base3.7.1
nltk3.7
numpy1.24.3
pandas1.3.5
scikit-learn1.3.1
scipy1.10.1
spacy3.5.3
torch2.0.1
transformers4.29.2
Table 5. Macro Metrics of Classification Models (“Laptops”, TF-IDF representations).
Table 5. Macro Metrics of Classification Models (“Laptops”, TF-IDF representations).
ModelPrecision (Macro)Recall (Macro)F1 (Macro)
AdaBoostClassifier0.3518520.3566810.333003
BaggingClassifier0.7343750.7257540.729989
BernoulliNB0.8886160.6135060.663075
CalibratedClassifierCV0.9687500.8750000.913765
DecisionTreeClassifier0.7500000.7343750.741935
DummyClassifier0.1972790.3333330.247863
ExtraTreeClassifier0.7500000.7343750.741935
ExtraTreesClassifier0.7500000.7343750.741935
GradientBoostingClassifier0.6486740.5420260.567858
KNeighborsClassifier0.6047250.5445400.560063
LinearSVC0.7250000.6562500.685855
LogisticRegression0.7831270.8010060.786207
LogisticRegressionCV0.6202020.5718390.584436
MLPClassifier0.7500000.7343750.741935
NearestCentroid0.4104170.3825430.377963
PassiveAggressiveClassifier0.7500000.7343750.741935
Perceptron0.7000000.7187500.705556
RandomForestClassifier0.7500000.7343750.741935
RidgeClassifier0.6713710.6406250.653305
SGDClassifier0.7250000.6562500.685855
SVC0.7403850.7176720.725551
Table 6. Weighted Metrics of Classification Models (“Laptops”, TF-IDF representations).
Table 6. Weighted Metrics of Classification Models (“Laptops”, TF-IDF representations).
ModelPrecision (Weighted)Recall (Weighted)F1 (Weighted)
AdaBoostClassifier0.6953890.6122450.615577
BaggingClassifier0.9795920.9591840.969209
BernoulliNB0.8263550.7959180.776692
CalibratedClassifierCV0.9445150.9387760.937463
DecisionTreeClassifier1.0000000.9795920.989467
DummyClassifier0.3502710.5918370.440084
ExtraTreeClassifier1.0000000.9795920.989467
ExtraTreesClassifier1.0000000.9795920.989467
GradientBoostingClassifier0.8566020.7959180.808693
KNeighborsClassifier0.6712020.6734690.668267
LinearSVC0.9585030.9387760.946707
LogisticRegression0.8598370.8571430.854516
LogisticRegressionCV0.7381570.7551020.734680
MLPClassifier1.0000000.9795920.989467
NearestCentroid0.6848640.5510200.602253
PassiveAggressiveClassifier1.0000000.9795920.989467
Perceptron0.9836730.9591840.969161
RandomForestClassifier1.0000000.9795920.989467
RidgeClassifier0.9414090.9183670.926085
SGDClassifier0.9585030.9387760.946707
SVC0.8349690.8367350.831855
Table 7. Macro Metrics of Classification Models (“Restaurants”, TF-IDF representations).
Table 7. Macro Metrics of Classification Models (“Restaurants”, TF-IDF representations).
ModelPrecision (Macro)Recall (Macro)F1 (Macro)
AdaBoostClassifier0.7983090.4136170.416818
BaggingClassifier0.9951690.9814810.988043
BernoulliNB0.8287500.6136170.672518
CalibratedClassifierCV0.9129130.7913940.836613
DecisionTreeClassifier1.0000001.0000001.000000
DummyClassifier0.2361110.3333330.276423
ExtraTreeClassifier1.0000001.0000001.000000
ExtraTreesClassifier1.0000001.0000001.000000
GradientBoostingClassifier0.9365080.6000000.674314
KNeighborsClassifier0.7245370.6742920.696110
LinearSVC0.9249010.9198260.920608
LogisticRegression0.8662340.7025050.756657
LogisticRegressionCV0.8735850.8198260.843531
MLPClassifier1.0000001.0000001.000000
NearestCentroid0.4821860.4508170.445678
PassiveAggressiveClassifier0.9309360.8395420.879074
Perceptron0.6885760.7009800.689108
RandomForestClassifier1.0000001.0000001.000000
RidgeClassifier0.9249840.8247280.865804
SGDClassifier0.9249010.9198260.920608
SVC0.9334550.7074070.771812
Table 8. Weighted Metrics of Classification Models (“Restaurants”, TF-IDF representations).
Table 8. Weighted Metrics of Classification Models (“Restaurants”, TF-IDF representations).
ModelPrecision (Weighted)Recall (Weighted)F1 (Weighted)
AdaBoostClassifier0.7727960.7291670.645017
BaggingClassifier0.9897340.9895830.989473
BernoulliNB0.8344850.8333330.810978
CalibratedClassifierCV0.9156340.9166670.910002
DecisionTreeClassifier1.0000001.0000001.000000
DummyClassifier0.5017360.7083330.587398
ExtraTreeClassifier1.0000001.0000001.000000
ExtraTreesClassifier1.0000001.0000001.000000
GradientBoostingClassifier0.8650790.8333330.806849
KNeighborsClassifier0.8039640.8125000.806575
LinearSVC0.9502640.9479170.948238
LogisticRegression0.8737820.8750000.862497
LogisticRegressionCV0.9130460.9166670.913535
MLPClassifier1.0000001.0000001.000000
NearestCentroid0.8246740.5625000.650863
PassiveAggressiveClassifier0.9287600.9270830.924641
Perceptron0.9653890.9479170.953627
RandomForestClassifier1.0000001.0000001.000000
RidgeClassifier0.9269000.9270830.923168
SGDClassifier0.9502640.9479170.948238
SVC0.8957950.8854170.871685
Table 9. Macro Metrics of Classification Models (“MTSC”, TF-IDF representations).
Table 9. Macro Metrics of Classification Models (“MTSC”, TF-IDF representations).
ModelPrecision (Macro)Recall (Macro)F1 (Macro)
AdaBoostClassifier0.5365690.5110310.511124
BaggingClassifier0.5462740.5398470.542036
BernoulliNB0.6203380.6093010.613077
CalibratedClassifierCV0.5986960.5901060.591621
DecisionTreeClassifier0.4839790.4838800.482779
DummyClassifier0.1247820.3333330.181587
ExtraTreeClassifier0.4498160.4463740.446124
ExtraTreesClassifier0.6103370.5978930.601791
GradientBoostingClassifier0.5849190.5477930.550139
KNeighborsClassifier0.4834020.4731720.474578
LinearSVC0.5892150.5845750.586015
LogisticRegression0.6056430.5944840.597691
LogisticRegressionCV0.6095290.6029390.605218
MLPClassifier0.5394250.5377600.538246
NearestCentroid0.4800830.4378830.416298
PassiveAggressiveClassifier0.5625940.5582420.559689
Perceptron0.5691770.5513410.554123
RandomForestClassifier0.5820650.5645240.567516
RidgeClassifier0.6025950.5976420.599191
SGDClassifier0.5999420.6025760.600953
SVC0.5788460.5625400.565638
Table 10. Weighted Metrics of Classification Models (“MTSC”, TF-IDF representations).
Table 10. Weighted Metrics of Classification Models (“MTSC”, TF-IDF representations).
ModelPrecision (Weighted)Recall (Weighted)F1 (Weighted)
AdaBoostClassifier0.5444590.5375220.527558
BaggingClassifier0.5569880.5602090.557818
BernoulliNB0.6267050.6291450.626609
CalibratedClassifierCV0.6106220.6151830.610546
DecisionTreeClassifier0.4972150.4965100.495555
DummyClassifier0.1401350.3743460.203929
ExtraTreeClassifier0.4633080.4659690.462460
ExtraTreesClassifier0.6175980.6204190.617156
GradientBoostingClassifier0.5863930.5863870.573521
KNeighborsClassifier0.4945590.5000000.494082
LinearSVC0.6020320.6055850.603095
LogisticRegression0.6147040.6186740.614889
LogisticRegressionCV0.6212670.6247820.622249
MLPClassifier0.5540070.5567190.555062
NearestCentroid0.5009530.4572430.430506
PassiveAggressiveClassifier0.5757610.5794070.577021
Perceptron0.5785660.5837700.576333
RandomForestClassifier0.5908950.5959860.588793
RidgeClassifier0.6158040.6195460.616933
SGDClassifier0.6177700.6151830.616206
SVC0.5877050.5924960.586000
Table 11. Macro Metrics of Classification Models (“Laptops”, BERT representations).
Table 11. Macro Metrics of Classification Models (“Laptops”, BERT representations).
ModelPrecision (Macro)Recall (Macro)F1 (Macro)
AdaBoostClassifier0.8111110.5977010.628514
BaggingClassifier0.7500000.7343750.741935
BernoulliNB0.5798080.5506470.563304
CalibratedClassifierCV0.9569440.8843390.914598
DecisionTreeClassifier0.7416670.7187500.729096
DummyClassifier0.1972790.3333330.247863
ExtraTreeClassifier0.7500000.7343750.741935
ExtraTreesClassifier0.7500000.7343750.741935
GradientBoostingClassifier0.7416670.6718750.701984
KNeighborsClassifier0.7072830.7133620.709995
LinearSVC0.9676720.9676720.967672
LogisticRegression0.9155350.7385060.794381
LogisticRegressionCV0.8660710.7780170.814312
MLPClassifier0.9888890.9791670.983598
NearestCentroid0.4435480.3943970.414286
PassiveAggressiveClassifier0.7523810.7291670.686800
Perceptron0.7036800.8081900.705096
RandomForestClassifier0.7500000.7343750.741935
RidgeClassifier0.9153230.7686780.813889
SGDClassifier0.7008930.6928880.696820
SVC0.8404040.6113510.649462
Table 12. Weighted Metrics of Classification Models (“Laptops”, BERT representations).
Table 12. Weighted Metrics of Classification Models (“Laptops”, BERT representations).
ModelPrecision (Weighted)Recall (Weighted)F1 (Weighted)
AdaBoostClassifier0.7530610.7346940.724577
BaggingClassifier1.0000000.9795920.989467
BernoulliNB0.7858710.7551020.767912
CalibratedClassifierCV0.9401360.9387760.937837
DecisionTreeClassifier0.9802720.9591840.968200
DummyClassifier0.3502710.5918370.440084
ExtraTreeClassifier1.0000000.9795920.989467
ExtraTreesClassifier1.0000000.9795920.989467
GradientBoostingClassifier0.9802720.9591840.967774
KNeighborsClassifier0.7978050.7959180.796472
LinearSVC0.9591840.9591840.959184
LogisticRegression0.8704400.8571430.850731
LogisticRegressionCV0.8190600.8163270.813943
MLPClassifier0.9802720.9795920.979436
NearestCentroid0.7030940.6734690.683382
PassiveAggressiveClassifier0.8518950.7959180.779637
Perceptron0.8282400.7755100.783489
RandomForestClassifier1.0000000.9795920.989467
RidgeClassifier0.8828180.8775510.872789
SGDClassifier0.9169100.8979590.907268
SVC0.7873840.7755100.760764
Table 13. Macro Metrics of Classification Models (“Restaurants”, BERT representations).
Table 13. Macro Metrics of Classification Models (“Restaurants”, BERT representations).
ModelPrecision (Macro)Recall (Macro)F1 (Macro)
AdaBoostClassifier0.5375940.4566180.486397
BaggingClassifier1.0000001.0000001.000000
BernoulliNB0.5122200.5124180.508235
CalibratedClassifierCV0.9094560.5851850.641217
DecisionTreeClassifier1.0000001.0000001.000000
DummyClassifier0.2361110.3333330.276423
ExtraTreeClassifier1.0000001.0000001.000000
ExtraTreesClassifier1.0000001.0000001.000000
GradientBoostingClassifier0.9688890.8111110.870047
KNeighborsClassifier0.7187040.6458610.674860
LinearSVC0.9261900.6407410.688251
LogisticRegression0.8911050.5988020.647447
LogisticRegressionCV0.9031750.5666670.623642
MLPClassifier0.9230770.9852940.949003
NearestCentroid0.5121380.3959150.425557
PassiveAggressiveClassifier0.8795840.5111110.563584
Perceptron0.7274310.7201530.716450
RandomForestClassifier1.0000001.0000001.000000
RidgeClassifier0.9187240.6370370.702541
SGDClassifier0.7805560.5296300.578348
SVC0.9000000.5333330.571188
Table 14. Weighted Metrics of Classification Models (“Restaurants”, BERT representations).
Table 14. Weighted Metrics of Classification Models (“Restaurants”, BERT representations).
ModelPrecision (Weighted)Recall (Weighted)F1 (Weighted)
AdaBoostClassifier0.7873200.7916670.781311
BaggingClassifier1.0000001.0000001.000000
BernoulliNB0.7787460.7395830.756588
CalibratedClassifierCV0.8549420.8333330.802002
DecisionTreeClassifier1.0000001.0000001.000000
DummyClassifier0.5017360.7083330.587398
ExtraTreeClassifier1.0000001.0000001.000000
ExtraTreesClassifier1.0000001.0000001.000000
GradientBoostingClassifier0.9338890.9270830.922239
KNeighborsClassifier0.8000640.8125000.802541
LinearSVC0.8803570.8645830.837967
LogisticRegression0.8487260.8333330.804811
LogisticRegressionCV0.8463290.8229170.789030
MLPClassifier0.9759620.9687500.970436
NearestCentroid0.8001870.6145830.682193
PassiveAggressiveClassifier0.8185210.7916670.746228
Perceptron0.8105470.7916670.797258
RandomForestClassifier1.0000001.0000001.000000
RidgeClassifier0.8706920.8541670.832109
SGDClassifier0.8001740.8020830.762642
SVC0.8395830.8125000.769105
Table 15. Macro Metrics of Classification Models (“MTSC”, BERT representations).
Table 15. Macro Metrics of Classification Models (“MTSC”, BERT representations).
ModelPrecision (Macro)Recall (Macro)F1 (Macro)
AdaBoostClassifier0.6175790.6076000.611387
BaggingClassifier0.5830020.5709000.574723
BernoulliNB0.5940070.5990980.594284
CalibratedClassifierCV0.7176180.6979040.704669
DecisionTreeClassifier0.5026830.5064140.503590
DummyClassifier0.1247820.3333330.181587
ExtraTreeClassifier0.4205600.4225480.420626
ExtraTreesClassifier0.6524500.6370240.641867
GradientBoostingClassifier0.6745040.6596060.664726
KNeighborsClassifier0.5750930.5523260.554446
LinearSVC0.7067730.6949610.699503
LogisticRegression0.7199550.7002260.707067
LogisticRegressionCV0.7268260.7000670.708159
MLPClassifier0.6927430.6862420.688107
NearestCentroid0.5742240.5742240.574224
PassiveAggressiveClassifier0.6573550.6297130.582214
Perceptron0.6308840.5473690.446289
RandomForestClassifier0.6708880.6599530.663928
RidgeClassifier0.7107430.6925090.698889
SGDClassifier0.6877350.6720530.659408
SVC0.7201930.6904420.698967
Table 16. Weighted Metrics of Classification Models (“MTSC”, BERT representations).
Table 16. Weighted Metrics of Classification Models (“MTSC”, BERT representations).
ModelPrecision (Weighted)Recall (Weighted)F1 (Weighted)
AdaBoostClassifier0.6196900.6204190.619124
BaggingClassifier0.5876810.5898780.587083
BernoulliNB0.6076920.6020940.602517
CalibratedClassifierCV0.7183110.7181500.715857
DecisionTreeClassifier0.5185980.5130890.515090
DummyClassifier0.1401350.3743460.203929
ExtraTreeClassifier0.4373340.4310650.433455
ExtraTreesClassifier0.6543670.6553230.652600
GradientBoostingClassifier0.6782320.6797560.677239
KNeighborsClassifier0.5785200.5759160.568944
LinearSVC0.7107570.7102970.709302
LogisticRegression0.7209050.7198950.717951
LogisticRegressionCV0.7265040.7242580.721086
MLPClassifier0.6964250.6963350.694988
NearestCentroid0.5794070.5794070.579407
PassiveAggressiveClassifier0.6745020.6082020.570665
Perceptron0.6485150.5218150.417314
RandomForestClassifier0.6768990.6788830.676761
RidgeClassifier0.7120000.7120420.709899
SGDClassifier0.6982380.6727750.661620
SVC0.7173850.7155320.711701
Table 17. Classification report for fine-tuned BERT (“Laptops”).
Table 17. Classification report for fine-tuned BERT (“Laptops”).
ModelPrecisionRecallF1-ScoreSupport
negative0.970.970.9729
neutral1.000.880.9316
positive0.671.000.804
conflict0.000.000.000
Accuracy--0.9449
Weighted Avg0.880.950.9049
Macro Avg0.950.940.9449
Table 18. Classification report for fine-tuned BERT (“Restaurants”).
Table 18. Classification report for fine-tuned BERT (“Restaurants”).
ModelPrecisionRecallF1-ScoreSupport
negative0.890.990.9468
neutral1.000.500.6718
positive0.640.700.6710
conflict0.000.000.000
Accuracy--0.8696
Weighted Avg0.630.550.5796
Macro Avg0.890.860.8696
Table 19. Classification report for fine-tuned BERT (“MTSC”).
Table 19. Classification report for fine-tuned BERT (“MTSC”).
ModelPrecisionRecallF1-ScoreSupport
negative0.570.790.66262
neutral0.930.680.79429
positive0.720.750.73455
Accuracy--0.731146
Weighted Avg0.740.740.731146
Macro Avg0.770.730.741146
Table 20. Classification Report for Majority Vote (“Laptops”, TF-IDF representations, all models).
Table 20. Classification Report for Majority Vote (“Laptops”, TF-IDF representations, all models).
ModelPrecisionRecallF1-ScoreSupport
conflict0.000.000.000
negative1.000.880.9316
neutral1.001.001.004
positive0.971.000.9829
Accuracy--0.9649
Weighted Avg0.980.960.9749
Macro Avg0.740.720.7349
Table 21. Top five models with respect to accuracy (“Laptops”, TF-IDF representations).
Table 21. Top five models with respect to accuracy (“Laptops”, TF-IDF representations).
ModelAccuracy
DecisionTreeClassifier0.9796
ExtraTreeClassifier0.9796
ExtraTreesClassifier0.9796
MLPClassifier0.9796
PassiveAggressiveClassifier0.9796
Table 22. Classification report for majority vote (“Laptops”, TF-IDF representations, top five models).
Table 22. Classification report for majority vote (“Laptops”, TF-IDF representations, top five models).
ModelPrecisionRecallF1-ScoreSupport
conflict0.000.000.000
negative1.000.940.9716
neutral1.001.001.004
positive1.001.001.0029
Accuracy--0.9849
Weighted Avg1.000.980.9949
Macro Avg0.750.730.7449
Table 23. Classification Report for Majority Vote (“Restaurants”, TF-IDF representations, all models).
Table 23. Classification Report for Majority Vote (“Restaurants”, TF-IDF representations, all models).
ModelPrecisionRecallF1-ScoreSupport
negative1.000.890.9418
neutral1.000.900.9510
positive0.961.000.9868
Accuracy--0.9796
Weighted Avg0.970.970.9796
Macro Avg0.990.930.9696
Table 24. Top five models with respect to accuracy (“Restaurants”, TF-IDF representations).
Table 24. Top five models with respect to accuracy (“Restaurants”, TF-IDF representations).
ModelAccuracy
DecisionTreeClassifier1.0000
ExtraTreeClassifier1.0000
ExtraTreesClassifier1.0000
MLPClassifier1.0000
RandomForestClassifier1.0000
Table 25. Classification report for majority vote (“Restaurants”, TF-IDF representations, top 5 models).
Table 25. Classification report for majority vote (“Restaurants”, TF-IDF representations, top 5 models).
ModelPrecisionRecallF1-ScoreSupport
negative1.001.001.0018
neutral1.001.001.0010
positive1.001.001.0068
Accuracy--1.0096
Weighted Avg1.001.001.0096
Macro Avg1.001.001.0096
Table 26. Classification Report for Majority Vote (“MTSC”, TF-IDF representations, all models).
Table 26. Classification Report for Majority Vote (“MTSC”, TF-IDF representations, all models).
ModelPrecisionRecallF1-ScoreSupport
negative0.640.670.66429
neutral0.620.690.65455
positive0.590.430.50262
Accuracy--0.621146
Weighted Avg0.620.620.621146
Macro Avg0.620.600.601146
Table 27. Top five models with respect to accuracy (“MTSC”, TF-IDF representations).
Table 27. Top five models with respect to accuracy (“MTSC”, TF-IDF representations).
ModelAccuracy
BernoulliNB0.6291
LogisticRegressionCV0.6248
ExtraTreesClassifier0.6204
RidgeClassifier0.6195
LogisticRegression0.6187
Table 28. Classification Report for Majority Vote (“MTSC”, TF-IDF representations, top 5 models).
Table 28. Classification Report for Majority Vote (“MTSC”, TF-IDF representations, top 5 models).
ModelPrecisionRecallF1-ScoreSupport
negative0.650.700.67429
neutral0.640.660.65455
positive0.560.450.50262
Accuracy--0.631146
Weighted Avg0.620.630.621146
Macro Avg0.610.600.611146
Table 29. Classification Report for Majority Vote (“Laptops”, BERT representations, all models).
Table 29. Classification Report for Majority Vote (“Laptops”, BERT representations, all models).
ModelPrecisionRecallF1-ScoreSupport
conflict0.000.000.000
negative1.000.810.9016
neutral1.000.750.864
positive0.911.000.9529
Accuracy--0.9249
Weighted Avg0.940.920.9349
Macro Avg0.730.640.6849
Table 30. Top five models with respect to accuracy (“Laptops”, BERT representations).
Table 30. Top five models with respect to accuracy (“Laptops”, BERT representations).
ModelAccuracy
ExtraTreesClassifier0.9796
RandomForestClassifier0.9796
DecisionTreeClassifier0.9592
ExtraTreeClassifier0.9592
GradientBoostingClassifier0.9592
Table 31. Classification Report for Majority Vote (“Laptops”, BERT representations, top five models).
Table 31. Classification Report for Majority Vote (“Laptops”, BERT representations, top five models).
ModelPrecisionRecallF1-ScoreSupport
conflict0.000.000.000
negative1.000.940.9716
neutral1.001.001.004
positive1.001.001.0029
Accuracy--0.9849
Weighted Avg1.000.980.9949
Macro Avg0.750.730.7449
Table 32. Classification Report for Majority Vote (“Restaurants”, BERT representations, all models).
Table 32. Classification Report for Majority Vote (“Restaurants”, BERT representations, all models).
ModelPrecisionRecallF1-ScoreSupport
negative0.920.610.7318
neutral1.000.300.4610
positive0.841.000.9168
Accuracy--0.8596
Weighted Avg0.870.850.8396
Macro Avg0.920.640.7096
Table 33. Top five models with respect to accuracy (“Restaurants”, BERT representations).
Table 33. Top five models with respect to accuracy (“Restaurants”, BERT representations).
ModelAccuracy
BaggingClassifier1.0000
DecisionTreeClassifier1.0000
ExtraTreeClassifier1.0000
ExtraTreesClassifier1.0000
RandomForestClassifier1.0000
Table 34. Classification Report for Majority Vote (“Restaurants”, BERT representations, top five models).
Table 34. Classification Report for Majority Vote (“Restaurants”, BERT representations, top five models).
ModelPrecisionRecallF1-ScoreSupport
negative1.001.001.0018
neutral1.001.001.0010
positive1.001.001.0068
Accuracy--1.0096
Weighted Avg1.001.001.0096
Macro Avg1.001.001.0096
Table 35. Classification Report for Majority Vote (“MTSC”, BERT representations, all models).
Table 35. Classification Report for Majority Vote (“MTSC”, BERT representations, all models).
ModelPrecisionRecallF1-ScoreSupport
negative0.710.840.77429
neutral0.750.700.73455
positive0.710.580.64262
Accuracy--0.731146
Weighted Avg0.730.730.721146
Macro Avg0.720.710.711146
Table 36. Top five models with respect to accuracy (“MTSC”, BERT representations).
Table 36. Top five models with respect to accuracy (“MTSC”, BERT representations).
ModelAccuracy
LogisticRegressionCV0.7243
LogisticRegression0.7199
CalibratedClassifierCV0.7182
SVC0.7155
RidgeClassifier0.7120
Table 37. Classification Report for Majority Vote (“MTSC”, BERT representations, top five models).
Table 37. Classification Report for Majority Vote (“MTSC”, BERT representations, top five models).
ModelPrecisionRecallF1-ScoreSupport
negative0.760.770.76429
neutral0.700.790.75455
positive0.730.570.64262
Accuracy--0.731146
Weighted Avg0.730.730.731146
Macro Avg0.730.710.721146
Table 38. Selected sentences from the “MTSC” data set that were wrongly classified by the majority vote, with ChatGPT predicted polarities. The wrong predictions are in bold.
Table 38. Selected sentences from the “MTSC” data set that were wrongly classified by the majority vote, with ChatGPT predicted polarities. The wrong predictions are in bold.
SentenceAspectTrue PolarityChatGPT Polarity
In an atmosphere where some delegates remain anti-Trump and party leaders like Paul Ryan are barely mentioning Trump in their speeches, Hillary Clinton is a unifying force.Hillary Clintonpositivepositive
A new presidential cabinet will be formed as well as a national reconciliation committee, which will include youth movements that have been behind anti-Morsi demonstrations.Morsinegativenegative
His persona is generally adult.Hispositiveneutral
The more left wing candidate for deputy leader—Julie Morgan, widow of former first minister of Wales, Rhodri—was defeated by Carolyn Harris at the party’s spring conference.Carolyn Harrispositivepositive
In a statement Saturday, London Moore, the president of the Theta Gamma chapter of Delta Delta Delta, condemned the “racist, offensive and disgraceful” behavior seen in the video.London Moore, the president of the Theta Gamma chapter of Delta Delta Deltapositivepositive
President Muhammadu Buhari was poised to win a second term despite falling short on promises to recharge the economy and defeat the Boko Haram insurgents.President Muhammadu Buharinegativeneutral
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chifu, A.-G.; Fournier, S. Sentiment Difficulty in Aspect-Based Sentiment Analysis. Mathematics 2023, 11, 4647. https://doi.org/10.3390/math11224647

AMA Style

Chifu A-G, Fournier S. Sentiment Difficulty in Aspect-Based Sentiment Analysis. Mathematics. 2023; 11(22):4647. https://doi.org/10.3390/math11224647

Chicago/Turabian Style

Chifu, Adrian-Gabriel, and Sébastien Fournier. 2023. "Sentiment Difficulty in Aspect-Based Sentiment Analysis" Mathematics 11, no. 22: 4647. https://doi.org/10.3390/math11224647

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop