Alert button

"Sentiment Analysis": models, code, and papers
Alert button

A Wide Evaluation of ChatGPT on Affective Computing Tasks

Aug 26, 2023
Mostafa M. Amin, Rui Mao, Erik Cambria, Björn W. Schuller

Figure 1 for A Wide Evaluation of ChatGPT on Affective Computing Tasks
Figure 2 for A Wide Evaluation of ChatGPT on Affective Computing Tasks

With the rise of foundation models, a new artificial intelligence paradigm has emerged, by simply using general purpose foundation models with prompting to solve problems instead of training a separate machine learning model for each problem. Such models have been shown to have emergent properties of solving problems that they were not initially trained on. The studies for the effectiveness of such models are still quite limited. In this work, we widely study the capabilities of the ChatGPT models, namely GPT-4 and GPT-3.5, on 13 affective computing problems, namely aspect extraction, aspect polarity classification, opinion extraction, sentiment analysis, sentiment intensity ranking, emotions intensity ranking, suicide tendency detection, toxicity detection, well-being assessment, engagement measurement, personality assessment, sarcasm detection, and subjectivity detection. We introduce a framework to evaluate the ChatGPT models on regression-based problems, such as intensity ranking problems, by modelling them as pairwise ranking classification. We compare ChatGPT against more traditional NLP methods, such as end-to-end recurrent neural networks and transformers. The results demonstrate the emergent abilities of the ChatGPT models on a wide range of affective computing problems, where GPT-3.5 and especially GPT-4 have shown strong performance on many problems, particularly the ones related to sentiment, emotions, or toxicity. The ChatGPT models fell short for problems with implicit signals, such as engagement measurement and subjectivity detection.

* 8 pages with references, 2 tables 
Viaarxiv icon

Support for Stock Trend Prediction Using Transformers and Sentiment Analysis

May 18, 2023
Harsimrat Kaeley, Ye Qiao, Nader Bagherzadeh

Figure 1 for Support for Stock Trend Prediction Using Transformers and Sentiment Analysis
Figure 2 for Support for Stock Trend Prediction Using Transformers and Sentiment Analysis
Figure 3 for Support for Stock Trend Prediction Using Transformers and Sentiment Analysis
Figure 4 for Support for Stock Trend Prediction Using Transformers and Sentiment Analysis

Stock trend analysis has been an influential time-series prediction topic due to its lucrative and inherently chaotic nature. Many models looking to accurately predict the trend of stocks have been based on Recurrent Neural Networks (RNNs). However, due to the limitations of RNNs, such as gradient vanish and long-term dependencies being lost as sequence length increases, in this paper we develop a Transformer based model that uses technical stock data and sentiment analysis to conduct accurate stock trend prediction over long time windows. This paper also introduces a novel dataset containing daily technical stock data and top news headline data spanning almost three years. Stock prediction based solely on technical data can suffer from lag caused by the inability of stock indicators to effectively factor in breaking market news. The use of sentiment analysis on top headlines can help account for unforeseen shifts in market conditions caused by news coverage. We measure the performance of our model against RNNs over sequence lengths spanning 5 business days to 30 business days to mimic different length trading strategies. This reveals an improvement in directional accuracy over RNNs as sequence length is increased, with the largest improvement being close to 18.63% at 30 business days.

* 8 pages, 3 figures, 1 table. To be published in IISES 18th Economics & Finance Conference, London 
Viaarxiv icon

Quantum Natural Language Processing based Sentiment Analysis using lambeq Toolkit

May 30, 2023
Srinjoy Ganguly, Sai Nandan Morapakula, Luis Miguel Pozo Coronado

Figure 1 for Quantum Natural Language Processing based Sentiment Analysis using lambeq Toolkit
Figure 2 for Quantum Natural Language Processing based Sentiment Analysis using lambeq Toolkit
Figure 3 for Quantum Natural Language Processing based Sentiment Analysis using lambeq Toolkit
Figure 4 for Quantum Natural Language Processing based Sentiment Analysis using lambeq Toolkit

Sentiment classification is one the best use case of classical natural language processing (NLP) where we can witness its power in various daily life domains such as banking, business and marketing industry. We already know how classical AI and machine learning can change and improve technology. Quantum natural language processing (QNLP) is a young and gradually emerging technology which has the potential to provide quantum advantage for NLP tasks. In this paper we show the first application of QNLP for sentiment analysis and achieve perfect test set accuracy for three different kinds of simulations and a decent accuracy for experiments ran on a noisy quantum device. We utilize the lambeq QNLP toolkit and $t|ket>$ by Cambridge Quantum (Quantinuum) to bring out the results.

* 10.1109/ICPC2T53885.2022.9776836  
* 6 pages, 9 figures 
Viaarxiv icon

Studying the impacts of pre-training using ChatGPT-generated text on downstream tasks

Sep 02, 2023
Sarthak Anand

Figure 1 for Studying the impacts of pre-training using ChatGPT-generated text on downstream tasks
Figure 2 for Studying the impacts of pre-training using ChatGPT-generated text on downstream tasks
Figure 3 for Studying the impacts of pre-training using ChatGPT-generated text on downstream tasks
Figure 4 for Studying the impacts of pre-training using ChatGPT-generated text on downstream tasks

In recent times, significant advancements have been witnessed in the field of language models, particularly with the emergence of Large Language Models (LLMs) that are trained on vast amounts of data extracted from internet archives. These LLMs, such as ChatGPT, have become widely accessible, allowing users to generate text for various purposes including articles, essays, jokes, and poetry. Given that LLMs are trained on a diverse range of text sources, encompassing platforms like Reddit and Twitter, it is foreseeable that future training datasets will also incorporate text generated by previous iterations of the models themselves. In light of this development, our research aims to investigate the influence of artificial text in the pre-training phase of language models. Specifically, we conducted a comparative analysis between a language model, RoBERTa, pre-trained using CNN/DailyMail news articles, and ChatGPT, which employed the same articles for its training and evaluated their performance on three downstream tasks as well as their potential gender bias, using sentiment analysis as a metric. Through a series of experiments, we demonstrate that the utilization of artificial text during pre-training does not have a significant impact on either the performance of the models in downstream tasks or their gender bias. In conclusion, our findings suggest that the inclusion of text generated by LLMs in their own pre-training process does not yield substantial effects on the subsequent performance of the models in downstream tasks or their potential gender bias.

* Master's thesis 
Viaarxiv icon

Reducing Computational Costs in Sentiment Analysis: Tensorized Recurrent Networks vs. Recurrent Networks

Jun 16, 2023
Gabriel Lopez, Anna Nguyen, Joe Kaul

Figure 1 for Reducing Computational Costs in Sentiment Analysis: Tensorized Recurrent Networks vs. Recurrent Networks
Figure 2 for Reducing Computational Costs in Sentiment Analysis: Tensorized Recurrent Networks vs. Recurrent Networks

Anticipating audience reaction towards a certain text is integral to several facets of society ranging from politics, research, and commercial industries. Sentiment analysis (SA) is a useful natural language processing (NLP) technique that utilizes lexical/statistical and deep learning methods to determine whether different-sized texts exhibit positive, negative, or neutral emotions. Recurrent networks are widely used in machine-learning communities for problems with sequential data. However, a drawback of models based on Long-Short Term Memory networks and Gated Recurrent Units is the significantly high number of parameters, and thus, such models are computationally expensive. This drawback is even more significant when the available data are limited. Also, such models require significant over-parameterization and regularization to achieve optimal performance. Tensorized models represent a potential solution. In this paper, we classify the sentiment of some social media posts. We compare traditional recurrent models with their tensorized version, and we show that with the tensorized models, we reach comparable performances with respect to the traditional models while using fewer resources for the training.

Viaarxiv icon

Aspect-oriented Opinion Alignment Network for Aspect-Based Sentiment Classification

Aug 22, 2023
Xueyi Liu, Rui Hou, Yanglei Gan, Da Luo, Changlin Li, Xiaojun Shi, Qiao Liu

Figure 1 for Aspect-oriented Opinion Alignment Network for Aspect-Based Sentiment Classification
Figure 2 for Aspect-oriented Opinion Alignment Network for Aspect-Based Sentiment Classification
Figure 3 for Aspect-oriented Opinion Alignment Network for Aspect-Based Sentiment Classification
Figure 4 for Aspect-oriented Opinion Alignment Network for Aspect-Based Sentiment Classification

Aspect-based sentiment classification is a crucial problem in fine-grained sentiment analysis, which aims to predict the sentiment polarity of the given aspect according to its context. Previous works have made remarkable progress in leveraging attention mechanism to extract opinion words for different aspects. However, a persistent challenge is the effective management of semantic mismatches, which stem from attention mechanisms that fall short in adequately aligning opinions words with their corresponding aspect in multi-aspect sentences. To address this issue, we propose a novel Aspect-oriented Opinion Alignment Network (AOAN) to capture the contextual association between opinion words and the corresponding aspect. Specifically, we first introduce a neighboring span enhanced module which highlights various compositions of neighboring words and given aspects. In addition, we design a multi-perspective attention mechanism that align relevant opinion information with respect to the given aspect. Extensive experiments on three benchmark datasets demonstrate that our model achieves state-of-the-art results. The source code is available at https://github.com/AONE-NLP/ABSA-AOAN.

* 8 pages, 5 figure, ECAI 2023 
Viaarxiv icon

An Ensemble Approach to Question Classification: Integrating Electra Transformer, GloVe, and LSTM

Aug 25, 2023
Sanad Aburass, Osama Dorgham

Figure 1 for An Ensemble Approach to Question Classification: Integrating Electra Transformer, GloVe, and LSTM
Figure 2 for An Ensemble Approach to Question Classification: Integrating Electra Transformer, GloVe, and LSTM
Figure 3 for An Ensemble Approach to Question Classification: Integrating Electra Transformer, GloVe, and LSTM

Natural Language Processing (NLP) has emerged as a crucial technology for understanding and generating human language, playing an essential role in tasks such as machine translation, sentiment analysis, and more pertinently, question classification. As a subfield within NLP, question classification focuses on determining the type of information being sought, a fundamental step for downstream applications like question answering systems. This study presents an innovative ensemble approach for question classification, combining the strengths of Electra, GloVe, and LSTM models. Rigorously tested on the well-regarded TREC dataset, the model demonstrates how the integration of these disparate technologies can lead to superior results. Electra brings in its transformer-based capabilities for complex language understanding, GloVe offers global vector representations for capturing word-level semantics, and LSTM contributes its sequence learning abilities to model long-term dependencies. By fusing these elements strategically, our ensemble model delivers a robust and efficient solution for the complex task of question classification. Through rigorous comparisons with well-known models like BERT, RoBERTa, and DistilBERT, the ensemble approach verifies its effectiveness by attaining an 80% accuracy score on the test dataset.

Viaarxiv icon

Overview of Memotion 3: Sentiment and Emotion Analysis of Codemixed Hinglish Memes

Sep 12, 2023
Shreyash Mishra, S Suryavardan, Megha Chakraborty, Parth Patwa, Anku Rani, Aman Chadha, Aishwarya Reganti, Amitava Das, Amit Sheth, Manoj Chinnakotla, Asif Ekbal, Srijan Kumar

Figure 1 for Overview of Memotion 3: Sentiment and Emotion Analysis of Codemixed Hinglish Memes
Figure 2 for Overview of Memotion 3: Sentiment and Emotion Analysis of Codemixed Hinglish Memes
Figure 3 for Overview of Memotion 3: Sentiment and Emotion Analysis of Codemixed Hinglish Memes
Figure 4 for Overview of Memotion 3: Sentiment and Emotion Analysis of Codemixed Hinglish Memes

Analyzing memes on the internet has emerged as a crucial endeavor due to the impact this multi-modal form of content wields in shaping online discourse. Memes have become a powerful tool for expressing emotions and sentiments, possibly even spreading hate and misinformation, through humor and sarcasm. In this paper, we present the overview of the Memotion 3 shared task, as part of the DeFactify 2 workshop at AAAI-23. The task released an annotated dataset of Hindi-English code-mixed memes based on their Sentiment (Task A), Emotion (Task B), and Emotion intensity (Task C). Each of these is defined as an individual task and the participants are ranked separately for each task. Over 50 teams registered for the shared task and 5 made final submissions to the test set of the Memotion 3 dataset. CLIP, BERT modifications, ViT etc. were the most popular models among the participants along with approaches such as Student-Teacher model, Fusion, and Ensembling. The best final F1 score for Task A is 34.41, Task B is 79.77 and Task C is 59.82.

* Defactify2 @AAAI 2023 
Viaarxiv icon

BanglaBook: A Large-scale Bangla Dataset for Sentiment Analysis from Book Reviews

May 26, 2023
Mohsinul Kabir, Obayed Bin Mahfuz, Syed Rifat Raiyan, Hasan Mahmud, Md Kamrul Hasan

Figure 1 for BanglaBook: A Large-scale Bangla Dataset for Sentiment Analysis from Book Reviews
Figure 2 for BanglaBook: A Large-scale Bangla Dataset for Sentiment Analysis from Book Reviews
Figure 3 for BanglaBook: A Large-scale Bangla Dataset for Sentiment Analysis from Book Reviews
Figure 4 for BanglaBook: A Large-scale Bangla Dataset for Sentiment Analysis from Book Reviews

The analysis of consumer sentiment, as expressed through reviews, can provide a wealth of insight regarding the quality of a product. While the study of sentiment analysis has been widely explored in many popular languages, relatively less attention has been given to the Bangla language, mostly due to a lack of relevant data and cross-domain adaptability. To address this limitation, we present BanglaBook, a large-scale dataset of Bangla book reviews consisting of 158,065 samples classified into three broad categories: positive, negative, and neutral. We provide a detailed statistical analysis of the dataset and employ a range of machine learning models to establish baselines including SVM, LSTM, and Bangla-BERT. Our findings demonstrate a substantial performance advantage of pre-trained models over models that rely on manually crafted features, emphasizing the necessity for additional training resources in this domain. Additionally, we conduct an in-depth error analysis by examining sentiment unigrams, which may provide insight into common classification errors in under-resourced languages like Bangla. Our codes and data are publicly available at https://github.com/mohsinulkabir14/BanglaBook.

* Accepted in ACL Findings 2023 
Viaarxiv icon

Social Media, Topic Modeling and Sentiment Analysis in Municipal Decision Support

Aug 08, 2023
Miloš Švaňa

Figure 1 for Social Media, Topic Modeling and Sentiment Analysis in Municipal Decision Support
Figure 2 for Social Media, Topic Modeling and Sentiment Analysis in Municipal Decision Support
Figure 3 for Social Media, Topic Modeling and Sentiment Analysis in Municipal Decision Support

Many cities around the world are aspiring to become. However, smart initiatives often give little weight to the opinions of average citizens. Social media are one of the most important sources of citizen opinions. This paper presents a prototype of a framework for processing social media posts with municipal decision-making in mind. The framework consists of a sequence of three steps: (1) determining the sentiment polarity of each social media post (2) identifying prevalent topics and mapping these topics to individual posts, and (3) aggregating these two pieces of information into a fuzzy number representing the overall sentiment expressed towards each topic. Optionally, the fuzzy number can be reduced into a tuple of two real numbers indicating the "amount" of positive and negative opinion expressed towards each topic. The framework is demonstrated on tweets published from Ostrava, Czechia over a period of about two months. This application illustrates how fuzzy numbers represent sentiment in a richer way and capture the diversity of opinions expressed on social media.

Viaarxiv icon