IEEE Transactions on Affective Computing
Word Representation, Language Representation, Pretrained Language Models, Affective Tasks, Text Preprocessing, Word Embeddings, Emotion Classification, Sentiment analysis, Sarcasm Detection
Affective tasks, including sentiment analysis, emotion classification, and sarcasm detection have drawn a lot of attention in recent years due to a broad range of useful applications in various domains. The main goal of affect detection tasks is to recognize states such as mood, sentiment, and emotions from textual data (e.g., news articles or product reviews). Despite the importance of utilizing preprocessing steps in different stages (i.e., word representation learning and building a classification model) of affect detection tasks, this topic has not been studied well. To that end, we explore whether applying various preprocessing methods (stemming, lemmatization, stopword removal, punctuation removal and so on) and their combinations in different stages of the affect detection pipeline can improve the model performance. The are many preprocessing approaches that can be utilized in affect detection tasks. However, their influence on the final performance depends on the type of preprocessing and the stages that they are applied. Moreover, the preprocessing impacts vary across different affective tasks. Our analysis provides thorough insights into how preprocessing steps can be applied in building an effect detection pipeline and their respective influence on performance.
© Copyright the author(s) (2023)
Locate the Document
Published as: Babanejad, N., Davoudi, H., Agrawal, A., An, A., & Papagelis, M. (2023). The Role of Preprocessing for Word Representation Learning in Affective Tasks. IEEE Transactions on Affective Computing.