Please use this identifier to cite or link to this item: http://202.28.34.124/dspace/handle123456789/3627
Title: A Method of Identifying Sentiment from Consumer Multimodality Reviews
A Method of Identifying Sentiment from Consumer Multimodality Reviews
Authors: Jun Wan
Jun Wan
Jantima Polpinij
จันทิมา พลพินิจ
Mahasarakham University
Jantima Polpinij
จันทิมา พลพินิจ
Jantima.p@msu.ac.th
Jantima.p@msu.ac.th
Keywords: Multimodal Sentiment Analysis
Sentiment Classification
Consumer Reviews
Text and Emoticon Fusion
Machine Learning
Deep Learning
BERT
Issue Date:  31
Publisher: Mahasarakham University
Abstract: The rapid expansion of e-commerce and digital media has led to an increasing reliance on online consumer reviews, which play a crucial role in shaping purchasing decisions. Traditional sentiment analysis methods often focus solely on textual data, overlooking the multimodal nature of consumer reviews, which frequently include emoticons and other non-textual elements. This study addresses the challenge of sentiment classification in multimodal consumer reviews by integrating text and emoticons to improve classification accuracy. The primary objective of this research is to develop a sentiment classification method capable of identifying sentiment from movie reviews containing both text and emoticons. Two classification tasks were considered: binary sentiment classification (positive and negative) and multiclass sentiment classification (positive, neutral, and negative). The dataset was collected from the Douban Film and Television Network, comprising Chinese-language movie reviews with embedded emoticons. Textual data was represented using five embedding techniques—Word2Vec, GloVe, FastText, Ada-002, and BERT—while emoticons were encoded using one-hot encoding. The fusion of textual and emoticon features was performed using concatenation. To evaluate model performance, machine learning classifiers such as Support Vector Machine (SVM) and Random Forest, as well as deep learning models like Convolutional Neural Networks (CNN) and Long Short-Term Memory (LSTM), were trained using 10-fold cross-validation. Experimental results demonstrated that deep learning models, particularly LSTM and CNN, outperformed traditional classifiers when combined with contextual embeddings such as BERT and Ada-002. In binary classification, CNN with Ada-002 achieved the highest accuracy, while LSTM with BERT exhibited superior performance in multiclass classification. The inclusion of emoticons enhanced classification results, particularly in deep learning models. For example, in experiments using the first dataset, the Word2Vec + CNN model achieved an accuracy of 0.80 with text alone, which increased to 0.83 when emoticons were included. Similarly, the GLOVE + CNN model improved from 0.80 to 0.83 with the addition of emoticons. Furthermore, two datasets were used to validate the model's performance under different conditions, testing whether the developed model can generalize and maintain good performance when encountering different data.This study highlights the importance of multimodal fusion in sentiment analysis, demonstrating that integrating emoticons with advanced text representations significantly improves sentiment classification accuracy. These findings provide valuable insights for enhancing sentiment analysis techniques in consumer review analysis.
-
URI: http://202.28.34.124/dspace/handle123456789/3627
Appears in Collections:The Faculty of Informatics

Files in This Item:
File Description SizeFormat 
65011263002.pdf8.62 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.