A comparison of Long Short-Term Memory, Convolutional Neural Network, Transformer, and Mamba models for sentiment analysis
Date
2024
Authors
Ruan, Hang
Journal Title
Journal ISSN
Volume Title
Publisher
Abstract
Sentiment analysis is a critical task in Natural Language Processing (NLP) that helps decode the emotions and opinions embedded in text. With applications spanning from market research and social media monitoring to political analysis and customer feedback evaluation, sentiment analysis provides invaluable insights into public opinion and consumer behavior. This project studies the evolution of sentiment analysis models, focusing on the advancements made by deep learning techniques such as Long Short-Term Memory (LSTM), Convolutional Neural Networks (CNNs), and transformer-based models like Bidirectional Encoder Representations from Transformers (BERT) and Generative Pre-trained Transformer (GPT). These models have set new benchmarks for accuracy, efficiency, and versatility. Additionally, this explores Mamba, a recent State Space Model (SSM) designed to overcome the computational challenges of transformers in handling long sequences and demonstrates state-of-the-art performance on language modeling tasks comparable to transformers twice its size. This study examines the strengths and limitations of these models, comparing their performance on sentiment analysis datasets to provide a comprehensive understanding of their applicability and efficacy in various contexts.
Description
Keywords
LSTM, CNN, Transformer, Mamba, sentiment analysis