Published August 10, 2024 | Version v1
Journal article Open

A Comparative Study of Feature Attention Mechanisms in Text Classification

  • 1. Faculty of Computer Science, Gujarat University, Ahmedabad, India
  • 2. K.S. School of Business Management and Information Technology, Gujarat University

Description

Abstract:
 Text classification is essential in natural language processing (NLP), involving categorizing text into predefined categories. Traditional machine learning algorithms, like Support Vector Machines (SVM) and Naive Bayes, require extensive feature engineering and struggle with high-dimensional data. Deep learning models, such as Convolutional Neural Networks (CNN) and Long Short-Term Memory (LSTM) networks, have improved text classification by learning feature representations. However, these models face challenges with long-range dependencies and word importance.
Attention mechanisms, including additive, multiplicative, self-attention, and multi-head attention, address these limitations by allowing models to focus on relevant input parts. This paper compares these mechanisms' impact on text classification performance using multiple datasets and neural network architectures.
Results show that self-attention and multi-head attention outperform additive and multiplicative attention, effectively capturing global dependencies and enhancing feature representation. The study offers insights into each mechanism's strengths and weaknesses, providing guidelines for selecting suitable mechanisms for specific text classification tasks.

Files

15...pdf

Files (359.7 kB)

Name Size Download all
md5:1cae5cd70ba7c63d53f3050dec153efd
359.7 kB Preview Download