Bi-LSTM model to increase accuracy in text classification: Combining word2vec CNN and attention mechanism

Beakcheol Jang, Myeonghwi Kim, Gaspard Harerimana, Sang Ug Kang, Jong Wook Kim

Research output: Contribution to journalArticlepeer-review

15 Citations (Scopus)

Abstract

There is a need to extract meaningful information from big data, classify it into different categories, and predict end-user behavior or emotions. Large amounts of data are generated from various sources such as social media and websites. Text classification is a representative research topic in the field of natural-language processing that categorizes unstructured text data into meaningful categorical classes. The long short-term memory (LSTM) model and the convolutional neural network for sentence classification produce accurate results and have been recently used in various natural-language processing (NLP) tasks. Convolutional neural network (CNN) models use convolutional layers and maximum pooling or max-overtime pooling layers to extract higher-level features, while LSTM models can capture long-term dependencies between word sequences hence are better used for text classification. However, even with the hybrid approach that leverages the powers of these two deep-learning models, the number of features to remember for classification remains huge, hence hindering the training process. In this study, we propose an attention-based Bi-LSTM+CNN hybrid model that capitalize on the advantages of LSTM and CNN with an additional attention mechanism. We trained the model using the Internet Movie Database (IMDB) movie review data to evaluate the performance of the proposed model, and the test results showed that the proposed hybrid attention Bi-LSTM+CNN model produces more accurate classification results, as well as higher recall and F1 scores, than individual multi-layer perceptron (MLP), CNN or LSTM models as well as the hybrid models.

Original languageEnglish
Article number5841
JournalApplied Sciences (Switzerland)
Volume10
Issue number17
DOIs
Publication statusPublished - 2020 Sep

Bibliographical note

Funding Information:
Funding: This work was supported by the National Research Foundation of Korea Grant funded by the Korea Government under Grant NRF-2019R1F1A1058058.

Publisher Copyright:
© 2020 by the authors.

All Science Journal Classification (ASJC) codes

  • Materials Science(all)
  • Instrumentation
  • Engineering(all)
  • Process Chemistry and Technology
  • Computer Science Applications
  • Fluid Flow and Transfer Processes

Fingerprint

Dive into the research topics of 'Bi-LSTM model to increase accuracy in text classification: Combining word2vec CNN and attention mechanism'. Together they form a unique fingerprint.

Cite this