Comparison of RNN Architectures and Non-RNN Architectures in Sentiment Analysis





1D ConvNets, Accuracy, Bidirectional Recurrent Neural Network, Gated Recurrent Unit, Long Short-Term Memory


This study compares the sentiment analysis performance of multiple Recurrent Neural Network architectures and One-Dimensional Convolutional Neural Networks. THE METHODS EVALUATED ARE simple Recurrent Neural Network, Long Short-Term Memory, Gated Recurrent Unit, Bidirectional Recurrent Neural Network, and 1D ConvNets. A dataset comprising text reviews with positive or negative sentiment labels was evaluated. All evaluated models demonstrated an extremely high accuracy, ranging from 99.81% to 99.99%. Apart from that, the loss generated by these models is also low, ranging from 0.0043 to 0.0021. However, there are minor performance differences between the evaluated architectures. The Long Short-Term Memory and Gated Recurrent Unit models mainly perform marginally better than the Simple Recurrent Neural Network, albeit with slightly lower accuracy and loss. In the meantime, the Bidirectional Recurrent Neural Network model demonstrates competitive performance, as it can effectively manage text context from both directions. Additionally, One-Dimensional Convolutional Neural Networks provide satisfactory results, indicating that convolution-based approaches are also effective in sentiment analysis. The findings of this study provide practitioners with essential insights for selecting an appropriate architecture for sentiment analysis tasks. While all models yield excellent performance, the choice of architecture can impact computational efficiency and training time. Therefore, a comprehensive comprehension of the respective characteristics of Recurrent Neural Network architectures and One-Dimensional Convolutional Neural Networks is essential for making more informed decisions when constructing sentiment analysis models.

GS Cited Analysis


Download data is not yet available.


Alshammari, H. H., & Alkhiri, H. (2023). Optimized recurrent neural network mechanism for olive leaf disease diagnosis based on wavelet transform. Alexandria Engineering Journal, 78(June), 149–161.

ArunKumar, K. E., Kalaga, D. V., Kumar, C. M. S., Kawaji, M., & Brenza, T. M. (2021). Forecasting of COVID-19 using deep layer Recurrent Neural Networks (RNNs) with Gated Recurrent Units (GRUs) and Long Short-Term Memory (LSTM) cells. Chaos, Solitons and Fractals, 146, 110861.

ArunKumar, K. E., Kalaga, D. V., Mohan Sai Kumar, C., Kawaji, M., & Brenza, T. M. (2022). Comparative analysis of Gated Recurrent Units (GRU), long Short-Term memory (LSTM) cells, autoregressive Integrated moving average (ARIMA), seasonal autoregressive Integrated moving average (SARIMA) for forecasting COVID-19 trends. Alexandria Engineering Journal, 61(10), 7585–7603.

Bacanin, N., Jovanovic, L., Zivkovic, M., Kandasamy, V., Antonijevic, M., Deveci, M., & Strumberger, I. (2023). Multivariate energy forecasting via metaheuristic tuned long-short term memory and gated recurrent unit neural networks. Information Sciences, 642(May), 119122.

Banerjee, I., Ling, Y., Chen, M. C., Hasan, S. A., Langlotz, C. P., Moradzadeh, N., Chapman, B., Amrhein, T., Mong, D., Rubin, D. L., Farri, O., & Lungren, M. P. (2019). Comparative effectiveness of convolutional neural network (CNN) and recurrent neural network (RNN) architectures for radiology text report classification. Artificial Intelligence in Medicine, 97(August 2018), 79–88.

Chandrasekaran, G., Antoanela, N., Andrei, G., Monica, C., & Hemanth, J. (2022). Visual Sentiment Analysis Using Deep Learning Models with Social Media Data. Applied Sciences (Switzerland), 12(3).

Doshi, U., Barot, V., & Gavhane, S. (2020). Emotion detection and sentiment analysis of static images. 2020 International Conference on Convergence to Digital World - Quo Vadis, ICCDW 2020, Iccdw.

Elfaik, H., & Nfaoui, E. H. (2023). Leveraging feature-level fusion representations and attentional bidirectional RNN-CNN deep models for Arabic affect analysis on Twitter. Journal of King Saud University - Computer and Information Sciences, 35(1), 462–482.

Gholami, H., Mohammadifar, A., Golzari, S., Song, Y., & Pradhan, B. (2023). Interpretability of simple RNN and GRU deep learning models used to map land susceptibility to gully erosion. Science of the Total Environment, 904(September).

Mirzaei, S., Kang, J. L., & Chu, K. Y. (2022). A comparative study on long short-term memory and gated recurrent unit neural networks in fault diagnosis for chemical processes using visualization. Journal of the Taiwan Institute of Chemical Engineers, 130, 104028.

Nugen, F., Vera Garcia, D. V., Sohn, S., Mickley, J. P., Wyles, C. C., Erickson, B. J., & Taunton, M. J. (2023). Application of Natural Language Processing in Total Joint Arthroplasty: Opportunities and Challenges. Journal of Arthroplasty, 38(10), 1948–1953.

Onan, A. (2022). Bidirectional convolutional recurrent neural network architecture with group-wise enhancement mechanism for text sentiment classification. Journal of King Saud University - Computer and Information Sciences, 34(5), 2098–2117.

Pino, P. A., Soto, C., Asún, R. A., & Guti, F. (2023). Profiling support in literacy development : Use of natural language processing to identify learning needs in higher education ☆. 58(September).

Rayhan Ahmed, M., Islam, S., Muzahidul Islam, A. K. M., & Shatabda, S. (2023). An ensemble 1D-CNN-LSTM-GRU model with data augmentation for speech emotion recognition. Expert Systems with Applications, 218(February), 119633.

Shastry, K. A., & Shastry, A. (2023). An integrated deep learning and natural language processing approach for continuous remote monitoring in digital health. Decision Analytics Journal, 8(May), 100301.

Sonsare, P. M., & C, G. (2021). Cascading 1D-Convnet Bidirectional Long Short Term Memory Network with Modified COCOB Optimizer: A Novel Approach for Protein Secondary Structure Prediction. Chaos, Solitons and Fractals, 153.

Yao, Z., Wang, Z., Wang, D., Wu, J., & Chen, L. (2023). An ensemble CNN-LSTM and GRU adaptive weighting model based improved sparrow search algorithm for predicting runoff using historical meteorological and runoff data as input. Journal of Hydrology, 625(April).


Crossmark Updates

How to Cite

Hindarto, D. (2023). Comparison of RNN Architectures and Non-RNN Architectures in Sentiment Analysis. Sinkron : Jurnal Dan Penelitian Teknik Informatika, 8(4), 2537-2546.

Most read articles by the same author(s)

1 2 3 > >>