Journal of Future Networks and Communications

Journal of Future Networks and Communications

SentCNN - A Lightweight and Efficient Deep Learning Model for Accurate Text Classification

School of Information and Software Engineering, University of Electronics Science and Technology of China, Hefei, Anhui, 230026, P.R.China.

Iram Javed

Mattu University, Metu, Ethiopia.

Arulmurugan Ramu

Journal of Future Networks and Communications

Received On : 25 October 2024

Revised On : 10 December 2024

Accepted On : 18 December 2024

Published On : 05 January 2025

Volume 01, Issue 01

Pages : 033-040


Article Views

Abstract


Text categorization plays a crucial role in various NLP applications, including sentiment analysis, subject labeling, and question answering. Machine-learning-based text classification is a key research subject with numerous applications, including spam detection, hate speech identification, reviews, rating summarization, sentiment analysis, and topic modeling. Machine learning algorithms for text classification have proven to be quite reliable and are widely employed. We introduce SentCNN, based on convolutional neural network (CNN) for text classification that addresses NLP challenges. SentCNN's simplified architecture ensures outstanding performance while lowering processing costs. In this work, Stanford Sentiment Treebank dataset is used for text classification. The proposed SentCNN achieved remarkable results with an accuracy of 95.2% with 95.6% recall, and an F1 score of 95.2% and compared with sophisticated models like DPCNN, RCNN, and TextCNN. The comparative results shows that the proposed SentCNN outperformed existing models and effective for text classification.


Keywords


Text Classification, Natural Language Processing, Convolutional Neural Networks and Stanford Sentiment Treebank.


  1. Palanivinayagam, C. Z. El-Bayeh, and R. Damaševiˇcius, “Twenty Years of Machine-Learning-Based Text Classification: A Systematic Review,” Algorithms, vol. 16, p. 236, 2023, doi: 10.3390/a16050236.
  2. F. Sebastiani, “Machine Learning in Automated Text Categorization,” ACM Comput. Surv., vol. 34, pp. 1–47, 2002.
  3. H. Kim and Y.-S. Jeong, “Sentiment Classification Using Convolutional Neural Networks,” Appl. Sci., vol. 9, p. 2347, 2019, doi: 10.3390/app9112347.
  4. K. Kowsari, K. Jafari Meimandi, M. Heidarysafa, S. Mendu, L. Barnes, and D. Brown, “Text Classification Algorithms: A Survey,” Information, vol. 10, p. 150, 2019.
  5. D. W. Otter, J. R. Medina, and J. K. Kalita, “A Survey of the Usages of Deep Learning for Natural Language Processing,” IEEE Trans. Neural Netw. Learn. Syst., vol. 32, no. 2, pp. 604–624, Feb. 2021, doi: 10.1109/TNNLS.2020.2979670.
  6. E. O. Arkhangelskaya and S. I. Nikolenko, “Deep Learning for Natural Language Processing: A Survey,” J. Math. Sci., vol. 273, pp. 533–582, 2023, doi: 10.1007/s10958-023-06519-6.
  7. L. He, S. Tan, F. Xiang, J. Wu, and L. Tan, “Research and development of deep learning-based text classification,” Comput. Eng., vol. 47, pp. 1–11, 2021.
  8. S. K. Prabhakar, H. Rajaguru, and D. O. Won, “Performance Analysis of Hybrid Deep Learning Models with Attention Mechanism Positioning and Focal Loss for Text Classification,” Sci. Program., vol. 2021, p. 2420254, 2021.
  9. L. Duan, Q. You, X. Wu, and J. Sun, “Multilabel Text Classification Algorithm Based on Fusion of Two-Stream Transformer,” Electronics, vol. 11, p. 2138, 2022.
  10. Q. Li, Z. Xiao, and Y. Zhao, “Research on the Classification of New Energy Industry Policy Texts Based on BERT Model,” Sustainability, vol. 15, p. 111, 2023.
  11. D. Wu, Z. Wang, and W. Zhao, “XLNet-CNN-GRU dual-channel aspect-level review text sentiment classification method,” Multimed. Tools Appl., vol. 83, pp. 5871–5892, 2024.
  12. T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean, “Distributed Representations of Words and Phrases and their Compositionality,” in Adv. Neural Inf. Process. Syst., vol. 26, 2013.
  13. Y. Kim, “Convolutional Neural Networks for Sentence Classification,” in Proc. 2014 Conf. Empirical Methods Natural Lang. Process., Doha, Qatar, Oct. 2014, pp. 1746–1751.
  14. P. Liu, X. Qiu, and X. Huang, “Recurrent Neural Network for Text Classification with Multi-Task Learning,” arXiv, 2016, arXiv:1605.05101.
  15. J. Y. Lee and F. Dernoncourt, “Sequential short-text classification with recurrent and convolutional neural networks,” arXiv, 2016, arXiv:1603.03827.
  16. A. Conneau, H. Schwenk, L. Barrault, and Y. Lecun, “Very deep convolutional networks for text classification,” arXiv, 2016, arXiv:1606.01781.
  17. S. Lai, L. Xu, K. Liu, and J. Zhao, “Recurrent convolutional neural networks for text classification,” in Proc. Twenty-Ninth AAAI Conf. Artif. Intell., Austin, TX, USA, Jan. 2015, pp. 2267–2273.
  18. X. Zhang, J. Zhao, and Y. LeCun, “Character-level convolutional networks for text classification,” Adv. Neural Inf. Process. Syst., vol. 2015, 2015.
  19. R. Socher, A. Perelygin, et al., “Recursive deep models for semantic compositionality over a sentiment treebank,” in Proc. Conf. Empirical Methods Natural Lang. Process., 2013, pp. 1631–1642.
  20. M. A. Parwez and M. J. I. A. Abulaish, “Multi-label classification of microblogging texts using convolution neural network,” IEEE Access, vol. 7, pp. 68678–68691, 2019.
  21. D. Tang, F. Wei, B. Qin, N. Yang, T. Liu, M. Zhou, and D. Engineering, “Sentiment embeddings with applications to sentiment analysis,” IEEE Trans. Knowl. Data Eng., vol. 28, pp. 496–509, 2015.
  22. “Stanford Sentiment Treebank Dataset,” [Online]. Available: https://nlp.stanford.edu/sentiment/code.html/. [Accessed: Nov. 11, 2024].
  23. W. Hu, J. Xiong, N. Wang, F. Liu, Y. Kong, and C. Yang, “Integrated Mod el Text Classification Based on Multineural Networks,” Electronics, vol. 13, p. 453, 2024, doi: 10.3390/electronics13020453.
  24. R. Johnson and T. Zhang, “Deep Pyramid Convolutional Neural Networks for Text Categorization,” in Proc. 55th Annu. Meeting Assoc. Comput. Linguist., 2017, pp. 562–570.

CRediT Author Statement


The authors confirm contribution to the paper as follows:

Conceptualization: IJ, AR; Methodology: IJ, AR; Software: AR; Data Curation: IJ; Writing- Original Draft Preparation: IJ, AR; Visualization: IJ; Supervision: AR; Validation: IJ, AR; Writing- Reviewing and Editing: IJ, AR; Writing- Original Draft: IJ, AR; All authors reviewed the results and approved the final version of the manuscript.


Acknowledgements


The author(s) received no financial support for the research, authorship, and/or publication of this article.


Funding


No funding was received for conducting this research.


Ethics declarations


Conflict of interest

The authors have no conflicts of interest to declare that are relevant to the content of this article.


Availability of data and materials


No data available for above study.


Author information


Contributions

All authors have equal contribution in the paper and all authors have read and agreed to the published version of the manuscript.

Corresponding author

Mattu University, Metu, Ethiopia.

Arulmurugan Ramu

Rights and permissions


Open Access This article is licensed under a Creative Commons Attribution NoDerivs is a more restrictive license. It allows you to redistribute the material commercially or non-commercially but the user cannot make any changes whatsoever to the original, i.e. no derivatives of the original work. To view a copy of this license, visit https://creativecommons.org/licenses/by-nc-nd/4.0/


Cite this article


Iram Javed and Arulmurugan Ramu, "SentCNN - A Lightweight and Efficient Deep Learning Model for Accurate Text Classification", Journal of Future Networks and Communications, vol.1, no.1, pp. 033-040, January 2025. doi: XXXX/XXXX/JFNC202501004.


Copyright


© 2025 Iram Javed and Arulmurugan Ramu. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.