Text categorization plays a crucial role in various NLP applications, including sentiment analysis, subject labeling, and question answering. Machine-learning-based text classification is a key research subject with numerous applications, including spam detection, hate speech identification, reviews, rating summarization, sentiment analysis, and topic modeling. Machine learning algorithms for text classification have proven to be quite reliable and are widely employed. We introduce SentCNN, based on convolutional neural network (CNN) for text classification that addresses NLP challenges. SentCNN's simplified architecture ensures outstanding performance while lowering processing costs. In this work, Stanford Sentiment Treebank dataset is used for text classification. The proposed SentCNN achieved remarkable results with an accuracy of 95.2% with 95.6% recall, and an F1 score of 95.2% and compared with sophisticated models like DPCNN, RCNN, and TextCNN. The comparative results shows that the proposed SentCNN outperformed existing models and effective for text classification.
Keywords
Text Classification, Natural Language Processing, Convolutional Neural Networks and Stanford Sentiment Treebank.
Palanivinayagam, C. Z. El-Bayeh, and R. Damaševiˇcius, “Twenty Years of Machine-Learning-Based Text Classification: A Systematic Review,” Algorithms, vol. 16, p. 236, 2023, doi: 10.3390/a16050236.
F. Sebastiani, “Machine Learning in Automated Text Categorization,” ACM Comput. Surv., vol. 34, pp. 1–47, 2002.
H. Kim and Y.-S. Jeong, “Sentiment Classification Using Convolutional Neural Networks,” Appl. Sci., vol. 9, p. 2347, 2019, doi: 10.3390/app9112347.
K. Kowsari, K. Jafari Meimandi, M. Heidarysafa, S. Mendu, L. Barnes, and D. Brown, “Text Classification Algorithms: A Survey,” Information, vol. 10, p. 150, 2019.
D. W. Otter, J. R. Medina, and J. K. Kalita, “A Survey of the Usages of Deep Learning for Natural Language Processing,” IEEE Trans. Neural Netw. Learn. Syst., vol. 32, no. 2, pp. 604–624, Feb. 2021, doi: 10.1109/TNNLS.2020.2979670.
E. O. Arkhangelskaya and S. I. Nikolenko, “Deep Learning for Natural Language Processing: A Survey,” J. Math. Sci., vol. 273, pp. 533–582, 2023, doi: 10.1007/s10958-023-06519-6.
L. He, S. Tan, F. Xiang, J. Wu, and L. Tan, “Research and development of deep learning-based text classification,” Comput. Eng., vol. 47, pp. 1–11, 2021.
S. K. Prabhakar, H. Rajaguru, and D. O. Won, “Performance Analysis of Hybrid Deep Learning Models with Attention Mechanism Positioning and Focal Loss for Text Classification,” Sci. Program., vol. 2021, p. 2420254, 2021.
L. Duan, Q. You, X. Wu, and J. Sun, “Multilabel Text Classification Algorithm Based on Fusion of Two-Stream Transformer,” Electronics, vol. 11, p. 2138, 2022.
Q. Li, Z. Xiao, and Y. Zhao, “Research on the Classification of New Energy Industry Policy Texts Based on BERT Model,” Sustainability, vol. 15, p. 111, 2023.
D. Wu, Z. Wang, and W. Zhao, “XLNet-CNN-GRU dual-channel aspect-level review text sentiment classification method,” Multimed. Tools Appl., vol. 83, pp. 5871–5892, 2024.
T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean, “Distributed Representations of Words and Phrases and their Compositionality,” in Adv. Neural Inf. Process. Syst., vol. 26, 2013.
Y. Kim, “Convolutional Neural Networks for Sentence Classification,” in Proc. 2014 Conf. Empirical Methods Natural Lang. Process., Doha, Qatar, Oct. 2014, pp. 1746–1751.
P. Liu, X. Qiu, and X. Huang, “Recurrent Neural Network for Text Classification with Multi-Task Learning,” arXiv, 2016, arXiv:1605.05101.
J. Y. Lee and F. Dernoncourt, “Sequential short-text classification with recurrent and convolutional neural networks,” arXiv, 2016, arXiv:1603.03827.
A. Conneau, H. Schwenk, L. Barrault, and Y. Lecun, “Very deep convolutional networks for text classification,” arXiv, 2016, arXiv:1606.01781.
S. Lai, L. Xu, K. Liu, and J. Zhao, “Recurrent convolutional neural networks for text classification,” in Proc. Twenty-Ninth AAAI Conf. Artif. Intell., Austin, TX, USA, Jan. 2015, pp. 2267–2273.
X. Zhang, J. Zhao, and Y. LeCun, “Character-level convolutional networks for text classification,” Adv. Neural Inf. Process. Syst., vol. 2015, 2015.
R. Socher, A. Perelygin, et al., “Recursive deep models for semantic compositionality over a sentiment treebank,” in Proc. Conf. Empirical Methods Natural Lang. Process., 2013, pp. 1631–1642.
M. A. Parwez and M. J. I. A. Abulaish, “Multi-label classification of microblogging texts using convolution neural network,” IEEE Access, vol. 7, pp. 68678–68691, 2019.
D. Tang, F. Wei, B. Qin, N. Yang, T. Liu, M. Zhou, and D. Engineering, “Sentiment embeddings with applications to sentiment analysis,” IEEE Trans. Knowl. Data Eng., vol. 28, pp. 496–509, 2015.
W. Hu, J. Xiong, N. Wang, F. Liu, Y. Kong, and C. Yang, “Integrated Mod el Text Classification Based on Multineural Networks,” Electronics, vol. 13, p. 453, 2024, doi: 10.3390/electronics13020453.
R. Johnson and T. Zhang, “Deep Pyramid Convolutional Neural Networks for Text Categorization,” in Proc. 55th Annu. Meeting Assoc. Comput. Linguist., 2017, pp. 562–570.
CRediT Author Statement
The authors confirm contribution to the paper as follows:
Conceptualization: IJ, AR;
Methodology: IJ, AR;
Software: AR;
Data Curation: IJ;
Writing- Original Draft Preparation: IJ, AR;
Visualization: IJ;
Supervision: AR;
Validation: IJ, AR;
Writing- Reviewing and Editing: IJ, AR;
Writing- Original Draft: IJ, AR; All authors reviewed the results and approved the final version of the manuscript.
Acknowledgements
The author(s) received no financial support for the research, authorship, and/or publication of this article.
Funding
No funding was received for conducting this research.
Ethics declarations
Conflict of interest
The authors have no conflicts of interest to declare that are relevant to the content of this article.
Availability of data and materials
No data available for above study.
Author information
Contributions
All authors have equal contribution in the paper and all authors have read and agreed to the published version of the manuscript.
Open Access This article is licensed under a Creative Commons Attribution NoDerivs is a more restrictive license. It allows you to redistribute the material commercially or non-commercially but the user cannot make any changes whatsoever to the original, i.e. no derivatives of the original work. To view a copy of this license, visit https://creativecommons.org/licenses/by-nc-nd/4.0/
Cite this article
Iram Javed and Arulmurugan Ramu, "SentCNN - A Lightweight and Efficient Deep Learning Model for Accurate Text Classification", Journal of Future Networks and Communications, vol.1, no.1, pp. 033-040, January 2025. doi: XXXX/XXXX/JFNC202501004.