Education, Science, Technology, Innovation and Life
Open Access
Sign In

Research on Text Classification Method Based on NLP

Download as PDF

DOI: 10.23977/acss.2023.070213 | Downloads: 45 | Views: 430

Author(s)

Mengnan Wang 1

Affiliation(s)

1 The University of Queensland, Brisbane St Lucia, 4072, QLD

Corresponding Author

Mengnan Wang

ABSTRACT

Natural language processing is a science that integrates computer knowledge, mathematical knowledge and linguistic knowledge, while text classification and recognition is considered an important research area and direction of natural language processing. In the context of the big data era, how to effectively classify text information in the face of a sea of text-based data is the focus of current research. This paper describes the theoretical knowledge of text classification concepts, text representation methods and text classifiers. Firstly, the basic concepts of text classification and the classification process are introduced. Then the model structures of convolutional and recurrent neural networks and their variants are introduced, followed by the structure and implementation principles of two classical word embedding models, Word2vec and BERT.

KEYWORDS

Natural Language Processing, Text Classification, Deep Learning

CITE THIS PAPER

Mengnan Wang. Research on Text Classification Method Based on NLP. Advances in Computer, Signals and Systems (2023) Vol. 7: 93-100. DOI: http://dx.doi.org/10.23977/acss.2023.070213.

REFERENCES

[1] Zhang Q, Gao T Z, Liu X Y, et al. Public environment emotion prediction model using LSTM network[J]. Sustainability, 2020, 12(4):1-16. 
[2] Zou Y, Zhao T D, Qian W B. An improved model for spam user identification[P]. DEStech Transactions on Computer Science and Engineering, 2018. 
[3] Nawangsari R P, Kusumaningrum R, Wibowo A. Word2Vec for indonesian sentiment analysis towards hotel reviews: an evaluation study [J]. Procedia Computer Science, 2019, 157: 360-366.
[4] Liu P, Qiu X, Huang X. Recurrent neural network for text classification with multi-task learning[J]. arXiv preprint arXiv:1605.05101, 2016. 
[5] Yang M, Zhao W, Ye J, et al. Investigating capsule networks with dynamic routing for text classification [C]//Proceedings of the 2018 conference on empirical methods in natural language processing. 2018: 3110-3119. 
[6] Lin R, Fu C, Mao C, et al. Academic News Text Classification Model Based on Attention Mechanism and RCNN [C]// Springer, Singapore. Springer, Singapore, 2018. 
[7] Feng G., Zhang X., Liu S. Research on Chinese text classification based on CapsNet [J]. Data Analysis and Knowledge Discovery, 2019, 2(12). 
[8] Zhao Q., Du Y. H., Lu T. L., et al. A text similarity analysis algorithm based on capsule-BiGRU [J]. Computer Engineering and Applications, 2020, 11(27):1-9. 
[9] Lei K., Fu Q., Yang M., et al. Tag Recommendation by Text Classification with Attention-Based Capsule Network [J]. Neurocomputing, 2020.
[10] Sel L., Karci A., D Hanbay. Feature Selection for Text Classification Using Mutual Information[C]// 2019 International Artificial Intelligence and Data Processing Symposium (IDAP). IEEE, 2019. 
[11] Asim M. N., Wasim M., Ali M. S., et al. Comparison of feature selection methods in text classification on highly skewed datasets[C]// International Conference on Latest Trends in Electrical Engineering & Computing Technologies. 2017:1-8.
[12] Wang Shuang. Research on automatic text classification based on machine learning [D]. University of Electronic Science and Technology, 2020.
[13] Devlin J, Chang M W, Lee K, et al. Bert: pre-training of deep bidirectional transformers for language understanding [J]. ar Xiv preprint ar Xiv:1810.04805, 2018. 
[14] Liu Wanjun, Liang Xuejian, Qu Haicheng. Study on the learning performance of convolutional neural networks with different pooling models [J]. Chinese Journal of Graphical Graphics, 2016, 21(9):1178-1190.
[15] Hochreiter S, Schmidhuber J. Long short-term memory [J]. Neural computation, 1997, 9(8): 1735-1780. 
[16] Chung J, Gulcehre C, Cho K H , et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling [J]. Eprint Arxiv, 2014. 

Downloads: 13684
Visits: 260436

Sponsors, Associates, and Links


All published work is licensed under a Creative Commons Attribution 4.0 International License.

Copyright © 2016 - 2031 Clausius Scientific Press Inc. All Rights Reserved.