英文摘要 |
As the amount of data increases, manually classifying texts is expensive. Therefore, automated text classification has become important, such as spam detection, news classification, and sentiment analysis. Recently, deep learning models in natural language are roughly divided into two categories: sequential and graph based. The sequential models usually use RNN and CNN, as well as the BERT model and its variants; In recent years, researchers started to apply the graph based deep learning model to NLP, using word co-occurrence and TF-IDF weights to build graphs in order to learn the features of words and documents for classification.
In the experiment, we use different datasets, MR, R8, R52 and Ohsumed for verification. Comparing with sequential and graph-based models, the accuracy of our proposed method on MR can achieve 0.79. |