Publication | Open Access
Incorporating Context-Relevant Knowledge into Convolutional Neural Networks for Short Text Classification
20
Citations
5
References
2019
Year
Convolutional Neural NetworkEngineeringMachine LearningNeural NetworkLanguage ProcessingText MiningWord EmbeddingsNatural Language ProcessingText Classification MethodsData ScienceComputational LinguisticsDocument ClassificationLanguage StudiesLanguage ModelsAutomatic ClassificationNlp TaskDeep LearningContext-relevant KnowledgeShort Text ClassificationConvolutional Neural NetworksText ProcessingLinguistics
Some text classification methods don’t work well on short texts due to the data sparsity. What’s more, they don’t fully exploit context-relevant knowledge. In order to tackle these problems, we propose a neural network to incorporate context-relevant knowledge into a convolutional neural network for short text classification. Our model consists of two modules. The first module utilizes two layers to extract concept and context features respectively and then employs an attention layer to extract those context-relevant concepts. The second module utilizes a convolutional neural network to extract high-level features from the word and the contextrelevant concept features. The experimental results on three datasets show that our proposed model outperforms the stateof-the-art models.
| Year | Citations | |
|---|---|---|
Page 1
Page 1