Publication | Open Access
A Unified Tagging Solution: Bidirectional LSTM Recurrent Neural Network with Word Embedding
74
Citations
30
References
2015
Year
EngineeringMachine LearningTaggingPart-of-speech TaggingUnified Tagging SolutionRecurrent Neural NetworkCorpus LinguisticsText MiningWord EmbeddingsNatural Language ProcessingSpeech RecognitionWord EmbeddingData ScienceComputational LinguisticsEntity RecognitionLanguage StudiesMachine TranslationSequence ModellingNlp TaskDeep LearningLinguisticsChunkingPo Tagging
Bidirectional Long Short-Term Memory Recurrent Neural Network (BLSTM-RNN) has been shown to be very effective for modeling and predicting sequential data, e.g. speech utterances or handwritten documents. In this study, we propose to use BLSTM-RNN for a unified tagging solution that can be applied to various tagging tasks including part-of-speech tagging, chunking and named entity recognition. Instead of exploiting specific features carefully optimized for each task, our solution only uses one set of task-independent features and internal representations learnt from unlabeled text for all tasks.Requiring no task specific knowledge or sophisticated feature engineering, our approach gets nearly state-of-the-art performance in all these three tagging tasks.
| Year | Citations | |
|---|---|---|
1997 | 93.8K | |
2014 | 33.2K | |
2013 | 18.1K | |
1997 | 9.6K | |
2013 | 8.7K | |
1967 | 6.7K | |
2010 | 5.4K | |
2011 | 5.2K | |
2008 | 5.2K | |
2012 | 2K |
Page 1
Page 1