Publication | Closed Access
Post-Processing of Word Representations via Variance Normalization and Dynamic Embedding
16
Citations
18
References
2019
Year
Unknown Venue
Structured PredictionEngineeringMachine LearningCross-lingual RepresentationNeurolinguisticsPsycholinguisticsLanguage ProcessingText MiningWord EmbeddingsNatural Language ProcessingSpeech RecognitionData ScienceComputational LinguisticsLanguage StudiesMachine TranslationSequence ModellingNlp TaskVariance NormalizationDistributional SemanticsEmbedded Vector RepresentationsDynamic EmbeddingLinguistics
Language processing becomes more and more important in multimedia processing. Although embedded vector representations of words offer impressive performance on many natural language processing (NLP) applications, the information of ordered input sequences is lost to some extent if only context-based samples are used in the training. For further performance improvement, two new post-processing techniques, called post-processing via variance normalization (PVN) and post-processing via dynamic embedding (PDE), are proposed in this work. The PVN method normalizes the variance of principal components of word vectors, while the PDE method learns orthogonal latent variables from ordered input sequences. The PVN and the PDE methods can be integrated to achieve better performance. We apply these post-processing techniques to several popular word embedding methods to yield their post-processed representations. Extensive experiments are conducted to demonstrate the effectiveness of the proposed post-processing techniques.
| Year | Citations | |
|---|---|---|
Page 1
Page 1