Skip to main navigation Skip to search Skip to main content

Learning sense-specificword embeddings by exploiting bilingual resources

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Recent work has shown success in learning word embeddings with neural network language models (NNLM). However, the majority of previous NNLMs represent each word with a single embedding, which fails to capture polysemy. In this paper, we address this problem by representing words with multiple and sense-specific embeddings, which are learned from bilingual parallel data. We evaluate our embeddings using the word similarity measurement and show that our approach is significantly better in capturing the sense-level word similarities. We further feed our embeddings as features in Chinese named entity recognition and obtain noticeable improvements against single embeddings.

Original languageEnglish
Title of host publicationCOLING 2014 - 25th International Conference on Computational Linguistics, Proceedings of COLING 2014
Subtitle of host publicationTechnical Papers
PublisherAssociation for Computational Linguistics, ACL Anthology
Pages497-507
Number of pages11
ISBN (Electronic)9781941643266
StatePublished - 2014
Event25th International Conference on Computational Linguistics, COLING 2014 - Dublin, Ireland
Duration: 23 Aug 201429 Aug 2014

Publication series

NameCOLING 2014 - 25th International Conference on Computational Linguistics, Proceedings of COLING 2014: Technical Papers

Conference

Conference25th International Conference on Computational Linguistics, COLING 2014
Country/TerritoryIreland
CityDublin
Period23/08/1429/08/14

Fingerprint

Dive into the research topics of 'Learning sense-specificword embeddings by exploiting bilingual resources'. Together they form a unique fingerprint.

Cite this