Skip to main navigation Skip to search Skip to main content

Decoupled word embeddings using latent topics

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

In this paper, we propose decoupled word embeddings (DWE) as a universal word representation that covers multiple senses of words. Toward this goal, our model represents each word as a combination of multiple word vectors that are associated with latent topics. Specifically, we decompose a word vector into multiple word vectors for multiple senses, according to the topic weight obtained from pre-trained topic models. Although this dynamic word representation is simple, the proposed model can leverage both local and global contexts. Through extensive experiments, including qualitative and quantitative analyses, we demonstrate that the proposed model is comparable to or better than state-of-the-art word embedding models. The code is publicly available at https://github.com/righ120/DWE.

Original languageEnglish
Title of host publication35th Annual ACM Symposium on Applied Computing, SAC 2020
PublisherAssociation for Computing Machinery
Pages875-882
Number of pages8
ISBN (Electronic)9781450368667
DOIs
StatePublished - 30 Mar 2020
Event35th Annual ACM Symposium on Applied Computing, SAC 2020 - Brno, Czech Republic
Duration: 30 Mar 20203 Apr 2020

Publication series

NameProceedings of the ACM Symposium on Applied Computing

Conference

Conference35th Annual ACM Symposium on Applied Computing, SAC 2020
Country/TerritoryCzech Republic
CityBrno
Period30/03/203/04/20

Keywords

  • Contextualized word embedding
  • Multi-sense word embedding
  • Topic modeling

Fingerprint

Dive into the research topics of 'Decoupled word embeddings using latent topics'. Together they form a unique fingerprint.

Cite this