JISE


  [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12]


Journal of Information Science and Engineering, Vol. 38 No. 2, pp. 413-427


Revisiting Supervised Word Embeddings


DIEU VU1, KHANG TRUONG, KHANH NGUYEN,
NGO VAN LINH AND KHOAT THAN+
School of Information and Communication Technology
Hanoi University of Science and Technology
Hanoi 100000, Vietnam
+E-mail: khoattq@soict.hust.edu.vn

1Faculty of Electrical and Electronic Engineering
Phenikaa University
Hanoi 12116, Vietnam


Word embeddings are playing a crucial role in a variety of applications. However, most previous works focus on word embeddings which are either non-discriminative or hardly interpretable. In this work, we investigate a novel approach, referred to as SWET, which learns supervised word embeddings using topic models from labeled corpora. SWET inherits the interpretability of topic models, the discriminativeness of supervised inference from labels. More importantly, SWET enables us to directly exploit a large class of existing unsupervised and supervised topic models to learn supervised word embeddings. Extensive experiments show that SWET outperforms unsupervised approaches by a large margin, and are highly competitive with supervised baselines.


Keywords: supervised word embeddings, topic models, supervised learning, supervised topic models, word vectors

  Retrieve PDF document (JISE_202202_08.pdf)