Loading…
NIPS 2013 has ended
Sunday, December 8 • 2:00pm - 6:00pm
Learning word embeddings efficiently with noise-contrastive estimation

Sign up or log in to save this to your schedule, view media, leave feedback and see who's attending!

Continuous-valued word embeddings learned by neural language models have recently been shown to capture semantic and syntactic information about words very well, setting performance records on several word similarity tasks. The best results are obtained by learning high-dimensional embeddings from very large quantities of data, which makes scalability of the training method a critical factor. We propose a simple and scalable new approach to learning word embeddings based on training log-bilinear models with noise-contrastive estimation. Our approach is simpler, faster, and produces better results than the current state-of-the art method of Mikolov et al. (2013a). We achieve results comparable to the best ones reported, which were obtained on a cluster, using four times less data and more than an order of magnitude less computing time. We also investigate several model types and find that the embeddings learned by the simpler models perform at least as well as those learned by the more complex ones.
None


Sunday December 8, 2013 2:00pm - 6:00pm PST
Harrah's Special Events Center, 2nd Floor
  Posters
  • posterid Sun30
  • location Poster# Sun30

Attendees (0)