Learning Composition Models for Phrase Embeddings
Author(s) -
Mo Yu,
Mark Dredze
Publication year - 2015
Publication title -
transactions of the association for computational linguistics
Language(s) - English
Resource type - Journals
ISSN - 2307-387X
DOI - 10.1162/tacl_a_00135
Subject(s) - phrase , computer science , natural language processing , artificial intelligence , word (group theory) , variety (cybernetics) , context (archaeology) , task (project management) , construct (python library) , similarity (geometry) , language model , phrase structure rules , linguistics , generative grammar , paleontology , philosophy , management , economics , image (mathematics) , biology , programming language
Lexical embeddings can serve as useful representations for words for a variety of NLP tasks, but learning embeddings for phrases can be challenging. While separate embeddings are learned for each word, this is infeasible for every phrase. We construct phrase embeddings by learning how to compose word embeddings using features that capture phrase structure and context. We propose efficient unsupervised and task-specific learning objectives that scale our model to large datasets. We demonstrate improvements on both language modeling and several phrase semantic similarity tasks with various phrase lengths. We make the implementation of our model and the datasets available for general use.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom