Regularized and Retrofitted models for Learning Sentence Representation with Context
Author(s) -
Tanay Kumar Saha,
Shafiq Joty,
Naeemul Hassan,
Mohammad Al Hasan
Publication year - 2017
Publication title -
iuscholarworks (indiana university)
Language(s) - English
Resource type - Conference proceedings
DOI - 10.1145/3132847.3133011
Subject(s) - computer science , sentence , artificial intelligence , representation (politics) , natural language processing , cluster analysis , context (archaeology) , ranking (information retrieval) , word (group theory) , scope (computer science) , context model , machine learning , mathematics , object (grammar) , paleontology , geometry , politics , political science , law , biology , programming language
Vector representation of sentences is important for many text processing tasks that involve classifying, clustering, or ranking sentences. For solving these tasks, bag-of-word based representation has been used for a long time. In recent years, distributed representation of sentences learned by neural models from unlabeled data has been shown to outperform traditional bag-of-words representations. However, most existing methods belonging to the neural models consider only the content of a sentence, and disregard its relations with other sentences in the context. In this paper, we first characterize two types of contexts depending on their scope and utility. We then propose two approaches to incorporate contextual information into content-based models. We evaluate our sentence representation models in a setup, where context is available to infer sentence vectors. Experimental results demonstrate that our proposed models outshine existing models on three fundamental tasks, such as, classifying, clustering, and ranking sentences.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom