z-logo
open-access-imgOpen Access
Grounded Compositional Semantics for Finding and Describing Images with Sentences
Author(s) -
Richard Socher,
Andrej Karpathy,
Quoc V. Le,
Christopher D. Manning,
Andrew Y. Ng
Publication year - 2014
Publication title -
transactions of the association for computational linguistics
Language(s) - English
Resource type - Journals
ISSN - 2307-387X
DOI - 10.1162/tacl_a_00177
Subject(s) - computer science , sentence , recurrent neural network , artificial intelligence , natural language processing , dependency (uml) , word (group theory) , feature (linguistics) , feature vector , semantics (computer science) , focus (optics) , word order , artificial neural network , linguistics , philosophy , physics , optics , programming language
Previous work on Recursive Neural Networks (RNNs) shows that these models can produce compositional feature vectors for accurately representing and classifying sentences or images. However, the sentence vectors of previous models cannot accurately represent visually grounded meaning. We introduce the DT-RNN model which uses dependency trees to embed sentences into a vector space in order to retrieve images that are described by those sentences. Unlike previous RNN-based models which use constituency trees, DT-RNNs naturally focus on the action and agents in a sentence. They are better able to abstract from the details of word order and syntactic expression. DT-RNNs outperform other recursive and recurrent neural networks, kernelized CCA and a bag-of-words baseline on the tasks of finding an image that fits a sentence description and vice versa. They also give more similar representations to sentences that describe the same image.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom