z-logo
open-access-imgOpen Access
Learning Typed Entailment Graphs with Global Soft Constraints
Author(s) -
Mohammad Javad Hosseini,
Nathanael Chambers,
Siva Reddy,
Xavier Holt,
Shay B. Cohen,
Mark Johnson,
Mark Steedman
Publication year - 2018
Publication title -
transactions of the association for computational linguistics
Language(s) - English
Resource type - Journals
ISSN - 2307-387X
DOI - 10.1162/tacl_a_00250
Subject(s) - logical consequence , textual entailment , computer science , paraphrase , artificial intelligence , predicate (mathematical logic) , natural language processing , scalability , similarity (geometry) , theoretical computer science , programming language , image (mathematics) , database
This paper presents a new method for learning typed entailment graphs from text. We extract predicate-argument structures from multiple-source news corpora, and compute local distributional similarity scores to learn entailments between predicates with typed arguments (e.g., {em person} contracted {em disease}). Previous work has used transitivity constraints to improve local decisions, but these constraints are intractable on large graphs. We instead propose a scalable method that learns globally consistent similarity scores based on new soft constraints that consider both the structures across typed entailment graphs and inside each graph. Learning takes only a few hours to run over $100$K predicates and our results show large improvements over local similarity scores on two entailment datasets. We further show improvements over paraphrases and entailments from the Paraphrase Database, and prior state-of-the-art entailment graphs. We show that the entailment graphs improve performance in a downstream task.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom