PATTY: A Large Resource of Relational Patterns

PATTY is a collection of semantically-typed relational patterns mined from large corpora. The patterns are organised into synonyms and subsumptions.

The taxonomy derived from Wikipedia contains over 350,000 pattern synsets.


HARPY: Hypernyms and Alignment of Relational Paraphrases

Collections of relational paraphrases have been automatically constructed from large text corpora, as a WordNet counterpart for the realm of binary predicates and their surface forms. However, these resources fall short in their coverage of hypernymy links (subsumptions) among the synsets of phrases. HARPY closes this gap by computing a high-quality alignment between the relational phrases of the Patty taxonomy, one of the largest collections of this kind, and the verb senses of WordNet. HARPY taxonomy of relational phrases and verb senses, contains 20,812 synsets organized into a DAG with 616,792 hypernymy links. 

RELLY: Inferring Hypernym Relationships Between Relational Phrases

Relational phrases (e.g., ``got married to'') and their hypernyms (e.g., ``is a relative of'') are central for many tasks including question answering, open information extraction, paraphrasing, and entailment detection. This has motivated the development of several linguistic resources (e.g. DIRT, PATTY, and WiseNet) which systematically collect and organize relational phrases. These resources have demonstrable practical benefits, but are each limited due to noise, sparsity, or size. We present a new general-purpose method, RELLY, for constructing a large hypernymy graph of relational phrases with high-quality subsumptions using collective probabilistic programming techniques. Our graph induction approach integrates small high-precision knowledge bases together with large automatically curated resources, and reasons collectively to combine these resources into a consistent graph. Using RELLY, we construct a high-coverage, high-precision hypernymy graph consisting of 20K relational phrases and 35K hypernymy links. Our evaluation indicates a hypernymy link precision of 78%, and demonstrates the value of this resource for a document-relevance ranking task.

POLY: Mining Relational Paraphrases from Multilingual Sentence

Language resources that systematically organize paraphrases for binary relations are of great value for various NLP tasks and have recently been advanced in projects like PATTY, WiseNet and DEFIE. This paper presents a new method for building such a resource and the resource itself, called POLY. Starting with a very large collection of multilingual sentences parsed into triples of phrases, our method clusters relational phrases using probabilistic measures. We judiciously leverage fine-grained semantic typing of relational arguments for identifying synonymous phrases. The evaluation of POLY shows significant improvements in precision and recall over the prior works on PATTY and DEFIE. An extrinsic use case demonstrates the benefits of POLY for question answering.




This data accompanies the publication Nakashole et. al EMNLP2012.
Please cite this publication if you use the data.



This data accompanies the publication Grycner et. al COLING2014.
Please cite this publication if you use the data.


This data accompanies the publication Grycner et. al EMNLP 2015. 
Please cite this publication if you use the data.


This data accompanies the publication Grycner et. al EMNLP 2016. 
Please cite this publication if you use the data.