3260 papers • 126 benchmarks • 313 datasets
In natural language processing, open information extraction is the task of generating a structured, machine-readable representation of the information in text, usually in the form of triples or n-ary propositions (Source: Wikipedia).
(Image credit: Papersgraph)
These leaderboards are used to track progress in open-information-extraction-8
Use these libraries to find open-information-extraction-8 models and implementations
It is found that most of the facts between entities present in OPIEC cannot be found in DBpedia and/or YAGO, that OIE facts often differ in the level of specificity compared to knowledge base facts, and that Oie open relations are generally highly polysemous.
This work develops a methodology that leverages the recent QA-SRL annotation to create a first independent and large scale Open IE annotation and uses it to automatically compare the most prominent Open IE systems.
CMVC is proposed, a novel unsupervised framework that leverages two views of knowledge jointly for canonicalizing OKBs without the need of manually annotated labels and demonstrates the superiority of the framework through extensive experiments on multiple real-world OKB data sets against state-of-the-art methods.
An effective multi-stage tuning framework called MT4CrossIE, designed for enhancing cross-lingual open information extraction by injecting language-specific knowledge into the shared model, is proposed.
The design and use of the Stanford CoreNLP toolkit is described, an extensible pipeline that provides core natural language analysis, and it is suggested that this follows from a simple, approachable design, straightforward interfaces, the inclusion of robust and good quality analysis components, and not requiring use of a large amount of associated baggage.
This paper proposes Schema Induction using Coupled Tensor Factorization (SICTF), a novel tensor factorization method for relation schema induction that factorizes Open Information Extraction triples extracted from a domain corpus along with additional side information in a principled way to induce relation schemas.
This work investigates knowledge-guided linguistic rewrites as a secondary source of evidence and finds that they can vastly improve the quality of inference rule corpora, obtaining 27 to 33 point precision improvement while retaining substantial recall.
This paper presents a straightforward approach for adapting PropS, a rule-based predicate-argument analysis for English, to a new language, German, and obtains an Open IE system for German covering 89% of the English rule set.
Adding a benchmark result helps the community track progress.