3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in explanation-generation
Use these libraries to find explanation-generation models and implementations
No subtasks available.
The results indicate that, by training on in-domain data, gains can be made in explainable, automated fact-checking for claims which require specific expertise.
AR-BERT is proposed, a novel two-level global-local entity embedding scheme that allows efficient joint training of KG-based aspect embeddings and ALSC models and introduces the problem of determining mode significance in multi-modal explanation generation, and proposes a two step solution.
The TextGraphs-13 Shared Task on Explanation Regeneration asked participants to develop methods to reconstruct gold explanations for elementary science questions, and Red Dragon AI’s entries used the language of the questions and explanation text directly, rather than a constructing a separate graph-like representation.
This work introduces interpretable autoprompting (iPrompt), an algorithm that generates a natural-language string explaining the data that iteratively alternates between generating explanations with an LLM and reranking them based on their performance when used as a prompt.
A large user study is described showing that the presence of QED explanations significantly improves the ability of untrained raters to spot errors made by a strong neural QA baseline.
TE2Rules (Tree Ensemble to Rules), a novel approach for explaining binary classification tree ensemble models through a list of rules, particularly focusing on explaining the minority class, is introduced.
This paper develops a general framework for interpretable natural language understanding that requires only a small set of human annotated explanations for training and proposes an explanation-based self-training method under this framework for semi-supervised learning.
A leakage-adjusted simulatability (LAS) metric is introduced for evaluating NL explanations, which measures how well explanations help an observer predict a model’s output, while controlling for how explanations can directly leak the output.
This work introduces Summarize and Score (SASC), a method that takes in a text module and returns a natural language explanation of the module's selectivity along with a score for how reliable the explanation is, and shows that SASC can generate explanations for the response of individual fMRI voxels to language stimuli.
This work introduces a new annotated dataset of 1.3K instances of elaborative simplification and analyzes how entities, ideas, and concepts are elaborated through the lens of contextual specificity, and establishes baselines for elaboration generation using large scale pre-trained language models.
Adding a benchmark result helps the community track progress.