3260 papers • 126 benchmarks • 313 datasets
Given a passage, a question, and an answer phrase, the goal of distractor generation (DG) is to generate context-related wrong options (i.e., distractor) for multiple-choice questions (MCQ).
(Image credit: Papersgraph)
These leaderboards are used to track progress in distractor-generation-6
Use these libraries to find distractor-generation-6 models and implementations
No subtasks available.
A hierarchical encoder-decoder framework with static and dynamic attention mechanisms to tackle the task of distractor generation for multiple choice reading comprehension questions from examinations and generates longer and semantic-rich distractors which are closer to distractors in real reading comprehension from examinations.
This work investigates how machine learning models, specifically ranking models, can be used to select useful distractors for multiple choice questions and empirically studies feature-based and neural net based ranking models with experiments on the recently released SciQ dataset and the MCQL dataset.
A new distractor generation scheme with multi-tasking and negative answer training strategies for effectively generating multiple distractors that shows strong distracting power for multiple choice question.
This work proposes a series of novel techniques for applying large pre-trained Transformer encoder-decoder models, namely PEGASUS and T5, to the tasks of question-answer generation and distractor generation, and shows that these models outperform strong baselines using both automated metrics and human raters.
This work proposes an unsupervised cross-lingual language generation framework (called ZmBART) that does not use any parallel or pseudo-parallel/back-translated data and is fine-tuned with task-specific supervised English data and directly evaluated with low-resource languages in the Zero-shot setting.
This paper presents a new BERTbased method for automatically generating distractors using only a small-scale dataset, releases a new such dataset of Swedish MCQs (used for training the model), and proposes a methodology for assessing the generated distractors.
Adding a benchmark result helps the community track progress.