3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in discourse-segmentation-10
No benchmarks available.
Use these libraries to find discourse-segmentation-10 models and implementations
No subtasks available.
It is shown that both the pairing and global features are useful on their own, and their combination achieved an F1 of 92.6% of identifying insentence discourse boundaries, which is a 17.8% error-rate reduction over the state of the art performance.
This paper proposes statistical discourse segmenters for five languages and three domains that do not rely on gold pre-annotations, and considers the problem of learning discourse segmenter when no labeled data is available for a language.
The results show that dependency information is less useful than expected, but the model provided is a fully scalable, robust model that only relies on part-of-speech information, and it performs well across languages in the absence of any gold-standard annotation.
This paper proposes an end-to-end neural segmenter based on BiLSTM-CRF framework that addresses the problem of data insufficiency by transferring a word representation model that is trained on a large corpus and proposes a restricted self-attention mechanism in order to capture useful information within a neighborhood.
This work annotates the first high-quality small-scale medical corpus in English with discourse segments and analyzes how well news-trained segmenters perform on this domain.
This work proposes three transformer-based architectures and provides comprehensive comparisons with previously proposed approaches on three standard datasets, and establishes a new state-of-the-art, reducing in particular the error rates by a large margin in all cases.
An accurate framework for carrying out multi-lingual discourse segmentation with BERT is described, trained to identify segments by casting the problem as a token classification problem and jointly learning syntactic features like part-of-speech tags and dependency relations.
This work proposes Discern, a discourse-aware entailment reasoning network to strengthen the connection and enhance the understanding for both document and dialog, and splits the document into clause-like elementary discourse units using a pre-trained discourse segmentation model.
MUDERN is proposed, a Multi-passage Discourse-aware Entailment Reasoning Network which extracts conditions in the rule texts through discourse segmentation, conducts multi-passages entailment reasoning to answer user questions directly, or asks clarification follow-up questions to inquiry more information.
Adding a benchmark result helps the community track progress.