3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in sentence-pair-classification-3
No benchmarks available.
Use these libraries to find sentence-pair-classification-3 models and implementations
No subtasks available.
DACCORD is introduced, an original dataset in French for automatic detection of contradictions between sentences and new, manually translated versions of two datasets, namely the well known dataset RTE3 and the recent dataset GQNLI, from English to French, for the task of natural language inference / recognising textual entailment.
This paper constructs an auxiliary sentence from the aspect and converts ABSA to a sentence-pair classification task, such as question answering (QA) and natural language inference (NLI), and fine-tune the pre-trained model from BERT.
The first large-scale Chinese Language Understanding Evaluation (CLUE) benchmark is introduced, an open-ended, community-driven project that brings together 9 tasks spanning several well-established single-sentence/sentence-pair classification tasks, as well as machine reading comprehension, all on original Chinese text.
Glyce, the glyph-vectors for Chinese character representations, is presented and it is shown that glyph-based models are able to consistently outperform word/char ID- based models in a wide range of Chinese NLP tasks.
The first Chinese Biomedical Language Understanding Evaluation Evaluation (CBLUE) benchmark is presented: a collection of natural language understanding tasks including named entity recognition, information extraction, clinical diagnosis normalization, single-sentence/sentence-pair classification, and an associated online platform for model evaluation, comparison, and analysis.
This work introduces a novel continual architecture search (CAS) approach, so as to continually evolve the model parameters during the sequential training of several tasks, without losing performance on previously learned tasks, thus enabling life-long learning.
This work develops a contrastive self-training framework, COSINE, to enable fine-tuning LMs with weak supervision, underpinned by contrastive regularization and confidence-based reweighting, which gradually improves model fitting while effectively suppressing error propagation.
It is shown that elastic weight consolidation (EWC) allows fine-tuning of models to mitigate biases while being less susceptible to catastrophic forgetting, yielding models with lower levels of forgetting on the original dataset for equivalent gains in accuracy on the fine- Tuning (unbiased) dataset.
This paper finds that it is possible to hack the model in a data-free way by modifying one single word embedding vector, with almost no accuracy sacrificed on clean samples.
Adding a benchmark result helps the community track progress.