3260 papers • 126 benchmarks • 313 datasets
Identify labeling errors in data
(Image credit: Papersgraph)
These leaderboards are used to track progress in label-error-detection-8
Use these libraries to find label-error-detection-8 models and implementations
No datasets available.
No subtasks available.
An extension of the Confident Learning framework is proposed to this setting, as well as a label quality score that ranks examples with label errors much higher than those which are correctly labeled.
WenetSpeech is the current largest open-source Mandarin speech corpus with transcriptions, which benefits research on production-level speech recognition, and a novel end-to-end label error detection approach is proposed.
This work proposes a novel framework, called CTRL11CTRL (Clustering TRaining Losses for label error detection), to detect label errors in multiclass datasets, and demonstrates state-of-the-art error detection accuracy on both image and tabular datasets under labeling noise.
This work presents for the first time a method for detecting label errors in image datasets with semantic segmentation, i.e., pixel-wise class labels by lifting the consideration of uncertainty to the level of predicted components and enabling the usage of DNNs together with component-level uncertainty quantification for the detection of label errors.
A benchmarking environment AQuA is proposed to rigorously evaluate methods that enable machine learning in the presence of label noise and a design space is introduced to delineate concrete design choices of label error detection models.
Adding a benchmark result helps the community track progress.