3260 papers • 126 benchmarks • 313 datasets
Bias detection is the task of detecting and measuring racism, sexism and otherwise discriminatory behavior in a model (Source: https://stereoset.mit.edu/)
(Image credit: Papersgraph)
These leaderboards are used to track progress in bias-detection-9
Use these libraries to find bias-detection-9 models and implementations
The usefulness of the automated dependence plots (ADP) across multiple use-cases and datasets including model selection, bias detection, understanding out-of-sample behavior, and exploring the latent space of a generative model is demonstrated.
StereoSet, a large-scale natural English dataset to measure stereotypical biases in four domains: gender, profession, race, and religion, is presented and it is shown that popular models like BERT, GPT-2, RoBERTa, and XLnet exhibit strong stereotypical biases.
B bipol, a novel multi-axes bias metric with explainability, is used to estimate and explain how much bias exists in five English NLP benchmark datasets and two Swedish datasets for bias, along multiple axes.
This paper applies the WEAT bias detection method to four sets of word embeddings trained on corpora from four different domains: news, social networking, biomedical and a gender-balanced corpus extracted from Wikipedia, and finds that some domains are definitely more prone to gender bias than others.
A multilingual method for the extraction of biased sentences from Wikipedia is proposed, and it is used to create corpora in Bulgarian, French and English and exploit the data with well-known classification methods to detect biased sentences.
The analysis shows that the use of acoustic signal helped to improve bias detection by more than 6% absolute over using text and metadata only, and the dataset is released to the research community, hoping to help advance the field of multi-modal political bias detection.
This work performs comprehensive experiments for detecting subjective bias using BERT-based models on the Wiki Neutrality Corpus (WNC), and proposes BERT -based ensembles that outperform state-of-the-art methods like BERTlarge by a margin of 5.6 F1 score.
Adding a benchmark result helps the community track progress.