3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in gender-bias-detection-17
No benchmarks available.
Use these libraries to find gender-bias-detection-17 models and implementations
No subtasks available.
This paper asks whether LLMs hold wide-reaching biases of positive or negative sentiment for specific social groups similar to the "what is beautiful is good" bias found in people in experimental psychology, and investigates bias along less-studied but still consequential, dimensions, such as age and beauty.
This paper applies the WEAT bias detection method to four sets of word embeddings trained on corpora from four different domains: news, social networking, biomedical and a gender-balanced corpus extracted from Wikipedia, and finds that some domains are definitely more prone to gender bias than others.
This work proposes a matching approach that selects a subset of images from the full dataset with balanced attribute distributions across protected attributes, and demonstrates its work in the context of gender bias in multiple open-source facial-recognition classifiers and finds that bias persists after removing key confounders via matching.
The results show that the bias of models increase as datasets become more imbalanced or datasets attributes become more correlated, the level of dominance of correlated sensitive dataset features impact bias, and the sensitive information remains in the latent representation even when bias-mitigation algorithms are applied.
It is found that, contrary to previous research, coverage and sentiment biases suggest equal public interest in female politicians, however, the results of the nominal and lexical analyses suggest this interest is not as professional or respectful as that expressed about male politicians.
In this paper, a simple logistic regression model using manually selected volumetric features and a convolutional neural network trained on 3D MRI data are compared, finding that CNN performance is generally improved for both male and female subjects when including more female subjects in the training dataset.
This paper proposes FairShap, a novel instance-level data re-weighting method for fair algorithmic decision-making through data valuation by means of Shapley Values, which is model-agnostic and easily interpretable.
An experimental framework developed to automatically detect gender biases in court decisions issued in Brazilian Portuguese is presented and features identified to be critical in such a technology are described to be critical in such a technology, given its proposed use as a support tool for research and assessment of court activity.
Adding a benchmark result helps the community track progress.