3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in selection-bias-4
No benchmarks available.
Use these libraries to find selection-bias-4 models and implementations
No subtasks available.
This work study the performance of segmentation and classification models when applied to the proposed dataset and demonstrate the application of models trained on PanNuke to whole-slide images.
This paper model CVR in a brand-new perspective by making good use of sequential pattern of user actions, i.e., impression -> click -> conversion, which is the first public dataset which contains samples with sequential dependence of click and conversion labels for CVR modeling.
This work develops a universal lower bound for single-node interventions that establishes that the largest clique is always a fundamental impediment to structure learning and presents a two-phase intervention design algorithm that matches the optimal number of interventions up to a multiplicative logarithmic factor in the number of maximal cliques.
A new analytical expression for MDI is derived, and based on this new expression, a debiased MDI feature importance measure is proposed, called MDI-oob, which achieves state-of-the-art performance in feature selection from Random Forests for both deep and shallow trees.
A neural network framework called Adversarial Balancing-based representation learning for Causal Effect Inference (ABCEI), based on recent advances in representation learning, is proposed that is robust against treatment selection bias, and matches/outperforms the state-of-the-art approaches.
This paper investigates the problem of selection bias on six NLSM datasets and finds that four out of them are significantly biased, and proposes a training and evaluation framework to alleviate the bias.
The results show that the choice between the methodologies is consequential and depends on the presence of selection bias, and the degree of position bias and interaction noise, and that counterfactual methods can obtain the highest ranking performance; however, in other circumstances their optimization can be detrimental to the user experience.
The usefulness of the automated dependence plots (ADP) across multiple use-cases and datasets including model selection, bias detection, understanding out-of-sample behavior, and exploring the latent space of a generative model is demonstrated.
A treatment-effect estimator is developed using algorithmic decisions as instruments for a class of stochastic and deterministic algorithms that evaluate the Coronavirus Aid, Relief, and Economic Security Act, which allocated many billions of dollars worth of relief funding to hospitals via an algorithmic rule.
A simple causal mechanism is introduced to describe the role underspecification plays in the generation of spurious correlations and directly informs the development of two lightweight black-box evaluation methods.
Adding a benchmark result helps the community track progress.