3260 papers • 126 benchmarks • 313 datasets
Hateful meme classification aims to detect harmful content within the text or images of memes.
(Image credit: Papersgraph)
These leaderboards are used to track progress in hateful-meme-classification-18
Use these libraries to find hateful-meme-classification-18 models and implementations
No subtasks available.
It is demonstrated that the simple pre-training task of predicting which caption goes with which image is an efficient and scalable way to learn SOTA image representations from scratch on a dataset of 400 million (image, text) pairs collected from the internet.
The Hate-CLIPper architecture is proposed, which explicitly models the cross-modal interactions between the image and text representations obtained using Contrastive Language-Image Pre-training (CLIP) encoders via a feature interaction matrix (FIM).
Hateful meme with Reasons Dataset (HatReD) is introduced, which is a new multimodal hateful meme dataset annotated with the underlying hateful contextual reasons and a new conditional generation task that aims to automatically generate underlying reasons to explain hateful memes.
A novel approach named ISSUES is proposed for multimodal hateful meme classification that leverages a pre-trained CLIP vision-language model and the textual inversion technique to effectively capture the multimodal semantic content of the memes.
This work demonstrates a retrieval-based hateful memes detection system, which is capable of identifying hatefulness based on data unseen in training, and allows developers to update the hateful memes detection system by simply adding new examples without retraining.
This paper proposes an explainable approach to detect harmful memes, inspired by the powerful capacity of Large Language Models on text generation and reasoning, and proposes to fine-tune a small language model as the debate judge for harmfulness inference.
This study introduces a novel dataset PrideMM comprising 5,063 text-embedded images associated with the LGBTQ+ Pride movement, thereby addressing a serious gap in existing resources and proposes a novel framework MemeCLIP for efficient downstream learning while preserving the knowledge of the pre-trained CLIP model.
Pen—a prompt-enhanced network framework based on the prompt learning approach, which surpasses manual prompt methods, showcasing superior generalization and classification accuracy in hateful meme classification tasks.
This work proposes a robust adaptation framework for hateful meme detection that enhances in-domain accuracy and cross-domain generalization while preserving the general vision-language capabilities of LMMs and achieves improved robustness under adversarial attacks compared to SFT models.
Adding a benchmark result helps the community track progress.