3260 papers • 126 benchmarks • 313 datasets
Humor detection is the task of identifying comical or amusing elements.
(Image credit: Papersgraph)
These leaderboards are used to track progress in humor-detection-3
Use these libraries to find humor-detection-3 models and implementations
No subtasks available.
This paper proposes a novel sparsity-aware algorithm for sparse data and weighted quantile sketch for approximate tree learning and provides insights on cache access patterns, data compression and sharding to build a scalable tree boosting system called XGBoost.
The evaluation performed on two contrasting settings confirm the strength and robustness of the model and suggests two important factors in achieving high accuracy in the current task: usage of sentence embeddings and utilizing the linguistic structure of humor in designing the proposed model.
XLNet is proposed, a generalized autoregressive pretraining method that enables learning bidirectional contexts by maximizing the expected likelihood over all permutations of the factorization order and overcomes the limitations of BERT thanks to its autore progressive formulation.
A novel framework, MISA, is proposed, which projects each modality to two distinct subspaces, which provide a holistic view of the multimodal data, which is used for fusion that leads to task predictions.
This paper builds a model that learns to identify humorous jokes based on ratings gleaned from Reddit pages, consisting of almost 16,000 labeled instances, and employs a Transformer architecture for its advantages in learning from sentence context.
A corpus of 27,000 tweets written in Spanish and crowd-annotated by their humor value and funniness score, with about four annotations per tweet, tagged by 1,300 people over the Internet is presented.
This study proposes a distant supervised learning approach where the training sentences are automatically annotated based on the emojis they have, and experimentally shows that training classifier on cheap, large and possibly erroneous data annotated using this approach leads to more accurate results compared with training the same classifiers on the more expensive, much smaller and error-free manually annotated training data.
Overall, this paper deepens the understanding of the syntactic and semantic structure of satirical news headlines and provides insights for building humor-producing systems.
Adding a benchmark result helps the community track progress.