3260 papers • 126 benchmarks • 313 datasets
Develop a classifier that could make a 3-way classification in-between ‘Overtly Aggressive’, ‘Covertly Aggressive’ and ‘Non-aggressive’ text data. For this, TRAC-2 dataset of 5,000 aggression-annotated data from social media each in Bangla (in both Roman and Bangla script), Hindi (in both Roman and Devanagari script) and English for training and validation is to be used.
(Image credit: Papersgraph)
These leaderboards are used to track progress in aggression-identification-5
No benchmarks available.
Use these libraries to find aggression-identification-5 models and implementations
No subtasks available.
The main features employed by the method are information extracted from word embeddings and the output of a sentiment analyser and it shows that despite its simplicity the method performs well when compared with more elaborated methods.
This system description paper presents a proposal to augment the provided dataset to increase the number of labeled comments from 15,000 to 60,000, and introduces linguistic variety into the dataset, to train a special deep neural net, which generalizes especially well to unseen data.
The superior performance of the SVM classifier was achieved mainly because of its better prediction of the majority class than the BERT based classifiers, which were found to predict the minority classes better.
An ensemble of multiple fine-tuned BERT models based on bootstrap aggregating (bagging) is proposed and presented, finding that the F1-score drastically increases when ensembling up to 15 models, but the returns diminish for more models.
It is found that only a few annotations of most controversial documents are enough for all the personalization methods to significantly outperform classic, generalized solutions.
An end-to-end neural model using attention on top of BERT that incorporates a multi-task learning paradigm to address both the sub-tasks simultaneously is proposed.
Adding a benchmark result helps the community track progress.