3260 papers • 126 benchmarks • 313 datasets
The main objective Semantic Similarity is to measure the distance between the semantic meanings of a pair of words, phrases, sentences, or documents. For example, the word “car” is more similar to “bus” than it is to “cat”. The two main approaches to measuring Semantic Similarity are knowledge-based approaches and corpus-based, distributional methods. Source: Visual and Semantic Knowledge Transfer for Large Scale Semi-supervised Object Detection
(Image credit: Papersgraph)
These leaderboards are used to track progress in semantic-similarity-8
Use these libraries to find semantic-similarity-8 models and implementations
Sentence-BERT (SBERT), a modification of the pretrained BERT network that use siamese and triplet network structures to derive semantically meaningful sentence embeddings that can be compared using cosine-similarity is presented.
The Tree-LSTM is introduced, a generalization of LSTMs to tree-structured network topologies that outperform all existing systems and strong LSTM baselines on two tasks: predicting the semantic relatedness of two sentences and sentiment classification.
Experimental results show that ERNIE outperforms other baseline methods, achieving new state-of-the-art results on five Chinese natural language processing tasks including natural language inference, semantic similarity, named entity recognition, sentiment analysis and question answering.
It is shown that introducing a pre-trained multilingual language model dramatically reduces the amount of parallel training data required to achieve good performance by 80%, and a model that achieves 83.7% bi-text retrieval accuracy over 112 languages on Tatoeba is released.
The proposed method follows an edge-based approach using a lexical database and gives highest correlation value for both word and sentence similarity outperforming other similar models.
This paper elaborates the efforts to assemble a resource for STS in the medical domain, MedSTS, which consists of a total of 174,629 sentence pairs gathered from a clinical corpus at Mayo Clinic, and analyzed the medical concepts in the Med STS corpus.
The Biomedical Language Understanding Evaluation (BLUE) benchmark is introduced to facilitate research in the development of pre-training language representations in the biomedicine domain and it is found that the BERT model pre-trained on PubMed abstracts and MIMIC-III clinical notes achieves the best results.
A global objective is formulated for learning the embeddings from text corpora and knowledge bases, which adopts a novel margin-based loss that is robust to noisy labels and faithfully models type correlation derived from knowledge bases.
Evaluating different word embedding models trained on a large Portuguese corpus, including both Brazilian and European variants, suggests that word analogies are not appropriate forword embedding evaluation; task-specific evaluations appear to be a better option.
Adding a benchmark result helps the community track progress.