3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in learning-word-embeddings-13
No benchmarks available.
Use these libraries to find learning-word-embeddings-13 models and implementations
No datasets available.
No subtasks available.
The proposed Speech2Vec model, a novel deep neural network architecture for learning fixed-length vector representations of audio segments excised from a speech corpus, is based on a RNN Encoder-Decoder framework, and borrows the methodology of skipgrams or continuous bag-of-words for training.
INSTRUCTOR is a single embedder that can generate text embeddings tailored to different downstream tasks and domains, without any further training, and achieves state-of-the-art performance.
This paper argues that word embedding can be naturally viewed as a ranking problem due to the ranking nature of the evaluation metrics, and proposes a novel framework WordRank that efficiently estimates word representations via robust ranking, in which the attention mechanism and robustness to noise are readily achieved via the DCG-like ranking losses.
The results demonstrate the potential of hyperbolic word embeddings, particularly in low dimensions, though without clear superiority over their Euclidean counterparts, and discuss problems in the formulation of the analogy task resulting from the curvature ofhyperbolic space.
GCNs are utilized for Document Timestamping problem and for learning word embeddings using dependency context of a word instead of sequential context and two limitations of existing GCN models are addressed.
Experimental results on four translation datasets of different languages show that MorphTE can compress word embedding parameters by about 20 times without performance loss and significantly outperforms related embedding compression methods.
This work proposes a new learning objective that incorporates both a neural language model objective (Mikolov et al., 2013) and prior knowledge from semantic resources to learn improved lexical semantic embeddings.
A comprehensive study of how the quality of embeddings changes according to hyper-parameter settings is presented, and it is observed that bigger corpora do not necessarily produce better biomedical domain word embedDings.
Dict2vec builds new word pairs from dictionary entries so that semantically-related words are moved closer, and negative sampling filters out pairs whose words are unrelated in dictionaries.
This work shows that with a step size, the Mixing method converges to the global optimum of the semidefinite program almost surely in a locally linear rate under random initialization, and is the first low-rank semideFinite programming method to achieve a global optimum on the spherical manifold without assumption.
Adding a benchmark result helps the community track progress.