3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in word-translation-11
No benchmarks available.
Use these libraries to find word-translation-11 models and implementations
No subtasks available.
It is shown that a bilingual dictionary can be built between two languages without using any parallel corpora, by aligning monolingual word embedding spaces in an unsupervised way.
A novel method is presented that first aligns the second moment of the word distributions of the two languages and then iteratively refines the alignment, which achieves better performance than recent state-of-the-art deep adversarial approaches and is competitive with the supervised baseline.
This paper proposes an unified formulation that directly optimizes a retrieval criterion in an end-to-end fashion for word translation, and shows that this approach outperforms the state of the art on word translation.
This work proposes a fully unsupervised framework for learning MWEs that directly exploits the relations between all language pairs and substantially outperforms previous approaches in the experiments on multilingual word translation and cross-lingual word similarity.
An extension of Structural Correspondence Learning (SCL), a recently proposed algorithm for domain adaptation, is described for cross-lingual adaptation in the context of text classification, showing a significant improvement of the proposed method over a machine translation baseline.
This work proposes a novel geometric approach for learning bilingual mappings given monolingual embeddings and a bilingual dictionary that outperforms previous approaches on the bilingual lexicon induction and cross-lingual word similarity tasks.
This work proposes a bilingual extension of the CBOW method which leverages sentence-aligned corpora to obtain robust cross-lingual word and sentence representations and significantly improves cross-lingsual sentence retrieval performance over all other approaches while maintaining parity with the current state-of-the-art methods on word-translation.
LexC-Gen serves as a potential solution to close the performance gap between open-source multilingual models, such as BLOOMZ and Aya-101, and state-of-the-art commercial models like GPT-4o on low-resource-language tasks.
An unsupervised and a very resource-light approach for measuring semantic similarity between texts in different languages, applicable to virtually any pair of languages for which there exists a sufficiently large corpus, required to learn monolingual word embeddings.
This work shows that a framework using representation learning, bilingual dictionary induction and statistical machine translation yields the best precision at 10 of 0.827 on professional-to-consumer word translation, and mean opinion scores of 4.10 and 4.28 out of 5 for clinical correctness and layperson readability, respectively, on sentence translation.
Adding a benchmark result helps the community track progress.