3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in analogical-similarity-8
Use these libraries to find analogical-similarity-8 models and implementations
No subtasks available.
This paper presents an analysis of Transformer-based language model performance across a wide range of model scales -- from models with tens of millions of parameters up to a 280 billion parameter model called Gopher.
An interpretable, scalable algorithm is developed that can extract analogies from a large dataset of procedural texts, achieving 79% precision, and it is demonstrated that the algorithm is robust to paraphrasing the input texts.
This paper shows that the conventional winning ticket is hard to find at weight level of ViTs by existing methods, and generalizes the LTH for ViTs to input data consisting of image patches inspired by the input dependence ofViTs, which verifies the integrity of the theory.
This work trains a predicted compute-optimal model, Chinchilla, that uses the same compute budget as Gopher but with 70B parameters and 4$\times$ more more data, and reaches a state-of-the-art average accuracy, greater than a 7% improvement over Gopher.
Adding a benchmark result helps the community track progress.