3260 papers • 126 benchmarks • 313 datasets
Language acquisition refers to tasks related to the learning of a second language.
(Image credit: Papersgraph)
These leaderboards are used to track progress in language-acquisition-2
Use these libraries to find language-acquisition-2 models and implementations
This work studies how a visually grounded speech model, trained on images of scenes paired with spoken captions, captures aspects of semantics, and introduces a newly collected data set of human semantic relevance judgements and an associated task, semantic speech retrieval, where the goal is to search for spoken utterances that are semantically relevant to a given text query.
A model is presented that incrementally updates a semantic network, with limited computational steps, and replicates many patterns found in human semantic fluency using a simple random walk, showing that a combination of both structural and semantic features are correlated with human performance patterns.
The proposed model significantly outperforms five comparison methods for interpreting zero-shot sentences and demonstrates human-interpretable intermediate outputs of the model in the appendix.
This paper introduces the Corpus of Linguistic Acceptability (CoLA), a set of 10,657 English sentences labeled as grammatical or ungrammatical from published linguistics literature, and trains several recurrent neural network models on acceptability classification, and finds that the authors' models outperform unsupervised models by Lau et al. (2016) on CoLA.
This system uses a logistic regression model to predict the likelihood of a student making a mistake while answering an exercise on Duolingo in all three language tracks - English/Spanish (en/es), Spanish/ English (es/en) and French/English (fr/en).
Experiments on transfer between natural languages show that zero-shot performance on a test language is highly correlated with typological syntactic similarity to the training language, suggesting that representations induced from natural languages correspond to the cross-linguistic syntactic properties studied in linguistic typology.
It is argued that the ability to imagine out-of-distribution goals is key to enable creative discoveries and open-ended learning, and Imagine, an intrinsically motivated deep reinforcement learning architecture that models this ability is introduced.
VisCOLL, a visually grounded language learning task, which simulates the continual acquisition of compositional phrases from streaming visual scenes, and reveals that SoTA continual learning approaches provide little to no improvements on VisCOLL.
A methodology for constructing corpora of CDS paired with sentential logical forms, and uses this method to create two corpora, in English and Hebrew, that provide syntactic and semantic annotation for two corpora from CHILDES.
Experiments on phonemic transcripts of spontaneous speech by parents to young children suggest that the model-based, unsupervised algorithm for recovering word boundaries in a natural-language text from which they have been deleted is more effective than other proposed algorithms, at least when utterance boundaries are given and the text includes a substantial number of short utterances.
Adding a benchmark result helps the community track progress.