3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in decipherment-11
No benchmarks available.
Use these libraries to find decipherment-11 models and implementations
No subtasks available.
A deep generative model for unsupervised text style transfer that unifies previously proposed non-generative techniques and demonstrates the effectiveness of the method on a wide range of unsuper supervised style transfer tasks, including sentiment transfer, formality transfer, word decipherment, author imitation, and related language translation.
This paper proposes a method that leverages refined alignment of latent representations to perform style transfer on the basis of non-parallel text, and demonstrates the effectiveness of this cross-alignment method on three tasks: sentiment modification, decipherment of word substitution ciphers, and recovery of word order.
This work describes an approach for low-resource unsupervised POS tagging that yields fully grounded output and requires no labeled training data and demonstrates the tagger’s utility by incorporating it into a true ‘zero-resource’ variant of the MALOPA dependency parser model.
This work develops unsupervised models for character segmentation, character-image clustering, and decipherment of cluster sequences for enciphered manuscripts and gives empirical results for multiple ciphers.
This paper proposes a new technique that uses a target domain language model as the discriminator, providing richer and more stable token-level feedback during the learning process, and shows that this approach leads to improved performance on three tasks: word substitution decipherment, sentiment modification, and related language translation.
This first attempt at using techniques from computational linguistics to analyze the undeciphered proto-Elamite script using hierarchical clustering, n-gram frequencies, and LDA topic models demonstrates the utility of these techniques as an aid to manual decipherment.
A noisy-channel WFST cascade model for deciphering the original non-Latin script from observed romanized text in an unsupervised fashion and demonstrates that adding inductive bias through phonetic and visual priors on character mappings substantially improves the model’s performance on both languages, yielding results much closer to the supervised skyline.
A novel neural approach for automatic decipherment of lost languages with first automatic results in deciphering Linear B, a syllabic language related to ancient Greek, where the model correctly translates 67.3% of cognates.
A decipherment model that handles both of these challenges by building on rich linguistic constraints reflecting consistent patterns in historical sound change is proposed, and a measure for language closeness which correctly identifies related languages for Gothic and Ugaritic is proposed.
Adding a benchmark result helps the community track progress.