3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in dialogue-understanding
Use these libraries to find dialogue-understanding models and implementations
The machine learning architecture of the Snips Voice Platform is presented, a software solution to perform Spoken Language Understanding on microprocessors typical of IoT devices that is fast and accurate while enforcing privacy by design, as no personal user data is ever collected.
An innovated contextualized attention-based deep neural network, SDNet, to fuse context into traditional MRC models, which leverages both inter-attention and self-att attention to comprehend conversation context and extract relevant information from passage.
The design of an embedded, private-by-design SLU system is outlined and it is shown that it has performance on-par with cloud-based commercial solutions.
This work proposes a more flexible, interpretable and customizable encoder alternative, Branchformer, with parallel branches for modeling various ranged dependencies in end-to-end speech processing, including the ability to have variable inference complexity in a single trained model.
The core architecture of SpeechBrain is described, designed to support several tasks of common interest, allowing users to naturally conceive, compare and share novel speech processing pipelines.
A slot gate that focuses on learning the relationship between intent and slot attention vectors in order to obtain better semantic frame results by the global optimization is proposed.
A novel Collaborative Memory Network (CM-Net) based on the well-designed block, named CM-block, which achieves the state-of-the-art results on the ATIS and SNIPS in most of criteria, and significantly outperforms the baseline models on the CAIS.
A novel bi-directional interrelated model for joint intent detection and slot filling is proposed and an SF-ID network is introduced to establish direct connections for the two tasks to help them promote each other mutually.
A novel framework for SLU to better incorporate the intent information, which further guiding the slot filling is proposed, which achieves the state-of-the-art performance and outperforms other previous methods by a large margin.
This work proposes a strategy to overcome this requirement in which speech synthesis is used to generate a large synthetic training dataset from several artificial speakers, and confirms the effectiveness of this approach with experiments on two open-source SLU datasets.
Adding a benchmark result helps the community track progress.