predict the type of given vocal bursts
3260 papers • 126 benchmarks • 313 datasets
predict the type of given vocal bursts
(Image credit: Papersgraph)
These leaderboards are used to track progress in speech-emotion-recognition
Use these libraries to find speech-emotion-recognition models and implementations
No subtasks available.
A global objective is formulated for learning the embeddings from text corpora and knowledge bases, which adopts a novel margin-based loss that is robust to noisy labels and faithfully models type correlation derived from knowledge bases.
SeizureNet, a deep learning framework which learns multi-spectral feature embeddings using an ensemble architecture for cross-patient seizure type classification, is presented and shown to improve the accuracy of smaller networks through knowledge distillation for applications with low-memory constraints.
This work proposes an end-to-end solution with a neural network model that uses a variant of cross-entropy loss function to handle out-of-context labels, and hierarchical loss normalization to cope with overly-specific ones in FETC.
An automatic toll tax collection framework designed for challenging conditions, consisting of three sequential steps: vehicle type recognition, license plate localization, and license plate reading, using state-of-the-art YOLO models.
A graph neural network model is presented that predicts types by probabilistically reasoning over a program’s structure, names, and patterns and can employ one-shot learning to predict an open vocabulary of types, including rare and user-defined ones.
A multi-task learning framework for BioNER to collectively use the training data of different types of entities and improve the performance on each of them, which achieves substantially better performance than state-of-the-art BioNER systems and baseline neural sequence labeling models.
Two data driven frameworks are considered: a deep neural network and a support vector machine using SIFT features for automatic recognition of cars of four types: Bus, Truck, Van and Small car.
This work proposes the task of context-dependent fine type tagging, where the set of acceptable labels for a mention is restricted to only those deducible from the local context (e.g. sentence or document).
Ludwig is a flexible, extensible and easy to use toolbox which allows users to train deep learning models and use them for obtaining predictions without writing code, and introduces a general modularized deep learning architecture called Encoder-Combiner-Decoder that can be instantiated to perform a vast amount of machine learning tasks.
This paper presents an approach to learning an image embedding that respects item type, and jointly learns notions of item similarity and compatibility in an end-to-end model.
Adding a benchmark result helps the community track progress.