3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in multi-speaker-source-separation-2
No benchmarks available.
Use these libraries to find multi-speaker-source-separation-2 models and implementations
No datasets available.
No subtasks available.
An algorithm based on the directional sparse filtering (DSF) framework that utilizes the Lehmer mean with learnable weights to adaptively account for source imbalance is proposed.
This paper analyzed how recurrent neural network (RNNs) cope with temporal dependencies by determining the relevant memory time span in a long short-term memory (LSTM) cell by leaking the state variable with a controlled lifetime and evaluating the task performance.
This work has shown that data of a specific scenario is relevant for solving another scenario, and concluded that a single model, trained on different scenarios is capable of matching performance of scenario specific models.
A deep clustering approach is used which trains on multichannel mixtures and learns to project spectrogram bins to source clusters that correlate with various spatial features, and shows that this system is capable of performing sound separation on monophonic inputs, despite having learned how to do so using multi-channel recordings.
Adding a benchmark result helps the community track progress.