1
Modeling Temporal Tonal Relations in Polyphonic Music Through Deep Networks With a Novel Image-Based Representation
2
An Overview of Lead and Accompaniment Separation in Music
3
Jointly Detecting and Separating Singing Voice: A Multi-Task Approach
4
Adversarial Semi-Supervised Audio Source Separation Applied to Singing Voice Extraction
5
SVSGAN: Singing Voice Separation Via Generative Adversarial Network
6
Singing Voice Separation with Deep U-Net Convolutional Networks
7
A Functional Taxonomy of Music Generation Systems
8
Sinusoidal Partials Tracking for Singing Analysis Using the Heuristic of the Minimal Frequency and Magnitude Difference
9
Improving music source separation based on deep neural networks through data augmentation and network blending
10
Monoaural Audio Source Separation Using Deep Convolutional Neural Networks
11
The 2016 Signal Separation Evaluation Campaign
12
Singing Voice Separation Using RPCA with Weighted l_1 -norm
13
Exploiting Saliency for Object Segmentation from Image Level Labels
14
Deep clustering and conventional networks for music separation: Stronger together
15
Densely Connected Convolutional Networks
16
Evaluation of audio source separation models using hypothesis-driven non-parametric statistical methods
17
Auditory Scene Analysis: The Perceptual Organization of Sound by Albert Bregman (review)
18
Neural networks and Deep Learning
19
Multichannel Audio Source Separation With Deep Neural Networks
20
Single Channel Audio Source Separation using Deep Neural Network Ensembles
21
Singing Voice Separation and Pitch Extraction from Monaural Polyphonic Audio Music via DNN and Adaptive Pitch Tracking
22
Singing Voice Separation and Vocal F0 Estimation Based on Mutual Combination of Robust Principal Component Analysis and Subharmonic Summation
23
Common fate model for unison source separation
24
Vocal activity informed singing voice separation with the iKala dataset
25
Deep neural network based instrument extraction from music
26
Scalable audio separation with light Kernel Additive Modelling
27
Deep Karaoke: Extracting Vocals from Musical Mixtures Using a Convolutional Deep Neural Network
28
Joint Optimization of Masks and Deep Recurrent Neural Networks for Monaural Source Separation
29
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
30
Adam: A Method for Stochastic Optimization
31
Modelling Mutual Information between Voiceprint and Optimal Number of Mel-Frequency Cepstral Coefficients in Voice Discrimination
32
Visualising Singing Style under Common Musical Events Using Pitch-Dynamics Trajectories and Modified TRACLUS Clustering
33
On Training Targets for Supervised Speech Separation
34
Fully convolutional networks for semantic segmentation
35
Singing-Voice Separation from Monaural Recordings using Deep Recurrent Neural Networks
36
MedleyDB: A Multitrack Dataset for Annotation-Intensive MIR Research
37
Vocal Separation from Monaural Music Using Temporal/Spectral Continuity and Sparsity Constraints
38
Deep content-based music recommendation
39
ImageNet classification with deep convolutional neural networks
40
Music/Voice Separation Using the Similarity Matrix
41
A General Flexible Framework for the Handling of Prior Information in Audio Source Separation
42
Singing-voice separation from monaural recordings using robust principal component analysis
43
Adaptive filtering for music/voice separation exploiting the repeating musical structure
44
A Musically Motivated Mid-Level Representation for Pitch Estimation and Musical Audio Source Separation
45
The Augmented Lagrange Multiplier Method for Exact Recovery of Corrupted Low-Rank Matrices
46
Musical instrument identification using multiscale Mel-frequency cepstral coefficients
47
Real-time Polyphonic Music Transcription with Non-negative Matrix Factorization and Beta-divergence
48
Understanding the difficulty of training deep feedforward neural networks
49
A Modeling of Singing Voice Robust to Accompaniment Sounds and Its Application to Singer Identification and Vocal-Timbre-Similarity-Based Music Information Retrieval
50
Nonnegative Matrix Factorization with the Itakura-Saito Divergence: With Application to Music Analysis
51
Monaural Sound Source Separation by Nonnegative Matrix Factorization With Temporal Continuity and Sparseness Criteria
52
Performance measurement in blind audio source separation
53
LyricAlly: automatic synchronization of acoustic musical signals and textual lyrics
55
Separation of Mixed Audio Sources By Independent Subspace Analysis
56
Book Review: Auditory Scene Analysis: The Perceptual Organization of Sound
57
Approximation capabilities of multilayer feedforward networks
58
Some Experiments on the Recognition of Speech, with One and with Two Ears
59
Mining Labeled Data from Web-Scale Collections for Vocal Activity Detection in Music
60
An Analysis/Synthesis Framework for Automatic F0 Annotation of Multitrack Datasets
61
Singing voice separation using rpca with weighted l1norm. In: International Conference on Latent Variable Analysis and Signal Separation (LVA/ICA)
62
Learning to Pinpoint Singing Voice from Weakly Labeled Examples
63
TensorFlow: Large-scale machine learning on heterogeneous systems
64
Dropout: a simple way to prevent neural networks from overfitting
65
Implementation and Evaluation of Real-Time Interactive User Interface Design in Self-learning Singing Pitch Training Apps
66
REpeating Pattern Extraction Technique (REPET): A Simple Method for Music/Voice Separation
67
Deep contentbased music recommendation. In: Advances in neural information processing systems
68
Timbre and Melody Features for the Recognition of Vocal Activity and Instrumental Solos in Polyphonic Music
69
Automatic Recognition of Lyrics in Singing
70
Single Channel Vocal Separation using Median Filtering and Factorisation Techniques
71
The Use of Mel-frequency Cepstral Coefficients in Musical Instrument Identification
72
Separation of Vocals from Polyphonic Audio Recordings
73
On Ideal Binary Mask As the Computational Goal of Auditory Scene Analysis
74
Algorithms for Non-negative Matrix Factorization
75
Discrete-time signal processing (2nd ed.)
77
Discrete-Time Signal Pro-cessing
78
For iKala, the GNSDRs for both singing voice and music accompaniment are 9.50 dB and 6.34 dB respectively; For DSD100, the SDRs for both singing voice and music accompaniment
79
Publisher's Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations