Sorry, you need to enable JavaScript to visit this website.

Improving Music Source Separation based on DNNs through Data Augmentation and Network Blending

Citation Author(s):
Stefan Uhlich, Marcello Porcu, Franck Giron, Michael Enenkl, Thomas Kemp, Naoya Takahashi, Yuki Mitsufuji
Submitted by:
Stefan Uhlich
Last updated:
28 February 2017 - 6:24am
Document Type:
Poster
Document Year:
2017
Event:
Presenters:
Stefan Uhlich, Yuki Mitsufuji
Paper Code:
AASP-P1.5
 

This paper deals with the separation of music into individual instrument tracks which is known to be a challenging problem. We describe two different deep neural network architectures for this task, a feed-forward and a recurrent one, and show that each of them yields themselves state-of-the art results on the SiSEC DSD100 dataset. For the recurrent network, we use data augmentation during training and show that even simple separation networks are prone to overfitting if no data augmentation is used. Furthermore, we propose a blending of both neural network systems where we linearly combine their raw outputs and then perform a multi-channel Wiener filter post-processing. This blending scheme yields the best results that have been reported to-date on the SiSEC DSD100 dataset.

up
0 users have voted: