Documents
Poster
Improving Music Source Separation based on DNNs through Data Augmentation and Network Blending
- Citation Author(s):
- Submitted by:
- Stefan Uhlich
- Last updated:
- 28 February 2017 - 6:24am
- Document Type:
- Poster
- Document Year:
- 2017
- Event:
- Presenters:
- Stefan Uhlich, Yuki Mitsufuji
- Paper Code:
- AASP-P1.5
- Categories:
- Log in to post comments
This paper deals with the separation of music into individual instrument tracks which is known to be a challenging problem. We describe two different deep neural network architectures for this task, a feed-forward and a recurrent one, and show that each of them yields themselves state-of-the art results on the SiSEC DSD100 dataset. For the recurrent network, we use data augmentation during training and show that even simple separation networks are prone to overfitting if no data augmentation is used. Furthermore, we propose a blending of both neural network systems where we linearly combine their raw outputs and then perform a multi-channel Wiener filter post-processing. This blending scheme yields the best results that have been reported to-date on the SiSEC DSD100 dataset.