Sorry, you need to enable JavaScript to visit this website.

Learning to Dequantize Speech Signals by Primal-Dual Networks: An Approach for Acoustic Sensor Networks

Citation Author(s):
Ziyue Zhao, Dirk Lorenz, Tim Fingscheidt
Submitted by:
Christoph Brauer
Last updated:
12 May 2019 - 6:42pm
Document Type:
Poster
Document Year:
2019
Event:
Presenters:
Christoph Brauer and Ziyue Zhao
Paper Code:
4919
 

We introduce a method to improve the quality of simple scalar quantization in the context of acoustic sensor networks by combining ideas from sparse reconstruction, artificial neural networks and weighting filters. We start from the observation that optimization methods based on sparse reconstruction resemble the structure of a neural network. Hence, building upon a successful enhancement method, we unroll the algorithms and use this to build a neural network which we train to obtain enhanced decoding. In addition, the weighting filter from code-excited linear predictive (CELP) speech coding is integrated into the loss function of the neural network, achieving perceptually improved reconstructed speech. Our experiments show that our proposed trained methods allow for better speech reconstruction than the reference optimization methods.

Link to paper: https://ieeexplore.ieee.org/document/8683341

Source code available under: https://github.com/chrbraue/primal_dual_networks

up
0 users have voted: