Sorry, you need to enable JavaScript to visit this website.

Short-time spectral aggregation for speaker embedding

Citation Author(s):
Youzhi Tu, Man-Wai Mak
Submitted by:
youzhi tu
Last updated:
23 June 2021 - 6:40am
Document Type:
Presentation Slides
Document Year:
Presenters Name:
Youzhi Tu
Paper Code:



State-of-the-art speaker verification systems take frame-level acoustics features as input and produce fixed-dimensional embeddings as utterance-level representations. Thus, how to aggregate information from frame-level features is vital for achieving high performance. This paper introduces short-time spectral pooling (STSP) for better aggregation of frame-level information. STSP transforms the temporal feature maps of a speaker embedding network into the spectral domain and extracts the lowest spectral components of the averaged spectrograms for aggregation. Benefiting from the low-pass characteristic of the averaged spectrograms, STSP is able to preserve most of the speaker information in the feature maps using a few spectral components only. We show that statistics pooling is a special case of STSP where only the DC spectral components are used. Experiments on VoxCeleb1 and VOiCES 2019 show that STSP outperforms statistics pooling and multi-head attentive pooling, which suggests that leveraging more spectral information in the CNN feature maps can produce highly discriminative speaker embeddings.

0 users have voted:

Dataset Files