- Read more about DNN-HA: A DNN-based hearing-aid strategy for real-time processing
- Log in to post comments
Although hearing aids (HAs) can compensate for elevated hearing thresholds using sound amplification, they often fail to restore auditory perception in adverse listening conditions. Here, we present a deep-neural-network (DNN) HA processing strategy that can provide individualised sound processing for the audiogram of a listener using a single model architecture. Our multi-purpose HA model can be used for different individuals and can process audio inputs of 3.2 ms in <0.5 ms, thus paving the way for precise DNN-based treatments of hearing loss that can be embedded in hearing devices.
- Categories:
- Read more about THE EFFECT OF PARTIAL TIME-FREQUENCY MASKING OF THE DIRECT SOUND ON THE PERCEPTION OF REVERBERANT SPEECH
- Log in to post comments
The perception of sound in real-life acoustic environments, such as enclosed rooms or open spaces with reflective objects, is affected by reverberation. Hence, reverberation is extensively studied in the context of auditory perception, with many studies highlighting the importance of the direct sound for perception. Based on this insight, speech processing methods often use time-frequency (TF) analysis to detect TF bins that are dominated by the direct sound, and then use the detected bins to reproduce or enhance the speech signals.
- Categories:
- Read more about MULTI-LINGUAL MULTI-TASK SPEECH EMOTION RECOGNITION USING WAV2VEC 2.0
- Log in to post comments
Speech Emotion Recognition (SER) has several use cases for
Digital Entertainment Content (DEC) in Over-the-top (OTT)
services, emotive Text-to-Speech (TTS) engines and voice
assistants. In this work, we present a Multi-Lingual (MLi) and
Multi-Task Learning (MTL) audio only SER system based on
the multi-lingual pre-trained wav2vec 2.0 model. The model
is fine-tuned on 25 open source datasets in 13 locales across
7 emotion categories. We show that, a) Our wav2vec 2.0
single task based model outperforms Pre-trained Audio Neural
- Categories:
- Read more about Deep Learning for Prominence Detection in Children's Read Speech
- Log in to post comments
Poster.pdf
- Categories:
- Read more about Internet Streaming Audio Based Speech Perception Threshold Measurement in Cochlear Implant Users
- Log in to post comments
Traditional face-to-face subjective listening test has become a challenge due to the COVID-19 pandemic. We developed a remote assessment system with Tencent Meeting, a video conferencing application, to address this issue. This paper
- Categories:
- Read more about Internet Streaming Audio Based Speech Perception Threshold Measurement in Cochlear Implant Users
- Log in to post comments
Traditional face-to-face subjective listening test has become a challenge due to the COVID-19 pandemic. We developed a remote assessment system with Tencent Meeting, a video conferencing application, to address this issue. This paper
- Categories:
- Read more about A NOISE-ROBUST SIGNAL PROCESSING STRATEGY FOR COCHLEAR IMPLANTS USING NEURAL NETWORKS
- Log in to post comments
- Categories:
- Read more about Effects of Spectral Tilt on Listeners' Preferences And Intelligibility
- Log in to post comments
- Categories:
- Read more about NON-INTRUSIVE SPEECH QUALITY ASSESSMENT USING NEURAL NETWORKS
- Log in to post comments
- Categories:
- Read more about DETECTION OF SPOKEN WORDS IN NOISE: COMPARISON OF HUMAN PERFORMANCE TO MAXIMUM LIKELIHOOD DETECTION
- Log in to post comments
In this work, we are interested in assessing the optimality of the human auditory system, when the input stimuli is natural speech that is affected by additive noise. In order to do this, we consider the DANTALE II listening test paradigm of Wagener et al., which has been used to evaluate the intelligibility of noisy speech by exposing human listeners to a selection of constructed noisy sentences. Inspired by this test, we propose a simple model for the communication and classification of noisy speech that takes place in the test.
- Categories: