Sorry, you need to enable JavaScript to visit this website.

Audio for Multimedia

FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING


Audio fingerprinting systems are often well designed to cope with a range of broadband noise types however they cope less well when presented with additive noise containing sinusoidal components. This is largely due to the fact that in a short-time signal representa- tion (over periods of ≈ 20ms) these noise components are largely indistinguishable from salient components of the desirable signal that is to be fingerprinted.

Draft_v2.pdf

PDF icon Draft_v2.pdf (63 downloads)

Paper Details

Authors:
Submitted On:
30 April 2018 - 7:27pm
Short Link:
Type:
Event:
Presenter's Name:
Paper Code:
Document Year:
Cite

Document Files

Draft_v2.pdf

(63 downloads)

Subscribe

[1] , "FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING", IEEE SigPort, 2018. [Online]. Available: http://sigport.org/3197. Accessed: Sep. 24, 2018.
@article{3197-18,
url = {http://sigport.org/3197},
author = { },
publisher = {IEEE SigPort},
title = {FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING},
year = {2018} }
TY - EJOUR
T1 - FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING
AU -
PY - 2018
PB - IEEE SigPort
UR - http://sigport.org/3197
ER -
. (2018). FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING. IEEE SigPort. http://sigport.org/3197
, 2018. FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING. Available at: http://sigport.org/3197.
. (2018). "FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING." Web.
1. . FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING [Internet]. IEEE SigPort; 2018. Available from : http://sigport.org/3197

FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING


Audio fingerprinting systems are often well designed to cope with a range of broadband noise types however they cope less well when presented with additive noise containing sinusoidal components. This is largely due to the fact that in a short-time signal representa- tion (over periods of ≈ 20ms) these noise components are largely indistinguishable from salient components of the desirable signal that is to be fingerprinted.

Draft_v2.pdf

PDF icon Draft_v2.pdf (74 downloads)

Paper Details

Authors:
Submitted On:
30 April 2018 - 7:27pm
Short Link:
Type:
Event:
Presenter's Name:
Paper Code:
Document Year:
Cite

Document Files

Draft_v2.pdf

(74 downloads)

Subscribe

[1] , "FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING", IEEE SigPort, 2018. [Online]. Available: http://sigport.org/3196. Accessed: Sep. 24, 2018.
@article{3196-18,
url = {http://sigport.org/3196},
author = { },
publisher = {IEEE SigPort},
title = {FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING},
year = {2018} }
TY - EJOUR
T1 - FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING
AU -
PY - 2018
PB - IEEE SigPort
UR - http://sigport.org/3196
ER -
. (2018). FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING. IEEE SigPort. http://sigport.org/3196
, 2018. FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING. Available at: http://sigport.org/3196.
. (2018). "FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING." Web.
1. . FOREGROUND HARMONIC NOISE REDUCTION FOR ROBUST AUDIO FINGERPRINTING [Internet]. IEEE SigPort; 2018. Available from : http://sigport.org/3196

Depression Speaks: Automatic Discrimination Between Depressed and Non-Depressed Speakers Based on Nonverbal Speech Features


This article proposes an automatic approach - based on nonverbal speech features - aimed at the automatic discrimination between depressed and non-depressed speakers. The experiments have been performed over one of the largest corpora collected for such a task in the literature ($62$ patients diagnosed with depression and $54$ healthy control subjects), especially when it comes to data where the depressed speakers have been diagnosed as such by professional psychiatrists.

icassp.pdf

PDF icon icassp.pdf (418 downloads)

Paper Details

Authors:
F.Scibelli, G.Roffo, M.Tayarani, L.Bartoli, G.De Mattia, A.Esposito and A.Vinciarelli
Submitted On:
19 April 2018 - 2:20pm
Short Link:
Type:
Event:
Presenter's Name:
Paper Code:
Document Year:
Cite

Document Files

icassp.pdf

(418 downloads)

Subscribe

[1] F.Scibelli, G.Roffo, M.Tayarani, L.Bartoli, G.De Mattia, A.Esposito and A.Vinciarelli, "Depression Speaks: Automatic Discrimination Between Depressed and Non-Depressed Speakers Based on Nonverbal Speech Features", IEEE SigPort, 2018. [Online]. Available: http://sigport.org/2992. Accessed: Sep. 24, 2018.
@article{2992-18,
url = {http://sigport.org/2992},
author = {F.Scibelli; G.Roffo; M.Tayarani; L.Bartoli; G.De Mattia; A.Esposito and A.Vinciarelli },
publisher = {IEEE SigPort},
title = {Depression Speaks: Automatic Discrimination Between Depressed and Non-Depressed Speakers Based on Nonverbal Speech Features},
year = {2018} }
TY - EJOUR
T1 - Depression Speaks: Automatic Discrimination Between Depressed and Non-Depressed Speakers Based on Nonverbal Speech Features
AU - F.Scibelli; G.Roffo; M.Tayarani; L.Bartoli; G.De Mattia; A.Esposito and A.Vinciarelli
PY - 2018
PB - IEEE SigPort
UR - http://sigport.org/2992
ER -
F.Scibelli, G.Roffo, M.Tayarani, L.Bartoli, G.De Mattia, A.Esposito and A.Vinciarelli. (2018). Depression Speaks: Automatic Discrimination Between Depressed and Non-Depressed Speakers Based on Nonverbal Speech Features. IEEE SigPort. http://sigport.org/2992
F.Scibelli, G.Roffo, M.Tayarani, L.Bartoli, G.De Mattia, A.Esposito and A.Vinciarelli, 2018. Depression Speaks: Automatic Discrimination Between Depressed and Non-Depressed Speakers Based on Nonverbal Speech Features. Available at: http://sigport.org/2992.
F.Scibelli, G.Roffo, M.Tayarani, L.Bartoli, G.De Mattia, A.Esposito and A.Vinciarelli. (2018). "Depression Speaks: Automatic Discrimination Between Depressed and Non-Depressed Speakers Based on Nonverbal Speech Features." Web.
1. F.Scibelli, G.Roffo, M.Tayarani, L.Bartoli, G.De Mattia, A.Esposito and A.Vinciarelli. Depression Speaks: Automatic Discrimination Between Depressed and Non-Depressed Speakers Based on Nonverbal Speech Features [Internet]. IEEE SigPort; 2018. Available from : http://sigport.org/2992

A First Attempt at Polyphonic Sound Event Detection Using Connectionist Temporal Classification


Sound event detection is the task of detecting the type, starting time, and ending time of sound events in audio streams. Recently, recurrent neural networks (RNNs) have become the mainstream solution for sound event detection. Because RNNs make a prediction at every frame, it is necessary to provide exact starting and ending times of the sound events in the training data, making data annotation an extremely time-consuming process.

Paper Details

Authors:
Florian Metze
Submitted On:
27 February 2017 - 5:12pm
Short Link:
Type:
Event:
Presenter's Name:
Paper Code:
Document Year:
Cite

Document Files

2017.03 Poster for ICASSP.pdf

(209 downloads)

Subscribe

[1] Florian Metze, "A First Attempt at Polyphonic Sound Event Detection Using Connectionist Temporal Classification", IEEE SigPort, 2017. [Online]. Available: http://sigport.org/1451. Accessed: Sep. 24, 2018.
@article{1451-17,
url = {http://sigport.org/1451},
author = {Florian Metze },
publisher = {IEEE SigPort},
title = {A First Attempt at Polyphonic Sound Event Detection Using Connectionist Temporal Classification},
year = {2017} }
TY - EJOUR
T1 - A First Attempt at Polyphonic Sound Event Detection Using Connectionist Temporal Classification
AU - Florian Metze
PY - 2017
PB - IEEE SigPort
UR - http://sigport.org/1451
ER -
Florian Metze. (2017). A First Attempt at Polyphonic Sound Event Detection Using Connectionist Temporal Classification. IEEE SigPort. http://sigport.org/1451
Florian Metze, 2017. A First Attempt at Polyphonic Sound Event Detection Using Connectionist Temporal Classification. Available at: http://sigport.org/1451.
Florian Metze. (2017). "A First Attempt at Polyphonic Sound Event Detection Using Connectionist Temporal Classification." Web.
1. Florian Metze. A First Attempt at Polyphonic Sound Event Detection Using Connectionist Temporal Classification [Internet]. IEEE SigPort; 2017. Available from : http://sigport.org/1451

Natural Sound Rendering for Headphones: Integration of signal processing techniques


With the strong growth of assistive and personal listening devices, natural sound rendering over headphones is becoming a necessity for prolonged listening in multimedia and virtual reality applications. The aim of natural sound rendering is to naturally recreate the sound scenes with the spatial and timbral quality as natural as possible, so as to achieve a truly immersive listening experience. However, rendering natural sound over headphones encounters many challenges. This tutorial article presents signal processing techniques to tackle these challenges to assist human listening.

Paper Details

Authors:
Kaushik Sunder, Ee-Leng Tan
Submitted On:
23 February 2016 - 1:44pm
Short Link:
Type:

Document Files

SPM2015manuscript-Natural Sound Rendering for Headphones.pdf

(495 downloads)

Subscribe

[1] Kaushik Sunder, Ee-Leng Tan, "Natural Sound Rendering for Headphones: Integration of signal processing techniques", IEEE SigPort, 2015. [Online]. Available: http://sigport.org/166. Accessed: Sep. 24, 2018.
@article{166-15,
url = {http://sigport.org/166},
author = {Kaushik Sunder; Ee-Leng Tan },
publisher = {IEEE SigPort},
title = {Natural Sound Rendering for Headphones: Integration of signal processing techniques},
year = {2015} }
TY - EJOUR
T1 - Natural Sound Rendering for Headphones: Integration of signal processing techniques
AU - Kaushik Sunder; Ee-Leng Tan
PY - 2015
PB - IEEE SigPort
UR - http://sigport.org/166
ER -
Kaushik Sunder, Ee-Leng Tan. (2015). Natural Sound Rendering for Headphones: Integration of signal processing techniques. IEEE SigPort. http://sigport.org/166
Kaushik Sunder, Ee-Leng Tan, 2015. Natural Sound Rendering for Headphones: Integration of signal processing techniques. Available at: http://sigport.org/166.
Kaushik Sunder, Ee-Leng Tan. (2015). "Natural Sound Rendering for Headphones: Integration of signal processing techniques." Web.
1. Kaushik Sunder, Ee-Leng Tan. Natural Sound Rendering for Headphones: Integration of signal processing techniques [Internet]. IEEE SigPort; 2015. Available from : http://sigport.org/166