P Demonte
Speech-to-screen : spatial separation of dialogue from noise towards improved speech intelligibility for the small screen
Demonte, P; Tang, Y; Hughes, RJ; Cox, TJ; Fazenda, BM; Shirley, BG
Authors
Y Tang
RJ Hughes
Prof Trevor Cox T.J.Cox@salford.ac.uk
Professor
Dr Bruno Fazenda B.M.Fazenda@salford.ac.uk
Reader
Dr Ben Shirley B.G.Shirley@salford.ac.uk
Reader
Abstract
Can externalizing dialogue when in the presence of stereo background noise improve speech intelligibility? This has been investigated for audio over headphones using head-tracking in order to explore potential future developments for small-screen devices. A quantitative listening experiment tasked participants with identifying target words in spoken sentences played in the presence of background noise via headphones. Sixteen different combinations of three independent variables were tested: speech and noise locations (internalized/externalized), video (on/off), and masking noise (stationary/fluctuating noise). The results revealed that the best improvements to speech intelligibility were generated by both the video-on condition and externalizing speech at the screen while retaining masking noise in the stereo mix.
Citation
Demonte, P., Tang, Y., Hughes, R., Cox, T., Fazenda, B., & Shirley, B. (2018, May). Speech-to-screen : spatial separation of dialogue from noise towards improved speech intelligibility for the small screen. Presented at 144th International Pro Audio Convention (AES Milan 2018), Milan, Italy
Presentation Conference Type | Other |
---|---|
Conference Name | 144th International Pro Audio Convention (AES Milan 2018) |
Conference Location | Milan, Italy |
Start Date | May 23, 2018 |
End Date | May 26, 2018 |
Acceptance Date | Feb 14, 2018 |
Online Publication Date | May 14, 2018 |
Publication Date | May 14, 2018 |
Deposit Date | May 30, 2018 |
Publicly Available Date | May 13, 2019 |
Publisher URL | http://www.aes.org/e-lib/browse.cfm?elib=19407 |
Related Public URLs | http://www.aes.org/ |
Additional Information | Event Type : Conference Funders : Engineering and Physical Sciences Research Council (EPSRC);BBC Projects : S3A: Future Spatial Audio for an Immersive Experience at Home Grant Number: EP/L000539/1 |
Files
AES 144 Paper71 DemonteEtAl 2018 200318 (002).pdf
(851 Kb)
PDF
You might also like
Cloud-based AI for automatic audio production for personalized immersive XR experiences
(2022)
Journal Article
Loudness differences for Voice-over-Voice audio in TV and streaming
(2020)
Journal Article
Downloadable Citations
About USIR
Administrator e-mail: library-research@salford.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search