A data driven approach to audiovisual speech mapping

Abel, Andrew and Marxer, Ricard and Barker, Jon and Watt, Roger and Whitmer, Bill and Derleth, Peter and Hussain, Amir; Liu, Cheng-Lin and Hussain, Amir and Luo, Bin and Tan, Kay Chen and Zeng, Yi and Zhang, Zhaoxiang, eds. (2016) A data driven approach to audiovisual speech mapping. In: Advances in Brain Inspired Cognitive Systems. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) . Springer-Verlag, CHN, pp. 331-342. ISBN 9783319496856 (https://doi.org/10.1007/978-3-319-49685-6_30)

[thumbnail of Abel-etal-BICS2016-A-data-driven-approach-to-audiovisual-speech-mapping]
Preview
Text. Filename: Abel_etal_BICS2016_A_data_driven_approach_to_audiovisual_speech_mapping.pdf
Accepted Author Manuscript
License: Strathprints license 1.0

Download (165kB)| Preview

Abstract

The concept of using visual information as part of audio speech processing has been of significant recent interest. This paper presents a data driven approach that considers estimating audio speech acoustics using only temporal visual information without considering linguistic features such as phonemes and visemes. Audio (log filterbank) and visual (2D-DCT) features are extracted, and various configurations of MLP and datasets are used to identify optimal results, showing that given a sequence of prior visual frames an equivalent reasonably accurate audio frame estimation can be mapped.