|Appears in Collections:||Psychology Book Chapters and Sections|
|Title:||A Data Driven Approach to Audiovisual Speech Mapping|
|Citation:||Abel A, Marxer R, Hussain A, Barker J, Watt R, Whitmer B & Derleth P (2016) A Data Driven Approach to Audiovisual Speech Mapping In: Liu CL, Hussain A, Luo B, Tan KC, Zeng Y, Zhang Z (ed.) Advances in Brain Inspired Cognitive Systems, Cham, Switzerland: Springer. BICS 2016: International Conference on Brain Inspired Cognitive Systems, 28.11.2016 - 30.11.2016, Beijing, China, pp. 331-342.|
|Series/Report no.:||Lecture Notes in Computer Science, 10023|
|Abstract:||The concept of using visual information as part of audio speech processing has been of significant recent interest. This paper presents a data driven approach that considers estimating audio speech acoustics using only temporal visual information without considering linguistic features such as phonemes and visemes. Audio (log filterbank) and visual (2D-DCT) features are extracted, and various configurations of MLP and datasets are used to identify optimal results, showing that given a sequence of prior visual frames an equivalent reasonably accurate audio frame estimation can be mapped.|
|Rights:||Publisher policy allows this work to be made available in this repository. Published in Liu CL., Hussain A., Luo B., Tan K., Zeng Y., Zhang Z. (eds) Advances in Brain Inspired Cognitive Systems. BICS 2016. Lecture Notes in Computer Science, vol 10023, published by Springer. The original publication is available at Springer via http://dx.doi.org/10.1007/978-3-319-49685-6_30|
|abelBics2016Paper-final-submitted.pdf||193.3 kB||Adobe PDF||View/Open|
This item is protected by original copyright
Items in the Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
If you believe that any material held in STORRE infringes copyright, please contact email@example.com providing details and we will remove the Work from public display in STORRE and investigate your claim.