Academic literature on the topic 'Visual speech model'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Visual speech model.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "Visual speech model"
Jia, Xi Bin, and Mei Xia Zheng. "Video Based Visual Speech Feature Model Construction." Applied Mechanics and Materials 182-183 (June 2012): 1367–71. http://dx.doi.org/10.4028/www.scientific.net/amm.182-183.1367.
Full textMishra, Saumya, Anup Kumar Gupta, and Puneet Gupta. "DARE: Deceiving Audio–Visual speech Recognition model." Knowledge-Based Systems 232 (November 2021): 107503. http://dx.doi.org/10.1016/j.knosys.2021.107503.
Full textBrahme, Aparna, and Umesh Bhadade. "Effect of Various Visual Speech Units on Language Identification Using Visual Speech Recognition." International Journal of Image and Graphics 20, no. 04 (October 2020): 2050029. http://dx.doi.org/10.1142/s0219467820500291.
Full textMetzger, Brian A. ,., John F. ,. Magnotti, Elizabeth Nesbitt, Daniel Yoshor, and Michael S. ,. Beauchamp. "Cross-modal suppression model of speech perception: Visual information drives suppressive interactions between visual and auditory speech in pSTG." Journal of Vision 20, no. 11 (October 20, 2020): 434. http://dx.doi.org/10.1167/jov.20.11.434.
Full textHazen, T. J. "Visual model structures and synchrony constraints for audio-visual speech recognition." IEEE Transactions on Audio, Speech and Language Processing 14, no. 3 (May 2006): 1082–89. http://dx.doi.org/10.1109/tsa.2005.857572.
Full textFagel, Sascha. "Merging methods of speech visualization." ZAS Papers in Linguistics 40 (January 1, 2005): 19–32. http://dx.doi.org/10.21248/zaspil.40.2005.255.
Full textLoh, Marco, Gabriele Schmid, Gustavo Deco, and Wolfram Ziegler. "Audiovisual Matching in Speech and Nonspeech Sounds: A Neurodynamical Model." Journal of Cognitive Neuroscience 22, no. 2 (February 2010): 240–47. http://dx.doi.org/10.1162/jocn.2009.21202.
Full textYu, Wentao, Steffen Zeiler, and Dorothea Kolossa. "Reliability-Based Large-Vocabulary Audio-Visual Speech Recognition." Sensors 22, no. 15 (July 23, 2022): 5501. http://dx.doi.org/10.3390/s22155501.
Full textHow, Chun Kit, Ismail Mohd Khairuddin, Mohd Azraai Mohd Razman, Anwar P. P. Abdul Majeed, and Wan Hasbullah Mohd Isa. "Development of Audio-Visual Speech Recognition using Deep-Learning Technique." MEKATRONIKA 4, no. 1 (June 27, 2022): 88–95. http://dx.doi.org/10.15282/mekatronika.v4i1.8625.
Full textHolubenko, Nataliia. "Cognitive and Intersemiotic Model of the Visual and Verbal Modes in a Screen Adaptation to Literary Texts." World Journal of English Language 12, no. 6 (July 18, 2022): 129. http://dx.doi.org/10.5430/wjel.v12n6p129.
Full textDissertations / Theses on the topic "Visual speech model"
Somasundaram, Arunachalam. "A facial animation model for expressive audio-visual speech." Columbus, Ohio : Ohio State University, 2006. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=osu1148973645.
Full textVan, Wassenhove Virginie. "Cortical dynamics of auditory-visual speech a forward model of multisensory integration /." College Park, Md. : University of Maryland, 2004. http://hdl.handle.net/1903/1871.
Full textThesis research directed by: Neuroscience and Cognitive Science. Title from t.p. of PDF. Includes bibliographical references. Published by UMI Dissertation Services, Ann Arbor, Mich. Also available in paper.
Cosker, Darren. "Animation of a hierarchical image based facial model and perceptual analysis of visual speech." Thesis, Cardiff University, 2005. http://orca.cf.ac.uk/56003/.
Full textTheobald, Barry-John. "Visual speech synthesis using shape and appearance models." Thesis, University of East Anglia, 2003. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.396720.
Full textDean, David Brendan. "Synchronous HMMs for audio-visual speech processing." Thesis, Queensland University of Technology, 2008. https://eprints.qut.edu.au/17689/3/David_Dean_Thesis.pdf.
Full textDean, David Brendan. "Synchronous HMMs for audio-visual speech processing." Queensland University of Technology, 2008. http://eprints.qut.edu.au/17689/.
Full textMukherjee, Niloy 1978. "Spontaneous speech recognition using visual context-aware language models." Thesis, Massachusetts Institute of Technology, 2003. http://hdl.handle.net/1721.1/62380.
Full textIncludes bibliographical references (p. 83-88).
The thesis presents a novel situationally-aware multimodal spoken language system called Fuse that performs speech understanding for visual object selection. An experimental task was created in which people were asked to refer, using speech alone, to objects arranged on a table top. During training, Fuse acquires a grammar and vocabulary from a "show-and-tell" procedure in which visual scenes are paired with verbal descriptions of individual objects. Fuse determines a set of visually salient words and phrases and associates them to a set of visual features. Given a new scene, Fuse uses the acquired knowledge to generate class-based language models conditioned on the objects present in the scene as well as a spatial language model that predicts the occurrences of spatial terms conditioned on target and landmark objects. The speech recognizer in Fuse uses a weighted mixture of these language models to search for more likely interpretations of user speech in context of the current scene. During decoding, the weights are updated using a visual attention model which redistributes attention over objects based on partially decoded utterances. The dynamic situationally-aware language models enable Fuse to jointly infer spoken language utterances underlying speech signals as well as the identities of target objects they refer to. In an evaluation of the system, visual situationally-aware language modeling shows significant , more than 30 %, decrease in speech recognition and understanding error rates. The underlying ideas of situation-aware speech understanding that have been developed in Fuse may may be applied in numerous areas including assistive and mobile human-machine interfaces.
by Niloy Mukherjee.
S.M.
Kalantari, Shahram. "Improving spoken term detection using complementary information." Thesis, Queensland University of Technology, 2015. https://eprints.qut.edu.au/90074/1/Shahram_Kalantari_Thesis.pdf.
Full textDeena, Salil Prashant. "Visual speech synthesis by learning joint probabilistic models of audio and video." Thesis, University of Manchester, 2012. https://www.research.manchester.ac.uk/portal/en/theses/visual-speech-synthesis-by-learning-joint-probabilistic-models-of-audio-and-video(bdd1a78b-4957-469e-8be4-34e83e676c79).html.
Full textAhmad, Nasir. "A motion based approach for audio-visual automatic speech recognition." Thesis, Loughborough University, 2011. https://dspace.lboro.ac.uk/2134/8564.
Full textBooks on the topic "Visual speech model"
G, Stork David, and Hennecke Marcus E, eds. Speechreading by humans and machines: Models, systems, and applications. Berlin: Springer, 1996.
Find full textHidden Markov Models for Visual Speech Synthesis in Limited Data Environments. Storming Media, 2001.
Find full textStork, David G., and Marcus E. Hennecke. Speechreading by Humans and Machines: Models, Systems, and Applications. Springer, 2010.
Find full textStork, David G., and Marcus E. Hennecke. Speechreading by Humans and Machines: Models, Systems, and Applications. Springer London, Limited, 2013.
Find full textAntrobus, John S. How Does the Waking and Sleeping Brain Produce Spontaneous Thought and Imagery, and Why? Edited by Kalina Christoff and Kieran C. R. Fox. Oxford University Press, 2018. http://dx.doi.org/10.1093/oxfordhb/9780190464745.013.36.
Full text(Editor), David G. Stork, and Marcus E. Hennecke (Editor), eds. Speechreading by Humans and Machines: Models, Systems, and Applications (NATO ASI Series / Computer and Systems Sciences). Springer, 1996.
Find full textRaine, Michael. A New Form of Silent Cinema. Oxford University Press, 2017. http://dx.doi.org/10.1093/oso/9780190254971.003.0007.
Full textTobin, Claudia. Modernism and Still Life. Edinburgh University Press, 2020. http://dx.doi.org/10.3366/edinburgh/9781474455138.001.0001.
Full textTitus, Barbara. Hearing Maskanda. Bloomsbury Publishing Inc, 2022. http://dx.doi.org/10.5040/9781501377792.
Full textBerressem, Hanjo. Felix Guattari's Schizoanalytic Ecology. Edinburgh University Press, 2020. http://dx.doi.org/10.3366/edinburgh/9781474450751.001.0001.
Full textBook chapters on the topic "Visual speech model"
Grant, Ken W., and Joshua G. W. Bernstein. "Toward a Model of Auditory-Visual Speech Intelligibility." In Multisensory Processes, 33–57. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-10461-0_3.
Full textMeiXia, Zheng, and Jia XiBin. "Joint LBP and DCT Model for Visual Speech." In Advances in Intelligent and Soft Computing, 101–7. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-27866-2_13.
Full textAkinpelu, Samson, and Serestina Viriri. "A Robust Deep Transfer Learning Model for Accurate Speech Emotion Classification." In Advances in Visual Computing, 419–30. Cham: Springer Nature Switzerland, 2022. http://dx.doi.org/10.1007/978-3-031-20716-7_33.
Full textDeena, Salil, and Aphrodite Galata. "Speech-Driven Facial Animation Using a Shared Gaussian Process Latent Variable Model." In Advances in Visual Computing, 89–100. Berlin, Heidelberg: Springer Berlin Heidelberg, 2009. http://dx.doi.org/10.1007/978-3-642-10331-5_9.
Full textBothe, Hans H. "A visual speech model based on fuzzy-neuro methods." In Image Analysis and Processing, 152–58. Berlin, Heidelberg: Springer Berlin Heidelberg, 1995. http://dx.doi.org/10.1007/3-540-60298-4_251.
Full textBaayen, R. Harald, Robert Schreuder, and Richard Sproat. "Morphology in the Mental Lexicon: A Computational Model for Visual Word Recognition." In Text, Speech and Language Technology, 267–93. Dordrecht: Springer Netherlands, 2000. http://dx.doi.org/10.1007/978-94-010-9458-0_9.
Full textSeong, Thum Wei, Mohd Zamri Ibrahim, Nurul Wahidah Binti Arshad, and D. J. Mulvaney. "A Comparison of Model Validation Techniques for Audio-Visual Speech Recognition." In IT Convergence and Security 2017, 112–19. Singapore: Springer Singapore, 2017. http://dx.doi.org/10.1007/978-981-10-6451-7_14.
Full textSun, Zhongbo, Yannan Wang, and Li Cao. "An Attention Based Speaker-Independent Audio-Visual Deep Learning Model for Speech Enhancement." In MultiMedia Modeling, 722–28. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-37734-2_60.
Full textXie, Lei, and Zhi-Qiang Liu. "Multi-stream Articulator Model with Adaptive Reliability Measure for Audio Visual Speech Recognition." In Advances in Machine Learning and Cybernetics, 994–1004. Berlin, Heidelberg: Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/11739685_104.
Full textMatthews, Iain, J. Andrew Bangham, Richard Harvey, and Stephen Cox. "A comparison of active shape model and scale decomposition based features for visual speech recognition." In Lecture Notes in Computer Science, 514–28. Berlin, Heidelberg: Springer Berlin Heidelberg, 1998. http://dx.doi.org/10.1007/bfb0054762.
Full textConference papers on the topic "Visual speech model"
Wang, Wupeng, Chao Xing, Dong Wang, Xiao Chen, and Fengyu Sun. "A Robust Audio-Visual Speech Enhancement Model." In ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020. http://dx.doi.org/10.1109/icassp40776.2020.9053033.
Full textTiippana, Kaisa, Ilmari Kurki, and Tarja Peromaa. "Applying the summation model in audiovisual speech perception." In The 14th International Conference on Auditory-Visual Speech Processing. ISCA: ISCA, 2017. http://dx.doi.org/10.21437/avsp.2017-28.
Full textFan, Heng, Jinhai Xiang, Guoliang Li, and Fuchuan Ni. "Robust visual tracking via deep discriminative model." In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2017. http://dx.doi.org/10.1109/icassp.2017.7952492.
Full textLiu, Li, Gang Feng, and Denis Beautemps. "Inner Lips Parameter Estimation based on Adaptive Ellipse Model." In The 14th International Conference on Auditory-Visual Speech Processing. ISCA: ISCA, 2017. http://dx.doi.org/10.21437/avsp.2017-15.
Full textIsrael Santos, Timothy, Andrew Abel, Nick Wilson, and Yan Xu. "Speaker-Independent Visual Speech Recognition with the Inception V3 Model." In 2021 IEEE Spoken Language Technology Workshop (SLT). IEEE, 2021. http://dx.doi.org/10.1109/slt48900.2021.9383540.
Full textEdge, J. D., A. Hilton, and P. Jackson. "Model-based synthesis of visual speech movements from 3D video." In SIGGRAPH '09: Posters. New York, New York, USA: ACM Press, 2009. http://dx.doi.org/10.1145/1599301.1599309.
Full textDinet, Eric, and Emmanuel Kubicki. "A selective attention model for predicting visual attractors." In ICASSP 2008 - 2008 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 2008. http://dx.doi.org/10.1109/icassp.2008.4517705.
Full textCosker, D. P. "Speaker-independent speech-driven facial animation using a hierarchical model." In International Conference on Visual Information Engineering (VIE 2003). Ideas, Applications, Experience. IEE, 2003. http://dx.doi.org/10.1049/cp:20030514.
Full textDai, Pingyang, Yanlong Luo, Weisheng Liu, Cuihua Li, and Yi Xie. "Robust visual tracking via part-based sparsity model." In ICASSP 2013 - 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2013. http://dx.doi.org/10.1109/icassp.2013.6637963.
Full textWu, Jinjian, Guangming Shi, Weisi Lin, and C. C. Jay Kuo. "Enhanced just noticeable difference model with visual regularity consideration." In 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2016. http://dx.doi.org/10.1109/icassp.2016.7471943.
Full text