Journal articles on the topic 'Visual speech model'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 journal articles for your research on the topic 'Visual speech model.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.
Jia, Xi Bin, and Mei Xia Zheng. "Video Based Visual Speech Feature Model Construction." Applied Mechanics and Materials 182-183 (June 2012): 1367–71. http://dx.doi.org/10.4028/www.scientific.net/amm.182-183.1367.
Full textMishra, Saumya, Anup Kumar Gupta, and Puneet Gupta. "DARE: Deceiving Audio–Visual speech Recognition model." Knowledge-Based Systems 232 (November 2021): 107503. http://dx.doi.org/10.1016/j.knosys.2021.107503.
Full textBrahme, Aparna, and Umesh Bhadade. "Effect of Various Visual Speech Units on Language Identification Using Visual Speech Recognition." International Journal of Image and Graphics 20, no. 04 (October 2020): 2050029. http://dx.doi.org/10.1142/s0219467820500291.
Full textMetzger, Brian A. ,., John F. ,. Magnotti, Elizabeth Nesbitt, Daniel Yoshor, and Michael S. ,. Beauchamp. "Cross-modal suppression model of speech perception: Visual information drives suppressive interactions between visual and auditory speech in pSTG." Journal of Vision 20, no. 11 (October 20, 2020): 434. http://dx.doi.org/10.1167/jov.20.11.434.
Full textHazen, T. J. "Visual model structures and synchrony constraints for audio-visual speech recognition." IEEE Transactions on Audio, Speech and Language Processing 14, no. 3 (May 2006): 1082–89. http://dx.doi.org/10.1109/tsa.2005.857572.
Full textFagel, Sascha. "Merging methods of speech visualization." ZAS Papers in Linguistics 40 (January 1, 2005): 19–32. http://dx.doi.org/10.21248/zaspil.40.2005.255.
Full textLoh, Marco, Gabriele Schmid, Gustavo Deco, and Wolfram Ziegler. "Audiovisual Matching in Speech and Nonspeech Sounds: A Neurodynamical Model." Journal of Cognitive Neuroscience 22, no. 2 (February 2010): 240–47. http://dx.doi.org/10.1162/jocn.2009.21202.
Full textYu, Wentao, Steffen Zeiler, and Dorothea Kolossa. "Reliability-Based Large-Vocabulary Audio-Visual Speech Recognition." Sensors 22, no. 15 (July 23, 2022): 5501. http://dx.doi.org/10.3390/s22155501.
Full textHow, Chun Kit, Ismail Mohd Khairuddin, Mohd Azraai Mohd Razman, Anwar P. P. Abdul Majeed, and Wan Hasbullah Mohd Isa. "Development of Audio-Visual Speech Recognition using Deep-Learning Technique." MEKATRONIKA 4, no. 1 (June 27, 2022): 88–95. http://dx.doi.org/10.15282/mekatronika.v4i1.8625.
Full textHolubenko, Nataliia. "Cognitive and Intersemiotic Model of the Visual and Verbal Modes in a Screen Adaptation to Literary Texts." World Journal of English Language 12, no. 6 (July 18, 2022): 129. http://dx.doi.org/10.5430/wjel.v12n6p129.
Full textKröger, Bernd J., Julia Gotto, Susanne Albert, and Christiane Neuschaefer-Rube. "visual articulatory model and its application to therapy of speech disorders: a pilot study." ZAS Papers in Linguistics 40 (January 1, 2005): 79–94. http://dx.doi.org/10.21248/zaspil.40.2005.259.
Full textLi, Dengshi, Yu Gao, Chenyi Zhu, Qianrui Wang, and Ruoxi Wang. "Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy." Sensors 23, no. 4 (February 11, 2023): 2053. http://dx.doi.org/10.3390/s23042053.
Full textYuan, Yuan, Chunlin Tian, and Xiaoqiang Lu. "Auxiliary Loss Multimodal GRU Model in Audio-Visual Speech Recognition." IEEE Access 6 (2018): 5573–83. http://dx.doi.org/10.1109/access.2018.2796118.
Full textSlaney, Malcolm, and Richard F. Lyon. "Visual representations of speech—A computer model based on correlation." Journal of the Acoustical Society of America 88, S1 (November 1990): S23. http://dx.doi.org/10.1121/1.2028916.
Full textEdge, James D., Adrian Hilton, and Philip Jackson. "Model-Based Synthesis of Visual Speech Movements from 3D Video." EURASIP Journal on Audio, Speech, and Music Processing 2009 (2009): 1–12. http://dx.doi.org/10.1155/2009/597267.
Full textSharma, Usha, Sushila Maheshkar, A. N. Mishra, and Rahul Kaushik. "Visual Speech Recognition Using Optical Flow and Hidden Markov Model." Wireless Personal Communications 106, no. 4 (September 10, 2018): 2129–47. http://dx.doi.org/10.1007/s11277-018-5930-z.
Full textSetyati, Endang, Mauridhi Hery Purnomo, Surya Sumpeno, and Joan Santoso. "HIDDEN MARKOV MODELS BASED INDONESIAN VISEME MODEL FOR NATURAL SPEECH WITH AFFECTION." Kursor 8, no. 3 (December 13, 2016): 102. http://dx.doi.org/10.28961/kursor.v8i3.61.
Full textSeo, Minji, and Myungho Kim. "Fusing Visual Attention CNN and Bag of Visual Words for Cross-Corpus Speech Emotion Recognition." Sensors 20, no. 19 (September 28, 2020): 5559. http://dx.doi.org/10.3390/s20195559.
Full textHertrich, Ingo, Susanne Dietrich, and Hermann Ackermann. "Cross-modal Interactions during Perception of Audiovisual Speech and Nonspeech Signals: An fMRI Study." Journal of Cognitive Neuroscience 23, no. 1 (January 2011): 221–37. http://dx.doi.org/10.1162/jocn.2010.21421.
Full textBlackburn, Catherine L., Pádraig T. Kitterick, Gary Jones, Christian J. Sumner, and Paula C. Stacey. "Visual Speech Benefit in Clear and Degraded Speech Depends on the Auditory Intelligibility of the Talker and the Number of Background Talkers." Trends in Hearing 23 (January 2019): 233121651983786. http://dx.doi.org/10.1177/2331216519837866.
Full textFleming, Luke. "Negating speech." Gesture 14, no. 3 (December 31, 2014): 263–96. http://dx.doi.org/10.1075/gest.14.3.01fle.
Full textNikolaus, Mitja, Afra Alishahi, and Grzegorz Chrupała. "Learning English with Peppa Pig." Transactions of the Association for Computational Linguistics 10 (2022): 922–36. http://dx.doi.org/10.1162/tacl_a_00498.
Full textRyumin, Dmitry, Denis Ivanko, and Elena Ryumina. "Audio-Visual Speech and Gesture Recognition by Sensors of Mobile Devices." Sensors 23, no. 4 (February 17, 2023): 2284. http://dx.doi.org/10.3390/s23042284.
Full textYang, Chih-Chun, Wan-Cyuan Fan, Cheng-Fu Yang, and Yu-Chiang Frank Wang. "Cross-Modal Mutual Learning for Audio-Visual Speech Recognition and Manipulation." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 3 (June 28, 2022): 3036–44. http://dx.doi.org/10.1609/aaai.v36i3.20210.
Full textWang, Dong, Bing Liu, Yong Zhou, Mingming Liu, Peng Liu, and Rui Yao. "Separate Syntax and Semantics: Part-of-Speech-Guided Transformer for Image Captioning." Applied Sciences 12, no. 23 (November 22, 2022): 11875. http://dx.doi.org/10.3390/app122311875.
Full textBiswas, Astik, P. K. Sahu, and Mahesh Chandra. "Multiple cameras audio visual speech recognition using active appearance model visual features in car environment." International Journal of Speech Technology 19, no. 1 (January 23, 2016): 159–71. http://dx.doi.org/10.1007/s10772-016-9332-x.
Full textLindborg, Alma, and Tobias S. Andersen. "Bayesian binding and fusion models explain illusion and enhancement effects in audiovisual speech perception." PLOS ONE 16, no. 2 (February 19, 2021): e0246986. http://dx.doi.org/10.1371/journal.pone.0246986.
Full textZeliang Zhang, Xiongfei Li, and Chengjia Yang. "Visual Speech Recognition based on Improved type of Hidden Markov Model." Journal of Convergence Information Technology 7, no. 13 (July 31, 2012): 119–26. http://dx.doi.org/10.4156/jcit.vol7.issue13.14.
Full textHONG, PENGYU, ZHEN WEN, and THOMAS S. HUANG. "iFACE: A 3D SYNTHETIC TALKING FACE." International Journal of Image and Graphics 01, no. 01 (January 2001): 19–26. http://dx.doi.org/10.1142/s0219467801000037.
Full textZhou, Hang, Yu Liu, Ziwei Liu, Ping Luo, and Xiaogang Wang. "Talking Face Generation by Adversarially Disentangled Audio-Visual Representation." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 9299–306. http://dx.doi.org/10.1609/aaai.v33i01.33019299.
Full textHe, Yibo, Kah Phooi Seng, and Li Minn Ang. "Multimodal Sensor-Input Architecture with Deep Learning for Audio-Visual Speech Recognition in Wild." Sensors 23, no. 4 (February 7, 2023): 1834. http://dx.doi.org/10.3390/s23041834.
Full textJeon, Sanghun, and Mun Sang Kim. "Noise-Robust Multimodal Audio-Visual Speech Recognition System for Speech-Based Interaction Applications." Sensors 22, no. 20 (October 12, 2022): 7738. http://dx.doi.org/10.3390/s22207738.
Full textHertrich, Ingo, Klaus Mathiak, Werner Lutzenberger, and Hermann Ackermann. "Time Course of Early Audiovisual Interactions during Speech and Nonspeech Central Auditory Processing: A Magnetoencephalography Study." Journal of Cognitive Neuroscience 21, no. 2 (February 2009): 259–74. http://dx.doi.org/10.1162/jocn.2008.21019.
Full textBielski, Lynn M., and Charissa R. Lansing. "Utility of the Baddeley and Hitch Model of Short-Term Working Memory To Investigate Spoken Language Understanding: A Tutorial." Perspectives on Aural Rehabilitation and Its Instrumentation 19, no. 1 (May 2012): 25–33. http://dx.doi.org/10.1044/arii19.1.25.
Full textAnwar, Miftahulkhairah, Fathiaty Murtadho, Endry Boeriswati, Gusti Yarmi, and Helvy Tiana Rosa. "analysis model of impolite Indonesian language use." Linguistics and Culture Review 5, S3 (December 5, 2021): 1426–41. http://dx.doi.org/10.21744/lingcure.v5ns3.1840.
Full textHanda, Anand, Rashi Agarwal, and Narendra Kohli. "Audio-Visual Emotion Recognition System Using Multi-Modal Features." International Journal of Cognitive Informatics and Natural Intelligence 15, no. 4 (October 2021): 1–14. http://dx.doi.org/10.4018/ijcini.20211001.oa34.
Full textMiller, Christi W., Erin K. Stewart, Yu-Hsiang Wu, Christopher Bishop, Ruth A. Bentler, and Kelly Tremblay. "Working Memory and Speech Recognition in Noise Under Ecologically Relevant Listening Conditions: Effects of Visual Cues and Noise Type Among Adults With Hearing Loss." Journal of Speech, Language, and Hearing Research 60, no. 8 (August 18, 2017): 2310–20. http://dx.doi.org/10.1044/2017_jslhr-h-16-0284.
Full textKusmana, Suherli, Endang Kasupardi, and Nunu Nurasa. "PENGARUH MODEL PEMBELAJARAN BERBASIS MASALAH MELALUI MEDIA AUDIO VISUAL TERHADAP PENINGKATAN KEMAMPUAN BERPIDATO SISWA KELAS IX SMP NEGERI 1 NUSAHERANG KABUPATEN KUNINGAN." Jurnal Tuturan 3, no. 1 (November 28, 2017): 419. http://dx.doi.org/10.33603/jt.v3i1.776.
Full textUhler, Kristin M., Rosalinda Baca, Emily Dudas, and Tammy Fredrickson. "Refining Stimulus Parameters in Assessing Infant Speech Perception Using Visual Reinforcement Infant Speech Discrimination: Sensation Level." Journal of the American Academy of Audiology 26, no. 10 (November 2015): 807–14. http://dx.doi.org/10.3766/jaaa.14093.
Full textGao, Ying, Yuqin Liu, and Chunyue Zhou. "Production and Interaction between Gesture and Speech: A Review." International Journal of English Linguistics 6, no. 2 (March 29, 2016): 131. http://dx.doi.org/10.5539/ijel.v6n2p131.
Full textMassaro, Dominic W., and Michael M. Cohen. "Perception of Synthesized Audible and Visible Speech." Psychological Science 1, no. 1 (January 1990): 55–63. http://dx.doi.org/10.1111/j.1467-9280.1990.tb00068.x.
Full textJ., Esra, and Diyar H. "Audio Visual Arabic Speech Recognition using KNN Model by Testing different Audio Features." International Journal of Computer Applications 180, no. 1 (December 15, 2017): 33–38. http://dx.doi.org/10.5120/ijca2017915901.
Full textLü, Guo-yun, Dong-mei Jiang, Yan-ning Zhang, Rong-chun Zhao, H. Sahli, Ilse Ravyse, and W. Verhelst. "DBN Based Multi-stream Multi-states Model for Continue Audio-Visual Speech Recognition." Journal of Electronics & Information Technology 30, no. 12 (April 22, 2011): 2906–11. http://dx.doi.org/10.3724/sp.j.1146.2007.00915.
Full textDeena, Salil, Shaobo Hou, and Aphrodite Galata. "Visual Speech Synthesis Using a Variable-Order Switching Shared Gaussian Process Dynamical Model." IEEE Transactions on Multimedia 15, no. 8 (December 2013): 1755–68. http://dx.doi.org/10.1109/tmm.2013.2279659.
Full textGogate, Mandar, Kia Dashtipour, Ahsan Adeel, and Amir Hussain. "CochleaNet: A robust language-independent audio-visual model for real-time speech enhancement." Information Fusion 63 (November 2020): 273–85. http://dx.doi.org/10.1016/j.inffus.2020.04.001.
Full textHazra, Sumon Kumar, Romana Rahman Ema, Syed Md Galib, Shalauddin Kabir, and Nasim Adnan. "Emotion recognition of human speech using deep learning method and MFCC features." Radioelectronic and Computer Systems, no. 4 (November 29, 2022): 161–72. http://dx.doi.org/10.32620/reks.2022.4.13.
Full textBrancazio, Lawrence, and Carol A. Fowler. "Merging auditory and visual phonetic information: A critical test for feedback?" Behavioral and Brain Sciences 23, no. 3 (June 2000): 327–28. http://dx.doi.org/10.1017/s0140525x00243240.
Full textVougioukas, Konstantinos, Stavros Petridis, and Maja Pantic. "Realistic Speech-Driven Facial Animation with GANs." International Journal of Computer Vision 128, no. 5 (October 13, 2019): 1398–413. http://dx.doi.org/10.1007/s11263-019-01251-8.
Full textSulubacak, Umut, Ozan Caglayan, Stig-Arne Grönroos, Aku Rouhe, Desmond Elliott, Lucia Specia, and Jörg Tiedemann. "Multimodal machine translation through visuals and speech." Machine Translation 34, no. 2-3 (August 13, 2020): 97–147. http://dx.doi.org/10.1007/s10590-020-09250-0.
Full textEt. al., D. N. V. S. L. S. Indira,. "An Enhanced CNN-2D for Audio-Visual Emotion Recognition (AVER) Using ADAM Optimizer." Turkish Journal of Computer and Mathematics Education (TURCOMAT) 12, no. 5 (April 11, 2021): 1378–88. http://dx.doi.org/10.17762/turcomat.v12i5.2030.
Full text