Letteratura scientifica selezionata sul tema "Visual speech information"

Cita una fonte nei formati APA, MLA, Chicago, Harvard e in molti altri stili

Scegli il tipo di fonte:

Consulta la lista di attuali articoli, libri, tesi, atti di convegni e altre fonti scientifiche attinenti al tema "Visual speech information".

Accanto a ogni fonte nell'elenco di riferimenti c'è un pulsante "Aggiungi alla bibliografia". Premilo e genereremo automaticamente la citazione bibliografica dell'opera scelta nello stile citazionale di cui hai bisogno: APA, MLA, Harvard, Chicago, Vancouver ecc.

Puoi anche scaricare il testo completo della pubblicazione scientifica nel formato .pdf e leggere online l'abstract (il sommario) dell'opera se è presente nei metadati.

Articoli di riviste sul tema "Visual speech information"

1

Miller, Rachel M., Kauyumari Sanchez e Lawrence D. Rosenblum. "Alignment to visual speech information". Attention, Perception, & Psychophysics 72, n. 6 (agosto 2010): 1614–25. http://dx.doi.org/10.3758/app.72.6.1614.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
2

Rosenblum, Lawrence D., Deborah A. Yakel, Naser Baseer, Anjani Panchal, Brynn C. Nodarse e Ryan P. Niehus. "Visual speech information for face recognition". Perception & Psychophysics 64, n. 2 (febbraio 2002): 220–29. http://dx.doi.org/10.3758/bf03195788.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
3

Yakel, Deborah A., e Lawrence D. Rosenblum. "Face identification using visual speech information". Journal of the Acoustical Society of America 100, n. 4 (ottobre 1996): 2570. http://dx.doi.org/10.1121/1.417401.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
4

Weinholtz, Chase, e James W. Dias. "Categorical perception of visual speech information". Journal of the Acoustical Society of America 139, n. 4 (aprile 2016): 2018. http://dx.doi.org/10.1121/1.4949950.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
5

HISANAGA, Satoko, Kaoru SEKIYAMA, Tomohiko IGASAKI e Nobuki MURAYAMA. "Effects of visual information on audio-visual speech processing". Proceedings of the Annual Convention of the Japanese Psychological Association 75 (15 settembre 2011): 2AM061. http://dx.doi.org/10.4992/pacjpa.75.0_2am061.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
6

Sell, Andrea J., e Michael P. Kaschak. "Does visual speech information affect word segmentation?" Memory & Cognition 37, n. 6 (settembre 2009): 889–94. http://dx.doi.org/10.3758/mc.37.6.889.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
7

Hall, Michael D., Paula M. T. Smeele e Patricia K. Kuhl. "Integration of auditory and visual speech information". Journal of the Acoustical Society of America 103, n. 5 (maggio 1998): 2985. http://dx.doi.org/10.1121/1.421677.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
8

McGiverin, Rolland. "Speech, Hearing and Visual". Behavioral & Social Sciences Librarian 8, n. 3-4 (16 aprile 1990): 73–78. http://dx.doi.org/10.1300/j103v08n03_12.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
9

Hollich, George J., Peter W. Jusczyk e Rochelle S. Newman. "Infants use of visual information in speech segmentation". Journal of the Acoustical Society of America 110, n. 5 (novembre 2001): 2703. http://dx.doi.org/10.1121/1.4777318.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
10

Tekin, Ender, James Coughlan e Helen Simon. "Improving speech enhancement algorithms by incorporating visual information". Journal of the Acoustical Society of America 134, n. 5 (novembre 2013): 4237. http://dx.doi.org/10.1121/1.4831575.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri

Tesi sul tema "Visual speech information"

1

Le, Cornu Thomas. "Reconstruction of intelligible audio speech from visual speech information". Thesis, University of East Anglia, 2016. https://ueaeprints.uea.ac.uk/67012/.

Testo completo
Abstract (sommario):
The aim of the work conducted in this thesis is to reconstruct audio speech signals using information which can be extracted solely from a visual stream of a speaker's face, with application for surveillance scenarios and silent speech interfaces. Visual speech is limited to that which can be seen of the mouth, lips, teeth, and tongue, where the visual articulators convey considerably less information than in the audio domain, leading to the task being difficult. Accordingly, the emphasis is on the reconstruction of intelligible speech, with less regard given to quality. A speech production model is used to reconstruct audio speech, where methods are presented in this work for generating or estimating the necessary parameters for the model. Three approaches are explored for producing spectral-envelope estimates from visual features as this parameter provides the greatest contribution to speech intelligibility. The first approach uses regression to perform the visual-to-audio mapping, and then two further approaches are explored using vector quantisation techniques and classification models, with long-range temporal information incorporated at the feature and model-level. Excitation information, namely fundamental frequency and aperiodicity, is generated using artificial methods and joint-feature clustering approaches. Evaluations are first performed using mean squared error analyses and objective measures of speech intelligibility to refine the various system configurations, and then subjective listening tests are conducted to determine word-level accuracy, giving real intelligibility scores, of reconstructed speech. The best performing visual-to-audio domain mapping approach, using a clustering-and-classification framework with feature-level temporal encoding, is able to achieve audio-only intelligibility scores of 77 %, and audiovisual intelligibility scores of 84 %, on the GRID dataset. Furthermore, the methods are applied to a larger and more continuous dataset, with less favourable results, but with the belief that extensions to the work presented will yield a further increase in intelligibility.
Gli stili APA, Harvard, Vancouver, ISO e altri
2

Andrews, Brandie. "Auditory and visual information facilitating speech integration". Connect to resource, 2007. http://hdl.handle.net/1811/25202.

Testo completo
Abstract (sommario):
Thesis (Honors)--Ohio State University, 2007.
Title from first page of PDF file. Document formatted into pages: contains 43 p.; also includes graphics. Includes bibliographical references (p. 27-28). Available online via Ohio State University's Knowledge Bank.
Gli stili APA, Harvard, Vancouver, ISO e altri
3

Fixmer, Eric Norbert Charles. "Grouping of auditory and visual information in speech". Thesis, University of Cambridge, 2008. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.612553.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
4

Keintz, Constance Kay. "Influence of visual information on the intelligibility of dysarthric speech". Diss., The University of Arizona, 2005. http://hdl.handle.net/10150/280714.

Testo completo
Abstract (sommario):
Purpose. The purpose of this study was to examine the influence of visual information on the intelligibility of dysarthric speech. The two research questions posed by this study were: (1) Does the presentation mode (auditory-only versus auditory-visual) influence the intelligibility of a homogeneous group of speakers with dysarthria? and (2) Does the experience of the listener (experienced versus inexperienced with dysarthric speech) influence the intelligibility scores of these speakers? Background. Investigations of speakers with hearing impairment and laryngectomy have indicated that intelligibility scores are higher in an auditory-visual mode compared to an auditory-only mode of presentation. Studies of speakers with dysarthria have resulted in mixed findings. Methodological issues such as heterogeneity of speaker groups and factors related to the stimuli may have contributed to these mixed findings. Method. Eight speakers with dysarthria related to Parkinson disease were audio and video tape-recorded reading sentences. Movie files were created in which an auditory-only condition containing the speaker's voice but no visual image of the speaker and an auditory-visual condition containing the speaker's voice and a view of his/her face. Two groups of listeners (experienced and inexperienced with dysarthric speech) completed listening sessions in which they listened to (auditory-only) and watched and listened to (auditory-visual) the movies and transcribed what they heard each speaker say. Results. Although auditory-visual scores were significantly higher than auditory-only intelligibility scores, the difference between these scores was influenced by the order in which the two conditions were presented. A speaker effect was found across presentation modes, with less intelligible speakers demonstrating greater benefit from the inclusion of visual information. No statistically significant difference was found between the two listener groups in this study. Conclusions. These findings suggest that clinicians should include assessment of both auditory-only and auditory-visual intelligibility measures in speakers with Parkinson disease. Management of intelligibility impairment in these individuals should consider whether visual information is beneficial to listeners.
Gli stili APA, Harvard, Vancouver, ISO e altri
5

Hagrot, Joel. "A Data-Driven Approach For Automatic Visual Speech In Swedish Speech Synthesis Applications". Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-246393.

Testo completo
Abstract (sommario):
This project investigates the use of artificial neural networks for visual speech synthesis. The objective was to produce a framework for animated chat bots in Swedish. A survey of the literature on the topic revealed that the state-of-the-art approach was using ANNs with either audio or phoneme sequences as input. Three subjective surveys were conducted, both in the context of the final product, and in a more neutral context with less post-processing. They compared the ground truth, captured using the deep-sensing camera of the iPhone X, against both the ANN model and a baseline model. The statistical analysis used mixed effects models to find any statistically significant differences. Also, the temporal dynamics and the error were analyzed. The results show that a relatively simple ANN was capable of learning a mapping from phoneme sequences to blend shape weight sequences with satisfactory results, except for the fact that certain consonant requirements were unfulfilled. The issues with certain consonants were also observed in the ground truth, to some extent. Post-processing with consonant-specific overlays made the ANN’s animations indistinguishable from the ground truth and the subjects perceived them as more realistic than the baseline model’s animations. The ANN model proved useful in learning the temporal dynamics and coarticulation effects for vowels, but may have needed more data to properly satisfy the requirements of certain consonants. For the purposes of the intended product, these requirements can be satisfied using consonant-specific overlays.
Detta projekt utreder hur artificiella neuronnät kan användas för visuell talsyntes. Ändamålet var att ta fram ett ramverk för animerade chatbotar på svenska. En översikt över litteraturen kom fram till att state-of-the-art-metoden var att använda artificiella neuronnät med antingen ljud eller fonemsekvenser som indata. Tre enkäter genomfördes, både i den slutgiltiga produktens kontext, samt i en mer neutral kontext med mindre bearbetning. De jämförde sanningsdatat, inspelat med iPhone X:s djupsensorkamera, med både neuronnätsmodellen och en grundläggande så kallad baselinemodell. Den statistiska analysen använde mixed effects-modeller för att hitta statistiskt signifikanta skillnader i resultaten. Även den temporala dynamiken analyserades. Resultaten visar att ett relativt enkelt neuronnät kunde lära sig att generera blendshapesekvenser utifrån fonemsekvenser med tillfredsställande resultat, förutom att krav såsom läppslutning för vissa konsonanter inte alltid uppfylldes. Problemen med konsonanter kunde också i viss mån ses i sanningsdatat. Detta kunde lösas med hjälp av konsonantspecifik bearbetning, vilket gjorde att neuronnätets animationer var oskiljbara från sanningsdatat och att de samtidigt upplevdes vara bättre än baselinemodellens animationer. Sammanfattningsvis så lärde sig neuronnätet vokaler väl, men hade antagligen behövt mer data för att på ett tillfredsställande sätt uppfylla kraven för vissa konsonanter. För den slutgiltiga produktens skull kan dessa krav ändå uppnås med hjälp av konsonantspecifik bearbetning.
Gli stili APA, Harvard, Vancouver, ISO e altri
6

Bergmann, Kirsten, e Stefan Kopp. "Verbal or visual? : How information is distributed across speech and gesture in spatial dialog". Universität Potsdam, 2006. http://opus.kobv.de/ubp/volltexte/2006/1037/.

Testo completo
Abstract (sommario):
In spatial dialog like in direction giving humans make frequent use of speechaccompanying gestures. Some gestures convey largely the same information as speech while others complement speech.
This paper reports a study on how speakers distribute meaning across speech and gesture, and depending on what factors. Utterance meaning and the wider dialog context were tested by statistically analyzing a corpus of direction-giving dialogs. Problems of speech production (as indicated by discourse markers and disfluencies), the communicative goals, and the information status were found to be influential, while feedback signals by the addressee do not have any influence.
Gli stili APA, Harvard, Vancouver, ISO e altri
7

Erdener, Vahit Doğu. "The effect of auditory, visual and orthographic information on second language acquisition /". View thesis View thesis, 2002. http://library.uws.edu.au/adt-NUWS/public/adt-NUWS20030408.114825/index.html.

Testo completo
Abstract (sommario):
Thesis (MA (Hons)) -- University of Western Sydney, 2002.
"A thesis submitted in partial fulfillment of the requirements for the degree of Masters of Arts (Honours), MARCS Auditory Laboratories & School of Psychology, University of Western Sydney, May 2002" Bibliography : leaves 83-93.
Gli stili APA, Harvard, Vancouver, ISO e altri
8

Patterson, Robert W. "The effects of inaccurate speech information on performance in a visual search and identification task". Thesis, Georgia Institute of Technology, 1987. http://hdl.handle.net/1853/30481.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
9

Erdener, Vahit Dogu, University of Western Sydney, of Arts Education and Social Sciences College e School of Psychology. "The effect of auditory, visual and orthographic information on second language acquisition". THESIS_CAESS_PSY_Erdener_V.xml, 2002. http://handle.uws.edu.au:8081/1959.7/685.

Testo completo
Abstract (sommario):
The current study investigates the effect of auditory and visual speech information and orthographic information on second/foreign language (L2) acquisition. To test this, native speakers of Turkish (a language with a transparent orthography) and native speakers of Australian English (a language with an opaque orthography) were exposed to Spanish (transparent orthography) and Irish (opaque orthography) legal non-word items in four experimental conditions: auditory-only, auditory-visual, auditory-orthographic, and auditory-visual-orthographic. On each trial, Turkish and Australian English speakers were asked to produce each Spanish and Irish legal non-words. In terms of phoneme errors it was found that Turkish participants generally made less errors in Spanish than their Australian counterparts, and visual speech information generally facilitated performance. Orthographic information had an overriding effect such that there was no visual advantage once it was provided. In the orthographic conditions, Turkish speakers performed better than their Australian English counterparts with Spanish items and worse with Irish terms. In terms of native speakers' ratings of participants' productions, it was found that orthographic input improved accent. Overall the results confirm findings that visual information enhances speech production in L2 and additionally show the facilitative effects of orthographic input in L2 acquisition as a function of orthographic depth. Inter-rater reliability measures revealed that the native speaker rating procedure may be prone to individual and socio-cultural influences that may stem from internal criteria for native accents. This suggests that native speaker ratings should be treated with caution.
Master of Arts (Hons)
Gli stili APA, Harvard, Vancouver, ISO e altri
10

Ostroff, Wendy Louise. "Non-linguistic Influences on Infants' Nonnative Phoneme Perception: Exaggerated prosody and Visual Speech Information Aid Discrimination". Diss., Virginia Tech, 2000. http://hdl.handle.net/10919/27640.

Testo completo
Abstract (sommario):
Research indicates that infants lose the capacity to perceive distinctions in nonnative sounds as they become sensitive to the speech sounds of their native language (i.e., by 10- to 12-months of age). However, investigations into the decline in nonnative phonetic perception have neglected to examine the role of non-linguistic information. Exaggerated prosodic intonation and facial input are prominent in the infantsâ language-learning environment, and both have been shown to ease the task of speech perception. The current investigation was designed to examine the impact of infant-directed (ID) speech and facial input on infantsâ ability to discriminate phonemes that do not contrast in their native language. Specifically, 11-month-old infants were tested for discrimination of both a native phoneme contrast and a nonnative phoneme contrast across four conditions, including an auditory manipulation (ID speech vs. AD speech) and a visual manipulation (Face vs. Geometric Form). The results indicated that infants could discriminate the native phonemes across any of the four conditions. Furthermore, the infants could discriminate the nonnative phonemes if they had enhanced auditory and visual information available to them (i.e., if they were presented in ID speech with a synchronous facial display), and if the nonnative discrimination task was the infantsâ first test session. These results suggest that infants do not lose the capacity to discriminate nonnative phonemes by the end of the first postnatal year, but that they rely on certain language-relevant and non-linguistic sources of information to discriminate nonnative sounds.
Ph. D.
Gli stili APA, Harvard, Vancouver, ISO e altri

Libri sul tema "Visual speech information"

1

Massaro, Dominic W. Speech perception by ear and eye: A paradigm for psychological inquiry. Hillsdale, N.J: Erlbaum Associates, 1987.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
2

Wayne, Cranton, Fihn Mark e SpringerLink (Online service), a cura di. Handbook of Visual Display Technology. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
3

Learning disabilities sourcebook: Basic consumer health information about dyslexia, dyscalculia, dysgraphia, speech and communication disorders, auditory and visual processing disorders, and other conditions that make learning difficult, including attention deficit hyperactivity disorder, down syndrome and other chromosomal disorders, fetal alcohol spectrum disorders, hearing and visual impairment, autism and other pervasive developmental disorders, and traumatic brain Injury; along with facts about diagnosing learning disabilities, early intervention, the special education process, legal protections, assistive technology, and accommodations, and guidelines for life-stage transitions, suggestions for coping with daily challenges, a glossary of related terms, and a directory of additional resources. 4a ed. Detroit, MI: Omnigraphics, 2012.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
4

Simpson, Jeffry A., e Dominic W. Massaro. Speech Perception by Ear and Eye: A Paradigm for Psychological Inquiry. Taylor & Francis Group, 2016.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
5

Massaro, Dominic W. Speech Perception by Ear and Eye: A Paradigm for Psychological Inquiry. Taylor & Francis Group, 2014.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
6

Massaro, Dominic W. Speech Perception by Ear and Eye: A Paradigm for Psychological Inquiry. Taylor & Francis Group, 2014.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
7

Simpson, Jeffry A., e Dominic W. Massaro. Speech Perception by Ear and Eye: A Paradigm for Psychological Inquiry. Taylor & Francis Group, 2014.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
8

Simpson, Jeffry A., e Dominic W. Massaro. Speech Perception by Ear and Eye: A Paradigm for Psychological Inquiry. Taylor & Francis Group, 2014.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
9

Massaro, Dominic W. Speech Perception by Ear and Eye: A Paradigm for Psychological Inquiry. Taylor & Francis Group, 2014.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
10

Massaro, Dominic W. Speech Perception by Ear and Eye: A Paradigm for Psychological Inquiry. Taylor & Francis Group, 2014.

Cerca il testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri

Capitoli di libri sul tema "Visual speech information"

1

Pondit, Ashish, Muhammad Eshaque Ali Rukon, Anik Das e Muhammad Ashad Kabir. "BenAV: a Bengali Audio-Visual Corpus for Visual Speech Recognition". In Neural Information Processing, 526–35. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-92270-2_45.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
2

Gupta, Deepika, Preety Singh, V. Laxmi e Manoj S. Gaur. "Boundary Descriptors for Visual Speech Recognition". In Computer and Information Sciences II, 307–13. London: Springer London, 2011. http://dx.doi.org/10.1007/978-1-4471-2155-8_39.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
3

Giachanou, Anastasia, Guobiao Zhang e Paolo Rosso. "Multimodal Fake News Detection with Textual, Visual and Semantic Information". In Text, Speech, and Dialogue, 30–38. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-58323-1_3.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
4

Wu, Shiow-yang, e Wen-Shen Chen. "Oral-Query-by-Sketch: An XML-Based Framework for Speech Access to Image Databases". In Visual and Multimedia Information Management, 341–55. Boston, MA: Springer US, 2002. http://dx.doi.org/10.1007/978-0-387-35592-4_24.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
5

Sagheer, Alaa, e Saleh Aly. "Integration of Face Detection and User Identification with Visual Speech Recognition". In Neural Information Processing, 479–87. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-34500-5_57.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
6

Gui, Jiaping, e Shilin Wang. "Shape Feature Analysis for Visual Speech and Speaker Recognition". In Communications in Computer and Information Science, 167–74. Berlin, Heidelberg: Springer Berlin Heidelberg, 2011. http://dx.doi.org/10.1007/978-3-642-23235-0_22.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
7

Nakamura, Satoshi. "Fusion of Audio-Visual Information for Integrated Speech Processing". In Lecture Notes in Computer Science, 127–43. Berlin, Heidelberg: Springer Berlin Heidelberg, 2001. http://dx.doi.org/10.1007/3-540-45344-x_20.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
8

Foo, Say Wei, e Liang Dong. "Recognition of Visual Speech Elements Using Hidden Markov Models". In Advances in Multimedia Information Processing — PCM 2002, 607–14. Berlin, Heidelberg: Springer Berlin Heidelberg, 2002. http://dx.doi.org/10.1007/3-540-36228-2_75.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
9

Bastanfard, Azam, Mohammad Aghaahmadi, Alireza Abdi kelishami, Maryam Fazel e Maedeh Moghadam. "Persian Viseme Classification for Developing Visual Speech Training Application". In Advances in Multimedia Information Processing - PCM 2009, 1080–85. Berlin, Heidelberg: Springer Berlin Heidelberg, 2009. http://dx.doi.org/10.1007/978-3-642-10467-1_104.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
10

Galuščáková, Petra, Pavel Pecina e Jan Hajič. "Penalty Functions for Evaluation Measures of Unsegmented Speech Retrieval". In Information Access Evaluation. Multilinguality, Multimodality, and Visual Analytics, 100–111. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-33247-0_12.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri

Atti di convegni sul tema "Visual speech information"

1

Akdemir, Eren, e Tolga Ciloglu. "Using visual information in automatic speech segmentation". In 2008 IEEE 16th Signal Processing, Communication and Applications Conference (SIU). IEEE, 2008. http://dx.doi.org/10.1109/siu.2008.4632641.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
2

Kawase, Saya, Jeesun Kim, Vincent Aubanel e Chris Davis. "Perceiving foreign-accented auditory-visual speech in noise: The influence of visual form and timing information". In Speech Prosody 2016. ISCA, 2016. http://dx.doi.org/10.21437/speechprosody.2016-99.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
3

Chen, Tsuhan, H. P. Graf, Homer H. Chen, Wu Chou, Barry G. Haskell, Eric D. Petajan e Yao Wang. "Lip synchronization in talking head video utilizing speech information". In Visual Communications and Image Processing '95, a cura di Lance T. Wu. SPIE, 1995. http://dx.doi.org/10.1117/12.206706.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
4

Jiang, Wei, Lexing Xie e Shih-Fu Chang. "Visual saliency with side information". In ICASSP 2009 - 2009 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 2009. http://dx.doi.org/10.1109/icassp.2009.4959946.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
5

Karabalkan, H., e H. Erdogan. "Information fusion techniques in Audio-Visual Speech Recognition". In 2009 IEEE 17th Signal Processing and Communications Applications Conference (SIU). IEEE, 2009. http://dx.doi.org/10.1109/siu.2009.5136443.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
6

Maulana, Muhammad Rizki Aulia Rahman, e Mohamad Ivan Fanany. "Indonesian audio-visual speech corpus for multimodal automatic speech recognition". In 2017 International Conference on Advanced Computer Science and Information Systems (ICACSIS). IEEE, 2017. http://dx.doi.org/10.1109/icacsis.2017.8355062.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
7

Aubanel, Vincent, Cassandra Masters, Jeesun Kim e Chris Davis. "Contribution of visual rhythmic information to speech perception in noise". In The 14th International Conference on Auditory-Visual Speech Processing. ISCA: ISCA, 2017. http://dx.doi.org/10.21437/avsp.2017-18.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
8

Hui Zhao e Chaojing Tang. "Visual speech synthesis based on Chinese dynamic visemes". In 2008 International Conference on Information and Automation (ICIA). IEEE, 2008. http://dx.doi.org/10.1109/icinfa.2008.4607983.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
9

Luo, Yiyu, Jing Wang, Xinyao Wang, Liang Wen e Lizhong Wang. "Audio-Visual Speech Separation Using I-Vectors". In 2019 IEEE 2nd International Conference on Information Communication and Signal Processing (ICICSP). IEEE, 2019. http://dx.doi.org/10.1109/icicsp48821.2019.8958547.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
10

Lu, Longbin, Xinman Zhang e Xuebin Xu. "Fusion of face and visual speech information for identity verification". In 2017 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS). IEEE, 2017. http://dx.doi.org/10.1109/ispacs.2017.8266530.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri

Rapporti di organizzazioni sul tema "Visual speech information"

1

Yatsymirska, Mariya. SOCIAL EXPRESSION IN MULTIMEDIA TEXTS. Ivan Franko National University of Lviv, febbraio 2021. http://dx.doi.org/10.30970/vjo.2021.49.11072.

Testo completo
Abstract (sommario):
The article investigates functional techniques of extralinguistic expression in multimedia texts; the effectiveness of figurative expressions as a reaction to modern events in Ukraine and their influence on the formation of public opinion is shown. Publications of journalists, broadcasts of media resonators, experts, public figures, politicians, readers are analyzed. The language of the media plays a key role in shaping the worldview of the young political elite in the first place. The essence of each statement is a focused thought that reacts to events in the world or in one’s own country. The most popular platform for mass information and social interaction is, first of all, network journalism, which is characterized by mobility and unlimited time and space. Authors have complete freedom to express their views in direct language, including their own word formation. Phonetic, lexical, phraseological and stylistic means of speech create expression of the text. A figurative word, a good aphorism or proverb, a paraphrased expression, etc. enhance the effectiveness of a multimedia text. This is especially important for headlines that simultaneously inform and influence the views of millions of readers. Given the wide range of issues raised by the Internet as a medium, research in this area is interdisciplinary. The science of information, combining language and social communication, is at the forefront of global interactions. The Internet is an effective source of knowledge and a forum for free thought. Nonlinear texts (hypertexts) – «branching texts or texts that perform actions on request», multimedia texts change the principles of information collection, storage and dissemination, involving billions of readers in the discussion of global issues. Mastering the word is not an easy task if the author of the publication is not well-read, is not deep in the topic, does not know the psychology of the audience for which he writes. Therefore, the study of media broadcasting is an important component of the professional training of future journalists. The functions of the language of the media require the authors to make the right statements and convincing arguments in the text. Journalism education is not only knowledge of imperative and dispositive norms, but also apodictic ones. In practice, this means that there are rules in media creativity that are based on logical necessity. Apodicticity is the first sign of impressive language on the platform of print or electronic media. Social expression is a combination of creative abilities and linguistic competencies that a journalist realizes in his activity. Creative self-expression is realized in a set of many important factors in the media: the choice of topic, convincing arguments, logical presentation of ideas and deep philological education. Linguistic art, in contrast to painting, music, sculpture, accumulates all visual, auditory, tactile and empathic sensations in a universal sign – the word. The choice of the word for the reproduction of sensory and semantic meanings, its competent use in the appropriate context distinguishes the journalist-intellectual from other participants in forums, round tables, analytical or entertainment programs. Expressive speech in the media is a product of the intellect (ability to think) of all those who write on socio-political or economic topics. In the same plane with him – intelligence (awareness, prudence), the first sign of which (according to Ivan Ogienko) is a good knowledge of the language. Intellectual language is an important means of organizing a journalistic text. It, on the one hand, logically conveys the author’s thoughts, and on the other – encourages the reader to reflect and comprehend what is read. The richness of language is accumulated through continuous self-education and interesting communication. Studies of social expression as an important factor influencing the formation of public consciousness should open up new facets of rational and emotional media broadcasting; to trace physical and psychological reactions to communicative mimicry in the media. Speech mimicry as one of the methods of disguise is increasingly becoming a dangerous factor in manipulating the media. Mimicry is an unprincipled adaptation to the surrounding social conditions; one of the most famous examples of an animal characterized by mimicry (change of protective color and shape) is a chameleon. In a figurative sense, chameleons are called adaptive journalists. Observations show that mimicry in politics is to some extent a kind of game that, like every game, is always conditional and artificial.
Gli stili APA, Harvard, Vancouver, ISO e altri
Offriamo sconti su tutti i piani premium per gli autori le cui opere sono incluse in raccolte letterarie tematiche. Contattaci per ottenere un codice promozionale unico!

Vai alla bibliografia