Dissertations / Theses on the topic 'Visual speech information'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 30 dissertations / theses for your research on the topic 'Visual speech information.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Le, Cornu Thomas. "Reconstruction of intelligible audio speech from visual speech information." Thesis, University of East Anglia, 2016. https://ueaeprints.uea.ac.uk/67012/.
Full textAndrews, Brandie. "Auditory and visual information facilitating speech integration." Connect to resource, 2007. http://hdl.handle.net/1811/25202.
Full textTitle from first page of PDF file. Document formatted into pages: contains 43 p.; also includes graphics. Includes bibliographical references (p. 27-28). Available online via Ohio State University's Knowledge Bank.
Fixmer, Eric Norbert Charles. "Grouping of auditory and visual information in speech." Thesis, University of Cambridge, 2008. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.612553.
Full textKeintz, Constance Kay. "Influence of visual information on the intelligibility of dysarthric speech." Diss., The University of Arizona, 2005. http://hdl.handle.net/10150/280714.
Full textHagrot, Joel. "A Data-Driven Approach For Automatic Visual Speech In Swedish Speech Synthesis Applications." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-246393.
Full textDetta projekt utreder hur artificiella neuronnät kan användas för visuell talsyntes. Ändamålet var att ta fram ett ramverk för animerade chatbotar på svenska. En översikt över litteraturen kom fram till att state-of-the-art-metoden var att använda artificiella neuronnät med antingen ljud eller fonemsekvenser som indata. Tre enkäter genomfördes, både i den slutgiltiga produktens kontext, samt i en mer neutral kontext med mindre bearbetning. De jämförde sanningsdatat, inspelat med iPhone X:s djupsensorkamera, med både neuronnätsmodellen och en grundläggande så kallad baselinemodell. Den statistiska analysen använde mixed effects-modeller för att hitta statistiskt signifikanta skillnader i resultaten. Även den temporala dynamiken analyserades. Resultaten visar att ett relativt enkelt neuronnät kunde lära sig att generera blendshapesekvenser utifrån fonemsekvenser med tillfredsställande resultat, förutom att krav såsom läppslutning för vissa konsonanter inte alltid uppfylldes. Problemen med konsonanter kunde också i viss mån ses i sanningsdatat. Detta kunde lösas med hjälp av konsonantspecifik bearbetning, vilket gjorde att neuronnätets animationer var oskiljbara från sanningsdatat och att de samtidigt upplevdes vara bättre än baselinemodellens animationer. Sammanfattningsvis så lärde sig neuronnätet vokaler väl, men hade antagligen behövt mer data för att på ett tillfredsställande sätt uppfylla kraven för vissa konsonanter. För den slutgiltiga produktens skull kan dessa krav ändå uppnås med hjälp av konsonantspecifik bearbetning.
Bergmann, Kirsten, and Stefan Kopp. "Verbal or visual? : How information is distributed across speech and gesture in spatial dialog." Universität Potsdam, 2006. http://opus.kobv.de/ubp/volltexte/2006/1037/.
Full textThis paper reports a study on how speakers distribute meaning across speech and gesture, and depending on what factors. Utterance meaning and the wider dialog context were tested by statistically analyzing a corpus of direction-giving dialogs. Problems of speech production (as indicated by discourse markers and disfluencies), the communicative goals, and the information status were found to be influential, while feedback signals by the addressee do not have any influence.
Erdener, Vahit Doğu. "The effect of auditory, visual and orthographic information on second language acquisition /." View thesis View thesis, 2002. http://library.uws.edu.au/adt-NUWS/public/adt-NUWS20030408.114825/index.html.
Full text"A thesis submitted in partial fulfillment of the requirements for the degree of Masters of Arts (Honours), MARCS Auditory Laboratories & School of Psychology, University of Western Sydney, May 2002" Bibliography : leaves 83-93.
Patterson, Robert W. "The effects of inaccurate speech information on performance in a visual search and identification task." Thesis, Georgia Institute of Technology, 1987. http://hdl.handle.net/1853/30481.
Full textErdener, Vahit Dogu, University of Western Sydney, of Arts Education and Social Sciences College, and School of Psychology. "The effect of auditory, visual and orthographic information on second language acquisition." THESIS_CAESS_PSY_Erdener_V.xml, 2002. http://handle.uws.edu.au:8081/1959.7/685.
Full textMaster of Arts (Hons)
Ostroff, Wendy Louise. "Non-linguistic Influences on Infants' Nonnative Phoneme Perception: Exaggerated prosody and Visual Speech Information Aid Discrimination." Diss., Virginia Tech, 2000. http://hdl.handle.net/10919/27640.
Full textPh. D.
Abdalla, Marwa. "Can participants extract subtle information from gesturelike visual stimuli that are coordinated with speech without using any other cues?" Thesis, University of Iowa, 2012. https://ir.uiowa.edu/etd/2805.
Full textAlharbi, Saad T. "Graphical and Non-speech Sound Metaphors in Email Browsing: An Empirical Approach. A Usability Based Study Investigating the Role of Incorporating Visual and Non-Speech Sound Metaphors to Communicate Email Data and Threads." Thesis, University of Bradford, 2009. http://hdl.handle.net/10454/4244.
Full textTaibah University in Medina and the Ministry of Higher Education in Saudi Arabia.
Alharbi, Saad Talal. "Graphical and non-speech sound metaphors in email browsing : an empirical approach : a usability based study investigating the role of incorporating visual and non-speech sound metaphors to communicate email data and threads." Thesis, University of Bradford, 2009. http://hdl.handle.net/10454/4244.
Full textKühnapfel, Thorsten. "Audio networks for speech enhancement and indexing." Thesis, Curtin University, 2009. http://hdl.handle.net/20.500.11937/206.
Full textNavarathna, Rajitha Dharshana Bandara. "Robust recognition of human behaviour in challenging environments." Thesis, Queensland University of Technology, 2014. https://eprints.qut.edu.au/66235/1/Rajitha%20Dharshana%20Bandara_Navarathna_Thesis.pdf.
Full textPonto, Jessica J. "Speech is a Mouth, Text is a Body." Miami University / OhioLINK, 2008. http://rave.ohiolink.edu/etdc/view?acc_num=miami1218076653.
Full textKalantari, Shahram. "Improving spoken term detection using complementary information." Thesis, Queensland University of Technology, 2015. https://eprints.qut.edu.au/90074/1/Shahram_Kalantari_Thesis.pdf.
Full textFong, Katherine KaYan. "IR-Depth Face Detection and Lip Localization Using Kinect V2." DigitalCommons@CalPoly, 2015. https://digitalcommons.calpoly.edu/theses/1425.
Full textVerma, Prabhat [Verfasser]. "Speech as Interface in Web Applications for Visually Challenged / Prabhat Verma." Munich : GRIN Verlag, 2015. http://d-nb.info/1097585689/34.
Full textCeder, Maria, and Camilla Hellström. "Det maskerande brusljudets påverkan på inlärningen av visuell information : om effekten av maskerande brusljud i öppna kontorslandskap." Thesis, Högskolan i Gävle, Avdelningen för socialt arbete och psykologi, 2012. http://urn.kb.se/resolve?urn=urn:nbn:se:hig:diva-12404.
Full textThis study examined if a masking white noise on irrelevant speech affects the encoding of visual information. An experiment was carried out in a laboratory with 32 participants. The participants were presented to a series of written words and were prompted to recall these words in any order. While the participants studied the written words, irrelevant speech from the same semantic category was presented with or without a masking noise. The participants were told to ignore the irrelevant speech. The results of this study showed that the number of intrusions from the irrelevant speech decreases and the number of recalled written words increases when the irrelevant speech is masked by a white noise compared to irrelevant speech without a masking noise. The findings of this study could be applied in the acoustic design of open-plan offices where cognitive tasks, such as reading comprehension and proofreading, are performed in a noisy environment. A white noise can reduce the intelligibility of office noise and irrelevant speech, which have positive effect on work performance.
Bannani, Rouaissia Sabrina. "Pour une prise en charge des difficultés de la compréhension orale en FLE : cas des collégiens tunisiens issus des milieux défavorisés." Thesis, Aix-Marseille, 2018. http://www.theses.fr/2018AIXM0466.
Full textThis research fits in the field of oral didactics and it aims to study the verbal interactions in the classes of Tunisian middle schools, for learners in difficulty, coming from underprivileged backgrounds.Despite of the efforts, invested by the teachers, and because they are individual and conceived anarchically, they are vain and the demotivation is of such a magnitude that it inhibits any act of learning however small it can be.Teachers are now aware of the need to develop the oral skill for these learners in difficulty, considering it as accessible but they sometimes forget that unlike the privileged areas, the majority of learners from underprivileged areas practice speaking a foreign language which they never use outside of school.What do the FFL methodologies tangibly offer, for teaching oral skills, taking into account students in difficulty?How to train students in oral comprehension and production skills, given the particular context of FFL classes in underprivileged areas?What support plan is proposed to prevent failure and to bring learners in difficulty to have objective and positive representations, vis-à-vis themselves, on one hand, the school alike learning in general, and French in particular on the other hand?In that way, we seek to determine which context can favor the emancipation of learners in difficulty by their speaking up in order to contribute to the oral didactics field and to propose some didactic paths that would make these learners active in the classroom, thus giving them the opportunity to prove their existence through participation, on one hand, and commitment to building the knowledge that is taught to them on the other hand
Simoncini, Claudio. "Intégration spatio-temporelle de l'information visuelle pour les mouvements oculaires et la perception : =Spatio-temporal integration of visual information for eye movements and perception." Thesis, Aix-Marseille, 2013. http://www.theses.fr/2013AIXM5065/document.
Full textWe focused on the impact of the statistical distributions of visual information on these various behavioral responses. We asked first how motion information is integrated to estimate speed in order to perform either a speed discrimination task or to control reflexive tracking eye movements. Next, we investigated how spatial distribution in textures affects both pattern recognition and fixational eye movements. To do so, we used a set of artificial stimuli that are naturalistic textures where we can maintain a tight control on their information contents as for instance their spatio-temporal frequency bandwidth. The first studies compared speed information decoding for ocular following eye movements and perceptual speed discrimination. We found a strong dissociation where ocular following take full advantage by the enlargement of the spatio-temporal frequency bandwidth while perceptual speed discrimination is largely impaired for large bandwidth stimuli. Such dissociation remains over a large temporal integration window. We propose an adaptive gain control mechanism to explain this opposite dependencies. The second series of experimental studies investigate the properties of fixation eye movements (microsaccade and saccade) as a function of the mean and variance of the spatial frequency content of visual static textures. We show that several characteristics of fixational saccades (location, direction and amplitude) varied systematically with the distribution of spatial frequencies. The spatial distribution of the fixation zones could be best predicted from the saliency maps of the stimuli
Hill, Brian, and 廖峻廷. "Robust Speech Recognition Integrating Visual Information." Thesis, 1997. http://ndltd.ncl.edu.tw/handle/97538191028447078081.
Full textKao, Jen-ching, and 高仁璟. "Effects of Audio-visual Information on the Intelligibility of Esophageal Speech." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/64143533157423675383.
Full text國立臺北護理健康大學
語言治療與聽力研究所
104
The purpose of this study was to determine the effects visual information on speech intelligibility for esophageal speakers, and further examining the influence of the degree of auditory speech intelligibility of the speaker on the effect of visual information. Besides, to invetigate the role of visual information on speech percetion, intelligibility scores for phonemes after visual information added were compared. The subjects of this study were 6 esophageal speakers and 60 listeners. Speakers were divided into two groups (3 per group), a Good esophgeal group whose auditoty speech intelligibility above 85%, and a Moderate group whose auditory speech intelligibility between 50%~75%. Speakers were recorded while they read sentences. Listeners transcribed sentences while watching and listening to videotapes of the speakers (audio-visual mode) and while only listening to the speakers (auditory-only mode). Scores of sentence intelligibility and phoneme intelligibility were determined baesd on listeners transcription. The results showed a statistically significant higher sentence intelligibility score for audio-visual mode compared to audio-only mode, as well as a significant interaction effect between mode of presentation and degree of auditory speech intelligibility. Within degree of auditory speech intelligibility, Good esophgeal group showed significantly greater benefit from the inclusion of visual information compared to Moderate esophgeal group. Besides, 17 out of 21 phonemes were benefited signigicantly from the inclusion of visual cues.Significant difference was found between 7 articulation places, and the greatest improvement was found for bilabial sounds. No significant difference was found between 7 articulation manners. The finding suggest that facial visual informataion increase the intelligibility of esophageal speech, and auditory speech intelligibility is an important variable that the less intelligible speakers’ scores can increase more when visual information was added. Besides, audio-visual processing is more effective than auditory processing for two reasons. First, some articulatory movement can be seen clearly by the mouth. Second, the audible and visible patterns are highly correlated. Many features of an utterance can be seen by the facial information.
Erdener, Vahit Dogu, University of Western Sydney, College of Arts, and School of Psychology. "Development of auditory-visual speech perception in young children." 2007. http://handle.uws.edu.au:8081/1959.7/13783.
Full textDoctor of Philosophy (PhD)
Erdener, Dogu. "Development of auditory-visual speech perception in young children." Thesis, 2007. http://handle.uws.edu.au:8081/1959.7/13783.
Full textErdener, Dogu. "The effect of auditory, visual and orthographic information on second language acquisition." Thesis, 2002. http://handle.uws.edu.au:8081/1959.7/685.
Full textLapchak, Marion Cone. "Exploring the effects of age, early-onset otitis media, and articulation errors on the integration of auditory and visual information in speech perception /." Diss., 2005. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:3188497.
Full textParsons, Brendan. "The neuroscience of cognitive enhancement : enhanced attention, working memory and visual information processing speed using 3D-MOT." Thèse, 2015. http://hdl.handle.net/1866/16316.
Full textCognitive enhancement is a domain of burgeoning interest in many domains including neuropsychology. While there are different methods that exist in order to achieve cognitive enhancement, there are few that are supported by research. The current work examines the state of cognitive enhancement interventions. It first outlines the weaknesses observed in these practices and then proposes a standard template for assessing cognitive enhancement tools. A research study is then presented that examines a novel cognitive enhancement tool, 3-dimensional multiple object tracking (3D-MOT), and weighs the current evidence for 3D-MOT against the proposed standard template. The results of the current work demonstrate that 3D-MOT is effective in enhancing attention, working memory and visual information processing speed, and represent a first step toward establishing 3D-MOT as a cognitive enhancement tool.
CHEN, TING-YU, and 陳亭妤. "The evaluation and design of the visual information discrimination for the transportation tickets : A case of Taiwan High Speed Rail." Thesis, 2018. http://ndltd.ncl.edu.tw/handle/rf77zf.
Full text逢甲大學
創意設計碩士學位學程
106
With the convenience and expansion of transport systems and the development trend of traffic operation requirements, plus the increase in foreign tourists in Taiwan tourism in recent years, the number of trains has gradually increased to meet more passengers. However, at the same time, travelers of many ages should also pay attention to and improve the design of their information content. In order to allow passengers to ride the train station accurately, the information on the transport ticket is also important. In view of this, this study uses the Taiwan high-speed rail transit ticket as a sample to carry out relevant discussions, user evaluation and design on the ticket. In the research process, the content analysis method was used to analyze the visual elements of the current high-speed rail vouchers with the graphic design theory; then the experience of different age groups for the use of the current high-speed rail tickets was obtained through questionnaire survey, and feedback was given. And use the Tobii Eye Tracker T120 eye tracker to do the task test, you can more accurately know the subjects in different situations, for the point of view of the current ticket; then, summarize the problems of the current ticket and design recommendations The improved design after finishing includes the visual architecture of the message framework and ticket design, and the task test of the second testee to check whether the problem of the current ticket is resolved. Finally, based on the analysis results, a new version of the high-speed railway transport ticket is proposed. , and its message architecture and interface visual design guidelines.