Dissertations / Theses on the topic 'Lipreading'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 45 dissertations / theses for your research on the topic 'Lipreading.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Lucey, Patrick Joseph. "Lipreading across multiple views." Thesis, Queensland University of Technology, 2007. https://eprints.qut.edu.au/16676/1/Patrick_Joseph_Lucey_Thesis.pdf.
Full textLucey, Patrick Joseph. "Lipreading across multiple views." Queensland University of Technology, 2007. http://eprints.qut.edu.au/16676/.
Full textMacLeod, A. "Effective methods for measuring lipreading skills." Thesis, University of Nottingham, 1987. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.233400.
Full textMacDermid, Catriona. "Lipreading and language processing by deaf children." Thesis, University of Surrey, 1991. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.291020.
Full textYuan, Hanfeng 1972. "Tactual display of consonant voicing to supplement lipreading." Thesis, Massachusetts Institute of Technology, 2003. http://hdl.handle.net/1721.1/87906.
Full textIncludes bibliographical references (p. 241-251).
This research is concerned with the development of tactual displays to supplement the information available through lipreading. Because voicing carries a high informational load in speech and is not well transmitted through lipreading, the efforts are focused on providing tactual displays of voicing to supplement the information available on the lips of the talker. This research includes exploration of 1) signal-processing schemes to extract information about voicing from the acoustic speech signal, 2) methods of displaying this information through a multi-finger tactual display, and 3) perceptual evaluations of voicing reception through the tactual display alone (T), lipreading alone (L), and the combined condition (L+T). Signal processing for the extraction of voicing information used amplitude-envelope signals derived from filtered bands of speech (i.e., envelopes derived from a lowpass-filtered band at 350 Hz and from a highpass-filtered band at 3000 Hz). Acoustic measurements made on the envelope signals of a set of 16 initial consonants represented through multiple tokens of C₁VC₂ syllables indicate that the onset-timing difference between the low- and high-frequency envelopes (EOA: envelope-onset asynchrony) provides a reliable and robust cue for distinguishing voiced from voiceless consonants. This acoustic cue was presented through a two-finger tactual display such that the envelope of the high-frequency band was used to modulate a 250-Hz carrier signal delivered to the index finger (250-I) and the envelope of the low-frequency band was used to modulate a 50-Hz carrier delivered to the thumb (50T).
(cont.) The temporal-onset order threshold for these two signals, measured with roving signal amplitude and duration, averaged 34 msec, sufficiently small for use of the EOA cue. Perceptual evaluations of the tactual display of EOA with speech signal indicated: 1) that the cue was highly effective for discrimination of pairs of voicing contrasts; 2) that the identification of 16 consonants was improved by roughly 15 percentage points with the addition of the tactual cue over L alone; and 3) that no improvements in L+T over L were observed for reception of words in sentences, indicating the need for further training on this task.
by Hanfeng Yuan.
Ph.D.
Chiou, Greg I. "Active contour models for distinct feature tracking and lipreading /." Thesis, Connect to this title online; UW restricted, 1995. http://hdl.handle.net/1773/6023.
Full textKaucic, Robert August. "Lip tracking for audio-visual speech recognition." Thesis, University of Oxford, 1997. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.360392.
Full textMatthews, Iain. "Features for audio-visual speech recognition." Thesis, University of East Anglia, 1998. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.266736.
Full textThangthai, Kwanchiva. "Computer lipreading via hybrid deep neural network hidden Markov models." Thesis, University of East Anglia, 2018. https://ueaeprints.uea.ac.uk/69215/.
Full textHiramatsu, Sandra. "Does lipreading help word reading? : an investigation of the relationship between visible speech and early reading achievement /." Thesis, Connect to this title online; UW restricted, 2005. http://hdl.handle.net/1773/7913.
Full textDivin, William. "The irrelevant speech effect, lipreading and theories of short-term memory." Thesis, University of Ulster, 1999. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.365401.
Full textAlness, Borg Axel, and Marcus Enström. "A study of the temporal resolution in lipreading using event vision." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-280325.
Full textMaskinell analys av visuella drag från läppar för att extrahera uttalade ord består av att hitta mönster i rörelser varpå maskin inlärning har använts i tidigare forskning för att hantera detta problem. I tidigare forskning har konventionella bildbaserade kameror använts med bra resultat. Att klassificera visuella drag är dyrt och att fånga precis tillräckligt med information kan vara viktigt. Eventkameror är en ny typ av kamera som är inspirerade av hur den mänskliga synen funkar och fångar bara förändringar i scenen och erbjuder väldigt hög temporär upplösning. I den här rapporten undersöker vi vikten av temporär upplösning vid läppläsning och om en eventkamera kan användas för läppläsning. En trend av initial ökning av noggrannhet som toppar på ett maximum för att sen minska när bildfrekvensen minskar kan observeras. Forskningen kan därför dra slutsatsen att när en bild baserad representation av eventdata används ökar den temporära upplösningen inte nödvändigtvis klassificeringsnoggrannheten. Det är däremot svårt att vara säker på den här slutsatsen eftersom det finns för många parametrar som kan påverka noggrannheten som att en ökande temporär upplösning kräver ett större dataset och parametrar för det neurala nätverket som använts.
Gray, Michael Stewart. "Unsupervised statistical methods for processing of image sequences /." Diss., Connect to a 24 p. preview or request complete full text in PDF format. Access restricted to UC campuses, 1998. http://wwwlib.umi.com/cr/ucsd/fullcit?p9901442.
Full textDupuis, Karine. "Bimodal cueing in aphasia : the influence of lipreading on speech discrimination and language comprehension." Thesis, University of British Columbia, 2011. http://hdl.handle.net/2429/33791.
Full textZhou, Yichao. "Lip password-based speaker verification system with unknown language alphabet." HKBU Institutional Repository, 2018. https://repository.hkbu.edu.hk/etd_oa/562.
Full textMontserrat, Maria Navarro. "The influence of situational cues on a standardized speechreading test." PDXScholar, 1985. https://pdxscholar.library.pdx.edu/open_access_etds/3546.
Full textNayfeh, Taysir H. "Multi-signal processing for voice recognition in noisy environments." Thesis, This resource online, 1991. http://scholar.lib.vt.edu/theses/available/etd-10222009-125021/.
Full textHo, Eve. "Speechreading abilities of Cantonese-speaking hearing-impaired children on consonants and words." Click to view the E-thesis via HKUTO, 1997. http://sunzi.lib.hku.hk/hkuto/record/B36209454.
Full text"A dissertation submitted in partial fulfilment of the requirements for the Bachelor of Science (Speech and Hearing Sciences), The University of Hong Kong, April 30, 1997." Also available in print.
Liu, Xin. "Lip motion tracking and analysis with application to lip-password based speaker verification." HKBU Institutional Repository, 2013. http://repository.hkbu.edu.hk/etd_ra/1538.
Full textGorman, Benjamin Millar. "A framework for speechreading acquisition tools." Thesis, University of Dundee, 2018. https://discovery.dundee.ac.uk/en/studentTheses/fc05921f-024e-471e-abd4-0d053634a2e7.
Full textLi, Meng. "On study of lip segmentation in color space." HKBU Institutional Repository, 2014. https://repository.hkbu.edu.hk/etd_oa/42.
Full textLees, Nicole C. "Vocalisations with a better view : hyperarticulation augments the auditory-visual advantage for the detection of speech in noise." Thesis, View thesis, 2007. http://handle.uws.edu.au:8081/1959.7/19576.
Full textLees, Nicole C. "Vocalisations with a better view hyperarticulation augments the auditory-visual advantage for the detection of speech in noise /." View thesis, 2007. http://handle.uws.edu.au:8081/1959.7/19576.
Full textA thesis submitted to the University of Western Sydney, College of Arts, in fulfilment of the requirements for the degree of Doctor of Philosophy. Includes bibliography.
Habermann, Barbara L. "Speechreading ability in elementary school-age children with and without functional articulation disorders." PDXScholar, 1990. https://pdxscholar.library.pdx.edu/open_access_etds/4087.
Full textEngelbrecht, Elizabeth M. "Die ontwikkeling van sosiale verhoudings van adolessente met ernstige gehoorverlies met hulle normaal horende portuurgroep." Pretoria : [s.n.], 2007. http://upetd.up.ac.za/thesis/available/etd-09122008-135458/.
Full textWagner, Jessica Lynn. "Exploration of Lip Shape Measures and their Association with Tongue Contact Patterns." Diss., CLICK HERE for online access, 2005. http://contentdm.lib.byu.edu/ETD/image/etd984.pdf.
Full textLidestam, Björn. "Semantic Framing of Speech : Emotional and Topical Cues in Perception of Poorly Specified Speech." Doctoral thesis, Linköpings universitet, Institutionen för beteendevetenskap, 2003. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-6344.
Full textShou, Virginia. "WHAT?: Visual Interpretations of the Miscommunication Between the Hearing and Deaf." VCU Scholars Compass, 2013. http://scholarscompass.vcu.edu/etd/3125.
Full textHoracio, Camila Paes. "Manifestações linguísticas em adultos com alterações no espectro da neuropatia auditiva." Universidade de São Paulo, 2010. http://www.teses.usp.br/teses/disponiveis/5/5143/tde-26082010-170001/.
Full textIntroduction: Post linguistic neural hearing loss in adults can lead to speech alterations and difficulties in auditory discrimination of sounds and comprehension of the message. Auditory neuropathy spectrum disorder (ANSD) is among the causes of neural hearing loss. Most studies on ANSD describe the standard for auditory diagnosis. However, the consequences of such hearing impairment in communication and its implication on speech therapy are scarce. Thus, it is necessary to identify the specific language aspects to be assessed in neurologically impaired individuals through a directed assessment protocol to allow the development of outlined treatment guidelines. Objective: This study aimed to describe the linguistic manifestations in adults with ANSD. Methods: The study included adults diagnosed with ANSD, who were literate and had no neurological or cognitive alterations. Data collection was carried out between 2007 and 2009 at the Speech, Language and Hearing service of the Clinic of Otorhinolaryngology of HCFMUSP. Twelve patients, eight males (66,7%) with ages ranging from 18 and 50 years of age were selected. An anamnesis protocol was designed. This protocol included data on education, use of hearing aids (HA) and specific hearing complaints. The assessment protocol consisted on tests of auditory reception and production of speech (phonemic identification; intelligibility; reading and text comprehension; and phonological awareness) and expression (speech and elaboration). The stimuli input were given in auditory only and in auditory plus visual mode (with lip reading). Results: The main characteristics observed in all participants were: male gender; incomplete primary school; use of hearing aids for less than three months in both ears; difficulty hearing in noisy environments; and dialogue, were the communicative situations that led to greater difficulty in expression. A significant improvement in speech perception was observed in all tests with lip reading. Conclusions: The language specificities of individuals with ANSD were: low educational level; speech rate alterations; difficulty in reading comprehension both by hearing and by reading; difficulty in phonological awareness; improvement of words and phrases repetition using LR.
Charlier, Brigitte. "Le développement des représentations phonologiques chez l'enfant sourd: étude comparative du langage parlé complété avec d'autres outils de communication." Doctoral thesis, Universite Libre de Bruxelles, 1994. http://hdl.handle.net/2013/ULB-DIPOT:oai:dipot.ulb.ac.be:2013/212631.
Full textBayard, Clémence. "Perception de la langue française parlée complétée: intégration du trio lèvres-main-son." Doctoral thesis, Universite Libre de Bruxelles, 2014. http://hdl.handle.net/2013/ULB-DIPOT:oai:dipot.ulb.ac.be:2013/209168.
Full textDans le cadre de cette thèse nous avons cherché à objectiver et caractériser l’intégration labio-manuelle dans la perception de la parole codée. Le poids accordé par le système perceptif aux informations manuelles, d’une part, et aux informations labiales, d’autre part, dépend-il de la qualité de chacune d’entre elles ?Varie-t-il en fonction du statut auditif ?Quand l’information auditive est disponible, comment le traitement de l’information manuelle est-il incorporé au traitement audio-visuel ?Pour tenter de répondre à cette série de questions, cinq paradigmes expérimentaux ont été créés et administrés à des adultes sourds et normo-entendants décodant la LPC.
Les trois premières études étaient focalisées sur la perception de la parole codée sans informations auditives. Dans l’étude n° 1, le but était d’objectiver l’intégration labio-manuelle ;l’impact de la qualité des informations labiales et du statut auditif sur cette intégration a également été investigué. L’objectif de l’étude n° 2 était d’examiner l’impact conjoint de la qualité des informations manuelles et labiales ;nous avons également comparé des décodeurs normo-entendants à des décodeurs sourds. Enfin, dans l’étude n° 3, nous avons examiné, chez des décodeurs normo-entendants et sourds, l’effet de l’incongruence entre les informations labiales et manuelles sur la perception de mots.
Les deux dernières études étaient focalisées sur la perception de la parole codée avec du son. L’objectif de l’étude n°4 était de comparer l’impact de la LPC sur l’intégration AV entre les sourds et les normo-entendants. Enfin, dans l’étude n°5, nous avons comparé l’impact de la LPC chez des décodeurs sourds présentant une récupération auditive faible ou forte.
Nos résultats ont permis de confirmer le véritable ancrage du code LPC sur la parole et de montrer que le poids de chaque information au sein du processus d’intégration est dépendant notamment de la qualité du stimulus manuel, de la qualité du stimulus labial et du niveau de performance auditive.
Doctorat en Sciences Psychologiques et de l'éducation
info:eu-repo/semantics/nonPublished
Huyse, Aurélie. "Intégration audio-visuelle de la parole: le poids de la vision varie-t-il en fonction de l'âge et du développement langagier?" Doctoral thesis, Universite Libre de Bruxelles, 2012. http://hdl.handle.net/2013/ULB-DIPOT:oai:dipot.ulb.ac.be:2013/209690.
Full textLe paradigme expérimental utilisé consistait toujours en une tâche d’identification de syllabes présentées dans trois modalités :auditive seule, visuelle seule et audio-visuelle (congruente et incongruente). Les cinq études avaient également comme point commun la présentation de stimuli visuels dont la qualité était réduite, visant à empêcher une lecture labiale de bonne qualité. Le but de chacune de ces études était non seulement d’examiner si les performances variaient en fonction des variables investiguées mais également de déterminer si les différences provenaient bien du processus d’intégration lui-même et non uniquement de différences au niveau de la perception unimodale. Pour cela, les scores des participants ont été comparés à des scores prédits sur base d’un modèle prenant en compte les variations individuelles des poids auditifs et visuels, le weighted fuzzy-logical model of perception.
L’ensemble des résultats, discuté dans la dernière partie de ce travail, fait pencher la balance en faveur de l’hypothèse d’une intégration dépendante du contexte. Nous proposons alors une nouvelle architecture de fusion bimodale, prenant en compte ces dernières données. Enfin, les implications sont aussi d’ordre pratique, suggérant la nécessité d’incorporer des évaluations et rééducations à la fois auditives et visuelles dans le cadre des programmes de revalidation de personnes âgées, dysphasiques ou avec implant cochléaire./During face-to-face conversation, perception of auditory speech is influenced by the visual speech cues contained in lip movements. Indeed, previous research has highlighted the ability of lip-reading to enhance and even modify speech perception. This phenomenon is known as audio-visual integration. The aim of this doctoral thesis is to study the possibility of modifying this audio-visual integration according to several variables. This work lies into the scope of an important debate between invariant versus subject-dependent audio-visual integration in speech processing. Each study of this dissertation investigates the impact of a specific variable on bimodal integration: the quality of the visual input, age of participants, the use of a cochlear implant, age at cochlear implantation and the presence of specific language impairments.
The paradigm used always consisted of a syllable identification task, where syllables were presented in three modalities: auditory only, visual only and audio-visual (congruent and incongruent). There was also a condition where the quality of the visual input was reduced, in order to prevent a lip-reading of good quality. The aim of each of the five studies was not only to examine whether performances were modified according to the variable under study but also to ascertain that differences were indeed issued from the integration process itself. Thereby, our results were analyzed in the framework of model predictive of audio-visual speech performance (weighted fuzzy-logical model of perception) in order to disentangle unisensory effects from audio-visual integration effects.
Taken together, our data suggest that speech integration is not automatic but rather depends on the context. We propose a new architecture of bimodal fusions, taking these considerations into account. Finally, there are also practical implications suggesting the need to incorporate not only auditory but also visual exercise in the rehabilitation programs of older adults and children with cochlear implants or with specific language impairements.
Doctorat en Sciences Psychologiques et de l'éducation
info:eu-repo/semantics/nonPublished
Fang-Chen, Chang, and 昌芳騁. "Lipreading System." Thesis, 1999. http://ndltd.ncl.edu.tw/handle/42032761223426994701.
Full textYuan, Hanfeng. "Tactual Display of Consonant Voicing to Supplement Lipreading." 2004. http://hdl.handle.net/1721.1/6568.
Full textThesis Supervisor: Nathaniel I. Durlach, Senior Research Scientist. Thesis Supervisor: Charlotte M. Reed, Senior Research Scientist.
Chang, Chih-Yu, and 張志瑜. "A Lipreading System Based on Hidden Markov Model." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/68212276335580923113.
Full text淡江大學
電機工程學系碩士班
97
Nowadays, the conventional speech recognition system has been used in many applications. However, the conventional speech recognition system would be interfered by the voice noise According to the disturbance, the recognition rate would be decreased in the noise condition. So, researchers proposed the singular visual feature speech recognition system, a lipreading system, to avoid the affection of voice noise. The lipreading system can be the assistance part of the conventional speech recognition system, to raise the speech recognition rate. In our research, we proposed a lipreading system which the lip image segmentation part is chromaticity color space combined with K-means algorithm. And taking the Hidden Markov Model as the recognition part to improve the recognition rate. In the experiment results, our method compared with other color based lip segmentation, and compared the recognition rate of different features.
Southard, Stuart D. Morris Richard. "Speechreading's benefit to the recognition of sentences as a function of signal-to-noise ratio." 2003. http://etd.lib.fsu.edu/theses/available/etd-11202003-175600/.
Full textAdvisor: Dr. Richard Morris, Florida State University, College of Communication, Dept. of Communication Disorders. Title and description from dissertation home page (viewed Mar. 3, 2004). Includes bibliographical references.
Lees, Nicole C., University of Western Sydney, and College of Arts. "Vocalisations with a better view : hyperarticulation augments the auditory-visual advantage for the detection of speech in noise." 2007. http://handle.uws.edu.au:8081/1959.7/19576.
Full textDoctor of Philosophy (PhD)
Mirus, Gene R. 1969. "The linguistic repertoire of deaf cuers: an ethnographic query on practice." Thesis, 2008. http://hdl.handle.net/2152/3889.
Full texttext
Lin, Wen-Chieh, and 林文杰. "A Space-Time Delay Neural Network for Motion Recognition and Its Application to Lipreading in Bimodal Speech Recognition." Thesis, 1996. http://ndltd.ncl.edu.tw/handle/30448892490229517714.
Full text國立交通大學
控制工程系
84
The researches of the motion recognition has received more and more attentions in recent years because the need for computer vision is increasing in many domains, such as the surveillance system, multimodal human computer interface, and traffic control system. Most of the existing approaches separate the recognition into the spatial feature extraction and time domai??cognition. However, we believe that the information of motion resides in the space-time domain, not restricted to the time domain or space domain only. Consequently, it seems more reasonable to integrate the feature extraction and classification in the space and time domains altogether. We propose a Space-Time Delay Neural Network (STDNN) that can deal with the 3-D dynamic information, such as motion recognition. For the motion recognition problem that we focus in this paper, the STDNN is an unified structure, in which the low-level spatiotemporal feature extraction and space-time recognition are embedded. It possesses the spatiotemporal shift-invariant recognition abilities that are inherited from the time delay neural network (TDNN) and space displacement neural network (SDNN). Unlike the multilayer perceptron (MLP), TDNN, and SDNN, the STDNN is constructed by the vector-type nodes and matrix-type links such that the spatiotemporal information can be gracefully represented in a neural network. Some experiments are done to evaluate the performance of the proposed STDNN. In the moving Arabic numerals (MAN) experiments, which simulate the object'smoving in the space-time domain by image sequences, the STDNN shows its generalization ability on spatiotemporal shift-invariance recognition. In the lipreading experiment, the STDNN recognizes the lip motions by the inputs of real image sequences. It shows that the STDNN has better performance than the existing TDNN- based system, especially on the generalization ability. Although the lipreading is a more specific application, the STDNN can be applied to other applications since no domain-dependentknowledge is used in the experiment.
Gritzman, Ashley Daniel. "Adaptive threshold optimisation for colour-based lip segmentation in automatic lip-reading systems." Thesis, 2016. http://hdl.handle.net/10539/22664.
Full textHaving survived the ordeal of a laryngectomy, the patient must come to terms with the resulting loss of speech. With recent advances in portable computing power, automatic lip-reading (ALR) may become a viable approach to voice restoration. This thesis addresses the image processing aspect of ALR, and focuses three contributions to colour-based lip segmentation. The rst contribution concerns the colour transform to enhance the contrast between the lips and skin. This thesis presents the most comprehensive study to date by measuring the overlap between lip and skin histograms for 33 di erent colour transforms. The hue component of HSV obtains the lowest overlap of 6:15%, and results show that selecting the correct transform can increase the segmentation accuracy by up to three times. The second contribution is the development of a new lip segmentation algorithm that utilises the best colour transforms from the comparative study. The algorithm is tested on 895 images and achieves percentage overlap (OL) of 92:23% and segmentation error (SE) of 7:39 %. The third contribution focuses on the impact of the histogram threshold on the segmentation accuracy, and introduces a novel technique called Adaptive Threshold Optimisation (ATO) to select a better threshold value. The rst stage of ATO incorporates -SVR to train the lip shape model. ATO then uses feedback of shape information to validate and optimise the threshold. After applying ATO, the SE decreases from 7:65% to 6:50%, corresponding to an absolute improvement of 1:15 pp or relative improvement of 15:1%. While this thesis concerns lip segmentation in particular, ATO is a threshold selection technique that can be used in various segmentation applications.
MT2017
Hochstrasser, Daniel. "Investigating the effect of visual phonetic cues on the auditory N1 & P2." Thesis, 2017. http://hdl.handle.net/1959.7/uws:44884.
Full textTan, Sok Hui (Jessica). "Seeing a talking face matters to infants, children and adults : behavioural and neurophysiological studies." Thesis, 2020. http://hdl.handle.net/1959.7/uws:59610.
Full textGoecke, Roland. "A stereo vision lip tracking algorithm and subsequent statistical analyses of the audio-video correlation in Australian English." Phd thesis, 2004. http://hdl.handle.net/1885/149999.
Full textFitzpatrick, Michael F. "Auditory and auditory-visual speech perception and production in noise in younger and older adults." Thesis, 2014. http://handle.uws.edu.au:8081/1959.7/uws:31936.
Full textBeadle, Julianne M. "Contributions of visual speech, visual distractors, and cognition to speech perception in noise for younger and older adults." Thesis, 2019. http://hdl.handle.net/1959.7/uws:55879.
Full text