Dissertations / Theses on the topic 'Automated speech Recognition'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Automated speech Recognition.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Davies, David Richard Llewellyn, and dave davies@canberra edu au. "Representing Time in Automated Speech Recognition." The Australian National University. Research School of Information Sciences and Engineering, 2003. http://thesis.anu.edu.au./public/adt-ANU20040602.163031.
Full textSooful, Jayren Jugpal. "Automated phoneme mapping for cross-language speech recognition." Diss., Pretoria [s.n.], 2004. http://upetd.up.ac.za/thesis/available/etd-01112005-131128.
Full textLAYOUSS, NIZAR GANDY ASSAF. "A critical examination of deep learningapproaches to automated speech recognition." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-153681.
Full textDookhoo, Raul. "AUTOMATED REGRESSION TESTING APPROACH TO EXPANSION AND REFINEMENT OF SPEECH RECOGNITION GRAMMARS." Master's thesis, University of Central Florida, 2008. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/2634.
Full textM.S.
School of Electrical Engineering and Computer Science
Engineering and Computer Science
Computer Science MS
Tsuchiya, Shinsuke. "Elicited Imitation and Automated Speech Recognition: Evaluating Differences among Learners of Japanese." BYU ScholarsArchive, 2011. https://scholarsarchive.byu.edu/etd/2782.
Full textBrashear, Helene Margaret. "Improving the efficacy of automated sign language practice tools." Diss., Georgia Institute of Technology, 2010. http://hdl.handle.net/1853/34703.
Full textMorton, Hazel. "A scenario based approach to speech-enabled computer assisted language learning based on automated speech recognition and virtual reality graphics." Thesis, University of Edinburgh, 2007. http://hdl.handle.net/1842/15438.
Full textGargett, Ross. "The Use of Automated Speech Recognition in Electronic Health Records in Rural Health Care Systems." Digital Commons @ East Tennessee State University, 2016. https://dc.etsu.edu/honors/340.
Full textZylich, Brian Matthew. "Training Noise-Robust Spoken Phrase Detectors with Scarce and Private Data: An Application to Classroom Observation Videos." Digital WPI, 2019. https://digitalcommons.wpi.edu/etd-theses/1289.
Full textAlcaraz, Meseguer Noelia. "Speech Analysis for Automatic Speech Recognition." Thesis, Norwegian University of Science and Technology, Department of Electronics and Telecommunications, 2009. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-9092.
Full textThe classical front end analysis in speech recognition is a spectral analysis which parametrizes the speech signal into feature vectors; the most popular set of them is the Mel Frequency Cepstral Coefficients (MFCC). They are based on a standard power spectrum estimate which is first subjected to a log-based transform of the frequency axis (mel- frequency scale), and then decorrelated by using a modified discrete cosine transform. Following a focused introduction on speech production, perception and analysis, this paper gives a study of the implementation of a speech generative model; whereby the speech is synthesized and recovered back from its MFCC representations. The work has been developed into two steps: first, the computation of the MFCC vectors from the source speech files by using HTK Software; and second, the implementation of the generative model in itself, which, actually, represents the conversion chain from HTK-generated MFCC vectors to speech reconstruction. In order to know the goodness of the speech coding into feature vectors and to evaluate the generative model, the spectral distance between the original speech signal and the one produced from the MFCC vectors has been computed. For that, spectral models based on Linear Prediction Coding (LPC) analysis have been used. During the implementation of the generative model some results have been obtained in terms of the reconstruction of the spectral representation and the quality of the synthesized speech.
Gabriel, Naveen. "Automatic Speech Recognition in Somali." Thesis, Linköpings universitet, Statistik och maskininlärning, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-166216.
Full textAl-Shareef, Sarah. "Conversational Arabic Automatic Speech Recognition." Thesis, University of Sheffield, 2015. http://etheses.whiterose.ac.uk/10145/.
Full textWang, Peidong. "Robust Automatic Speech Recognition By Integrating Speech Separation." The Ohio State University, 2021. http://rave.ohiolink.edu/etdc/view?acc_num=osu1619099401042668.
Full textUebler, Ulla. "Multilingual speech recognition /." Berlin : Logos Verlag, 2000. http://bvbr.bib-bvb.de:8991/F?func=service&doc_library=BVB01&doc_number=009117880&line_number=0001&func_code=DB_RECORDS&service_type=MEDIA.
Full textSeward, Alexander. "Efficient Methods for Automatic Speech Recognition." Doctoral thesis, KTH, Tal, musik och hörsel, 2003. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-3675.
Full textQC 20100811
Vipperla, Ravichander. "Automatic Speech Recognition for ageing voices." Thesis, University of Edinburgh, 2011. http://hdl.handle.net/1842/5725.
Full textGuzy, Julius Jonathan. "Automatic speech recognition : a refutation approach." Thesis, De Montfort University, 1988. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.254196.
Full textDeterding, David Henry. "Speaker normalisation for automatic speech recognition." Thesis, University of Cambridge, 1990. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.359822.
Full textBadr, Ibrahim. "Pronunciation learning for automatic speech recognition." Thesis, Massachusetts Institute of Technology, 2011. http://hdl.handle.net/1721.1/66022.
Full textCataloged from PDF version of thesis.
Includes bibliographical references (p. 99-101).
In many ways, the lexicon remains the Achilles heel of modern automatic speech recognizers (ASRs). Unlike stochastic acoustic and language models that learn the values of their parameters from training data, the baseform pronunciations of words in an ASR vocabulary are typically specified manually, and do not change, unless they are edited by an expert. Our work presents a novel generative framework that uses speech data to learn stochastic lexicons, thereby taking a step towards alleviating the need for manual intervention and automnatically learning high-quality baseform pronunciations for words. We test our model on a variety of domains: an isolated-word telephone speech corpus, a weather query corpus and an academic lecture corpus. We show significant improvements of 25%, 15% and 2% over expert-pronunciation lexicons, respectively. We also show that further improvements can be made by combining our pronunciation learning framework with acoustic model training.
by Ibrahim Badr.
S.M.
Chen, Chia-Ping. "Noise robustness in automatic speech recognition /." Thesis, Connect to this title online; UW restricted, 2004. http://hdl.handle.net/1773/5829.
Full textRagni, Anton. "Discriminative models for speech recognition." Thesis, University of Cambridge, 2014. https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.707926.
Full textEvans, N. W. D. "Spectral subtraction for speech enhancement and automatic speech recognition." Thesis, Swansea University, 2004. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.636935.
Full textThambiratnam, David P. "Speech recognition in adverse environments." Thesis, Queensland University of Technology, 1999. https://eprints.qut.edu.au/36099/1/36099_Thambiratnam_1999.pdf.
Full textLebart, Katia. "Speech dereverberation applied to automatic speech recognition and hearing aids." Thesis, University of Sussex, 1999. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.285064.
Full textLEBART, KATIA. "Speech dereverberation applied to automatic speech recognition and hearing aids." Rennes 1, 1999. http://www.theses.fr/1999REN10033.
Full textCouper, Kenney Fiona. "Automatic determination of sub-word units for automatic speech recognition." Thesis, University of Edinburgh, 2008. http://hdl.handle.net/1842/2788.
Full textGillespie, Bradford W. "Strategies for improving audible quality and speech recognition accuracy of reverberant speech /." Thesis, Connect to this title online; UW restricted, 2002. http://hdl.handle.net/1773/5930.
Full textJohnston, Samuel John Charles, and Samuel John Charles Johnston. "An Approach to Automatic and Human Speech Recognition Using Ear-Recorded Speech." Diss., The University of Arizona, 2017. http://hdl.handle.net/10150/625626.
Full textKleinschmidt, Tristan Friedrich. "Robust speech recognition using speech enhancement." Thesis, Queensland University of Technology, 2010. https://eprints.qut.edu.au/31895/1/Tristan_Kleinschmidt_Thesis.pdf.
Full textTabani, Hamid. "Low-power architectures for automatic speech recognition." Doctoral thesis, Universitat Politècnica de Catalunya, 2018. http://hdl.handle.net/10803/462249.
Full textEl reconocimiento automático de voz (ASR) es una de las aplicaciones más importantes en el área de la computación cognitiva. ASR rápido y preciso se está convirtiendo en una aplicación clave para dispositivos móviles y portátiles. Estos dispositivos, como los Smartphones, han incorporado el reconocimiento de voz como una de las principales interfaces de usuario. Es probable que esta tendencia hacia las interfaces de usuario basadas en voz continúe en los próximos años, lo que está cambiando la forma de interacción humano-máquina. Los sistemas de reconocimiento de voz efectivos requieren un reconocimiento en tiempo real, que es un desafío para los dispositivos móviles debido a la naturaleza de cálculo intensivo del problema y las limitaciones de potencia de dichos sistemas y supone un gran esfuerzo para las arquitecturas de CPU. Las arquitecturas GPU ofrecen capacidades de paralelización que pueden aprovecharse para aumentar el rendimiento de los sistemas de reconocimiento de voz. Sin embargo, la utilización eficiente de los recursos de la GPU para el reconocimiento de voz también es un desafío, ya que las implementaciones de software presentan accesos de memoria irregulares e impredecibles y una localidad temporal deficiente. El propósito de esta tesis es estudiar las características de los sistemas ASR que se ejecutan en dispositivos móviles de baja potencia para proponer diferentes técnicas para mejorar el rendimiento y el consumo de energía. Proponemos varias optimizaciones a nivel de software impulsadas por el análisis de potencia y rendimiento. A diferencia de las propuestas anteriores que intercambian precisión por el rendimiento al reducir el número de gaussianas evaluadas, mantenemos la precisión y mejoramos el rendimiento mediante el uso efectivo de la microarquitectura subyacente de la CPU. Usamos una implementación refactorizada del código de evaluación de GMM para reducir el impacto de las instrucciones de salto. Explotamos la unidad vectorial disponible en la mayoría de las CPU modernas para impulsar el cálculo de GMM. Además, calculamos las gaussianas para múltiples frames en paralelo, lo que reduce significativamente el uso de ancho de banda de memoria. Nuestros resultados experimentales muestran que las optimizaciones propuestas proporcionan un speedup de 2.68x sobre el decodificador Pocketsphinx en una CPU Intel Skylake de alta gama, mientras que logra un ahorro de energía del 61%. En segundo lugar, proponemos una técnica de renombrado de registros que explota la reutilización de registros físicos para reducir la presión sobre el banco de registros. Nuestra técnica aprovecha el uso compartido de registros físicos mediante la introducción de cambios en la tabla de renombrado de registros y la issue queue. Evaluamos nuestra técnica de renombrado sobre un procesador moderno. El esquema propuesto admite excepciones precisas y da como resultado mejoras de rendimiento del 9.5% para la evaluación GMM. Nuestros resultados experimentales muestran que el esquema de renombrado de registros propuesto proporciona un 6% de aceleración en promedio para SPEC2006. Finalmente, proponemos un acelerador para la evaluación de GMM que reduce el consumo de energía en tres órdenes de magnitud en comparación con soluciones basadas en CPU y GPU. El acelerador propuesto implementa un esquema de evaluación perezosa donde las GMMs se calculan bajo demanda, evitando el 50% de los cálculos. Finalmente, incluye un esquema de memorización que evita el 74.88% de las operaciones de coma flotante. El diseño final proporciona una aceleración de 164x y una reducción de energía de 3532x en comparación con una implementación altamente optimizada que se ejecuta en una CPU móvil moderna. Comparado con una GPU móvil de última generación, el acelerador de GMM logra un speedup de 5.89x sobre una implementación CUDA optimizada, mientras que reduce la energía en 241x.
Martínez, del Hoyo Canterla Alfonso. "Design of Detectors for Automatic Speech Recognition." Doctoral thesis, Norges teknisk-naturvitenskapelige universitet, Institutt for elektronikk og telekommunikasjon, 2012. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-16548.
Full textBengio, Yoshua. "Connectionist models applied to automatic speech recognition." Thesis, McGill University, 1987. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=63920.
Full textPrager, Richard William. "Parallel processing networks for automatic speech recognition." Thesis, University of Cambridge, 1987. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.238443.
Full textAustin, Stephen Christopher. "Hidden Markov models for automatic speech recognition." Thesis, University of Cambridge, 1988. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.292913.
Full textFrankel, Joe. "Linear dynamic models for automatic speech recognition." Thesis, University of Edinburgh, 2004. http://hdl.handle.net/1842/1087.
Full textGu, Y. "Perceptually-based features in automatic speech recognition." Thesis, Swansea University, 1991. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.637182.
Full textBaothman, Fatmah bint Abdul Rahman. "Phonology-based automatic speech recognition for Arabic." Thesis, University of Huddersfield, 2002. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.273720.
Full textHolmes, Wendy Jane. "Modelling segmental variability for automatic speech recognition." Thesis, University College London (University of London), 1997. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.267859.
Full textChan, Carlos Chun Ming. "Speaker model adaptation in automatic speech recognition." Thesis, Robert Gordon University, 1993. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.339307.
Full textDuchnowski, Paul. "A new structure for automatic speech recognition." Thesis, Massachusetts Institute of Technology, 1993. http://hdl.handle.net/1721.1/17333.
Full textIncludes bibliographical references (leaves 102-110).
by Paul Duchnowski.
Sc.D.
Wang, Stanley Xinlei. "Using graphone models in automatic speech recognition." Thesis, Massachusetts Institute of Technology, 2009. http://hdl.handle.net/1721.1/53114.
Full textIncludes bibliographical references (p. 87-90).
This research explores applications of joint letter-phoneme subwords, known as graphones, in several domains to enable detection and recognition of previously unknown words. For these experiments, graphones models are integrated into the SUMMIT speech recognition framework. First, graphones are applied to automatically generate pronunciations of restaurant names for a speech recognizer. Word recognition evaluations show that graphones are effective for generating pronunciations for these words. Next, a graphone hybrid recognizer is built and tested for searching song lyrics by voice, as well as transcribing spoken lectures in a open vocabulary scenario. These experiments demonstrate significant improvement over traditional word-only speech recognizers. Modifications to the flat hybrid model such as reducing the graphone set size are also considered. Finally, a hierarchical hybrid model is built and compared with the flat hybrid model on the lecture transcription task.
by Stanley Xinlei Wang.
M.Eng.
Seigel, Matthew Stephen. "Confidence estimation for automatic speech recognition hypotheses." Thesis, University of Cambridge, 2014. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.648633.
Full textAbdelhamied, Kadry A. "Automatic identification and recognition of deaf speech /." The Ohio State University, 1986. http://rave.ohiolink.edu/etdc/view?acc_num=osu1487266691094027.
Full textCherri, Mona Youssef 1956. "Automatic Speech Recognition Using Finite Inductive Sequences." Thesis, University of North Texas, 1996. https://digital.library.unt.edu/ark:/67531/metadc277749/.
Full textColton, Larry Don. "Confidence and rejection in automatic speech recognition /." Full text open access at:, 1997. http://content.ohsu.edu/u?/etd,21.
Full textLi, Jinyu. "Soft margin estimation for automatic speech recognition." Diss., Atlanta, Ga. : Georgia Institute of Technology, 2008. http://hdl.handle.net/1853/26613.
Full textCommittee Chair: Dr. Chin-Hui Lee; Committee Member: Dr. Anthony Joseph Yezzi; Committee Member: Dr. Biing-Hwang (Fred) Juang; Committee Member: Dr. Mark Clements; Committee Member: Dr. Ming Yuan. Part of the SMARTech Electronic Thesis and Dissertation Collection.
Arrowood, Jon A. "Using observation uncertainty for robust speech recognition." Diss., Available online, Georgia Institute of Technology, 2004:, 2003. http://etd.gatech.edu/theses/available/etd-04082004-180005/unrestricted/arrowood%5Fjon%5Fa%5F200312%5Fphd.pdf.
Full textBell, Peter. "Full covariance modelling for speech recognition." Thesis, University of Edinburgh, 2010. http://hdl.handle.net/1842/4912.
Full textWu, Jian, and 武健. "Discriminative speaker adaptation and environmental robustness in automatic speech recognition." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2004. http://hub.hku.hk/bib/B31246138.
Full textWrede, Britta. "Modelling the effects of speech rate variation for automatic speech recognition." [S.l. : s.n.], 2002. http://deposit.ddb.de/cgi-bin/dokserv?idn=969765304.
Full text