Literatura académica sobre el tema "Speech synthesis"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte las listas temáticas de artículos, libros, tesis, actas de conferencias y otras fuentes académicas sobre el tema "Speech synthesis".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Artículos de revistas sobre el tema "Speech synthesis"
Hirose, Yoshifumi. "Speech synthesis apparatus and speech synthesis method". Journal of the Acoustical Society of America 128, n.º 1 (2010): 515. http://dx.doi.org/10.1121/1.3472332.
Texto completoMurthy, Savitha y Dinkar Sitaram. "Low Resource Kannada Speech Recognition using Lattice Rescoring and Speech Synthesis". Indian Journal Of Science And Technology 16, n.º 4 (29 de enero de 2023): 282–91. http://dx.doi.org/10.17485/ijst/v16i4.2371.
Texto completoSilverman, Kim E. A. "Speech synthesis". Journal of the Acoustical Society of America 90, n.º 6 (diciembre de 1991): 3391. http://dx.doi.org/10.1121/1.401356.
Texto completoTakagi, Tohru. "Speech Synthesis." Journal of the Institute of Television Engineers of Japan 46, n.º 2 (1992): 163–71. http://dx.doi.org/10.3169/itej1978.46.163.
Texto completoKuusisto, Finn. "Speech synthesis". XRDS: Crossroads, The ACM Magazine for Students 21, n.º 1 (14 de octubre de 2014): 63. http://dx.doi.org/10.1145/2667637.
Texto completoKamai, Takahiro y Yumiko Kato. "Speech Synthesis Method And Speech Synthesizer". Journal of the Acoustical Society of America 129, n.º 4 (2011): 2356. http://dx.doi.org/10.1121/1.3582212.
Texto completoKagoshima, Takehiko y Masami Akamine. "Speech synthesis method and speech synthesizer". Journal of the Acoustical Society of America 125, n.º 6 (2009): 4108. http://dx.doi.org/10.1121/1.3155494.
Texto completoSuckle, Leonard I. "Speech synthesis system". Journal of the Acoustical Society of America 84, n.º 4 (octubre de 1988): 1580. http://dx.doi.org/10.1121/1.397209.
Texto completoSharman, Richard Anthony. "Speech synthesis system". Journal of the Acoustical Society of America 103, n.º 6 (junio de 1998): 3136. http://dx.doi.org/10.1121/1.423023.
Texto completoKagoshima, Takehiko y Masami Akamine. "Speech synthesis method". Journal of the Acoustical Society of America 124, n.º 5 (2008): 2678. http://dx.doi.org/10.1121/1.3020583.
Texto completoTesis sobre el tema "Speech synthesis"
Donovan, R. E. "Trainable speech synthesis". Thesis, University of Cambridge, 1996. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.598598.
Texto completoGreenwood, Andrew Richard. "Articulatory speech synthesis". Thesis, University of Liverpool, 1993. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.386773.
Texto completoTsukanova, Anastasiia. "Articulatory speech synthesis". Electronic Thesis or Diss., Université de Lorraine, 2019. http://www.theses.fr/2019LORR0166.
Texto completoThe thesis is set in the domain of articulatory speech synthesis and consists of three major parts: the first two are dedicated to the development of two articulatory speech synthesizers and the third addresses how we can relate them to each other. The first approach results from a rule-based approach to articulatory speech synthesis that aimed to have a comprehensive control over the articulators (the jaw, the tongue, the lips, the velum, the larynx and the epiglottis). This approach used a dataset of static mid-sagittal magnetic resonance imaging (MRI) captures showing blocked articulation of French vowels and a set of consonant-vowel syllables; that dataset was encoded with a PCA-based vocal tract model. Then the system comprised several components: using the recorded articulatory configurations to drive a rule-based articulatory speech synthesizer as a source of target positions to attain (which is the main contribution of this first part); adjusting the obtained vocal tract shapes from the phonetic perspective; running an acoustic simulation unit to obtain the sound. The results of this synthesis were evaluated visually, acoustically and perceptually, and the problems encountered were broken down by their origin: the dataset, its modeling, the algorithm for managing the vocal tract shapes, their translation to the area functions, and the acoustic simulation. We concluded that, among our test examples, the articulatory strategies for vowels and stops are most correct, followed by those of nasals and fricatives. The second explored approach started off a baseline deep feed-forward neural network-based speech synthesizer trained with the standard recipe of Merlin on the audio recorded during real-time MRI (RT-MRI) acquisitions: denoised (and yet containing a considerable amount of noise of the MRI machine) speech in French and force-aligned state labels encoding phonetic and linguistic information. This synthesizer was augmented with eight parameters representing articulatory information---the lips opening and protrusion, the distance between the tongue and the velum, the velum and the pharyngeal wall and the tongue and the pharyngeal wall---that were automatically extracted from the captures and aligned with the audio signal and the linguistic specification. The jointly synthesized speech and articulatory sequences were evaluated objectively with dynamic time warping (DTW) distance, mean mel-cepstrum distortion (MCD), BAP (band aperiodicity prediction error), and three measures for F0: RMSE (root mean square error), CORR (correlation coefficient) and V/UV (frame-level voiced/unvoiced error). The consistency of articulatory parameters with the phonetic label was analyzed as well. I concluded that the generated articulatory parameter sequences matched the original ones acceptably closely, despite struggling more at attaining a contact between the articulators, and that the addition of articulatory parameters did not hinder the original acoustic model. The two approaches above are linked through the use of two different kinds of MRI speech data. This motivated a search for such coarticulation-aware targets as those that we had in the static case to be present or absent in the real-time data. To compare static and real-time MRI captures, the measures of structural similarity, Earth mover's distance, and SIFT were utilized; having analyzed these measures for validity and consistency, I qualitatively and quantitatively studied their temporal behavior, interpreted it and analyzed the identified similarities. I concluded that SIFT and structural similarity did capture some articulatory information and that their behavior, overall, validated the static MRI dataset. [...]
Sun, Felix (Felix W. ). "Speech Representation Models for Speech Synthesis and Multimodal Speech Recognition". Thesis, Massachusetts Institute of Technology, 2016. http://hdl.handle.net/1721.1/106378.
Texto completoThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 59-63).
The field of speech recognition has seen steady advances over the last two decades, leading to the accurate, real-time recognition systems available on mobile phones today. In this thesis, I apply speech modeling techniques developed for recognition to two other speech problems: speech synthesis and multimodal speech recognition with images. In both problems, there is a need to learn a relationship between speech sounds and another source of information. For speech synthesis, I show that using a neural network acoustic model results in a synthesizer that is more tolerant of noisy training data than previous work. For multimodal recognition, I show how information from images can be effectively integrated into the recognition search framework, resulting in improved accuracy when image data is available.
by Felix Sun.
M. Eng.
Morton, K. "Speech production and synthesis". Thesis, University of Essex, 1987. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.377930.
Texto completoJin, Yi-Xuan. "A HIGH SPEED DIGITAL IMPLEMENTATION OF LPC SPEECH SYNTHESIZER USING THE TMS320". Thesis, The University of Arizona, 1985. http://hdl.handle.net/10150/275309.
Texto completoWong, Chun-ho Eddy. "Reliability of rating synthesized hypernasal speech signals in connected speech and vowels". Click to view the E-thesis via HKU Scholars Hub, 2007. http://lookup.lib.hku.hk/lookup/bib/B4200617X.
Texto completo"A dissertation submitted in partial fulfilment of the requirements for the Bachelor of Science (Speech and Hearing Sciences), The University of Hong Kong, June 30, 2007." Includes bibliographical references (p. 28-30). Also available in print.
Peng, Antai. "Speech expression modeling and synthesis". Diss., Georgia Institute of Technology, 1996. http://hdl.handle.net/1853/13560.
Texto completoBrierton, Richard A. "Variable frame-rate speech synthesis". Thesis, University of Liverpool, 1993. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.357363.
Texto completoKlompje, Gideon. "A parametric monophone speech synthesis system". Thesis, Link to online version, 2006. http://hdl.handle.net/10019/561.
Texto completoLibros sobre el tema "Speech synthesis"
Eric, Keller y European Cooperation in the Field of Scientific and Technical Research (Organization). COST 258., eds. Improvements in speech synthesis: COST 258: the naturalness of synthetic speech. Chichester, West Sussex: J. Wiley, 2002.
Buscar texto completoHolmes, J. N. Speech synthesis and recognition. 2a ed. New York: Taylor & Francis, 2001.
Buscar texto completoKeller, E., G. Bailly, A. Monaghan, J. Terken y M. Huckvale, eds. Improvements in Speech Synthesis. Chichester, UK: John Wiley & Sons, Ltd, 2001. http://dx.doi.org/10.1002/0470845945.
Texto completoKeller, E., G. Bailly, A. Monaghan, J. Terken y M. Huckvale, eds. Improvements in Speech Synthesis. Chichester, UK: John Wiley & Sons, Ltd, 2001. http://dx.doi.org/10.1002/0470845945.
Texto completovan Santen, Jan P. H., Joseph P. Olive, Richard W. Sproat y Julia Hirschberg, eds. Progress in Speech Synthesis. New York, NY: Springer New York, 1997. http://dx.doi.org/10.1007/978-1-4612-1894-4.
Texto completoWendy, Holmes, ed. Speech synthesis and recognition. 2a ed. London: Taylor & Francis, 2002.
Buscar texto completoTaylor, Paul. Text-to-speech synthesis. Cambridge, UK: Cambridge University Press, 2009.
Buscar texto completo1956-, Kleijn W. B. y Paliwal K. K, eds. Speech coding and synthesis. Amsterdam: Elsevier, 1995.
Buscar texto completoinc, International Resource Development, ed. Speech recognition & voice synthesis. Norwalk, Conn., U.S.A. (6 Prowitt St., Norwalk 06855): International Resource Development, 1985.
Buscar texto completoVan Santen, Jan P. H., ed. Progress in speech synthesis. New York: Springer, 1997.
Buscar texto completoCapítulos de libros sobre el tema "Speech synthesis"
Scully, Celia. "Articulatory Synthesis". En Speech Production and Speech Modelling, 151–86. Dordrecht: Springer Netherlands, 1990. http://dx.doi.org/10.1007/978-94-009-2037-8_7.
Texto completoSchroeder, Manfred R. "Speech Synthesis". En Computer Speech, 85–90. Berlin, Heidelberg: Springer Berlin Heidelberg, 1999. http://dx.doi.org/10.1007/978-3-662-03861-1_5.
Texto completoSchroeder, Manfred R. "Speech Synthesis". En Computer Speech, 129–34. Berlin, Heidelberg: Springer Berlin Heidelberg, 2004. http://dx.doi.org/10.1007/978-3-662-06384-2_6.
Texto completoOwens, F. J. "Speech Synthesis". En Signal Processing of Speech, 88–121. London: Macmillan Education UK, 1993. http://dx.doi.org/10.1007/978-1-349-22599-6_5.
Texto completoDutoit, Thierry y Baris Bozkurt. "Speech Synthesis". En Handbook of Signal Processing in Acoustics, 557–85. New York, NY: Springer New York, 2008. http://dx.doi.org/10.1007/978-0-387-30441-0_30.
Texto completoSinha, Priyabrata. "Speech Synthesis". En Speech Processing in Embedded Systems, 157–64. Boston, MA: Springer US, 2009. http://dx.doi.org/10.1007/978-0-387-75581-6_11.
Texto completoHinterleitner, Florian. "Speech Synthesis". En Quality of Synthetic Speech, 5–18. Singapore: Springer Singapore, 2017. http://dx.doi.org/10.1007/978-981-10-3734-4_2.
Texto completoKurematsu, Akira y Tsuyoshi Morimoto. "Speech Synthesis". En Automatic Speech Translation, 71–85. London: CRC Press, 2023. http://dx.doi.org/10.1201/9780429333385-4.
Texto completoBeckman, Mary E. "Speech Models and Speech Synthesis". En Progress in Speech Synthesis, 185–209. New York, NY: Springer New York, 1997. http://dx.doi.org/10.1007/978-1-4612-1894-4_15.
Texto completoSuendermann, David, Harald Höge y Alan Black. "Challenges in Speech Synthesis". En Speech Technology, 19–32. New York, NY: Springer US, 2010. http://dx.doi.org/10.1007/978-0-387-73819-2_2.
Texto completoActas de conferencias sobre el tema "Speech synthesis"
Taylor, P. "Speech synthesis". En IEE Colloquium Speech and Language Engineering - State of the Art. IEE, 1998. http://dx.doi.org/10.1049/ic:19980957.
Texto completoBREEN, AP. "SPEECH SYNTHESIS". En Autumn Conference 1998. Institute of Acoustics, 2024. http://dx.doi.org/10.25144/18952.
Texto completoLouw, Johannes A., Daniel R. van Niekerk y Georg I. Schlünz. "Introducing the Speect speech synthesis platform". En The Blizzard Challenge 2010. ISCA: ISCA, 2010. http://dx.doi.org/10.21437/blizzard.2010-4.
Texto completoHuckvale, Mark. "Speech synthesis, speech simulation and speech science". En 7th International Conference on Spoken Language Processing (ICSLP 2002). ISCA: ISCA, 2002. http://dx.doi.org/10.21437/icslp.2002-388.
Texto completoKarlsson, Inger y Lennart Neovius. "Speech synthesis experiments with the glove synthesiser". En 3rd European Conference on Speech Communication and Technology (Eurospeech 1993). ISCA: ISCA, 1993. http://dx.doi.org/10.21437/eurospeech.1993-213.
Texto completoValentini-Botinhao, Cassia, Xin Wang, Shinji Takaki y Junichi Yamagishi. "Investigating RNN-based speech enhancement methods for noise-robust Text-to-Speech". En 9th ISCA Speech Synthesis Workshop. ISCA, 2016. http://dx.doi.org/10.21437/ssw.2016-24.
Texto completoÁlvarez, David, Santiago Pascual y Antonio Bonafonte. "Problem-Agnostic Speech Embeddings for Multi-Speaker Text-to-Speech with SampleRNN". En 10th ISCA Speech Synthesis Workshop. ISCA: ISCA, 2019. http://dx.doi.org/10.21437/ssw.2019-7.
Texto completoSagisaka, Yoshinori, Takumi Yamashita y Yoko Kokenawa. "Speech synthesis with attitude". En Speech Prosody 2004. ISCA: ISCA, 2004. http://dx.doi.org/10.21437/speechprosody.2004-91.
Texto completoKrishna, Gautam, Co Tran, Yan Han, Mason Carnahan y Ahmed H. Tewfik. "Speech Synthesis Using EEG". En ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020. http://dx.doi.org/10.1109/icassp40776.2020.9053340.
Texto completoBlack, Alan W., Heiga Zen y Keiichi Tokuda. "Statistical Parametric Speech Synthesis". En 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07. IEEE, 2007. http://dx.doi.org/10.1109/icassp.2007.367298.
Texto completoInformes sobre el tema "Speech synthesis"
Greenberg, Steven. Speech Synthesis Using Perceptually Motivated Features. Fort Belvoir, VA: Defense Technical Information Center, enero de 2012. http://dx.doi.org/10.21236/ada567193.
Texto completoOre, Brian M. Speech Recognition, Articulatory Feature Detection, and Speech Synthesis in Multiple Languages. Fort Belvoir, VA: Defense Technical Information Center, noviembre de 2009. http://dx.doi.org/10.21236/ada519140.
Texto completoJohnson, W. L., Shrikanth Narayanan, Richard Whitney, Rajat Das y Catherine LaBore. Limited Domain Synthesis of Expressive Military Speech for Animated Characters. Fort Belvoir, VA: Defense Technical Information Center, enero de 2002. http://dx.doi.org/10.21236/ada459392.
Texto completoJohnson, W. L., S. Narayanan, R. Whitney, R. Das, M. Bulut y C. LaBore. Limited Domain Synthesis of Expressive Military Speech for Animated Characters. Fort Belvoir, VA: Defense Technical Information Center, enero de 2002. http://dx.doi.org/10.21236/ada459395.
Texto completoGordon, Jane. Use of synthetic speech in tests of speech discrimination. Portland State University Library, enero de 2000. http://dx.doi.org/10.15760/etd.5327.
Texto completoCastan, Diego, Md Rahman, Sarah Bakst, Chris Cobo-Kroenke, Mitchell McLaren, Martin Graciarena y Aaron Lawson. Speaker-targeted Synthetic Speech Detection. Office of Scientific and Technical Information (OSTI), febrero de 2022. http://dx.doi.org/10.2172/1844063.
Texto completoMathew, Jijo K. Speed Enforcement in Work Zones and Synthesis on Cost-Benefit Assessment of Installing Speed Enforcement Cameras on INDOT Road Network. Purdue University, 2023. http://dx.doi.org/10.5703/1288284317639.
Texto completoKostova, Maya. Synthesis of PSA Inhibitors as SPECT- and PET-Based Imaging Agents for Prostate Cancer. Fort Belvoir, VA: Defense Technical Information Center, junio de 2011. http://dx.doi.org/10.21236/ada548605.
Texto completoKabalka, G. W. Boron in nuclear medicine: New synthetic approaches to PET and SPECT. Office of Scientific and Technical Information (OSTI), septiembre de 1992. http://dx.doi.org/10.2172/7199090.
Texto completoKabalka, G. W. Boron in nuclear medicine: New synthetic approaches to PET, SPECT, and BNCT agents. Office of Scientific and Technical Information (OSTI), octubre de 1989. http://dx.doi.org/10.2172/5516333.
Texto completo