Academic literature on the topic 'Speech synthesis'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Speech synthesis.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "Speech synthesis"
Hirose, Yoshifumi. "Speech synthesis apparatus and speech synthesis method." Journal of the Acoustical Society of America 128, no. 1 (2010): 515. http://dx.doi.org/10.1121/1.3472332.
Full textMurthy, Savitha, and Dinkar Sitaram. "Low Resource Kannada Speech Recognition using Lattice Rescoring and Speech Synthesis." Indian Journal Of Science And Technology 16, no. 4 (January 29, 2023): 282–91. http://dx.doi.org/10.17485/ijst/v16i4.2371.
Full textSilverman, Kim E. A. "Speech synthesis." Journal of the Acoustical Society of America 90, no. 6 (December 1991): 3391. http://dx.doi.org/10.1121/1.401356.
Full textTakagi, Tohru. "Speech Synthesis." Journal of the Institute of Television Engineers of Japan 46, no. 2 (1992): 163–71. http://dx.doi.org/10.3169/itej1978.46.163.
Full textKuusisto, Finn. "Speech synthesis." XRDS: Crossroads, The ACM Magazine for Students 21, no. 1 (October 14, 2014): 63. http://dx.doi.org/10.1145/2667637.
Full textKamai, Takahiro, and Yumiko Kato. "Speech Synthesis Method And Speech Synthesizer." Journal of the Acoustical Society of America 129, no. 4 (2011): 2356. http://dx.doi.org/10.1121/1.3582212.
Full textKagoshima, Takehiko, and Masami Akamine. "Speech synthesis method and speech synthesizer." Journal of the Acoustical Society of America 125, no. 6 (2009): 4108. http://dx.doi.org/10.1121/1.3155494.
Full textSuckle, Leonard I. "Speech synthesis system." Journal of the Acoustical Society of America 84, no. 4 (October 1988): 1580. http://dx.doi.org/10.1121/1.397209.
Full textSharman, Richard Anthony. "Speech synthesis system." Journal of the Acoustical Society of America 103, no. 6 (June 1998): 3136. http://dx.doi.org/10.1121/1.423023.
Full textKagoshima, Takehiko, and Masami Akamine. "Speech synthesis method." Journal of the Acoustical Society of America 124, no. 5 (2008): 2678. http://dx.doi.org/10.1121/1.3020583.
Full textDissertations / Theses on the topic "Speech synthesis"
Donovan, R. E. "Trainable speech synthesis." Thesis, University of Cambridge, 1996. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.598598.
Full textGreenwood, Andrew Richard. "Articulatory speech synthesis." Thesis, University of Liverpool, 1993. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.386773.
Full textTsukanova, Anastasiia. "Articulatory speech synthesis." Electronic Thesis or Diss., Université de Lorraine, 2019. http://www.theses.fr/2019LORR0166.
Full textThe thesis is set in the domain of articulatory speech synthesis and consists of three major parts: the first two are dedicated to the development of two articulatory speech synthesizers and the third addresses how we can relate them to each other. The first approach results from a rule-based approach to articulatory speech synthesis that aimed to have a comprehensive control over the articulators (the jaw, the tongue, the lips, the velum, the larynx and the epiglottis). This approach used a dataset of static mid-sagittal magnetic resonance imaging (MRI) captures showing blocked articulation of French vowels and a set of consonant-vowel syllables; that dataset was encoded with a PCA-based vocal tract model. Then the system comprised several components: using the recorded articulatory configurations to drive a rule-based articulatory speech synthesizer as a source of target positions to attain (which is the main contribution of this first part); adjusting the obtained vocal tract shapes from the phonetic perspective; running an acoustic simulation unit to obtain the sound. The results of this synthesis were evaluated visually, acoustically and perceptually, and the problems encountered were broken down by their origin: the dataset, its modeling, the algorithm for managing the vocal tract shapes, their translation to the area functions, and the acoustic simulation. We concluded that, among our test examples, the articulatory strategies for vowels and stops are most correct, followed by those of nasals and fricatives. The second explored approach started off a baseline deep feed-forward neural network-based speech synthesizer trained with the standard recipe of Merlin on the audio recorded during real-time MRI (RT-MRI) acquisitions: denoised (and yet containing a considerable amount of noise of the MRI machine) speech in French and force-aligned state labels encoding phonetic and linguistic information. This synthesizer was augmented with eight parameters representing articulatory information---the lips opening and protrusion, the distance between the tongue and the velum, the velum and the pharyngeal wall and the tongue and the pharyngeal wall---that were automatically extracted from the captures and aligned with the audio signal and the linguistic specification. The jointly synthesized speech and articulatory sequences were evaluated objectively with dynamic time warping (DTW) distance, mean mel-cepstrum distortion (MCD), BAP (band aperiodicity prediction error), and three measures for F0: RMSE (root mean square error), CORR (correlation coefficient) and V/UV (frame-level voiced/unvoiced error). The consistency of articulatory parameters with the phonetic label was analyzed as well. I concluded that the generated articulatory parameter sequences matched the original ones acceptably closely, despite struggling more at attaining a contact between the articulators, and that the addition of articulatory parameters did not hinder the original acoustic model. The two approaches above are linked through the use of two different kinds of MRI speech data. This motivated a search for such coarticulation-aware targets as those that we had in the static case to be present or absent in the real-time data. To compare static and real-time MRI captures, the measures of structural similarity, Earth mover's distance, and SIFT were utilized; having analyzed these measures for validity and consistency, I qualitatively and quantitatively studied their temporal behavior, interpreted it and analyzed the identified similarities. I concluded that SIFT and structural similarity did capture some articulatory information and that their behavior, overall, validated the static MRI dataset. [...]
Sun, Felix (Felix W. ). "Speech Representation Models for Speech Synthesis and Multimodal Speech Recognition." Thesis, Massachusetts Institute of Technology, 2016. http://hdl.handle.net/1721.1/106378.
Full textThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 59-63).
The field of speech recognition has seen steady advances over the last two decades, leading to the accurate, real-time recognition systems available on mobile phones today. In this thesis, I apply speech modeling techniques developed for recognition to two other speech problems: speech synthesis and multimodal speech recognition with images. In both problems, there is a need to learn a relationship between speech sounds and another source of information. For speech synthesis, I show that using a neural network acoustic model results in a synthesizer that is more tolerant of noisy training data than previous work. For multimodal recognition, I show how information from images can be effectively integrated into the recognition search framework, resulting in improved accuracy when image data is available.
by Felix Sun.
M. Eng.
Morton, K. "Speech production and synthesis." Thesis, University of Essex, 1987. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.377930.
Full textJin, Yi-Xuan. "A HIGH SPEED DIGITAL IMPLEMENTATION OF LPC SPEECH SYNTHESIZER USING THE TMS320." Thesis, The University of Arizona, 1985. http://hdl.handle.net/10150/275309.
Full textWong, Chun-ho Eddy. "Reliability of rating synthesized hypernasal speech signals in connected speech and vowels." Click to view the E-thesis via HKU Scholars Hub, 2007. http://lookup.lib.hku.hk/lookup/bib/B4200617X.
Full text"A dissertation submitted in partial fulfilment of the requirements for the Bachelor of Science (Speech and Hearing Sciences), The University of Hong Kong, June 30, 2007." Includes bibliographical references (p. 28-30). Also available in print.
Peng, Antai. "Speech expression modeling and synthesis." Diss., Georgia Institute of Technology, 1996. http://hdl.handle.net/1853/13560.
Full textBrierton, Richard A. "Variable frame-rate speech synthesis." Thesis, University of Liverpool, 1993. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.357363.
Full textKlompje, Gideon. "A parametric monophone speech synthesis system." Thesis, Link to online version, 2006. http://hdl.handle.net/10019/561.
Full textBooks on the topic "Speech synthesis"
Eric, Keller, and European Cooperation in the Field of Scientific and Technical Research (Organization). COST 258., eds. Improvements in speech synthesis: COST 258: the naturalness of synthetic speech. Chichester, West Sussex: J. Wiley, 2002.
Find full textHolmes, J. N. Speech synthesis and recognition. 2nd ed. New York: Taylor & Francis, 2001.
Find full textKeller, E., G. Bailly, A. Monaghan, J. Terken, and M. Huckvale, eds. Improvements in Speech Synthesis. Chichester, UK: John Wiley & Sons, Ltd, 2001. http://dx.doi.org/10.1002/0470845945.
Full textKeller, E., G. Bailly, A. Monaghan, J. Terken, and M. Huckvale, eds. Improvements in Speech Synthesis. Chichester, UK: John Wiley & Sons, Ltd, 2001. http://dx.doi.org/10.1002/0470845945.
Full textvan Santen, Jan P. H., Joseph P. Olive, Richard W. Sproat, and Julia Hirschberg, eds. Progress in Speech Synthesis. New York, NY: Springer New York, 1997. http://dx.doi.org/10.1007/978-1-4612-1894-4.
Full textWendy, Holmes, ed. Speech synthesis and recognition. 2nd ed. London: Taylor & Francis, 2002.
Find full textTaylor, Paul. Text-to-speech synthesis. Cambridge, UK: Cambridge University Press, 2009.
Find full text1956-, Kleijn W. B., and Paliwal K. K, eds. Speech coding and synthesis. Amsterdam: Elsevier, 1995.
Find full textinc, International Resource Development, ed. Speech recognition & voice synthesis. Norwalk, Conn., U.S.A. (6 Prowitt St., Norwalk 06855): International Resource Development, 1985.
Find full textVan Santen, Jan P. H., ed. Progress in speech synthesis. New York: Springer, 1997.
Find full textBook chapters on the topic "Speech synthesis"
Scully, Celia. "Articulatory Synthesis." In Speech Production and Speech Modelling, 151–86. Dordrecht: Springer Netherlands, 1990. http://dx.doi.org/10.1007/978-94-009-2037-8_7.
Full textSchroeder, Manfred R. "Speech Synthesis." In Computer Speech, 85–90. Berlin, Heidelberg: Springer Berlin Heidelberg, 1999. http://dx.doi.org/10.1007/978-3-662-03861-1_5.
Full textSchroeder, Manfred R. "Speech Synthesis." In Computer Speech, 129–34. Berlin, Heidelberg: Springer Berlin Heidelberg, 2004. http://dx.doi.org/10.1007/978-3-662-06384-2_6.
Full textOwens, F. J. "Speech Synthesis." In Signal Processing of Speech, 88–121. London: Macmillan Education UK, 1993. http://dx.doi.org/10.1007/978-1-349-22599-6_5.
Full textDutoit, Thierry, and Baris Bozkurt. "Speech Synthesis." In Handbook of Signal Processing in Acoustics, 557–85. New York, NY: Springer New York, 2008. http://dx.doi.org/10.1007/978-0-387-30441-0_30.
Full textSinha, Priyabrata. "Speech Synthesis." In Speech Processing in Embedded Systems, 157–64. Boston, MA: Springer US, 2009. http://dx.doi.org/10.1007/978-0-387-75581-6_11.
Full textHinterleitner, Florian. "Speech Synthesis." In Quality of Synthetic Speech, 5–18. Singapore: Springer Singapore, 2017. http://dx.doi.org/10.1007/978-981-10-3734-4_2.
Full textKurematsu, Akira, and Tsuyoshi Morimoto. "Speech Synthesis." In Automatic Speech Translation, 71–85. London: CRC Press, 2023. http://dx.doi.org/10.1201/9780429333385-4.
Full textBeckman, Mary E. "Speech Models and Speech Synthesis." In Progress in Speech Synthesis, 185–209. New York, NY: Springer New York, 1997. http://dx.doi.org/10.1007/978-1-4612-1894-4_15.
Full textSuendermann, David, Harald Höge, and Alan Black. "Challenges in Speech Synthesis." In Speech Technology, 19–32. New York, NY: Springer US, 2010. http://dx.doi.org/10.1007/978-0-387-73819-2_2.
Full textConference papers on the topic "Speech synthesis"
Taylor, P. "Speech synthesis." In IEE Colloquium Speech and Language Engineering - State of the Art. IEE, 1998. http://dx.doi.org/10.1049/ic:19980957.
Full textBREEN, AP. "SPEECH SYNTHESIS." In Autumn Conference 1998. Institute of Acoustics, 2024. http://dx.doi.org/10.25144/18952.
Full textLouw, Johannes A., Daniel R. van Niekerk, and Georg I. Schlünz. "Introducing the Speect speech synthesis platform." In The Blizzard Challenge 2010. ISCA: ISCA, 2010. http://dx.doi.org/10.21437/blizzard.2010-4.
Full textHuckvale, Mark. "Speech synthesis, speech simulation and speech science." In 7th International Conference on Spoken Language Processing (ICSLP 2002). ISCA: ISCA, 2002. http://dx.doi.org/10.21437/icslp.2002-388.
Full textKarlsson, Inger, and Lennart Neovius. "Speech synthesis experiments with the glove synthesiser." In 3rd European Conference on Speech Communication and Technology (Eurospeech 1993). ISCA: ISCA, 1993. http://dx.doi.org/10.21437/eurospeech.1993-213.
Full textValentini-Botinhao, Cassia, Xin Wang, Shinji Takaki, and Junichi Yamagishi. "Investigating RNN-based speech enhancement methods for noise-robust Text-to-Speech." In 9th ISCA Speech Synthesis Workshop. ISCA, 2016. http://dx.doi.org/10.21437/ssw.2016-24.
Full textÁlvarez, David, Santiago Pascual, and Antonio Bonafonte. "Problem-Agnostic Speech Embeddings for Multi-Speaker Text-to-Speech with SampleRNN." In 10th ISCA Speech Synthesis Workshop. ISCA: ISCA, 2019. http://dx.doi.org/10.21437/ssw.2019-7.
Full textSagisaka, Yoshinori, Takumi Yamashita, and Yoko Kokenawa. "Speech synthesis with attitude." In Speech Prosody 2004. ISCA: ISCA, 2004. http://dx.doi.org/10.21437/speechprosody.2004-91.
Full textKrishna, Gautam, Co Tran, Yan Han, Mason Carnahan, and Ahmed H. Tewfik. "Speech Synthesis Using EEG." In ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020. http://dx.doi.org/10.1109/icassp40776.2020.9053340.
Full textBlack, Alan W., Heiga Zen, and Keiichi Tokuda. "Statistical Parametric Speech Synthesis." In 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07. IEEE, 2007. http://dx.doi.org/10.1109/icassp.2007.367298.
Full textReports on the topic "Speech synthesis"
Greenberg, Steven. Speech Synthesis Using Perceptually Motivated Features. Fort Belvoir, VA: Defense Technical Information Center, January 2012. http://dx.doi.org/10.21236/ada567193.
Full textOre, Brian M. Speech Recognition, Articulatory Feature Detection, and Speech Synthesis in Multiple Languages. Fort Belvoir, VA: Defense Technical Information Center, November 2009. http://dx.doi.org/10.21236/ada519140.
Full textJohnson, W. L., Shrikanth Narayanan, Richard Whitney, Rajat Das, and Catherine LaBore. Limited Domain Synthesis of Expressive Military Speech for Animated Characters. Fort Belvoir, VA: Defense Technical Information Center, January 2002. http://dx.doi.org/10.21236/ada459392.
Full textJohnson, W. L., S. Narayanan, R. Whitney, R. Das, M. Bulut, and C. LaBore. Limited Domain Synthesis of Expressive Military Speech for Animated Characters. Fort Belvoir, VA: Defense Technical Information Center, January 2002. http://dx.doi.org/10.21236/ada459395.
Full textGordon, Jane. Use of synthetic speech in tests of speech discrimination. Portland State University Library, January 2000. http://dx.doi.org/10.15760/etd.5327.
Full textCastan, Diego, Md Rahman, Sarah Bakst, Chris Cobo-Kroenke, Mitchell McLaren, Martin Graciarena, and Aaron Lawson. Speaker-targeted Synthetic Speech Detection. Office of Scientific and Technical Information (OSTI), February 2022. http://dx.doi.org/10.2172/1844063.
Full textMathew, Jijo K. Speed Enforcement in Work Zones and Synthesis on Cost-Benefit Assessment of Installing Speed Enforcement Cameras on INDOT Road Network. Purdue University, 2023. http://dx.doi.org/10.5703/1288284317639.
Full textKostova, Maya. Synthesis of PSA Inhibitors as SPECT- and PET-Based Imaging Agents for Prostate Cancer. Fort Belvoir, VA: Defense Technical Information Center, June 2011. http://dx.doi.org/10.21236/ada548605.
Full textKabalka, G. W. Boron in nuclear medicine: New synthetic approaches to PET and SPECT. Office of Scientific and Technical Information (OSTI), September 1992. http://dx.doi.org/10.2172/7199090.
Full textKabalka, G. W. Boron in nuclear medicine: New synthetic approaches to PET, SPECT, and BNCT agents. Office of Scientific and Technical Information (OSTI), October 1989. http://dx.doi.org/10.2172/5516333.
Full text