Literatura académica sobre el tema "Speech and audio signals"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte las listas temáticas de artículos, libros, tesis, actas de conferencias y otras fuentes académicas sobre el tema "Speech and audio signals".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Artículos de revistas sobre el tema "Speech and audio signals"
Rao*, G. Manmadha, Raidu Babu D.N, Krishna Kanth P.S.L, Vinay B. y Nikhil V. "Reduction of Impulsive Noise from Speech and Audio Signals by using Sd-Rom Algorithm". International Journal of Recent Technology and Engineering 10, n.º 1 (30 de mayo de 2021): 265–68. http://dx.doi.org/10.35940/ijrte.a5943.0510121.
Texto completoS. Ashwin, J. y N. Manoharan. "Audio Denoising Based on Short Time Fourier Transform". Indonesian Journal of Electrical Engineering and Computer Science 9, n.º 1 (1 de enero de 2018): 89. http://dx.doi.org/10.11591/ijeecs.v9.i1.pp89-92.
Texto completoKacur, Juraj, Boris Puterka, Jarmila Pavlovicova y Milos Oravec. "Frequency, Time, Representation and Modeling Aspects for Major Speech and Audio Processing Applications". Sensors 22, n.º 16 (22 de agosto de 2022): 6304. http://dx.doi.org/10.3390/s22166304.
Texto completoNittrouer, Susan y Joanna H. Lowenstein. "Beyond Recognition: Visual Contributions to Verbal Working Memory". Journal of Speech, Language, and Hearing Research 65, n.º 1 (12 de enero de 2022): 253–73. http://dx.doi.org/10.1044/2021_jslhr-21-00177.
Texto completoB, Nagesh y Dr M. Uttara Kumari. "A Review on Machine Learning for Audio Applications". Journal of University of Shanghai for Science and Technology 23, n.º 07 (30 de junio de 2021): 62–70. http://dx.doi.org/10.51201/jusst/21/06508.
Texto completoKubanek, M., J. Bobulski y L. Adrjanowicz. "Characteristics of the use of coupled hidden Markov models for audio-visual polish speech recognition". Bulletin of the Polish Academy of Sciences: Technical Sciences 60, n.º 2 (1 de octubre de 2012): 307–16. http://dx.doi.org/10.2478/v10175-012-0041-6.
Texto completoTimmermann, Johannes, Florian Ernst y Delf Sachau. "Speech enhancement for helicopter headsets with an integrated ANC-system for FPGA-platforms". INTER-NOISE and NOISE-CON Congress and Conference Proceedings 265, n.º 5 (1 de febrero de 2023): 2720–30. http://dx.doi.org/10.3397/in_2022_0382.
Texto completoAbdallah, Hanaa A. y Souham Meshoul. "A Multilayered Audio Signal Encryption Approach for Secure Voice Communication". Electronics 12, n.º 1 (20 de diciembre de 2022): 2. http://dx.doi.org/10.3390/electronics12010002.
Texto completoYin, Shu Hua. "Design of the Auxiliary Speech Recognition System of Super-Short-Range Reconnaissance Radar". Applied Mechanics and Materials 556-562 (mayo de 2014): 4830–34. http://dx.doi.org/10.4028/www.scientific.net/amm.556-562.4830.
Texto completoMoore, Brian C. J. "Binaural sharing of audio signals". Hearing Journal 60, n.º 11 (noviembre de 2007): 46–48. http://dx.doi.org/10.1097/01.hj.0000299172.13153.6f.
Texto completoTesis sobre el tema "Speech and audio signals"
Mason, Michael. "Hybrid coding of speech and audio signals". Thesis, Queensland University of Technology, 2001.
Buscar texto completoTrinkaus, Trevor R. "Perceptual coding of audio and diverse speech signals". Diss., Georgia Institute of Technology, 1999. http://hdl.handle.net/1853/13883.
Texto completoMészáros, Tomáš. "Speech Analysis for Processing of Musical Signals". Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2015. http://www.nusl.cz/ntk/nusl-234974.
Texto completoChoi, Hyung Keun. "Blind source separation of the audio signals in a real world". Thesis, Georgia Institute of Technology, 2002. http://hdl.handle.net/1853/14986.
Texto completoLucey, Simon. "Audio-visual speech processing". Thesis, Queensland University of Technology, 2002. https://eprints.qut.edu.au/36172/7/SimonLuceyPhDThesis.pdf.
Texto completoAnderson, David Verl. "Audio signal enhancement using multi-resolution sinusoidal modeling". Diss., Georgia Institute of Technology, 1999. http://hdl.handle.net/1853/15394.
Texto completoZeghidour, Neil. "Learning representations of speech from the raw waveform". Thesis, Paris Sciences et Lettres (ComUE), 2019. http://www.theses.fr/2019PSLEE004/document.
Texto completoWhile deep neural networks are now used in almost every component of a speech recognition system, from acoustic to language modeling, the input to such systems are still fixed, handcrafted, spectral features such as mel-filterbanks. This contrasts with computer vision, in which a deep neural network is now trained on raw pixels. Mel-filterbanks contain valuable and documented prior knowledge from human auditory perception as well as signal processing, and are the input to state-of-the-art speech recognition systems that are now on par with human performance in certain conditions. However, mel-filterbanks, as any fixed representation, are inherently limited by the fact that they are not fine-tuned for the task at hand. We hypothesize that learning the low-level representation of speech with the rest of the model, rather than using fixed features, could push the state-of-the art even further. We first explore a weakly-supervised setting and show that a single neural network can learn to separate phonetic information and speaker identity from mel-filterbanks or the raw waveform, and that these representations are robust across languages. Moreover, learning from the raw waveform provides significantly better speaker embeddings than learning from mel-filterbanks. These encouraging results lead us to develop a learnable alternative to mel-filterbanks, that can be directly used in replacement of these features. In the second part of this thesis we introduce Time-Domain filterbanks, a lightweight neural network that takes the waveform as input, can be initialized as an approximation of mel-filterbanks, and then learned with the rest of the neural architecture. Across extensive and systematic experiments, we show that Time-Domain filterbanks consistently outperform melfilterbanks and can be integrated into a new state-of-the-art speech recognition system, trained directly from the raw audio signal. Fixed speech features being also used for non-linguistic classification tasks for which they are even less optimal, we perform dysarthria detection from the waveform with Time-Domain filterbanks and show that it significantly improves over mel-filterbanks or low-level descriptors. Finally, we discuss how our contributions fall within a broader shift towards fully learnable audio understanding systems
Bando, Yoshiaki. "Robust Audio Scene Analysis for Rescue Robots". Kyoto University, 2018. http://hdl.handle.net/2433/232410.
Texto completoMoghimi, Amir Reza. "Array-based Spectro-temporal Masking For Automatic Speech Recognition". Research Showcase @ CMU, 2014. http://repository.cmu.edu/dissertations/334.
Texto completoBrangers, Kirstin M. "Perceptual Ruler for Quantifying Speech Intelligibility in Cocktail Party Scenarios". UKnowledge, 2013. http://uknowledge.uky.edu/ece_etds/31.
Texto completoLibros sobre el tema "Speech and audio signals"
Gold, Ben, Nelson Morgan y Dan Ellis. Speech and Audio Signal Processing. Hoboken, NJ, USA: John Wiley & Sons, Inc., 2011. http://dx.doi.org/10.1002/9781118142882.
Texto completoDeller, John R. Discrete-time processing of speech signals. New York: Institute of Electrical and Electronics Engineers, 2000.
Buscar texto completoG, Proakis John y Hansen John H. L, eds. Discrete-time processing of speech signals. New York: Macmillan Pub. Co., 1993.
Buscar texto completoV, Madisetti, ed. Video, speech, and audio signal processing and associated standards. Boca Raton, FL: CRC Press, 2009.
Buscar texto completoV, Madisetti, ed. Video, speech, and audio signal processing and associated standards. Boca Raton, FL: CRC Press, 2009.
Buscar texto completoGold, Bernard. Speech and audio signal processing: Processing and perception of speech and music. 2a ed. Hoboken, N.J: Wiley, 2011.
Buscar texto completoNelson, Morgan, ed. Speech and audio signal processing: Processing and perception of speech and music. New York: John Wiley, 2000.
Buscar texto completoMadisetti, V. Video, speech, and audio signal processing and associated standards. 2a ed. Boca Raton, FL: CRC Press, 2010.
Buscar texto completoMadisetti, V. Video, speech, and audio signal processing and associated standards. 2a ed. Boca Raton, FL: CRC Press, 2010.
Buscar texto completoS, Atal Bishnu, Cuperman Vladimir y Gersho Allen, eds. Speech and audio coding for wireless and network applications. Boston: Kluwer Academic Publishers, 1993.
Buscar texto completoCapítulos de libros sobre el tema "Speech and audio signals"
Buchanan, William J. "Speech and Audio Signals". En Advanced Data Communications and Networks, 111–27. Boston, MA: Springer US, 1997. http://dx.doi.org/10.1007/978-1-4419-8670-2_8.
Texto completoBuchanan, Bill. "Speech and Audio Signals". En Handbook of Data Communications and Networks, 96–109. Boston, MA: Springer US, 1999. http://dx.doi.org/10.1007/978-1-4757-0905-6_9.
Texto completoBuchanan, W. J. "Speech and Audio Signals". En The Handbook of Data Communications and Networks, 359–72. Boston, MA: Springer US, 2004. http://dx.doi.org/10.1007/978-1-4020-7870-5_19.
Texto completoBuchanan, W. "Speech and Audio Signals". En Advanced Data Communications and Networks, 111–27. London: CRC Press, 2023. http://dx.doi.org/10.1201/9781003420415-8.
Texto completoRichter, Michael M., Sheuli Paul, Veton Këpuska y Marius Silaghi. "Audio Signals and Speech Recognition". En Signal Processing and Machine Learning with Applications, 345–68. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-319-45372-9_18.
Texto completoBuchner, Herbert y Walter Kellermann. "TRINICON for Dereverberation of Speech and Audio Signals". En Speech Dereverberation, 311–85. London: Springer London, 2010. http://dx.doi.org/10.1007/978-1-84996-056-4_10.
Texto completoDouglas, Scott C. y Malay Gupta. "Convolutive Blind Source Separation for Audio Signals". En Blind Speech Separation, 3–45. Dordrecht: Springer Netherlands, 2007. http://dx.doi.org/10.1007/978-1-4020-6479-1_1.
Texto completoHerre, Jürgen y Manfred Lutzky. "Perceptual Audio Coding of Speech Signals". En Springer Handbook of Speech Processing, 393–410. Berlin, Heidelberg: Springer Berlin Heidelberg, 2008. http://dx.doi.org/10.1007/978-3-540-49127-9_18.
Texto completoKellermann, Walter. "Beamforming for Speech and Audio Signals". En Handbook of Signal Processing in Acoustics, 691–702. New York, NY: Springer New York, 2008. http://dx.doi.org/10.1007/978-0-387-30441-0_35.
Texto completoShamsi, Meysam, Nelly Barbot, Damien Lolive y Jonathan Chevelu. "Mixing Synthetic and Recorded Signals for Audio-Book Generation". En Speech and Computer, 479–89. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-60276-5_46.
Texto completoActas de conferencias sobre el tema "Speech and audio signals"
Shriver, Stefanie, Alan W. Black y Ronald Rosenfeld. "Audio signals in speech interfaces". En 6th International Conference on Spoken Language Processing (ICSLP 2000). ISCA: ISCA, 2000. http://dx.doi.org/10.21437/icslp.2000-35.
Texto completo"Signal Processing. Speech and Audio Processing". En 2022 29th International Conference on Systems, Signals and Image Processing (IWSSIP). IEEE, 2022. http://dx.doi.org/10.1109/iwssip55020.2022.9854416.
Texto completoPhan, Huy, Lars Hertel, Marco Maass, Radoslaw Mazur y Alfred Mertins. "Representing nonspeech audio signals through speech classification models". En Interspeech 2015. ISCA: ISCA, 2015. http://dx.doi.org/10.21437/interspeech.2015-682.
Texto completoKammerl, Julius, Neil Birkbeck, Sasi Inguva, Damien Kelly, A. J. Crawford, Hugh Denman, Anil Kokaram y Caroline Pantofaru. "Temporal synchronization of multiple audio signals". En ICASSP 2014 - 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2014. http://dx.doi.org/10.1109/icassp.2014.6854474.
Texto completo"Main Track: Speech and Audio Processing". En 2020 International Conference on Systems, Signals and Image Processing (IWSSIP). IEEE, 2020. http://dx.doi.org/10.1109/iwssip48289.2020.9145083.
Texto completoMovassagh, Mahmood, Joachim Thiemann y Peter Kabal. "Joint entropy-scalable coding of audio signals". En ICASSP 2012 - 2012 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 2012. http://dx.doi.org/10.1109/icassp.2012.6288537.
Texto completovan der Waal, R. G. y R. N. J. Veldhuis. "Subband coding of stereophonic digital audio signals". En [Proceedings] ICASSP 91: 1991 International Conference on Acoustics, Speech, and Signal Processing. IEEE, 1991. http://dx.doi.org/10.1109/icassp.1991.151053.
Texto completoRajagopalan, R. y B. Subramanian. "Removal of impulse noise from audio and speech signals". En International Symposium on Signals, Circuits and Systems, 2003. SCS 2003. IEEE, 2003. http://dx.doi.org/10.1109/scs.2003.1226973.
Texto completoZiolko, Mariusz, Bartosz Ziolko y Rafal Samborski. "Dual-Microphone Speech Extraction from Signals with Audio Background". En 2009 Fifth International Conference on Intelligent Information Hiding and Multimedia Signal Processing (IIH-MSP). IEEE, 2009. http://dx.doi.org/10.1109/iih-msp.2009.34.
Texto completoBraun, Jerome J. y Haim Levkowitz. "Internet-oriented visualization with audio presentation of speech signals". En Photonics West '98 Electronic Imaging, editado por Robert F. Erbacher y Alex Pang. SPIE, 1998. http://dx.doi.org/10.1117/12.309555.
Texto completoInformes sobre el tema "Speech and audio signals"
DeLeon, Phillip L. Techniques for Preprocessing Speech Signals for More Effective Audio Interfaces. Fort Belvoir, VA: Defense Technical Information Center, diciembre de 2001. http://dx.doi.org/10.21236/ada412195.
Texto completoMammone, Richard J., Khaled Assaleh, Kevin Farrell, Ravi Ramachandran y Mihailo Zilovic. A Modulation Model for Characterizing Speech Signals. Fort Belvoir, VA: Defense Technical Information Center, marzo de 1996. http://dx.doi.org/10.21236/ada311661.
Texto completoSpittka, J. y K. Vos. RTP Payload Format for the Opus Speech and Audio Codec. RFC Editor, junio de 2015. http://dx.doi.org/10.17487/rfc7587.
Texto completoHerrnstein, A. Start/End Delays of Voiced and Unvoiced Speech Signals. Office of Scientific and Technical Information (OSTI), septiembre de 1999. http://dx.doi.org/10.2172/15006006.
Texto completoChan, A. D., K. Englehart, B. Hudgins y D. F. Lovely. Hidden Markov Model Classification of Myoelectric Signals in Speech. Fort Belvoir, VA: Defense Technical Information Center, octubre de 2001. http://dx.doi.org/10.21236/ada410037.
Texto completoSTANDARD OBJECT SYSTEMS INC. Advanced Audio Interface for Phonetic Speech Recognition in a High Noise Environment. Fort Belvoir, VA: Defense Technical Information Center, enero de 2000. http://dx.doi.org/10.21236/ada373461.
Texto completoNelson, W. T., Robert S. Bolia, Mark A. Ericson y Richard L. McKinley. Spatial Audio Displays for Speech Communications: A Comparison of Free Field and Virtual Acoustic Environments. Fort Belvoir, VA: Defense Technical Information Center, enero de 1999. http://dx.doi.org/10.21236/ada430289.
Texto completoNelson, W. T., Robert S. Bolia, Mark A. Ericson y Richard L. McKinley. Monitoring the Simultaneous Presentation of Spatialized Speech Signals in a Virtual Acoustic Environment. Fort Belvoir, VA: Defense Technical Information Center, enero de 1998. http://dx.doi.org/10.21236/ada430284.
Texto completoNelson, W. T., Robert S. Bolia, Mark A. Ericson y Richard L. McKinley. Monitoring the Simultaneous Presentation of Multiple Spatialized Speech Signals in the Free Field. Fort Belvoir, VA: Defense Technical Information Center, enero de 1998. http://dx.doi.org/10.21236/ada430298.
Texto completoHamlin, Alexandra, Erik Kobylarz, James Lever, Susan Taylor y Laura Ray. Assessing the feasibility of detecting epileptic seizures using non-cerebral sensor. Engineer Research and Development Center (U.S.), diciembre de 2021. http://dx.doi.org/10.21079/11681/42562.
Texto completo