Academic literature on the topic 'Audio synthesi'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Audio synthesi.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "Audio synthesi"
Mikulicz, Szymon. "Precise Inter-Device Audio Playback Synchronization for Linux." International Journal of Signal Processing Systems 9, no. 3 (September 2021): 17–21. http://dx.doi.org/10.18178/ijsps.9.3.17-21.
Full textVOITKO, Viktoriia, Svitlana BEVZ, Sergii BURBELO, and Pavlo STAVYTSKYI. "AUDIO GENERATION TECHNOLOGY OF A SYSTEM OF SYNTHESIS AND ANALYSIS OF MUSIC COMPOSITIONS." Herald of Khmelnytskyi National University 305, no. 1 (February 23, 2022): 64–67. http://dx.doi.org/10.31891/2307-5732-2022-305-1-64-67.
Full textGeorge, E. Bryan, and Mark J. T. Smith. "Audio analysis/synthesis system." Journal of the Acoustical Society of America 97, no. 3 (March 1995): 2016. http://dx.doi.org/10.1121/1.412041.
Full textLi, Naihan, Yanqing Liu, Yu Wu, Shujie Liu, Sheng Zhao, and Ming Liu. "RobuTrans: A Robust Transformer-Based Text-to-Speech Model." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 05 (April 3, 2020): 8228–35. http://dx.doi.org/10.1609/aaai.v34i05.6337.
Full textWang, Cheng-i., and Shlomo Dubnov. "Guided Music Synthesis with Variable Markov Oracle." Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment 10, no. 5 (June 29, 2021): 55–62. http://dx.doi.org/10.1609/aiide.v10i5.12767.
Full textCabrera, Andrés, JoAnn Kuchera-Morin, and Curtis Roads. "The Evolution of Spatial Audio in the AlloSphere." Computer Music Journal 40, no. 4 (December 2016): 47–61. http://dx.doi.org/10.1162/comj_a_00382.
Full textPark, Se Jin, Minsu Kim, Joanna Hong, Jeongsoo Choi, and Yong Man Ro. "SyncTalkFace: Talking Face Generation with Precise Lip-Syncing via Audio-Lip Memory." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 2 (June 28, 2022): 2062–70. http://dx.doi.org/10.1609/aaai.v36i2.20102.
Full textKuntz, Matthieu, and Bernhard U. Seeber. "Spatial audio for interactive hearing research." INTER-NOISE and NOISE-CON Congress and Conference Proceedings 265, no. 2 (February 1, 2023): 5120–27. http://dx.doi.org/10.3397/in_2022_0741.
Full textLoy, D. Gareth. "The Systems Concepts Digital Synthesizer: An Architectural Retrospective." Computer Music Journal 37, no. 3 (September 2013): 49–67. http://dx.doi.org/10.1162/comj_a_00193.
Full textBessell, David. "Dynamic Convolution Modeling, a Hybrid Synthesis Strategy." Computer Music Journal 37, no. 1 (March 2013): 44–51. http://dx.doi.org/10.1162/comj_a_00159.
Full textDissertations / Theses on the topic "Audio synthesi"
CHEMLA, ROMEU SANTOS AXEL CLAUDE ANDRE'. "MANIFOLD REPRESENTATIONS OF MUSICAL SIGNALS AND GENERATIVE SPACES." Doctoral thesis, Università degli Studi di Milano, 2020. http://hdl.handle.net/2434/700444.
Full textAmong the diverse research fields within computer music, synthesis and generation of audio signals epitomize the cross-disciplinarity of this domain, jointly nourishing both scientific and artistic practices since its creation. Inherent in computer music since its genesis, audio generation has inspired numerous approaches, evolving both with musical practices and scientific/technical advances. Moreover, some syn- thesis processes also naturally handle the reverse process, named analysis, such that synthesis parameters can also be partially or totally extracted from actual sounds, and providing an alternative representation of the analyzed audio signals. On top of that, the recent rise of machine learning algorithms earnestly questioned the field of scientific research, bringing powerful data-centred methods that raised several epistemological questions amongst researchers, in spite of their efficiency. Especially, a family of machine learning methods, called generative models, are focused on the generation of original content using features extracted from an existing dataset. In that case, such methods not only questioned previous approaches in generation, but also the way of integrating this methods into existing creative processes. While these new generative frameworks are progressively introduced in the domain of image generation, the application of such generative techniques in audio synthesis is still marginal. In this work, we aim to propose a new audio analysis-synthesis framework based on these modern generative models, enhanced by recent advances in machine learning. We first review existing approaches, both in sound synthesis and in generative machine learning, and focus on how our work inserts itself in both practices and what can be expected from their collation. Subsequently, we focus a little more on generative models, and how modern advances in the domain can be exploited to allow us learning complex sound distributions, while being sufficiently flexible to be integrated in the creative flow of the user. We then propose an inference / generation process, mirroring analysis/synthesis paradigms that are natural in the audio processing domain, using latent models that are based on a continuous higher-level space, that we use to control the generation. We first provide preliminary results of our method applied on spectral information, extracted from several datasets, and evaluate both qualitatively and quantitatively the obtained results. Subsequently, we study how to make these methods more suitable for learning audio data, tackling successively three different aspects. First, we propose two different latent regularization strategies specifically designed for audio, based on and signal / symbol translation and perceptual constraints. Then, we propose different methods to address the inner temporality of musical signals, based on the extraction of multi-scale representations and on prediction, that allow the obtained generative spaces that also model the dynamics of the signal. As a last chapter, we swap our scientific approach to a more research & creation-oriented point of view: first, we describe the architecture and the design of our open-source library, vsacids, aiming to be used by expert and non-expert music makers as an integrated creation tool. Then, we propose an first musical use of our system by the creation of a real-time performance, called aego, based jointly on our framework vsacids and an explorative agent using reinforcement learning to be trained during the performance. Finally, we draw some conclusions on the different manners to improve and reinforce the proposed generation method, as well as possible further creative applications.
À travers les différents domaines de recherche de la musique computationnelle, l’analysie et la génération de signaux audio sont l’exemple parfait de la trans-disciplinarité de ce domaine, nourrissant simultanément les pratiques scientifiques et artistiques depuis leur création. Intégrée à la musique computationnelle depuis sa création, la synthèse sonore a inspiré de nombreuses approches musicales et scientifiques, évoluant de pair avec les pratiques musicales et les avancées technologiques et scientifiques de son temps. De plus, certaines méthodes de synthèse sonore permettent aussi le processus inverse, appelé analyse, de sorte que les paramètres de synthèse d’un certain générateur peuvent être en partie ou entièrement obtenus à partir de sons donnés, pouvant ainsi être considérés comme une représentation alternative des signaux analysés. Parallèlement, l’intérêt croissant soulevé par les algorithmes d’apprentissage automatique a vivement questionné le monde scientifique, apportant de puissantes méthodes d’analyse de données suscitant de nombreux questionnements épistémologiques chez les chercheurs, en dépit de leur effectivité pratique. En particulier, une famille de méthodes d’apprentissage automatique, nommée modèles génératifs, s’intéressent à la génération de contenus originaux à partir de caractéristiques extraites directement des données analysées. Ces méthodes n’interrogent pas seulement les approches précédentes, mais aussi sur l’intégration de ces nouvelles méthodes dans les processus créatifs existants. Pourtant, alors que ces nouveaux processus génératifs sont progressivement intégrés dans le domaine la génération d’image, l’application de ces techniques en synthèse audio reste marginale. Dans cette thèse, nous proposons une nouvelle méthode d’analyse-synthèse basés sur ces derniers modèles génératifs, depuis renforcés par les avancées modernes dans le domaine de l’apprentissage automatique. Dans un premier temps, nous examinerons les approches existantes dans le domaine des systèmes génératifs, sur comment notre travail peut s’insérer dans les pratiques de synthèse sonore existantes, et que peut-on espérer de l’hybridation de ces deux approches. Ensuite, nous nous focaliserons plus précisément sur comment les récentes avancées accomplies dans ce domaine dans ce domaine peuvent être exploitées pour l’apprentissage de distributions sonores complexes, tout en étant suffisamment flexibles pour être intégrées dans le processus créatif de l’utilisateur. Nous proposons donc un processus d’inférence / génération, reflétant les paradigmes d’analyse-synthèse existant dans le domaine de génération audio, basé sur l’usage de modèles latents continus que l’on peut utiliser pour contrôler la génération. Pour ce faire, nous étudierons déjà les résultats préliminaires obtenus par cette méthode sur l’apprentissage de distributions spectrales, prises d’ensembles de données diversifiés, en adoptant une approche à la fois quantitative et qualitative. Ensuite, nous proposerons d’améliorer ces méthodes de manière spécifique à l’audio sur trois aspects distincts. D’abord, nous proposons deux stratégies de régularisation différentes pour l’analyse de signaux audio : une basée sur la traduction signal/ symbole, ainsi qu’une autre basée sur des contraintes perceptives. Nous passerons par la suite à la dimension temporelle de ces signaux audio, proposant de nouvelles méthodes basées sur l’extraction de représentations temporelles multi-échelle et sur une tâche supplémentaire de prédiction, permettant la modélisation de caractéristiques dynamiques par les espaces génératifs obtenus. En dernier lieu, nous passerons d’une approche scientifique à une approche plus orientée vers un point de vue recherche & création. Premièrement, nous présenterons notre librairie open-source, vsacids, visant à être employée par des créateurs experts et non-experts comme un outil intégré. Ensuite, nous proposons une première utilisation musicale de notre système par la création d’une performance temps réel, nommée ægo, basée à la fois sur notre librarie et sur un agent d’exploration appris dynamiquement par renforcement au cours de la performance. Enfin, nous tirons les conclusions du travail accompli jusqu’à maintenant, concernant les possibles améliorations et développements de la méthode de synthèse proposée, ainsi que sur de possibles applications créatives.
Lundberg, Anton. "Data-Driven Procedural Audio : Procedural Engine Sounds Using Neural Audio Synthesis." Thesis, KTH, Datavetenskap, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-280132.
Full textDet i dagsläget dominerande tillvägagångssättet för rendering av ljud i interaktivamedia, såsom datorspel och virtual reality, innefattar uppspelning av statiska ljudfiler. Detta tillvägagångssätt saknar flexibilitet och kräver hantering av stora mängder ljuddata. Ett alternativt tillvägagångssätt är procedurellt ljud, vari ljudmodeller styrs för att generera ljud i realtid. Trots sina många fördelar används procedurellt ljud ännu inte i någon vid utsträckning inom kommersiella produktioner, delvis på grund av att det genererade ljudet från många föreslagna modeller inte når upp till industrins standarder. Detta examensarbete undersöker hur procedurellt ljud kan utföras med datadrivna metoder. Vi gör detta genom att specifikt undersöka metoder för syntes av bilmotorljud baserade på neural ljudsyntes. Genom att bygga på en nyligen publicerad metod som integrerar digital signalbehandling med djupinlärning, kallad Differentiable Digital Signal Processing (DDSP), kan vår metod skapa ljudmodeller genom att träna djupa neurala nätverk att rekonstruera inspelade ljudexempel från tolkningsbara latenta prediktorer. Vi föreslår en metod för att använda fasinformation från motorers förbränningscykler, samt en differentierbar metod för syntes av transienter. Våra resultat visar att DDSP kan användas till procedurella motorljud, men mer arbete krävs innan våra modeller kan generera motorljud utan oönskade artefakter samt innan de kan användas i realtidsapplikationer. Vi diskuterar hur vårt tillvägagångssätt kan vara användbart inom procedurellt ljud i mer generella sammanhang, samt hur vår metod kan tillämpas på andra ljudkällor
Elfitri, I. "Analysis by synthesis spatial audio coding." Thesis, University of Surrey, 2013. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.590657.
Full textWood, Steven Gregory. "Objective Test Methods for Waveguide Audio Synthesis." BYU ScholarsArchive, 2007. https://scholarsarchive.byu.edu/etd/853.
Full textUstun, Selen. "Audio browsing of automaton-based hypertext." Thesis, Texas A&M University, 2003. http://hdl.handle.net/1969.1/33.
Full textJehan, Tristan 1974. "Perceptual synthesis engine : an audio-driven timbre generator." Thesis, Massachusetts Institute of Technology, 2001. http://hdl.handle.net/1721.1/61543.
Full textIncludes bibliographical references (leaves 68-75).
A real-time synthesis engine which models and predicts the timbre of acoustic instruments based on perceptual features extracted from an audio stream is presented. The thesis describes the modeling sequence including the analysis of natural sounds, the inference step that finds the mapping between control and output parameters, the timbre prediction step, and the sound synthesis. The system enables applications such as cross-synthesis, pitch shifting or compression of acoustic instruments, and timbre morphing between instrument families. It is fully implemented in the Max/MSP environment. The Perceptual Synthesis Engine was developed for the Hyperviolin as a novel, generic and perceptually meaningful synthesis technique for non-discretely pitched instruments.
by Tristan Jehan.
S.M.
Payne, R. G. "Digital techniques for the analysis and synthesis of audio signals." Thesis, Bucks New University, 1988. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.234706.
Full textCoulibaly, Patrice Yefoungnigui. "Codage audio à bas débit avec synthèse sinusoïdale." Mémoire, Université de Sherbrooke, 2000. http://savoirs.usherbrooke.ca/handle/11143/1078.
Full textCoulibaly, Patrice Yefoungnigui. "Codage audio à bas débit avec synthèse sinusoïdale." Sherbrooke : Université de Sherbrooke, 2001.
Find full textAndreux, Mathieu. "Foveal autoregressive neural time-series modeling." Electronic Thesis or Diss., Paris Sciences et Lettres (ComUE), 2018. http://www.theses.fr/2018PSLEE073.
Full textThis dissertation studies unsupervised time-series modelling. We first focus on the problem of linearly predicting future values of a time-series under the assumption of long-range dependencies, which requires to take into account a large past. We introduce a family of causal and foveal wavelets which project past values on a subspace which is adapted to the problem, thereby reducing the variance of the associated estimators. We then investigate under which conditions non-linear predictors exhibit better performances than linear ones. Time-series which admit a sparse time-frequency representation, such as audio ones, satisfy those requirements, and we propose a prediction algorithm using such a representation. The last problem we tackle is audio time-series synthesis. We propose a new generation method relying on a deep convolutional neural network, with an encoder-decoder architecture, which allows to synthesize new realistic signals. Contrary to state-of-the-art methods, we explicitly use time-frequency properties of sounds to define an encoder with the scattering transform, while the decoder is trained to solve an inverse problem in an adapted metric
Books on the topic "Audio synthesi"
Kunow, Kristian. Rundfunk und Internet: These, Antithese, Synthese? Edited by Arbeitsgemeinschaft der Landesmedienanstalten in der Bundesrepublik Deutschland. Berlin: Vistas, 2013.
Find full textAudio system for technical readings. Berlin: Springer, 1998.
Find full text1944-, Kamajou François, and Cameroon. Ministry of Scientific Research., eds. Audit scientifique de la recherche agricole au Cameroun: Synthèse de l'audit, rapport général. [Yaoundé]: République du Cameroun, Ministère de la recherche scientifique et technique, 1993.
Find full textJoachim, Hornegger, ed. Pattern recognition and image processing in C [plus] [plus]. Wiesbaden: Vieweg, 1995.
Find full textAudio Engineering Society. International Conference. Virtual, synthetic and entertainment audio: The proceedings of the AES 22nd international conference 2002, June 15-17, Espoo, Finland. New York: AES, 2002.
Find full textNakagawa, Seiichi. Speech, hearing and neural network models. Tokyo: Ohmsha, 1995.
Find full textJunqua, Jean-Claude. Robustness in automatic speech recognition: Fundamentals and applications. Boston: Kluwer Academic Publishers, 1996.
Find full textRick, Beasley, ed. Voice application development with VoiceXML. Indianapolis, Ind: Sams, 2002.
Find full textIEEE/RSJ International Conference on Intelligent Robots and Systems (2001 Maui, Hawaii). Proceedings 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2001): Expanding the societal role of robotics in the next millennium, October 29-November 3, 2001, Outrigger Wailea Resort, Maui, Hawaii, USA. Piscataway, N.J: IEEE, 2001.
Find full textIEEE Workshop on Automatic Speech Recognition and Understanding (1997 Santa Barbara, Calif.). 1997 IEEE Workshop on Automatic Speech Recognition and Understanding proceedings. Piscataway, NJ: Published under the sponsorship of the IEEE Signal Processing Society, 1997.
Find full textBook chapters on the topic "Audio synthesi"
Weik, Martin H. "audio synthesis." In Computer Science and Communications Dictionary, 77. Boston, MA: Springer US, 2000. http://dx.doi.org/10.1007/1-4020-0613-6_1022.
Full textTarr, Eric. "Introduction to Signal Synthesis." In Hack Audio, 79–101. New York, NY : Routledge, 2019. | Series: Audio Engineering Society presents: Routledge, 2018. http://dx.doi.org/10.4324/9781351018463-7.
Full textBode, Peer D. "Analog Audio Synthesis." In The Routledge Companion to Media Technology and Obsolescence, 148–63. New York : Routledge/Taylor & Francis Group, 2019.: Routledge, 2018. http://dx.doi.org/10.4324/9781315442686-11.
Full textVan Every, Shawn. "Audio Synthesis and Analysis." In Pro Android Media, 179–93. Berkeley, CA: Apress, 2009. http://dx.doi.org/10.1007/978-1-4302-3268-1_8.
Full textJackson, Wallace. "The Synthesis of Digital Audio: Tone Generation." In Digital Audio Editing Fundamentals, 93–105. Berkeley, CA: Apress, 2015. http://dx.doi.org/10.1007/978-1-4842-1648-4_11.
Full textJackson, Wallace. "The History of Digital Audio: MIDI and Synthesis." In Digital Audio Editing Fundamentals, 11–17. Berkeley, CA: Apress, 2015. http://dx.doi.org/10.1007/978-1-4842-1648-4_2.
Full textChen, Jiashu. "3D Audio and Virtual Acoustical Environment Synthesis." In Acoustic Signal Processing for Telecommunication, 283–301. Boston, MA: Springer US, 2000. http://dx.doi.org/10.1007/978-1-4419-8644-3_13.
Full textZáviška, Pavel, Pavel Rajmic, Zdeněk Průša, and Vítězslav Veselý. "Revisiting Synthesis Model in Sparse Audio Declipper." In Latent Variable Analysis and Signal Separation, 429–45. Cham: Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-319-93764-9_40.
Full textBernardes, Gilberto, and Diogo Cocharro. "Dynamic Music Generation, Audio Analysis-Synthesis Methods." In Encyclopedia of Computer Graphics and Games, 1–4. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-319-08234-9_211-1.
Full textHuzaifah, Muhammad, and Lonce Wyse. "Deep Generative Models for Musical Audio Synthesis." In Handbook of Artificial Intelligence for Music, 639–78. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-72116-9_22.
Full textConference papers on the topic "Audio synthesi"
Huang, Mincong (Jerry), Samuel Chabot, and Jonas Braasch. "Panoptic Reconstruction of Immersive Virtual Soundscapes Using Human-Scale Panoramic Imagery with Visual Recognition." In ICAD 2021: The 26th International Conference on Auditory Display. icad.org: International Community for Auditory Display, 2021. http://dx.doi.org/10.21785/icad2021.043.
Full textFox, K. Michael, Jeremy Stewart, and Rob Hamilton. "madBPM: Musical and Auditory Display for Biological Predictive Modeling." In The 23rd International Conference on Auditory Display. Arlington, Virginia: The International Community for Auditory Display, 2017. http://dx.doi.org/10.21785/icad2017.045.
Full textChatziioannou, Vasileios. "Digital synthesis of impact sounds." In the Audio Mostly 2015. New York, New York, USA: ACM Press, 2015. http://dx.doi.org/10.1145/2814895.2814908.
Full textLie Lu, Yi Mao, Liu Wenyin, and Hong-Jiang Zhang. "Audio restoration by constrained audio texture synthesis." In 2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698). IEEE, 2003. http://dx.doi.org/10.1109/icme.2003.1221334.
Full textSchimbinschi, Florin, Christian Walder, Sarah M. Erfani, and James Bailey. "SynthNet: Learning to Synthesize Music End-to-End." In Twenty-Eighth International Joint Conference on Artificial Intelligence {IJCAI-19}. California: International Joint Conferences on Artificial Intelligence Organization, 2019. http://dx.doi.org/10.24963/ijcai.2019/467.
Full textZhu, Hao, Huaibo Huang, Yi Li, Aihua Zheng, and Ran He. "Arbitrary Talking Face Generation via Attentional Audio-Visual Coherence Learning." In Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20}. California: International Joint Conferences on Artificial Intelligence Organization, 2020. http://dx.doi.org/10.24963/ijcai.2020/327.
Full textSkinner, Martha. "Audio and Video Drawings Mapping Temporality." In ACADIA 2006: Synthetic Landscapes. ACADIA, 2006. http://dx.doi.org/10.52842/conf.acadia.2006.178.
Full textSkinner, Martha. "Audio and Video Drawings Mapping Temporality." In ACADIA 2006: Synthetic Landscapes. ACADIA, 2006. http://dx.doi.org/10.52842/conf.acadia.2006.178.
Full textYe, Zhenhui, Zhou Zhao, Yi Ren, and Fei Wu. "SyntaSpeech: Syntax-Aware Generative Adversarial Text-to-Speech." In Thirty-First International Joint Conference on Artificial Intelligence {IJCAI-22}. California: International Joint Conferences on Artificial Intelligence Organization, 2022. http://dx.doi.org/10.24963/ijcai.2022/620.
Full textMoore, Carl, and William Brent. "Interactive Real-time Concatenative Synthesis in Virtual Reality." In ICAD 2019: The 25th International Conference on Auditory Display. Newcastle upon Tyne, United Kingdom: Department of Computer and Information Sciences, Northumbria University, 2019. http://dx.doi.org/10.21785/icad2019.068.
Full textReports on the topic "Audio synthesi"
Murad, M. Hassan, Stephanie M. Chang, Celia Fiordalisi, Jennifer S. Lin, Timothy J. Wilt, Amy Tsou, Brian Leas, et al. Improving the Utility of Evidence Synthesis for Decision Makers in the Face of Insufficient Evidence. Agency for Healthcare Research and Quality (AHRQ), April 2021. http://dx.doi.org/10.23970/ahrqepcwhitepaperimproving.
Full textKiianovska, N. M. The development of theory and methods of using cloud-based information and communication technologies in teaching mathematics of engineering students in the United States. Видавничий центр ДВНЗ «Криворізький національний університет», December 2014. http://dx.doi.org/10.31812/0564/1094.
Full textBaluk, Nadia, Natalia Basij, Larysa Buk, and Olha Vovchanska. VR/AR-TECHNOLOGIES – NEW CONTENT OF THE NEW MEDIA. Ivan Franko National University of Lviv, February 2021. http://dx.doi.org/10.30970/vjo.2021.49.11074.
Full text