Índice
Literatura académica sobre el tema "Classification de scènes sonores"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte las listas temáticas de artículos, libros, tesis, actas de conferencias y otras fuentes académicas sobre el tema "Classification de scènes sonores".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Artículos de revistas sobre el tema "Classification de scènes sonores"
Saraczynska, Maja. "Théâtre otobiographique*. Entendre le récit de soi sur les scènes au XXe siècle". Mnemosyne, n.º 3 (11 de octubre de 2018): 10. http://dx.doi.org/10.14428/mnemosyne.v0i3.12093.
Texto completoCorteel, Etienne. "Création et manipulation de scènes sonores pour la wave field synthesis". Cahier Louis-Lumière 2, n.º 1 (2004): 62–83. http://dx.doi.org/10.3406/cllum.2004.863.
Texto completoLarrue, Jean-Marc y Marie-Madeleine Mervant-Roux. "Théâtre : le lieu où l’on entend". L’Annuaire théâtral, n.º 56-57 (30 de agosto de 2016): 17–45. http://dx.doi.org/10.7202/1037326ar.
Texto completoMagnat, Virginie. "Jean-Marc Larrue, Giusy Pisano and Jean-Paul Quéinnec (dirs), Dispositifs sonores: Corps, scènes, atmosphères". Theatre Research in Canada 43, n.º 1 (1 de abril de 2022): 146–49. http://dx.doi.org/10.3138/tric.43.1.b04.
Texto completoChomienne, Loïc, Cédric Goulon, Gaëtan Parseihian y Lionel Bringoux. "Perception de la verticale en présence d’indices d’orientation visuels ou sonores : vers une dépendance allocentrée ?" Movement & Sport Sciences - Science & Motricité, n.º 108 (2020): 33–37. http://dx.doi.org/10.1051/sm/2019036.
Texto completoZaaboub, Wala y Zouhour Ben Dhiaf. "Approche de détermination de signature de texture - application à la classification de couverts forestiers d'image satellitaire à haute résolution". Revue Française de Photogrammétrie et de Télédétection, n.º 207 (24 de septiembre de 2014): 45–58. http://dx.doi.org/10.52638/rfpt.2014.209.
Texto completoJedrzejewski, Franck. "Nœuds Polychromes et Entrelacs Sonores : Vers de Nouvelles Catégories Musicales". Musicae Scientiae 7, n.º 1_suppl (septiembre de 2003): 73–83. http://dx.doi.org/10.1177/10298649040070s104.
Texto completoKPEDENOU, Koffi Djagnikpo y Zakariyao KOUMOI. "Cartographie et analyse spatiale de la dégradation des terres dans le sud-est du Togo : une approche basée sur la télédétection". Annales de l’Université de Parakou - Série Sciences Naturelles et Agronomie 9, n.º 1 (30 de junio de 2019): 67–78. http://dx.doi.org/10.56109/aup-sna.v9i1.64.
Texto completoAraújo, Alan Nunes y Wanessa Pinheiro Prates. "MODELAGEM MATEMÁTICO-ESPACIAL NA IDENTIFICAÇÃO DE FRAGILIDADES AMBIENTAIS DA MICROBACIA DO RIO JARUCU, MUNICÍPIO DE BRASIL NOVO – PA". InterEspaço: Revista de Geografia e Interdisciplinaridade 4, n.º 12 (22 de marzo de 2018): 207. http://dx.doi.org/10.18764/2446-6549.v4n12p207-226.
Texto completoCARPENTIER, Thibaut. "Spatialisation sonore - Perception, captation et diffusion de scènes sonores". Bruit et vibrations, noviembre de 2022. http://dx.doi.org/10.51257/a-v1-br1150.
Texto completoTesis sobre el tema "Classification de scènes sonores"
Bisot, Victor. "Apprentissage de représentations pour l'analyse de scènes sonores". Electronic Thesis or Diss., Paris, ENST, 2018. http://www.theses.fr/2018ENST0016.
Texto completoThis thesis work focuses on the computational analysis of environmental sound scenes and events. The objective of such tasks is to automatically extract information about the context in which a sound has been recorded. The interest for this area of research has been rapidly increasing in the last few years leading to a constant growth in the number of works and proposed approaches. We explore and contribute to the main families of approaches to sound scene and event analysis, going from feature engineering to deep learning. Our work is centered at representation learning techniques based on nonnegative matrix factorization, which are particularly suited to analyse multi-source environments such as acoustic scenes. As a first approach, we propose a combination of image processing features with the goal of confirming that spectrograms contain enough information to discriminate sound scenes and events. From there, we leave the world of feature engineering to go towards automatically learning the features. The first step we take in that direction is to study the usefulness of matrix factorization for unsupervised feature learning techniques, especially by relying on variants of NMF. Several of the compared approaches allow us indeed to outperform feature engineering approaches to such tasks. Next, we propose to improve the learned representations by introducing the TNMF model, a supervised variant of NMF. The proposed TNMF models and algorithms are based on jointly learning nonnegative dictionaries and classifiers by minimising a target classification cost. The last part of our work highlights the links and the compatibility between NMF and certain deep neural network systems by proposing and adapting neural network architectures to the use of NMF as an input representation. The proposed models allow us to get state of the art performance on scene classification and overlapping event detection tasks. Finally we explore the possibility of jointly learning NMF and neural networks parameters, grouping the different stages of our systems in one optimisation problem
Olvera, Zambrano Mauricio Michel. "Robust sound event detection". Electronic Thesis or Diss., Université de Lorraine, 2022. http://www.theses.fr/2022LORR0324.
Texto completoFrom industry to general interest applications, computational analysis of sound scenes and events allows us to interpret the continuous flow of everyday sounds. One of the main degradations encountered when moving from lab conditions to the real world is due to the fact that sound scenes are not composed of isolated events but of multiple simultaneous events. Differences between training and test conditions also often arise due to extrinsic factors such as the choice of recording hardware and microphone positions, as well as intrinsic factors of sound events, such as their frequency of occurrence, duration and variability. In this thesis, we investigate problems of practical interest for audio analysis tasks to achieve robustness in real scenarios.Firstly, we explore the separation of ambient sounds in a practical scenario in which multiple short duration sound events with fast varying spectral characteristics (i.e., foreground sounds) occur simultaneously with background stationary sounds. We introduce the foreground-background ambient sound separation task and investigate whether a deep neural network with auxiliary information about the statistics of the background sound can differentiate between rapidly- and slowly-varying spectro-temporal characteristics. Moreover, we explore the use of per-channel energy normalization (PCEN) as a suitable pre-processing and the ability of the separation model to generalize to unseen sound classes. Results on mixtures of isolated sounds from the DESED and Audioset datasets demonstrate the generalization capability of the proposed separation system, which is mainly due to PCEN.Secondly, we investigate how to improve the robustness of audio analysis systems under mismatched training and test conditions. We explore two distinct tasks: acoustic scene classification (ASC) with mismatched recording devices and training of sound event detection (SED) systems with synthetic and real data.In the context of ASC, without assuming the availability of recordings captured simultaneously by mismatched training and test recording devices, we assess the impact of moment normalization and matching strategies and their integration with unsupervised adversarial domain adaptation. Our results show the benefits and limitations of these adaptation strategies applied at different stages of the classification pipeline. The best strategy matches source domain performance in the target domain.In the context of SED, we propose a PCEN based acoustic front-end with learned parameters. Then, we study the joint training of SED with auxiliary classification branches that categorize sounds as foreground or background according to their spectral properties. We also assess the impact of aligning the distributions of synthetic and real data at the frame or segment level based on optimal transport. Finally, we integrate an active learning strategy in the adaptation procedure. Results on the DESED dataset indicate that these methods are beneficial for the SED task and that their combination further improves performance on real sound scenes
Gontier, Félix. "Analyse et synthèse de scènes sonores urbaines par approches d'apprentissage profond". Thesis, Ecole centrale de Nantes, 2020. http://www.theses.fr/2020ECDN0042.
Texto completoThe advent of the Internet of Things (IoT) has enabled the development of largescale acoustic sensor networks to continuously monitor sound environments in urban areas. In the soundscape approach, perceptual quality attributes are associated with the activity of sound sources, quantities of importance to better account for the human perception of its acoustic environment. With recent success in acoustic scene analysis, deep learning approaches are uniquely suited to predict these quantities. Though, annotations necessary to the training process of supervised deep learning models are not easily obtainable, partly due to the fact that the information content of sensor measurements is limited by privacy constraints. To address this issue, a method is proposed for the automatic annotation of perceived source activity in large datasets of simulated acoustic scenes. On simulated data, trained deep learning models achieve state-of-the-art performances in the estimation of sourcespecific perceptual attributes and sound pleasantness. Semi-supervised transfer learning techniques are further studied to improve the adaptability of trained models by exploiting knowledge from the large amounts of unlabelled sensor data. Evaluations on annotated in situ recordings show that learning latent audio representations of sensor measurements compensates for the limited ecological validity of simulated sound scenes. In a second part, the use of deep learning methods for the synthesis of time domain signals from privacy-aware sensor measurements is investigated. Two spectral convolutional approaches are developed and evaluated against state-of-the-art methods designed for speech synthesis
Lafay, Grégoire. "Simulation de scènes sonores environnementales : Application à l’analyse sensorielle et l’analyse automatique". Thesis, Ecole centrale de Nantes, 2016. http://www.theses.fr/2016ECDN0007/document.
Texto completoThis thesis deals with environmental scene analysis, the auditory result of mixing separate but concurrent emitting sources. The sound environment is a complex object, which opens the field of possible research beyond the specific areas that are speech or music. For a person to make sense of its sonic environment, the involved process relies on both the perceived data and its context. For each experiment, one must be, as much as possible,in control of the evaluated stimuli, whether the field of investigation is perception or machine learning. Nevertheless, the sound environment needs to be studied in an ecological framework, using real recordings of sounds as stimuli rather than synthetic pure tones. We therefore propose a model of sound scenes allowing us to simulate complex sound environments from isolated sound recordings. The high level structural properties of the simulated scenes -- such as the type of sources, their sound levels or the event density -- are set by the experimenter. Based on knowledge of the human auditory system, the model abstracts the sound environment as a composite object, a sum of soundsources. The usefulness of the proposed model is assessed on two areas of investigation. The first is related to the soundscape perception issue, where the model is used to propose an innovative experimental protocol to study pleasantness perception of urban soundscape. The second tackles the major issue of evaluation in machine listening, for which we consider simulated data in order to powerfully assess the generalization capacities of automatic sound event detection systems
Moussallam, Manuel. "Représentations redondantes et hiérarchiques pour l'archivage et la compression de scènes sonores". Phd thesis, Télécom ParisTech, 2012. http://pastel.archives-ouvertes.fr/pastel-00834272.
Texto completoMoussallam, Manuel. "Représentations redondantes et hiérarchiques pour l'archivage et la compression de scènes sonores". Electronic Thesis or Diss., Paris, ENST, 2012. http://www.theses.fr/2012ENST0079.
Texto completoThe main goal of this work is automated processing of large volumes of audio data. Most specifically, one is interested in archiving, a process that encompass at least two distinct problems: data compression and data indexing. Jointly addressing these problems is a difficult task since many of their objectives may be concurrent. Therefore, building a consistent framework for audio archival is the matter of this thesis. Sparse representations of signals in redundant dictionaries have recently been found of interest for many sub-problems of the archival task. Sparsity is a desirable property both for compression and for indexing. Methods and algorithms to build such representations are the first topic of this thesis. Given the dimensionality of the considered data, greedy algorithms will be particularly studied. A first contribution of this thesis is the proposal of a variant of the famous Matching Pursuit algorithm, that exploits randomness and sub-sampling of very large time frequency dictionaries. We show that audio compression (especially at low bit-rate) can be improved using this method. This new algorithms comes with an original modeling of asymptotic pursuit behaviors, using order statistics and tools from extreme values theory. Other contributions deal with the second member of the archival problem: indexing. The same framework is used and applied to different layers of signal structures. First, redundancies and musical repetition detection is addressed. At larger scale, we investigate audio fingerprinting schemes and apply it to radio broadcast on-line segmentation. Performances have been evaluated during an international campaign within the QUAERO project. Finally, the same framework is used to perform source separation informed by the redundancy. All these elements validate the proposed framework for the audio archiving task. The layered structures of audio data are accessed hierarchically by greedy decomposition algorithms and allow processing the different objectives of archival at different steps, thus addressing them within the same framework
Baskind, Alexis. "Modèles et méthodes de description spatiale de scènes sonores : application aux enregistrements binauraux". Paris 6, 2003. http://www.theses.fr/2003PA066407.
Texto completoRompré, Louis. "Vers une méthode de classification de fichiers sonores /". Thèse, Trois-Rivières : Université du Québec à Trois-Rivières, 2007. http://www.uqtr.ca/biblio/notice/resume/30024804R.pdf.
Texto completoRompré, Louis. "Vers une méthode de classification de fichiers sonores". Thèse, Université du Québec à Trois-Rivières, 2007. http://depot-e.uqtr.ca/2022/1/030024804.pdf.
Texto completoPerotin, Lauréline. "Localisation et rehaussement de sources de parole au format Ambisonique : analyse de scènes sonores pour faciliter la commande vocale". Thesis, Université de Lorraine, 2019. http://www.theses.fr/2019LORR0124/document.
Texto completoThis work was conducted in the fast-growing context of hands-free voice command. In domestic environments, smart devices are usually laid in a fixed position, while the human speaker gives orders from anywhere, not necessarily next to the device, or nor even facing it. This adds difficulties compared to the problem of near-field voice command (typically for mobile phones) : strong reverberation, early reflections on furniture around the device, and surrounding noises can degrade the signal. Moreover, other speakers may interfere, which make the understanding of the target speaker quite difficult. In order to facilitate speech recognition in such adverse conditions, several preprocessing methods are introduced here. We use a spatialized audio format suitable for audio scene analysis : the Ambisonic format. We first propose a sound source localization method that relies on a convolutional and recurrent neural network. We define an input feature vector inspired by the acoustic intensity vector which improves the localization performance, in particular in real conditions involving several speakers and a microphone array laid on a table. We exploit the visualization technique called layerwise relevance propagation (LRP) to highlight the time-frequency zones that are correlate positively with the network output. This analysis is of paramount importance to establish the validity of a neural network. In addition, it shows that the neural network essentially relies on time-frequency zones where direct sound dominates reverberation and background noise. We then present a method to enhance the voice of the main speaker and ease its recognition. We adopt a mask-based beamforming framework based on a time-frequency mask estimated by a neural network. To deal with the situation of multiple speakers with similar loudness, we first use a wideband beamformer to enhance the target speaker thanks to the associated localization information. We show that this additional information is not enough for the network when two speakers are close to each other. However, if we also give an enhanced version of the interfering speaker as input to the network, it returns much better masks. The filters generated from those masks greatly improve speech recognition performance. We evaluate this algorithm in various environments, including real ones, with a black-box automatic speech recognition system. Finally, we combine the proposed localization and enhancement systems and evaluate the robustness of the latter to localization errors in real environments
Libros sobre el tema "Classification de scènes sonores"
Dispositifs sonores: Corps, scènes, atmosphères. Montréal: Presses de l'Université de Montréal, 2019.
Buscar texto completo