Contents
Academic literature on the topic 'Classification de scènes sonores'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Classification de scènes sonores.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "Classification de scènes sonores"
Saraczynska, Maja. "Théâtre otobiographique*. Entendre le récit de soi sur les scènes au XXe siècle." Mnemosyne, no. 3 (October 11, 2018): 10. http://dx.doi.org/10.14428/mnemosyne.v0i3.12093.
Full textCorteel, Etienne. "Création et manipulation de scènes sonores pour la wave field synthesis." Cahier Louis-Lumière 2, no. 1 (2004): 62–83. http://dx.doi.org/10.3406/cllum.2004.863.
Full textLarrue, Jean-Marc, and Marie-Madeleine Mervant-Roux. "Théâtre : le lieu où l’on entend." L’Annuaire théâtral, no. 56-57 (August 30, 2016): 17–45. http://dx.doi.org/10.7202/1037326ar.
Full textMagnat, Virginie. "Jean-Marc Larrue, Giusy Pisano and Jean-Paul Quéinnec (dirs), Dispositifs sonores: Corps, scènes, atmosphères." Theatre Research in Canada 43, no. 1 (April 1, 2022): 146–49. http://dx.doi.org/10.3138/tric.43.1.b04.
Full textChomienne, Loïc, Cédric Goulon, Gaëtan Parseihian, and Lionel Bringoux. "Perception de la verticale en présence d’indices d’orientation visuels ou sonores : vers une dépendance allocentrée ?" Movement & Sport Sciences - Science & Motricité, no. 108 (2020): 33–37. http://dx.doi.org/10.1051/sm/2019036.
Full textZaaboub, Wala, and Zouhour Ben Dhiaf. "Approche de détermination de signature de texture - application à la classification de couverts forestiers d'image satellitaire à haute résolution." Revue Française de Photogrammétrie et de Télédétection, no. 207 (September 24, 2014): 45–58. http://dx.doi.org/10.52638/rfpt.2014.209.
Full textJedrzejewski, Franck. "Nœuds Polychromes et Entrelacs Sonores : Vers de Nouvelles Catégories Musicales." Musicae Scientiae 7, no. 1_suppl (September 2003): 73–83. http://dx.doi.org/10.1177/10298649040070s104.
Full textKPEDENOU, Koffi Djagnikpo, and Zakariyao KOUMOI. "Cartographie et analyse spatiale de la dégradation des terres dans le sud-est du Togo : une approche basée sur la télédétection." Annales de l’Université de Parakou - Série Sciences Naturelles et Agronomie 9, no. 1 (June 30, 2019): 67–78. http://dx.doi.org/10.56109/aup-sna.v9i1.64.
Full textAraújo, Alan Nunes, and Wanessa Pinheiro Prates. "MODELAGEM MATEMÁTICO-ESPACIAL NA IDENTIFICAÇÃO DE FRAGILIDADES AMBIENTAIS DA MICROBACIA DO RIO JARUCU, MUNICÍPIO DE BRASIL NOVO – PA." InterEspaço: Revista de Geografia e Interdisciplinaridade 4, no. 12 (March 22, 2018): 207. http://dx.doi.org/10.18764/2446-6549.v4n12p207-226.
Full textCARPENTIER, Thibaut. "Spatialisation sonore - Perception, captation et diffusion de scènes sonores." Bruit et vibrations, November 2022. http://dx.doi.org/10.51257/a-v1-br1150.
Full textDissertations / Theses on the topic "Classification de scènes sonores"
Bisot, Victor. "Apprentissage de représentations pour l'analyse de scènes sonores." Electronic Thesis or Diss., Paris, ENST, 2018. http://www.theses.fr/2018ENST0016.
Full textThis thesis work focuses on the computational analysis of environmental sound scenes and events. The objective of such tasks is to automatically extract information about the context in which a sound has been recorded. The interest for this area of research has been rapidly increasing in the last few years leading to a constant growth in the number of works and proposed approaches. We explore and contribute to the main families of approaches to sound scene and event analysis, going from feature engineering to deep learning. Our work is centered at representation learning techniques based on nonnegative matrix factorization, which are particularly suited to analyse multi-source environments such as acoustic scenes. As a first approach, we propose a combination of image processing features with the goal of confirming that spectrograms contain enough information to discriminate sound scenes and events. From there, we leave the world of feature engineering to go towards automatically learning the features. The first step we take in that direction is to study the usefulness of matrix factorization for unsupervised feature learning techniques, especially by relying on variants of NMF. Several of the compared approaches allow us indeed to outperform feature engineering approaches to such tasks. Next, we propose to improve the learned representations by introducing the TNMF model, a supervised variant of NMF. The proposed TNMF models and algorithms are based on jointly learning nonnegative dictionaries and classifiers by minimising a target classification cost. The last part of our work highlights the links and the compatibility between NMF and certain deep neural network systems by proposing and adapting neural network architectures to the use of NMF as an input representation. The proposed models allow us to get state of the art performance on scene classification and overlapping event detection tasks. Finally we explore the possibility of jointly learning NMF and neural networks parameters, grouping the different stages of our systems in one optimisation problem
Olvera, Zambrano Mauricio Michel. "Robust sound event detection." Electronic Thesis or Diss., Université de Lorraine, 2022. http://www.theses.fr/2022LORR0324.
Full textFrom industry to general interest applications, computational analysis of sound scenes and events allows us to interpret the continuous flow of everyday sounds. One of the main degradations encountered when moving from lab conditions to the real world is due to the fact that sound scenes are not composed of isolated events but of multiple simultaneous events. Differences between training and test conditions also often arise due to extrinsic factors such as the choice of recording hardware and microphone positions, as well as intrinsic factors of sound events, such as their frequency of occurrence, duration and variability. In this thesis, we investigate problems of practical interest for audio analysis tasks to achieve robustness in real scenarios.Firstly, we explore the separation of ambient sounds in a practical scenario in which multiple short duration sound events with fast varying spectral characteristics (i.e., foreground sounds) occur simultaneously with background stationary sounds. We introduce the foreground-background ambient sound separation task and investigate whether a deep neural network with auxiliary information about the statistics of the background sound can differentiate between rapidly- and slowly-varying spectro-temporal characteristics. Moreover, we explore the use of per-channel energy normalization (PCEN) as a suitable pre-processing and the ability of the separation model to generalize to unseen sound classes. Results on mixtures of isolated sounds from the DESED and Audioset datasets demonstrate the generalization capability of the proposed separation system, which is mainly due to PCEN.Secondly, we investigate how to improve the robustness of audio analysis systems under mismatched training and test conditions. We explore two distinct tasks: acoustic scene classification (ASC) with mismatched recording devices and training of sound event detection (SED) systems with synthetic and real data.In the context of ASC, without assuming the availability of recordings captured simultaneously by mismatched training and test recording devices, we assess the impact of moment normalization and matching strategies and their integration with unsupervised adversarial domain adaptation. Our results show the benefits and limitations of these adaptation strategies applied at different stages of the classification pipeline. The best strategy matches source domain performance in the target domain.In the context of SED, we propose a PCEN based acoustic front-end with learned parameters. Then, we study the joint training of SED with auxiliary classification branches that categorize sounds as foreground or background according to their spectral properties. We also assess the impact of aligning the distributions of synthetic and real data at the frame or segment level based on optimal transport. Finally, we integrate an active learning strategy in the adaptation procedure. Results on the DESED dataset indicate that these methods are beneficial for the SED task and that their combination further improves performance on real sound scenes
Gontier, Félix. "Analyse et synthèse de scènes sonores urbaines par approches d'apprentissage profond." Thesis, Ecole centrale de Nantes, 2020. http://www.theses.fr/2020ECDN0042.
Full textThe advent of the Internet of Things (IoT) has enabled the development of largescale acoustic sensor networks to continuously monitor sound environments in urban areas. In the soundscape approach, perceptual quality attributes are associated with the activity of sound sources, quantities of importance to better account for the human perception of its acoustic environment. With recent success in acoustic scene analysis, deep learning approaches are uniquely suited to predict these quantities. Though, annotations necessary to the training process of supervised deep learning models are not easily obtainable, partly due to the fact that the information content of sensor measurements is limited by privacy constraints. To address this issue, a method is proposed for the automatic annotation of perceived source activity in large datasets of simulated acoustic scenes. On simulated data, trained deep learning models achieve state-of-the-art performances in the estimation of sourcespecific perceptual attributes and sound pleasantness. Semi-supervised transfer learning techniques are further studied to improve the adaptability of trained models by exploiting knowledge from the large amounts of unlabelled sensor data. Evaluations on annotated in situ recordings show that learning latent audio representations of sensor measurements compensates for the limited ecological validity of simulated sound scenes. In a second part, the use of deep learning methods for the synthesis of time domain signals from privacy-aware sensor measurements is investigated. Two spectral convolutional approaches are developed and evaluated against state-of-the-art methods designed for speech synthesis
Lafay, Grégoire. "Simulation de scènes sonores environnementales : Application à l’analyse sensorielle et l’analyse automatique." Thesis, Ecole centrale de Nantes, 2016. http://www.theses.fr/2016ECDN0007/document.
Full textThis thesis deals with environmental scene analysis, the auditory result of mixing separate but concurrent emitting sources. The sound environment is a complex object, which opens the field of possible research beyond the specific areas that are speech or music. For a person to make sense of its sonic environment, the involved process relies on both the perceived data and its context. For each experiment, one must be, as much as possible,in control of the evaluated stimuli, whether the field of investigation is perception or machine learning. Nevertheless, the sound environment needs to be studied in an ecological framework, using real recordings of sounds as stimuli rather than synthetic pure tones. We therefore propose a model of sound scenes allowing us to simulate complex sound environments from isolated sound recordings. The high level structural properties of the simulated scenes -- such as the type of sources, their sound levels or the event density -- are set by the experimenter. Based on knowledge of the human auditory system, the model abstracts the sound environment as a composite object, a sum of soundsources. The usefulness of the proposed model is assessed on two areas of investigation. The first is related to the soundscape perception issue, where the model is used to propose an innovative experimental protocol to study pleasantness perception of urban soundscape. The second tackles the major issue of evaluation in machine listening, for which we consider simulated data in order to powerfully assess the generalization capacities of automatic sound event detection systems
Moussallam, Manuel. "Représentations redondantes et hiérarchiques pour l'archivage et la compression de scènes sonores." Phd thesis, Télécom ParisTech, 2012. http://pastel.archives-ouvertes.fr/pastel-00834272.
Full textMoussallam, Manuel. "Représentations redondantes et hiérarchiques pour l'archivage et la compression de scènes sonores." Electronic Thesis or Diss., Paris, ENST, 2012. http://www.theses.fr/2012ENST0079.
Full textThe main goal of this work is automated processing of large volumes of audio data. Most specifically, one is interested in archiving, a process that encompass at least two distinct problems: data compression and data indexing. Jointly addressing these problems is a difficult task since many of their objectives may be concurrent. Therefore, building a consistent framework for audio archival is the matter of this thesis. Sparse representations of signals in redundant dictionaries have recently been found of interest for many sub-problems of the archival task. Sparsity is a desirable property both for compression and for indexing. Methods and algorithms to build such representations are the first topic of this thesis. Given the dimensionality of the considered data, greedy algorithms will be particularly studied. A first contribution of this thesis is the proposal of a variant of the famous Matching Pursuit algorithm, that exploits randomness and sub-sampling of very large time frequency dictionaries. We show that audio compression (especially at low bit-rate) can be improved using this method. This new algorithms comes with an original modeling of asymptotic pursuit behaviors, using order statistics and tools from extreme values theory. Other contributions deal with the second member of the archival problem: indexing. The same framework is used and applied to different layers of signal structures. First, redundancies and musical repetition detection is addressed. At larger scale, we investigate audio fingerprinting schemes and apply it to radio broadcast on-line segmentation. Performances have been evaluated during an international campaign within the QUAERO project. Finally, the same framework is used to perform source separation informed by the redundancy. All these elements validate the proposed framework for the audio archiving task. The layered structures of audio data are accessed hierarchically by greedy decomposition algorithms and allow processing the different objectives of archival at different steps, thus addressing them within the same framework
Baskind, Alexis. "Modèles et méthodes de description spatiale de scènes sonores : application aux enregistrements binauraux." Paris 6, 2003. http://www.theses.fr/2003PA066407.
Full textRompré, Louis. "Vers une méthode de classification de fichiers sonores /." Thèse, Trois-Rivières : Université du Québec à Trois-Rivières, 2007. http://www.uqtr.ca/biblio/notice/resume/30024804R.pdf.
Full textRompré, Louis. "Vers une méthode de classification de fichiers sonores." Thèse, Université du Québec à Trois-Rivières, 2007. http://depot-e.uqtr.ca/2022/1/030024804.pdf.
Full textPerotin, Lauréline. "Localisation et rehaussement de sources de parole au format Ambisonique : analyse de scènes sonores pour faciliter la commande vocale." Thesis, Université de Lorraine, 2019. http://www.theses.fr/2019LORR0124/document.
Full textThis work was conducted in the fast-growing context of hands-free voice command. In domestic environments, smart devices are usually laid in a fixed position, while the human speaker gives orders from anywhere, not necessarily next to the device, or nor even facing it. This adds difficulties compared to the problem of near-field voice command (typically for mobile phones) : strong reverberation, early reflections on furniture around the device, and surrounding noises can degrade the signal. Moreover, other speakers may interfere, which make the understanding of the target speaker quite difficult. In order to facilitate speech recognition in such adverse conditions, several preprocessing methods are introduced here. We use a spatialized audio format suitable for audio scene analysis : the Ambisonic format. We first propose a sound source localization method that relies on a convolutional and recurrent neural network. We define an input feature vector inspired by the acoustic intensity vector which improves the localization performance, in particular in real conditions involving several speakers and a microphone array laid on a table. We exploit the visualization technique called layerwise relevance propagation (LRP) to highlight the time-frequency zones that are correlate positively with the network output. This analysis is of paramount importance to establish the validity of a neural network. In addition, it shows that the neural network essentially relies on time-frequency zones where direct sound dominates reverberation and background noise. We then present a method to enhance the voice of the main speaker and ease its recognition. We adopt a mask-based beamforming framework based on a time-frequency mask estimated by a neural network. To deal with the situation of multiple speakers with similar loudness, we first use a wideband beamformer to enhance the target speaker thanks to the associated localization information. We show that this additional information is not enough for the network when two speakers are close to each other. However, if we also give an enhanced version of the interfering speaker as input to the network, it returns much better masks. The filters generated from those masks greatly improve speech recognition performance. We evaluate this algorithm in various environments, including real ones, with a black-box automatic speech recognition system. Finally, we combine the proposed localization and enhancement systems and evaluate the robustness of the latter to localization errors in real environments
Books on the topic "Classification de scènes sonores"
Dispositifs sonores: Corps, scènes, atmosphères. Montréal: Presses de l'Université de Montréal, 2019.
Find full text