Literatura académica sobre el tema "Diarisation de la parole"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte las listas temáticas de artículos, libros, tesis, actas de conferencias y otras fuentes académicas sobre el tema "Diarisation de la parole".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Artículos de revistas sobre el tema "Diarisation de la parole"
Sun, Guangzhi, Chao Zhang y Philip C. Woodland. "Combination of deep speaker embeddings for diarisation". Neural Networks 141 (septiembre de 2021): 372–84. http://dx.doi.org/10.1016/j.neunet.2021.04.020.
Texto completoZelenák, M. y J. Hernando. "Speaker overlap detection with prosodic features for speaker diarisation". IET Signal Processing 6, n.º 8 (1 de octubre de 2012): 798–804. http://dx.doi.org/10.1049/iet-spr.2011.0233.
Texto completoCassetta, Michele. "Parole, parole, parole". Dental Cadmos 87, n.º 01 (septiembre de 2019): 588. http://dx.doi.org/10.19256/d.cadmos.09.2019.08.
Texto completoChiari, Alexis. "Parole parole". Feuillets psychanalytiques N° 2, n.º 1 (21 de septiembre de 2017): 65–77. http://dx.doi.org/10.3917/fpsy.002.0065.
Texto completoRichmond, J. L. y B. J. Halkon. "Speaker Diarisation of Vibroacoustic Intelligence from Drone Mounted Laser Doppler Vibrometers". Journal of Physics: Conference Series 2041, n.º 1 (1 de octubre de 2021): 012011. http://dx.doi.org/10.1088/1742-6596/2041/1/012011.
Texto completoTyszler, Jean-Jacques. "Parole vide, parole pleine, parole imposée". Journal français de psychiatrie 45, n.º 1 (2017): 70. http://dx.doi.org/10.3917/jfp.045.0070.
Texto completoSchiavinato, Jacques. "Parole égarée, parole retrouvée". Revue de psychothérapie psychanalytique de groupe 28, n.º 1 (1997): 115–27. http://dx.doi.org/10.3406/rppg.1997.1366.
Texto completoBouville, Jean-Marc. "Parole d’enfant, parole à l’enfant, parole sur l’enfant". Revue de l'enfance et de l'adolescence 94, n.º 2 (2016): 7. http://dx.doi.org/10.3917/read.094.0007.
Texto completoAguiar, Flavio. "Macounaïma : parole perdue, parole retrouvée". Études françaises 28, n.º 2-3 (1992): 59. http://dx.doi.org/10.7202/035881ar.
Texto completoTournon, André. "Parole de badin, parole irrécusable". Réforme, Humanisme, Renaissance 76, n.º 1 (2013): 107–18. http://dx.doi.org/10.3406/rhren.2013.3294.
Texto completoTesis sobre el tema "Diarisation de la parole"
Cui, Can. "Séparation, diarisation et reconnaissance de la parole conjointes pour la transcription automatique de réunions". Electronic Thesis or Diss., Université de Lorraine, 2024. http://www.theses.fr/2024LORR0103.
Texto completoFar-field microphone-array meeting transcription is particularly challenging due to overlapping speech, ambient noise, and reverberation. To address these issues, we explored three approaches. First, we employ a multichannel speaker separation model to isolate individual speakers, followed by a single-channel, single-speaker automatic speech recognition (ASR) model to transcribe the separated and enhanced audio. This method effectively enhances speech quality for ASR. Second, we propose an end-to-end multichannel speaker-attributed ASR (MC-SA-ASR) model, which builds on an existing single-channel SA-ASR model and incorporates a multichannel Conformer-based encoder with multi-frame cross-channel attention (MFCCA). Unlike traditional approaches that require a multichannel front-end speech enhancement model, the MC-SA-ASR model handles far-field microphones in an end-to-end manner. We also experimented with different input features, including Mel filterbank and phase features, for that model. Lastly, we incorporate a multichannel beamforming and enhancement model as a front-end processing step, followed by a single-channel SA-ASR model to process the enhanced multi-speaker speech signals. We tested different fixed, hybrid, and fully neural network-based beamformers and proposed to jointly optimize the neural beamformer and SA-ASR models using the training objective for the latter. In addition to these methods, we developed a meeting transcription pipeline that integrates voice activity detection, speaker diarization, and SA-ASR to process real meeting recordings effectively. Experimental results indicate that, while using a speaker separation model can enhance speech quality, separation errors can propagate to ASR, resulting in suboptimal performance. A guided speaker separation approach proves to be more effective. Our proposed MC-SA-ASR model demonstrates efficiency in integrating multichannel information and the shared information between the ASR and speaker blocks. Experiments with different input features reveal that models trained with Mel filterbank features perform better in terms of word error rate (WER) and speaker error rate (SER) when the number of channels and speakers is low (2 channels with 1 or 2 speakers). However, for settings with 3 or 4 channels and 3 speakers, models trained with additional phase information outperform those using only Mel filterbank features. This suggests that phase information can enhance ASR by leveraging localization information from multiple channels. Although MFCCA-based MC-SA-ASR outperforms the single-channel SA-ASR and MC-ASR models without a speaker block, the joint beamforming and SA-ASR model further improves the performance. Specifically, joint training of the neural beamformer and SA-ASR yields the best performance, indicating that improving speech quality might be a more direct and efficient approach than using an end-to-end MC-SA-ASR model for multichannel meeting transcription. Furthermore, the study of the real meeting transcription pipeline underscores the potential for better end-to-end models. In our investigation on improving speaker assignment in SA-ASR, we found that the speaker block does not effectively help improve the ASR performance. This highlights the need for improved architectures that more effectively integrate ASR and speaker information
Soldi, Giovanni. "Diarisation du locuteur en temps réel pour les objets intelligents". Electronic Thesis or Diss., Paris, ENST, 2016. http://www.theses.fr/2016ENST0061.
Texto completoOn-line speaker diarization aims to detect “who is speaking now" in a given audio stream. The majority of proposed on-line speaker diarization systems has focused on less challenging domains, such as broadcast news and plenary speeches, characterised by long speaker turns and low spontaneity. The first contribution of this thesis is the development of a completely unsupervised adaptive on-line diarization system for challenging and highly spontaneous meeting data. Due to the obtained high diarization error rates, a semi-supervised approach to on-line diarization, whereby speaker models are seeded with a modest amount of manually labelled data and adapted by an efficient incremental maximum a-posteriori adaptation (MAP) procedure, is proposed. Obtained error rates may be low enough to support practical applications. The second part of the thesis addresses instead the problem of phone normalisation when dealing with short-duration speaker modelling. First, Phone Adaptive Training (PAT), a recently proposed technique, is assessed and optimised at the speaker modelling level and in the context of automatic speaker verification (ASV) and then is further developed towards a completely unsupervised system using automatically generated acoustic class transcriptions, whose number is controlled by regression tree analysis. PAT delivers significant improvements in the performance of a state-of-the-art iVector ASV system even when accurate phonetic transcriptions are not available
Mariotte, Théo. "Traitement automatique de la parole en réunion par dissémination de capteurs". Electronic Thesis or Diss., Le Mans, 2024. http://www.theses.fr/2024LEMA1001.
Texto completoThis thesis work focuses on automatic speech processing, and more specifically on speaker diarization. This task requires the signal to be segmented to identify events such as voice activity, overlapped speech, or speaker changes. This work tackles the scenario where the signal is recorded by a device located in the center of a group of speakers, as in meetings. These conditions lead to a degradation in signal quality due to the distance between the speakers (distant speech).To mitigate this degradation, one approach is to record the signal using a microphone array. The resulting multichannel signal provides information on the spatial distribution of the acoustic field. Two lines of research are being explored for speech segmentation using microphone arrays.The first introduces a method combining acoustic features with spatial features. We propose a new set of features based on the circular harmonics expansion. This approach improves segmentation performance under distant speech conditions while reducing the number of model parameters and improving robustness in case of change in the array geometry.The second proposes several approaches that combine channels using self-attention. Different models, inspired by an existing architecture, are developed. Combining channels also improves segmentation under distant speech conditions. Two of these approaches make feature extraction more interpretable. The proposed distant speech segmentation systems also improve speaker diarization.Channel combination shows poor robustness to changes in the array geometry during inference. To avoid this behavior, a learning procedure is proposed, which improves the robustness in case of array mismatch.Finally, we identified a gap in the public datasets available for distant multichannel automatic speech processing. An acquisition protocol is introduced to build a new dataset, integrating speaker position annotation in addition to speaker diarization.Thus, this work aims to improve the quality of multichannel distant speech segmentation. The proposed methods exploit the spatial information provided by microphone arrays while improving the robustness in case of array mismatch
Milner, Rosanna Margaret. "Using deep neural networks for speaker diarisation". Thesis, University of Sheffield, 2016. http://etheses.whiterose.ac.uk/16567/.
Texto completoSinclair, Mark. "Speech segmentation and speaker diarisation for transcription and translation". Thesis, University of Edinburgh, 2016. http://hdl.handle.net/1842/20970.
Texto completoKounadis-Bastian, Dionyssos. "Quelques contributions pour la séparation et la diarisation de sources audio dans des mélanges multicanaux convolutifs". Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAM012/document.
Texto completoIn this thesis we address the problem of audio source separation (ASS) for multichannel and underdetermined convolutive mixtures through probabilistic modeling. We focus on three aspects of the problem and make three contributions. Firstly, inspired from the empirically well validated representation of an audio signal, that is know as local Gaussian signal model (LGM) with non-negative matrix factorization (NMF), we propose a Bayesian extension to this, that overcomes some of the limitations of the NMF. We incorporate this representation in a multichannel ASS framework and compare it with the state of the art in ASS, yielding promising results.Secondly, we study how to separate mixtures of moving sources and/or of moving microphones.Movements make the acoustic path between sources and microphones become time-varying.Addresing time-varying audio mixtures appears is not so popular in the ASS literature.Thus, we begin from a state of the art LGM-with-NMF method designed for separating time-invariant audiomixtures and propose an extension that uses a Kalman smoother to track the acoustic path across time.The proposed method is benchmarked against a block-wise adaptation of that state of the art (ran on time segments),and delivers competitive results on both simulated and real-world mixtures.Lastly, we investigate the link between ASS and the task of audio diarisation.Audio diarisation is the recognition of the time intervals of activity of every speaker/source in the mix.Most state of the art ASS methods consider the sources ceaselssly emitting; A hypothesis that can result in spurious signal estimates for a source, in intervals where that source was not emitting.Our aim is that diarisation can aid ASS by indicating the emitting sources at each time frame.To that extent we design a joint framework for simultaneous diarization and ASS,that incorporates a hidden Markov model (HMM) to track the temporal activity of the sources, within a state of the art LGM-with-NMF ASS framework.We compare the proposed method with the state of the art in ASS and audio diarisation tasks.We obtain performances comparable, with the state of the art, in terms of separation and outperformant in terms of diarisation
Tevissen, Yannis. "Diarisation multimodale : vers des modèles robustes et justes en contexte réel". Electronic Thesis or Diss., Institut polytechnique de Paris, 2023. http://www.theses.fr/2023IPPAS014.
Texto completoSpeaker diarization, or the task of automatically determining "who spoke, when?" in an audio or video recording, is one of the pillars of modern conversation analysis systems. On television, the content broadcasted is very diverse and covers about every type of conversation, from calm discussions between two people to impassioned debates and wartime interviews. The archiving and indexing of this content, carried out by the Newsbridge company, requires robust and fair processing methods. In this work, we present two new methods for improving systems' robustness via fusion approaches. The first method focuses on voice activity detection, a necessary pre-processing step for every diarization system. The second is a multimodal approach that takes advantage of the latest advances in natural language processing. We also show that recent advances in diarization systems make the use of speaker diarization realistic, even in critical sectors such as the analysis of large audiovisual archives or the home care of the elderly. Finally, this work shows a new method for evaluating the algorithmic fairness of speaker diarization, with the objective to make its use more responsible
Ouni, Slim. "Parole Multimodale : de la parole articulatoire à la parole audiovisuelle". Habilitation à diriger des recherches, Université de Lorraine, 2013. http://tel.archives-ouvertes.fr/tel-00927119.
Texto completoZwyssig, Erich Paul. "Speech processing using digital MEMS microphones". Thesis, University of Edinburgh, 2013. http://hdl.handle.net/1842/8287.
Texto completoVermigli, Vania <1975>. "Parole parole parole… On connait la chanson omaggio ad Alain Resnais e alla musica francese del ‘900". Master's Degree Thesis, Università Ca' Foscari Venezia, 2020. http://hdl.handle.net/10579/17114.
Texto completoLibros sobre el tema "Diarisation de la parole"
Pozzi, Antonia. Parole. Milano: Garzanti, 1989.
Buscar texto completoOffice, National Audit. Parole. London: Stationery Office, 2000.
Buscar texto completoAlessandra, Cenni y Dino Onorina, eds. Parole. [Milan, Italy]: Garzanti, 2001.
Buscar texto completoCommittee, Connecticut General Assembly Legislative Program Review and Investigations. Board of Parole and parole services. Hartford, CT: The Committee, 1993.
Buscar texto completoNew York (State). Dept. of Audit and Control. Division of Parole, field parole services. [Albany, N.Y.]: The Office, 1990.
Buscar texto completoCattani, Adelino. Come dirlo?: Parole giuste, parole belle. Casoria: Loffredo, 2008.
Buscar texto completoCavalleri, Cesare. Persone & parole. Milano: Ares, 1989.
Buscar texto completoLazzara, Vito. Parole monche. Torino: Genesi editrice, 1992.
Buscar texto completoKästner, Erich. Parole Emil. München: Carl Hanser, 1998.
Buscar texto completoFuschini, Francesco. Parole poverette. 2a ed. Venezia: Marsilio, 1996.
Buscar texto completoCapítulos de libros sobre el tema "Diarisation de la parole"
Levesque, Roger J. R. "Parole". En Encyclopedia of Adolescence, 2036–37. New York, NY: Springer New York, 2011. http://dx.doi.org/10.1007/978-1-4419-1695-2_685.
Texto completoAntolak-Saper, Natalia. "Parole". En The Role of the Media in Criminal Justice Policy, 110–45. London: Routledge, 2022. http://dx.doi.org/10.4324/9781003220299-5.
Texto completoLevesque, Roger J. R. "Parole". En Encyclopedia of Adolescence, 2711–12. Cham: Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-319-33228-4_685.
Texto completoMitford, Jessica. "Parole". En The American Prison Business, 216–27. London: Routledge, 2023. http://dx.doi.org/10.4324/9781003327424-12.
Texto completoGottfredson, Michael R. y Don M. Gottfredson. "Parole Decisions". En Decision Making in Criminal Justice, 229–55. Boston, MA: Springer US, 1988. http://dx.doi.org/10.1007/978-1-4757-9954-5_9.
Texto completoNdiaye, Christiane. "Parole ouverte". En Introduction aux littératures francophones, 269–70. Montréal: Presses de l’Université de Montréal, 2004. http://dx.doi.org/10.4000/books.pum.10663.
Texto completoDanon-Boileau, Laurent. "Parole associative, parole compulsive". En Des psychanalystes en séance, 28–34. Gallimard, 2016. http://dx.doi.org/10.3917/gall.tamet.2016.01.0028.
Texto completoSoumahoro, Maboula. "Parole noire/Noire parole". En Racismes de France, 276–91. La Découverte, 2020. http://dx.doi.org/10.3917/dec.slaou.2020.01.0276.
Texto completo"Parole". En Briefs of Leading Cases in Corrections, 259–86. 6th edition. | New York: Routledge, 2016.: Routledge, 2016. http://dx.doi.org/10.4324/9781315531694-10.
Texto completo"Parole". En Benchmark. I.B.Tauris, 2003. http://dx.doi.org/10.5040/9780755622566.ch-013.
Texto completoActas de conferencias sobre el tema "Diarisation de la parole"
Bissig, Pascal, Klaus-Tycho Foerster, Simon Tanner y Roger Wattenhofer. "Distributed discussion diarisation". En 2017 14th IEEE Annual Consumer Communications & Networking Conference (CCNC). IEEE, 2017. http://dx.doi.org/10.1109/ccnc.2017.7983281.
Texto completoZhang, Yue, Felix Weninger, Boqing Liu, Maximilian Schmitt, Florian Eyben y Björn Schuller. "A Paralinguistic Approach To Speaker Diarisation". En MM '17: ACM Multimedia Conference. New York, NY, USA: ACM, 2017. http://dx.doi.org/10.1145/3123266.3123338.
Texto completoGarau, Giulia, Alfred Dielmann y Hervé Bourlard. "Audio-visual synchronisation for speaker diarisation". En Interspeech 2010. ISCA: ISCA, 2010. http://dx.doi.org/10.21437/interspeech.2010-704.
Texto completoKwon, Youngki, Jee-weon Jung, Hee-Soo Heo, You Jin Kim, Bong-Jin Lee y Joon Son Chung. "Adapting Speaker Embeddings for Speaker Diarisation". En Interspeech 2021. ISCA: ISCA, 2021. http://dx.doi.org/10.21437/interspeech.2021-448.
Texto completoLi, Qiujia, Florian L. Kreyssig, Chao Zhang y Philip C. Woodland. "Discriminative Neural Clustering for Speaker Diarisation". En 2021 IEEE Spoken Language Technology Workshop (SLT). IEEE, 2021. http://dx.doi.org/10.1109/slt48900.2021.9383617.
Texto completoMilner, Rosanna y Thomas Hain. "Segment-oriented evaluation of speaker diarisation performance". En 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2016. http://dx.doi.org/10.1109/icassp.2016.7472721.
Texto completoMilner, Rosanna y Thomas Hain. "DNN-Based Speaker Clustering for Speaker Diarisation". En Interspeech 2016. ISCA, 2016. http://dx.doi.org/10.21437/interspeech.2016-126.
Texto completoSun, G., D. Liu, C. Zhang y P. C. Woodland. "Content-Aware Speaker Embeddings for Speaker Diarisation". En ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021. http://dx.doi.org/10.1109/icassp39728.2021.9414390.
Texto completoAlbanie, Samuel, Gul Varol, Liliane Momeni, Triantafyllos Afouras, Andrew Brown, Chuhan Zhang, Ernesto Coto et al. "SeeHear: Signer Diarisation and a New Dataset". En ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021. http://dx.doi.org/10.1109/icassp39728.2021.9414856.
Texto completoHeo, Hee-Soo, Youngki Kwon, Bong-Jin Lee, You Jin Kim y Jee-Weon Jung. "High-Resolution Embedding Extractor for Speaker Diarisation". En ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2023. http://dx.doi.org/10.1109/icassp49357.2023.10097190.
Texto completoInformes sobre el tema "Diarisation de la parole"
DEPARTMENT OF THE ARMY WASHINGTON DC. Boards, Commissions, and Committees: Army Clemency and Parole Board. Fort Belvoir, VA: Defense Technical Information Center, octubre de 1998. http://dx.doi.org/10.21236/ada401997.
Texto completoPolinsky, A. Mitchell y Paul Riskind. Deterrence and the Optimal Use of Prison, Parole, and Probation. Cambridge, MA: National Bureau of Economic Research, mayo de 2017. http://dx.doi.org/10.3386/w23436.
Texto completoAnwar, Shamena y Hanming Fang. Testing for Racial Prejudice in the Parole Board Release Process: Theory and Evidence. Cambridge, MA: National Bureau of Economic Research, julio de 2012. http://dx.doi.org/10.3386/w18239.
Texto completoKuziemko, Ilyana. Going Off Parole: How the Elimination of Discretionary Prison Release Affects the Social Cost of Crime. Cambridge, MA: National Bureau of Economic Research, septiembre de 2007. http://dx.doi.org/10.3386/w13380.
Texto completoHealth hazard evaluation report: HETA-92-0271-2349, District of Columbia, Board of Parole, Washington, D.C. U.S. Department of Health and Human Services, Public Health Service, Centers for Disease Control and Prevention, National Institute for Occupational Safety and Health, septiembre de 1993. http://dx.doi.org/10.26616/nioshheta9202712349.
Texto completo