Academic literature on the topic 'Diarisation de la parole'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Diarisation de la parole.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "Diarisation de la parole"
Sun, Guangzhi, Chao Zhang, and Philip C. Woodland. "Combination of deep speaker embeddings for diarisation." Neural Networks 141 (September 2021): 372–84. http://dx.doi.org/10.1016/j.neunet.2021.04.020.
Full textZelenák, M., and J. Hernando. "Speaker overlap detection with prosodic features for speaker diarisation." IET Signal Processing 6, no. 8 (October 1, 2012): 798–804. http://dx.doi.org/10.1049/iet-spr.2011.0233.
Full textCassetta, Michele. "Parole, parole, parole." Dental Cadmos 87, no. 01 (September 2019): 588. http://dx.doi.org/10.19256/d.cadmos.09.2019.08.
Full textChiari, Alexis. "Parole parole." Feuillets psychanalytiques N° 2, no. 1 (September 21, 2017): 65–77. http://dx.doi.org/10.3917/fpsy.002.0065.
Full textRichmond, J. L., and B. J. Halkon. "Speaker Diarisation of Vibroacoustic Intelligence from Drone Mounted Laser Doppler Vibrometers." Journal of Physics: Conference Series 2041, no. 1 (October 1, 2021): 012011. http://dx.doi.org/10.1088/1742-6596/2041/1/012011.
Full textTyszler, Jean-Jacques. "Parole vide, parole pleine, parole imposée." Journal français de psychiatrie 45, no. 1 (2017): 70. http://dx.doi.org/10.3917/jfp.045.0070.
Full textSchiavinato, Jacques. "Parole égarée, parole retrouvée." Revue de psychothérapie psychanalytique de groupe 28, no. 1 (1997): 115–27. http://dx.doi.org/10.3406/rppg.1997.1366.
Full textBouville, Jean-Marc. "Parole d’enfant, parole à l’enfant, parole sur l’enfant." Revue de l'enfance et de l'adolescence 94, no. 2 (2016): 7. http://dx.doi.org/10.3917/read.094.0007.
Full textAguiar, Flavio. "Macounaïma : parole perdue, parole retrouvée." Études françaises 28, no. 2-3 (1992): 59. http://dx.doi.org/10.7202/035881ar.
Full textTournon, André. "Parole de badin, parole irrécusable." Réforme, Humanisme, Renaissance 76, no. 1 (2013): 107–18. http://dx.doi.org/10.3406/rhren.2013.3294.
Full textDissertations / Theses on the topic "Diarisation de la parole"
Cui, Can. "Séparation, diarisation et reconnaissance de la parole conjointes pour la transcription automatique de réunions." Electronic Thesis or Diss., Université de Lorraine, 2024. http://www.theses.fr/2024LORR0103.
Full textFar-field microphone-array meeting transcription is particularly challenging due to overlapping speech, ambient noise, and reverberation. To address these issues, we explored three approaches. First, we employ a multichannel speaker separation model to isolate individual speakers, followed by a single-channel, single-speaker automatic speech recognition (ASR) model to transcribe the separated and enhanced audio. This method effectively enhances speech quality for ASR. Second, we propose an end-to-end multichannel speaker-attributed ASR (MC-SA-ASR) model, which builds on an existing single-channel SA-ASR model and incorporates a multichannel Conformer-based encoder with multi-frame cross-channel attention (MFCCA). Unlike traditional approaches that require a multichannel front-end speech enhancement model, the MC-SA-ASR model handles far-field microphones in an end-to-end manner. We also experimented with different input features, including Mel filterbank and phase features, for that model. Lastly, we incorporate a multichannel beamforming and enhancement model as a front-end processing step, followed by a single-channel SA-ASR model to process the enhanced multi-speaker speech signals. We tested different fixed, hybrid, and fully neural network-based beamformers and proposed to jointly optimize the neural beamformer and SA-ASR models using the training objective for the latter. In addition to these methods, we developed a meeting transcription pipeline that integrates voice activity detection, speaker diarization, and SA-ASR to process real meeting recordings effectively. Experimental results indicate that, while using a speaker separation model can enhance speech quality, separation errors can propagate to ASR, resulting in suboptimal performance. A guided speaker separation approach proves to be more effective. Our proposed MC-SA-ASR model demonstrates efficiency in integrating multichannel information and the shared information between the ASR and speaker blocks. Experiments with different input features reveal that models trained with Mel filterbank features perform better in terms of word error rate (WER) and speaker error rate (SER) when the number of channels and speakers is low (2 channels with 1 or 2 speakers). However, for settings with 3 or 4 channels and 3 speakers, models trained with additional phase information outperform those using only Mel filterbank features. This suggests that phase information can enhance ASR by leveraging localization information from multiple channels. Although MFCCA-based MC-SA-ASR outperforms the single-channel SA-ASR and MC-ASR models without a speaker block, the joint beamforming and SA-ASR model further improves the performance. Specifically, joint training of the neural beamformer and SA-ASR yields the best performance, indicating that improving speech quality might be a more direct and efficient approach than using an end-to-end MC-SA-ASR model for multichannel meeting transcription. Furthermore, the study of the real meeting transcription pipeline underscores the potential for better end-to-end models. In our investigation on improving speaker assignment in SA-ASR, we found that the speaker block does not effectively help improve the ASR performance. This highlights the need for improved architectures that more effectively integrate ASR and speaker information
Soldi, Giovanni. "Diarisation du locuteur en temps réel pour les objets intelligents." Electronic Thesis or Diss., Paris, ENST, 2016. http://www.theses.fr/2016ENST0061.
Full textOn-line speaker diarization aims to detect “who is speaking now" in a given audio stream. The majority of proposed on-line speaker diarization systems has focused on less challenging domains, such as broadcast news and plenary speeches, characterised by long speaker turns and low spontaneity. The first contribution of this thesis is the development of a completely unsupervised adaptive on-line diarization system for challenging and highly spontaneous meeting data. Due to the obtained high diarization error rates, a semi-supervised approach to on-line diarization, whereby speaker models are seeded with a modest amount of manually labelled data and adapted by an efficient incremental maximum a-posteriori adaptation (MAP) procedure, is proposed. Obtained error rates may be low enough to support practical applications. The second part of the thesis addresses instead the problem of phone normalisation when dealing with short-duration speaker modelling. First, Phone Adaptive Training (PAT), a recently proposed technique, is assessed and optimised at the speaker modelling level and in the context of automatic speaker verification (ASV) and then is further developed towards a completely unsupervised system using automatically generated acoustic class transcriptions, whose number is controlled by regression tree analysis. PAT delivers significant improvements in the performance of a state-of-the-art iVector ASV system even when accurate phonetic transcriptions are not available
Mariotte, Théo. "Traitement automatique de la parole en réunion par dissémination de capteurs." Electronic Thesis or Diss., Le Mans, 2024. http://www.theses.fr/2024LEMA1001.
Full textThis thesis work focuses on automatic speech processing, and more specifically on speaker diarization. This task requires the signal to be segmented to identify events such as voice activity, overlapped speech, or speaker changes. This work tackles the scenario where the signal is recorded by a device located in the center of a group of speakers, as in meetings. These conditions lead to a degradation in signal quality due to the distance between the speakers (distant speech).To mitigate this degradation, one approach is to record the signal using a microphone array. The resulting multichannel signal provides information on the spatial distribution of the acoustic field. Two lines of research are being explored for speech segmentation using microphone arrays.The first introduces a method combining acoustic features with spatial features. We propose a new set of features based on the circular harmonics expansion. This approach improves segmentation performance under distant speech conditions while reducing the number of model parameters and improving robustness in case of change in the array geometry.The second proposes several approaches that combine channels using self-attention. Different models, inspired by an existing architecture, are developed. Combining channels also improves segmentation under distant speech conditions. Two of these approaches make feature extraction more interpretable. The proposed distant speech segmentation systems also improve speaker diarization.Channel combination shows poor robustness to changes in the array geometry during inference. To avoid this behavior, a learning procedure is proposed, which improves the robustness in case of array mismatch.Finally, we identified a gap in the public datasets available for distant multichannel automatic speech processing. An acquisition protocol is introduced to build a new dataset, integrating speaker position annotation in addition to speaker diarization.Thus, this work aims to improve the quality of multichannel distant speech segmentation. The proposed methods exploit the spatial information provided by microphone arrays while improving the robustness in case of array mismatch
Milner, Rosanna Margaret. "Using deep neural networks for speaker diarisation." Thesis, University of Sheffield, 2016. http://etheses.whiterose.ac.uk/16567/.
Full textSinclair, Mark. "Speech segmentation and speaker diarisation for transcription and translation." Thesis, University of Edinburgh, 2016. http://hdl.handle.net/1842/20970.
Full textKounadis-Bastian, Dionyssos. "Quelques contributions pour la séparation et la diarisation de sources audio dans des mélanges multicanaux convolutifs." Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAM012/document.
Full textIn this thesis we address the problem of audio source separation (ASS) for multichannel and underdetermined convolutive mixtures through probabilistic modeling. We focus on three aspects of the problem and make three contributions. Firstly, inspired from the empirically well validated representation of an audio signal, that is know as local Gaussian signal model (LGM) with non-negative matrix factorization (NMF), we propose a Bayesian extension to this, that overcomes some of the limitations of the NMF. We incorporate this representation in a multichannel ASS framework and compare it with the state of the art in ASS, yielding promising results.Secondly, we study how to separate mixtures of moving sources and/or of moving microphones.Movements make the acoustic path between sources and microphones become time-varying.Addresing time-varying audio mixtures appears is not so popular in the ASS literature.Thus, we begin from a state of the art LGM-with-NMF method designed for separating time-invariant audiomixtures and propose an extension that uses a Kalman smoother to track the acoustic path across time.The proposed method is benchmarked against a block-wise adaptation of that state of the art (ran on time segments),and delivers competitive results on both simulated and real-world mixtures.Lastly, we investigate the link between ASS and the task of audio diarisation.Audio diarisation is the recognition of the time intervals of activity of every speaker/source in the mix.Most state of the art ASS methods consider the sources ceaselssly emitting; A hypothesis that can result in spurious signal estimates for a source, in intervals where that source was not emitting.Our aim is that diarisation can aid ASS by indicating the emitting sources at each time frame.To that extent we design a joint framework for simultaneous diarization and ASS,that incorporates a hidden Markov model (HMM) to track the temporal activity of the sources, within a state of the art LGM-with-NMF ASS framework.We compare the proposed method with the state of the art in ASS and audio diarisation tasks.We obtain performances comparable, with the state of the art, in terms of separation and outperformant in terms of diarisation
Tevissen, Yannis. "Diarisation multimodale : vers des modèles robustes et justes en contexte réel." Electronic Thesis or Diss., Institut polytechnique de Paris, 2023. http://www.theses.fr/2023IPPAS014.
Full textSpeaker diarization, or the task of automatically determining "who spoke, when?" in an audio or video recording, is one of the pillars of modern conversation analysis systems. On television, the content broadcasted is very diverse and covers about every type of conversation, from calm discussions between two people to impassioned debates and wartime interviews. The archiving and indexing of this content, carried out by the Newsbridge company, requires robust and fair processing methods. In this work, we present two new methods for improving systems' robustness via fusion approaches. The first method focuses on voice activity detection, a necessary pre-processing step for every diarization system. The second is a multimodal approach that takes advantage of the latest advances in natural language processing. We also show that recent advances in diarization systems make the use of speaker diarization realistic, even in critical sectors such as the analysis of large audiovisual archives or the home care of the elderly. Finally, this work shows a new method for evaluating the algorithmic fairness of speaker diarization, with the objective to make its use more responsible
Ouni, Slim. "Parole Multimodale : de la parole articulatoire à la parole audiovisuelle." Habilitation à diriger des recherches, Université de Lorraine, 2013. http://tel.archives-ouvertes.fr/tel-00927119.
Full textZwyssig, Erich Paul. "Speech processing using digital MEMS microphones." Thesis, University of Edinburgh, 2013. http://hdl.handle.net/1842/8287.
Full textVermigli, Vania <1975>. "Parole parole parole… On connait la chanson omaggio ad Alain Resnais e alla musica francese del ‘900." Master's Degree Thesis, Università Ca' Foscari Venezia, 2020. http://hdl.handle.net/10579/17114.
Full textBooks on the topic "Diarisation de la parole"
Pozzi, Antonia. Parole. Milano: Garzanti, 1989.
Find full textOffice, National Audit. Parole. London: Stationery Office, 2000.
Find full textAlessandra, Cenni, and Dino Onorina, eds. Parole. [Milan, Italy]: Garzanti, 2001.
Find full textCommittee, Connecticut General Assembly Legislative Program Review and Investigations. Board of Parole and parole services. Hartford, CT: The Committee, 1993.
Find full textNew York (State). Dept. of Audit and Control. Division of Parole, field parole services. [Albany, N.Y.]: The Office, 1990.
Find full textCattani, Adelino. Come dirlo?: Parole giuste, parole belle. Casoria: Loffredo, 2008.
Find full textCavalleri, Cesare. Persone & parole. Milano: Ares, 1989.
Find full textLazzara, Vito. Parole monche. Torino: Genesi editrice, 1992.
Find full textKästner, Erich. Parole Emil. München: Carl Hanser, 1998.
Find full textFuschini, Francesco. Parole poverette. 2nd ed. Venezia: Marsilio, 1996.
Find full textBook chapters on the topic "Diarisation de la parole"
Levesque, Roger J. R. "Parole." In Encyclopedia of Adolescence, 2036–37. New York, NY: Springer New York, 2011. http://dx.doi.org/10.1007/978-1-4419-1695-2_685.
Full textAntolak-Saper, Natalia. "Parole." In The Role of the Media in Criminal Justice Policy, 110–45. London: Routledge, 2022. http://dx.doi.org/10.4324/9781003220299-5.
Full textLevesque, Roger J. R. "Parole." In Encyclopedia of Adolescence, 2711–12. Cham: Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-319-33228-4_685.
Full textMitford, Jessica. "Parole." In The American Prison Business, 216–27. London: Routledge, 2023. http://dx.doi.org/10.4324/9781003327424-12.
Full textGottfredson, Michael R., and Don M. Gottfredson. "Parole Decisions." In Decision Making in Criminal Justice, 229–55. Boston, MA: Springer US, 1988. http://dx.doi.org/10.1007/978-1-4757-9954-5_9.
Full textNdiaye, Christiane. "Parole ouverte." In Introduction aux littératures francophones, 269–70. Montréal: Presses de l’Université de Montréal, 2004. http://dx.doi.org/10.4000/books.pum.10663.
Full textDanon-Boileau, Laurent. "Parole associative, parole compulsive." In Des psychanalystes en séance, 28–34. Gallimard, 2016. http://dx.doi.org/10.3917/gall.tamet.2016.01.0028.
Full textSoumahoro, Maboula. "Parole noire/Noire parole." In Racismes de France, 276–91. La Découverte, 2020. http://dx.doi.org/10.3917/dec.slaou.2020.01.0276.
Full text"Parole." In Briefs of Leading Cases in Corrections, 259–86. 6th edition. | New York: Routledge, 2016.: Routledge, 2016. http://dx.doi.org/10.4324/9781315531694-10.
Full text"Parole." In Benchmark. I.B.Tauris, 2003. http://dx.doi.org/10.5040/9780755622566.ch-013.
Full textConference papers on the topic "Diarisation de la parole"
Bissig, Pascal, Klaus-Tycho Foerster, Simon Tanner, and Roger Wattenhofer. "Distributed discussion diarisation." In 2017 14th IEEE Annual Consumer Communications & Networking Conference (CCNC). IEEE, 2017. http://dx.doi.org/10.1109/ccnc.2017.7983281.
Full textZhang, Yue, Felix Weninger, Boqing Liu, Maximilian Schmitt, Florian Eyben, and Björn Schuller. "A Paralinguistic Approach To Speaker Diarisation." In MM '17: ACM Multimedia Conference. New York, NY, USA: ACM, 2017. http://dx.doi.org/10.1145/3123266.3123338.
Full textGarau, Giulia, Alfred Dielmann, and Hervé Bourlard. "Audio-visual synchronisation for speaker diarisation." In Interspeech 2010. ISCA: ISCA, 2010. http://dx.doi.org/10.21437/interspeech.2010-704.
Full textKwon, Youngki, Jee-weon Jung, Hee-Soo Heo, You Jin Kim, Bong-Jin Lee, and Joon Son Chung. "Adapting Speaker Embeddings for Speaker Diarisation." In Interspeech 2021. ISCA: ISCA, 2021. http://dx.doi.org/10.21437/interspeech.2021-448.
Full textLi, Qiujia, Florian L. Kreyssig, Chao Zhang, and Philip C. Woodland. "Discriminative Neural Clustering for Speaker Diarisation." In 2021 IEEE Spoken Language Technology Workshop (SLT). IEEE, 2021. http://dx.doi.org/10.1109/slt48900.2021.9383617.
Full textMilner, Rosanna, and Thomas Hain. "Segment-oriented evaluation of speaker diarisation performance." In 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2016. http://dx.doi.org/10.1109/icassp.2016.7472721.
Full textMilner, Rosanna, and Thomas Hain. "DNN-Based Speaker Clustering for Speaker Diarisation." In Interspeech 2016. ISCA, 2016. http://dx.doi.org/10.21437/interspeech.2016-126.
Full textSun, G., D. Liu, C. Zhang, and P. C. Woodland. "Content-Aware Speaker Embeddings for Speaker Diarisation." In ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021. http://dx.doi.org/10.1109/icassp39728.2021.9414390.
Full textAlbanie, Samuel, Gul Varol, Liliane Momeni, Triantafyllos Afouras, Andrew Brown, Chuhan Zhang, Ernesto Coto, et al. "SeeHear: Signer Diarisation and a New Dataset." In ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021. http://dx.doi.org/10.1109/icassp39728.2021.9414856.
Full textHeo, Hee-Soo, Youngki Kwon, Bong-Jin Lee, You Jin Kim, and Jee-Weon Jung. "High-Resolution Embedding Extractor for Speaker Diarisation." In ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2023. http://dx.doi.org/10.1109/icassp49357.2023.10097190.
Full textReports on the topic "Diarisation de la parole"
DEPARTMENT OF THE ARMY WASHINGTON DC. Boards, Commissions, and Committees: Army Clemency and Parole Board. Fort Belvoir, VA: Defense Technical Information Center, October 1998. http://dx.doi.org/10.21236/ada401997.
Full textPolinsky, A. Mitchell, and Paul Riskind. Deterrence and the Optimal Use of Prison, Parole, and Probation. Cambridge, MA: National Bureau of Economic Research, May 2017. http://dx.doi.org/10.3386/w23436.
Full textAnwar, Shamena, and Hanming Fang. Testing for Racial Prejudice in the Parole Board Release Process: Theory and Evidence. Cambridge, MA: National Bureau of Economic Research, July 2012. http://dx.doi.org/10.3386/w18239.
Full textKuziemko, Ilyana. Going Off Parole: How the Elimination of Discretionary Prison Release Affects the Social Cost of Crime. Cambridge, MA: National Bureau of Economic Research, September 2007. http://dx.doi.org/10.3386/w13380.
Full textHealth hazard evaluation report: HETA-92-0271-2349, District of Columbia, Board of Parole, Washington, D.C. U.S. Department of Health and Human Services, Public Health Service, Centers for Disease Control and Prevention, National Institute for Occupational Safety and Health, September 1993. http://dx.doi.org/10.26616/nioshheta9202712349.
Full text