Dissertations / Theses on the topic 'Machine à noyaux'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Machine à noyaux.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Dehlinger, Nicolas. "Étude des performances d'une machine à flux transverse à noyaux ferromagnétiques amorphes." Thesis, Université Laval, 2007. http://www.theses.ulaval.ca/2007/24934/24934.pdf.
Full textTransverse flux machines (TFM) are known for their excellent torque-to-mass and torque-to-volume ratio when compared to conventional machines. Despite this advantage, they have some serious shortcomings like complex construction and high cost, explaining why TFM that can be found in the literature are usually only prototypes. Moreover, the TFM shows a dependence of its force density upon its pole pitch and airgap thickness, which leads to high electrical frequencies and thus to high core losses. For all these reasons, this type of machine could be considered in high-torque low-speed applications such as wind turbines or electrical traction drives. The work presented in this document contributes to the development of a new TFM configuration: the claw-pole TFM with hybrid stator (CPTFMHS). Such a stator built from a combination of Fe-Si laminations and powdered iron (SMC), enables reducing iron losses significantly and improving the ease of manufacturing of the machine. The concept of the hybrid stator can be further developed by using new magnetic materials with lower specific losses. The substitution of Fe-Si laminations by amorphous cores in the stator of the CPTFMHS is studied in this work and presented here. Experimental measurements are conducted on a one-pole pair CPTFMHS machine with an amorphous core: the results show a reduction of the total iron losses, thus proving benefits of amorphous cores used in the machine. Finite element simulations coupled with experimental measurements lead to the following conclusion: the efficiency of a CPTFMHS machine can be maintained to a high level at frequencies above 400 Hz, thanks to the use of amorphous cores, which may not be possible with Fe-Si laminations.
Bietti, Alberto. "Méthodes à noyaux pour les réseaux convolutionnels profonds." Thesis, Université Grenoble Alpes (ComUE), 2019. http://www.theses.fr/2019GREAM051.
Full textThe increased availability of large amounts of data, from images in social networks, speech waveforms from mobile devices, and large text corpuses, to genomic and medical data, has led to a surge of machine learning techniques. Such methods exploit statistical patterns in these large datasets for making accurate predictions on new data. In recent years, deep learning systems have emerged as a remarkably successful class of machine learning algorithms, which rely on gradient-based methods for training multi-layer models that process data in a hierarchical manner. These methods have been particularly successful in tasks where the data consists of natural signals such as images or audio; this includes visual recognition, object detection or segmentation, and speech recognition.For such tasks, deep learning methods often yield the best known empirical performance; yet, the high dimensionality of the data and large number of parameters of these models make them challenging to understand theoretically. Their success is often attributed in part to their ability to exploit useful structure in natural signals, such as local stationarity or invariance, for instance through choices of network architectures with convolution and pooling operations. However, such properties are still poorly understood from a theoretical standpoint, leading to a growing gap between the theory and practice of machine learning. This thesis is aimed towards bridging this gap, by studying spaces of functions which arise from given network architectures, with a focus on the convolutional case. Our study relies on kernel methods, by considering reproducing kernel Hilbert spaces (RKHSs) associated to certain kernels that are constructed hierarchically based on a given architecture. This allows us to precisely study smoothness, invariance, stability to deformations, and approximation properties of functions in the RKHS. These representation properties are also linked with optimization questions when training deep networks with gradient methods in some over-parameterized regimes where such kernels arise. They also suggest new practical regularization strategies for obtaining better generalization performance on small datasets, and state-of-the-art performance for adversarial robustness on image tasks
Giffon, Luc. "Approximations parcimonieuses et méthodes à noyaux pour la compression de modèles d'apprentissage." Electronic Thesis or Diss., Aix-Marseille, 2020. http://www.theses.fr/2020AIXM0354.
Full textThis thesis aims at studying and experimentally validating the benefits, in terms of amount of computation and data needed, that kernel methods and sparse approximation methods can bring to existing machine learning algorithms. In a first part of this thesis, we propose a new type of neural architecture that uses a kernel function to reduce the number of learnable parameters, thus making it robust to overfiting in a regime where few labeled observations are available. In a second part of this thesis, we seek to reduce the complexity of existing machine learning models by including sparse approximations. First, we propose an alternative algorithm to the K-means algorithm which allows to speed up the inference phase by expressing the centroids as a product of sparse matrices. In addition to the convergence guarantees of the proposed algorithm, we provide an experimental validation of both the quality of the centroids thus expressed and their benefit in terms of computational cost. Then, we explore the compression of neural networks by replacing the matrices that constitute its layers with sparse matrix products. Finally, we hijack the Orthogonal Matching Pursuit (OMP) sparse approximation algorithm to make a weighted selection of decisiontrees from a random forest, we analyze the effect of the weights obtained and we propose a non-negative alternative to the method that outperforms all other tree selectiontechniques considered on a large panel of data sets
Tian, Xilan. "Apprentissage et noyau pour les interfaces cerveau-machine." Phd thesis, INSA de Rouen, 2012. http://tel.archives-ouvertes.fr/tel-00735891.
Full textXilan, Tian. "Apprentissage et Noyau pour les Interfaces Cerveau-machine." Phd thesis, INSA de Rouen, 2012. http://tel.archives-ouvertes.fr/tel-00699659.
Full textLouradour, Jérôme. "Noyaux de séquences pour la vérification du locuteur par machines à vecteurs de support." Toulouse 3, 2007. http://www.theses.fr/2007TOU30004.
Full textThis thesis is focused on the application of Support Vector Machines (SVM) to Automatic Text-Independent Speaker Verification. This speech processing task consists in determining whether a speech utterance was pronounced or not by a target speaker, without any constraint on the speech content. In order to apply a kernel method such as SVM to this binary classification of variable-length sequences, an appropriate approach is to use kernels that can handle sequences, and not acoustic vectors within sequences. As explained in the thesis report, both theoretical and practical reasons justify the effort of searching such kernels. The present study concentrates in exploring several aspects of kernels for sequences, and in applying them to a very large database speaker verification problem under realistic recording conditions. After reviewing emergent methods to conceive sequence kernels and presenting them in a unified framework, we propose a new family of such kernels : the Feature Space Normalized Sequence (FSNS) kernels. These kernels are a generalization of the GLDS kernel, which is now well-known for its efficiency in speaker verification. A theoretical and algorithmic study of FSNS kernels is carried out. In particular, several forms are introduced and justified, and a sparse greedy matrix approximation method is used to suggest an efficient and suitable implementation of FSNS kernels for speaker verification. .
Palazzo, Martin. "Dimensionality Reduction of Biomedical Tumor Profiles : a Machine Learning Approach." Thesis, Troyes, 2021. http://www.theses.fr/2021TROY0031.
Full textThe increasing pace of data generation from tumor profiles profiles during the last decade has enable the development of statistical learning algorithms to explore and analyze the landscape of tumor types, subtypes and patient survival from a biomolecular point of view. Tumor data is mainly described by trasncriptomic features and the level of expression of a given gene-transcript in the tumor cell, therefore these features can be used to learn statistical rules that improves the understanding about the state and type of a cancer cell. Nevertheless transcriptomic tumor data is high dimensional and each tumor can be described by thousands of gene features making it difficult to perform a machine learning task and to understand the underlying biological mechanisms. This thesis studies how to reduce dimensionality and to gain interpretability about which genes encode signals of the data distribution by proposing dimension reduction methods based on Feature Selection and Feature Extraction pipelines. The proposed methods are based on Latent Variable Models and Kernel Methods with the idea to explore the connection between pair-wise similarity functions of tumor samples and low dimensional latent spaces that captures the inner structure of the training data. Proposed methods have shown improvements in supervised and unsupervised feature selection tasks when compared with benchmark methods to classify and learn subgroups of tumors respectively
Abdallah, Fahed. "Noyaux reproduisants et critères de contraste pour l'élaboration de détecteurs à structure imposée." Troyes, 2004. http://www.theses.fr/2004TROY0002.
Full textIn this thesis, we consider statistical learning machines with try to infer rules from a given set or observations in order to make correct predictions on unseen examples. Building upon the theory of reproducing kernels, we develop a generalized linear detector in transformed spaces of high dimension, without explicitly doing any calculus in these spaces. The method is based on the optimization of the best second-order criterion with respect to the problem to solve. In fact, theoretical results show that second-order criteria are able, under some mild conditions, to guarantee the best solution in the sense of classical detection theories. Achieving a good generalisation performance with a receiver requires matching its complexity to the amount of available training data. This problem, known as the curse of dimensionality, has been studied theoretically by Vapnik and Chervonenkis. In this dissertation, we propose complexity control procedures in order to improve the performance of these receivers when few training data are available. Simulation results on real and synthetic data show clearly the competitiveness of our approach compared with other state of the art existing kernel methods like Support Vector Machines
Labbé, Benjamin. "Machines à noyaux pour le filtrage d'alarmes : application à la discrimination multiclasse en environnement maritime." Thesis, Rouen, INSA, 2011. http://www.theses.fr/2011ISAM0002.
Full textInfrared systems are keys to provide automatic control of threats to military forces. Such operational systems are constrained to real-time processing and high efficiency (low false-alarm rate) implying the recognition of threats among numerous irrelevant objects.In this document, we combine OneClass Support Vector Machines (SVM) to discriminate in the multiclass framework and to reject unknown objects (preserving the false-alarm rate).While learning, we perform variable selection to control the sparsity of the decision functions. We also introduce a new classifier, the Discriminative OneClass-SVM. It combines properties of both the biclass-SVM and the OneClass-SVM in a multiclass framework. This classifier detects novelty and has no dependency to the amount of categories, allowing to tackle large scale problems. Numerical experiments, on real world infrared datasets, demonstrate the relevance of our proposals for highly constrained systems, when compared to standard methods
Brault, Romain. "Large-scale operator-valued kernel regression." Thesis, Université Paris-Saclay (ComUE), 2017. http://www.theses.fr/2017SACLE024/document.
Full textMany problems in Machine Learning can be cast into vector-valued approximation. Operator-Valued Kernels and vector-valued Reproducing Kernel Hilbert Spaces provide a theoretical and practical framework to address that issue, extending nicely the well-known setting of scalar-valued kernels. However large scale applications are usually not affordable with these tools that require an important computational power along with a large memory capacity. In this thesis, we propose and study scalable methods to perform regression with Operator-Valued Kernels. To achieve this goal, we extend Random Fourier Features, an approximation technique originally introduced for scalar-valued kernels, to Operator-Valued Kernels. The idea is to take advantage of an approximated operator-valued feature map in order to come up with a linear model in a finite-dimensional space. This thesis is structured as follows. First we develop a general framework devoted to the approximation of shift-invariant MErcer kernels on Locally Compact Abelian groups and study their properties along with the complexity of the algorithms based on them. Second we show theoretical guarantees by bounding the error due to the approximation, with high probability. Third, we study various applications of Operator Random Fourier Features (ORFF) to different tasks of Machine learning such as multi-class classification, multi-task learning, time serie modelling, functionnal regression and anomaly detection. We also compare the proposed framework with other state of the art methods. Fourth, we conclude by drawing short-term and mid-term perspectives of this work
Chen, Dexiong. "Modélisation de données structurées avec des machines profondes à noyaux et des applications en biologie computationnelle." Thesis, Université Grenoble Alpes, 2020. http://www.theses.fr/2020GRALM070.
Full textDeveloping efficient algorithms to learn appropriate representations of structured data, including sequences or graphs, is a major and central challenge in machine learning. To this end, deep learning has become popular in structured data modeling. Deep neural networks have drawn particular attention in various scientific fields such as computer vision, natural language understanding or biology. For instance, they provide computational tools for biologists to possibly understand and uncover biological properties or relationships among macromolecules within living organisms. However, most of the success of deep learning methods in these fields essentially relies on the guidance of empirical insights as well as huge amounts of annotated data. Exploiting more data-efficient models is necessary as labeled data is often scarce.Another line of research is kernel methods, which provide a systematic and principled approach for learning non-linear models from data of arbitrary structure. In addition to their simplicity, they exhibit a natural way to control regularization and thus to avoid overfitting.However, the data representations provided by traditional kernel methods are only defined by simply designed hand-crafted features, which makes them perform worse than neural networks when enough labeled data are available. More complex kernels inspired by prior knowledge used in neural networks have thus been developed to build richer representations and thus bridge this gap. Yet, they are less scalable. By contrast, neural networks are able to learn a compact representation for a specific learning task, which allows them to retain the expressivity of the representation while scaling to large sample size.Incorporating complementary views of kernel methods and deep neural networks to build new frameworks is therefore useful to benefit from both worlds.In this thesis, we build a general kernel-based framework for modeling structured data by leveraging prior knowledge from classical kernel methods and deep networks. Our framework provides efficient algorithmic tools for learning representations without annotations as well as for learning more compact representations in a task-driven way. Our framework can be used to efficiently model sequences and graphs with simple interpretation of predictions. It also offers new insights about designing more expressive kernels and neural networks for sequences and graphs
Loquin, Kevin. "De l'utilisation des noyaux maxitifs en traitement de l'information." Phd thesis, Montpellier 2, 2008. http://www.theses.fr/2008MON20108.
Full textBeaucoup d'algorithmes en traitement du signal ou en statistiques utilisent, de façon plus ou moins explicite, la notion d'espérance mathématique associée à une représentation probabiliste du voisinage d'un point, que nous appelons noyau sommatif. Nous regroupons ainsi, sous la dénomination d'extraction sommative d'informations, des méthodes aussi diverses que la modélisation de la mesure, le filtrage linéaire, les processus d'échantillonnage, de reconstruction et de dérivation d'un signal numérique, l'estimation de densité de probabilité et de fonction de répartition par noyau ou par histogramme,...
Comme alternative à l'extraction sommative d'informations, nous présentons la méthode d'extraction maxitive d'informations qui utilise l'intégrale de Choquet associée à une représentation possibiliste du voisinage d'un point, que nous appelons noyau maxitif. La méconnaissance sur le noyau sommatif est prise en compte par le fait qu'un noyau maxitif représente une famille de noyaux sommatifs. De plus, le résultat intervalliste de l'extraction maxitive d'informations est l'ensemble des résultats ponctuels des extractions sommatives d'informations obtenues avec les noyaux sommatifs de la famille représentée par le noyau maxitif utilisé. En plus de cette justification théorique, nous présentons une série d'applications de l'extraction maxitive d'informations en statistiques et en traitement du signal qui constitue une boîte à outils à enrichir et à utiliser sur des cas réels.
Loquin, Kevin. "De l'utilisation des noyaux maxitifs en traitement de l'information." Phd thesis, Université Montpellier II - Sciences et Techniques du Languedoc, 2008. http://tel.archives-ouvertes.fr/tel-00356477.
Full textBeaucoup d'algorithmes en traitement du signal ou en statistiques utilisent, de façon plus ou moins explicite, la notion d'espérance mathématique associée à une représentation probabiliste du voisinage d'un point, que nous appelons noyau sommatif. Nous regroupons ainsi, sous la dénomination d'extraction sommative d'informations, des méthodes aussi diverses que la modélisation de la mesure, le filtrage linéaire, les processus d'échantillonnage, de reconstruction et de dérivation d'un signal numérique, l'estimation de densité de probabilité et de fonction de répartition par noyau ou par histogramme,...
Comme alternative à l'extraction sommative d'informations, nous présentons la méthode d'extraction maxitive d'informations qui utilise l'intégrale de Choquet associée à une représentation possibiliste du voisinage d'un point, que nous appelons noyau maxitif. La méconnaissance sur le noyau sommatif est prise en compte par le fait qu'un noyau maxitif représente une famille de noyaux sommatifs. De plus, le résultat intervalliste de l'extraction maxitive d'informations est l'ensemble des résultats ponctuels des extractions sommatives d'informations obtenues avec les noyaux sommatifs de la famille représentée par le noyau maxitif utilisé. En plus de cette justification théorique, nous présentons une série d'applications de l'extraction maxitive d'informations en statistiques et en traitement du signal qui constitue une boîte à outils à enrichir et à utiliser sur des cas réels.
Chen, Xiaoyi. "Transfer Learning with Kernel Methods." Thesis, Troyes, 2018. http://www.theses.fr/2018TROY0005.
Full textTransfer Learning aims to take advantage of source data to help the learning task of related but different target data. This thesis contributes to homogeneous transductive transfer learning where no labeled target data is available. In this thesis, we relax the constraint on conditional probability of labels required by covariate shift to be more and more general, based on which the alignment of marginal probabilities of source and target observations renders source and target similar. Thus, firstly, a maximum likelihood based approach is proposed. Secondly, SVM is adapted to transfer learning with an extra MMD-like constraint where Maximum Mean Discrepancy (MMD) measures this similarity. Thirdly, KPCA is used to align data in a RKHS on minimizing MMD. We further develop the KPCA based approach so that a linear transformation in the input space is enough for a good and robust alignment in the RKHS. Experimentally, our proposed approaches are very promising
Gaüzère, Benoit. "Application des méthodes à noyaux sur graphes pour la prédiction des propriétés des molécules." Phd thesis, Université de Caen, 2013. http://tel.archives-ouvertes.fr/tel-00933187.
Full textFranchi, Gianni. "Machine learning spatial appliquée aux images multivariées et multimodales." Thesis, Paris Sciences et Lettres (ComUE), 2016. http://www.theses.fr/2016PSLEM071/document.
Full textThis thesis focuses on multivariate spatial statistics and machine learning applied to hyperspectral and multimodal and images in remote sensing and scanning electron microscopy (SEM). In this thesis the following topics are considered:Fusion of images:SEM allows us to acquire images from a given sample using different modalities. The purpose of these studies is to analyze the interest of fusion of information to improve the multimodal SEM images acquisition. We have modeled and implemented various techniques of image fusion of information, based in particular on spatial regression theory. They have been assessed on various datasets.Spatial classification of multivariate image pixels:We have proposed a novel approach for pixel classification in multi/hyper-spectral images. The aim of this technique is to represent and efficiently describe the spatial/spectral features of multivariate images. These multi-scale deep descriptors aim at representing the content of the image while considering invariances related to the texture and to its geometric transformations.Spatial dimensionality reduction:We have developed a technique to extract a feature space using morphological principal component analysis. Indeed, in order to take into account the spatial and structural information we used mathematical morphology operators
Flamary, Rémi. "Apprentissage statistique pour le signal : applications aux interfaces cerveau-machine." Phd thesis, Université de Rouen, 2011. http://tel.archives-ouvertes.fr/tel-00687501.
Full textFlamary, Rémi. "Apprentissage statistique pour le signal : applications aux interfaces cerveau-machine." Phd thesis, Rouen, 2011. http://www.theses.fr/2011ROUES044.
Full textBrain Computer Interfaces (BCI) require the use of statistical learning methods for signal recognition. In this thesis we propose a general approach using prior knowledge on the problem at hand through regularization. To this end, we learn jointly the classifier and the feature extraction step in a unique optimization problem. We focus on the problem of sensor selection, and propose several regularization terms adapted to the problem. Our first contribution is a filter learning method called large margin filtering. It consists in learning a filtering maximizing the margin between samples of each classe so as to adapt to the properties of the features. In addition, this approach is easy to interpret and can lead to the selection of the most relevant sensors. Numerical experiments on a real life BCI problem and a 2D image classification show the good behaviour of our method both in terms of performance and interpretability. The second contribution is a general sparse multitask learning approach. Several classifiers are learned jointly and discriminant kernels for all the tasks are automatically selected. We propose some efficient algorithms and numerical experiments have shown the interest of our approach. Finally, the third contribution is a direct application of the sparse multitask learning to a BCI event-related potential classification problem. We propose an adapted regularization term that promotes both sensor selection and similarity between the classifiers. Numerical experiments show that the calibration time of a BCI can be drastically reduced thanks to the proposed multitask approach
Nguyen, Van Toi. "Visual interpretation of hand postures for human-machine interaction." Thesis, La Rochelle, 2015. http://www.theses.fr/2015LAROS035/document.
Full textNowadays, people want to interact with machines more naturally. One of the powerful communication channels is hand gesture. Vision-based approach has involved many researchers because this approach does not require any extra device. One of the key problems we need to resolve is hand posture recognition on RGB images because it can be used directly or integrated into a multi-cues hand gesture recognition. The main challenges of this problem are illumination differences, cluttered background, background changes, high intra-class variation, and high inter-class similarity. This thesis proposes a hand posture recognition system consists two phases that are hand detection and hand posture recognition. In hand detection step, we employed Viola-Jones detector with proposed concept Internal Haar-like feature. The proposed hand detection works in real-time within frames captured from real complex environments and avoids unexpected effects of background. The proposed detector outperforms original Viola-Jones detector using traditional Haar-like feature. In hand posture recognition step, we proposed a new hand representation based on a good generic descriptor that is kernel descriptor (KDES). When applying KDES into hand posture recognition, we proposed three improvements to make it more robust that are adaptive patch, normalization of gradient orientation in patches, and hand pyramid structure. The improvements make KDES invariant to scale change, patch-level feature invariant to rotation, and final hand representation suitable to hand structure. Based on these improvements, the proposed method obtains better results than original KDES and a state of the art method
Freyssinet, André. "Architecture et réalisation d'un système réparti à objets." Grenoble 1, 1991. http://www.theses.fr/1991GRE10078.
Full textGautheron, Léo. "Construction de Représentation de Données Adaptées dans le Cadre de Peu d'Exemples Étiquetés." Thesis, Lyon, 2020. http://www.theses.fr/2020LYSES044.
Full textMachine learning consists in the study and design of algorithms that build models able to handle non trivial tasks as well as or better than humans and hopefully at a lesser cost.These models are typically trained from a dataset where each example describes an instance of the same task and is represented by a set of characteristics and an expected outcome or label which we usually want to predict.An element required for the success of any machine learning algorithm is related to the quality of the set of characteristics describing the data, also referred as data representation or features.In supervised learning, the more the features describing the examples are correlated with the label, the more effective the model will be.There exist three main families of features: the ``observable'', the ``handcrafted'' and the ``latent'' features that are usually automatically learned from the training data.The contributions of this thesis fall into the scope of this last category. More precisely, we are interested in the specific setting of learning a discriminative representation when the number of data of interest is limited.A lack of data of interest can be found in different scenarios.First, we tackle the problem of imbalanced learning with a class of interest composed of a few examples by learning a metric that induces a new representation space where the learned models do not favor the majority examples.Second, we propose to handle a scenario with few available examples by learning at the same time a relevant data representation and a model that generalizes well through boosting models using kernels as base learners approximated by random Fourier features.Finally, to address the domain adaptation scenario where the target set contains no label while the source examples are acquired in different conditions, we propose to reduce the discrepancy between the two domains by keeping only the most similar features optimizing the solution of an optimal transport problem between the two domains
Szames, Esteban Alejandro. "Few group cross section modeling by machine learning for nuclear reactor." Thesis, université Paris-Saclay, 2020. http://www.theses.fr/2020UPASS134.
Full textModern nuclear reactors utilize core calculations that implement a thermo-hydraulic feedback requiring accurate homogenized few-group cross sections.They describe the interactions of neutrons with matter, and are endowed with the properties of smoothness and regularity, steaming from their underling physical phenomena. This thesis is devoted to the modeling of these functions by industry state-of-theart and innovative machine learning techniques. Mathematically, the subject can be defined as the analysis of convenient mapping techniques from one multi-dimensional space to another, conceptualize as the aggregated sum of these functions, whose quantity and domain depends on the simulations objectives. Convenient is intended in terms of computational performance, such as the model’s size, evaluation speed, accuracy, robustness to numerical noise, complexity,etc; always with respect to the engineering modeling objectives that specify the multidimensional spaces of interest. In this thesis, a standard UO₂ PWR fuel assembly is analyzed for three state-variables, burnup,fuel temperature, and boron concentration.Library storage requirements are optimized meeting the evaluation speed and accuracy targets in view of microscopic, macroscopic cross sections and the infinite multiplication factor. Three approximation techniques are studied: The state-of-the-art spline interpolation using computationally convenient B-spline basis, that generate high order local approximations. A full grid is used as usually donein the industry. Kernel methods, that are a very general machine learning framework able to pose in a normed vector space, a large variety of regression or classification problems. Kernel functions can reproduce different function spaces using an unstructured support,which is optimized with pool active learning techniques. The approximations are found through a convex optimization process simplified by the kernel trick. The intrinsic modular character of the method facilitates segregating the modeling phases: function space selection, application of numerical routines and support optimization through active learning. Artificial neural networks which are“model free” universal approximators able Artificial neural networks which are“model free” universal approximators able to approach continuous functions to an arbitrary degree without formulating explicit relations among the variables. With adequate training settings, intrinsically parallelizable multi-output networks minimize storage requirements offering the highest evaluation speed. These strategies are compared to each other and to multi-linear interpolation in a Cartesian grid, the industry standard in core calculations. The data set, the developed tools, and scripts are freely available under aMIT license
Sangnier, Maxime. "Outils d'apprentissage automatique pour la reconnaissance de signaux temporels." Rouen, 2015. http://www.theses.fr/2015ROUES064.
Full textThe work presented here tackles two different subjects in the wide thematic of how to build a numerical system to recognize temporal signals, mainly from limited observations. The first one is automatic feature extraction. For this purpose, we present a column generation algorithm, which is able to jointly learn a discriminative Time-Frequency (TF) transform, cast as a filter bank, with a support vector machine. This algorithm extends the state of the art on multiple kernel learning by non-linearly combining an infinite amount of kernels. The second direction of research is the way to handle the temporal nature of the signals. While our first contribution pointed out the importance of correctly choosing the time resolution to get a discriminative TF representation, the role of the time is clearly enlightened in early recognition of signals. Our second contribution lies in this field and introduces a methodological framework for early detection of a special event in a time-series, that is detecting an event before it ends. This framework builds upon multiple instance learning and similarity spaces by fitting them to the particular case of temporal sequences. Furthermore, our early detector comes with an efficient learning algorithm and theoretical guarantees on its generalization ability. Our two contributions have been empirically evaluated with brain-computer interface signals, soundscapes and human actions movies
Carriere, Mathieu. "On Metric and Statistical Properties of Topological Descriptors for geometric Data." Thesis, Université Paris-Saclay (ComUE), 2017. http://www.theses.fr/2017SACLS433/document.
Full textIn the context of supervised Machine Learning, finding alternate representations, or descriptors, for data is of primary interest since it can greatly enhance the performance of algorithms. Among them, topological descriptors focus on and encode the topological information contained in geometric data. One advantage of using these descriptors is that they enjoy many good and desireable properties, due to their topological nature. For instance, they are invariant to continuous deformations of data. However, the main drawback of these descriptors is that they often lack the structure and operations required by most Machine Learning algorithms, such as a means or scalar products. In this thesis, we study the metric and statistical properties of the most common topological descriptors, the persistence diagrams and the Mappers. In particular, we show that the Mapper, which is empirically instable, can be stabilized with an appropriate metric, that we use later on to conpute confidence regions and automatic tuning of its parameters. Concerning persistence diagrams, we show that scalar products can be defined with kernel methods by defining two kernels, or embeddings, into finite and infinite dimensional Hilbert spaces
Muller, Bruno. "Transfer Learning through Kernel Alignment : Application to Adversary Data Shifts in Automatic Sleep Staging." Thesis, Troyes, 2021. http://www.theses.fr/2021TROY0037.
Full textThis doctoral project aims at improving an automatic sleep staging system by taking into account inter-and-intra-individual variabilities, the latter having adversary effects on the classification. We focus on the detection of Rapid-Eye Movement periods during sleep. The core of our research is transfer learning and the selection of suitable detector(s) among a set, allowing the individualisation of the analysis by the exploitation of the observed data properties. We focus on the application of kernel alignment methods, firstly through the use of kernel-target alignment, studied here in a dual way, i.e. the kernel is fixed and the criterion is optimised with respect to the sought target labels. In a second step, we introduced kernel-cross alignment, allowing to take more efficiently advantage of the information contained in the training data. The ideas developed in the framework of this work have been extended to automatically selecting one or more efficient training sets for a given test set. The contributions of this work are both methodological and algorithmic, general in scope, but also focused on the application
Bonin, Jean-Jacques. "Contribution à la détection automatique de traits prosodiques dans les dialogues oraux homme-machine : utilisation dans un système de reconnaissance automatique de la parole continue." Nancy 1, 1991. http://www.theses.fr/1991NAN10154.
Full textNasser, Alissar. "Contribution à la classification non supervisée par machines à noyaux." Littoral, 2007. http://www.theses.fr/2007DUNK0182.
Full textUnsupervised classification has emerged as a popular technique for pattern recognition, image processing, and data mining. This is due to the development of advanced data measurements tools and data storage devices resulting in a huge quantity of data. This makes it necessary to analyze these data in order to extract some useful information. Unsupervised classification is one of the well-studied techniques, which concerns the partitioning of similar objects into clusters without any prior knowledge, such that objects in the same cluster share some unique properties. Two main categories of methods exist : (1) clustering methods in the multidimensional space and (2) projection methods for exploratory data analysis. The first category seeks zones/groups of high densities whereas the second category provides an accurate image on the plane of the multidimensional data. One of convenient lethods is by combining these two categories together in a way that involves a human operator into the process of structure analysis. Recently, Kernel machines gained a success in unsupervised classification. Instead, of projecting or classifying data directly in their input space, one transforms it into a high dimensional space called feature space and then applies any traditional projection technique such as Principal Components Analysis (PCA) or any clustering method such as K-means algorithm. The logic behind kernel is to enhance those features of the input data which make distinct pattern classes separate from each other. The present thesis shows the contribution of kernel machines in unsupervised classification, particularly in projection and classification methods. It first presents traditional projection methods and then present kernel Principal Components Analysis (kPCA). Spectral classification and kernel K-means clustering algortihm. The problems of adjusting kernel parameters and estimating the number of classes are studied. More over samples on synthetic and real data are executed ; results from various presented methods are compared. These clustering approaches are finally applied for the assistance to the detection of audio events in public transport
Lachaud, Antoine. "Discrimination robuste par méthode à noyaux." Thesis, Rouen, INSA, 2015. http://www.theses.fr/2015ISAM0015/document.
Full textThis thesis aims at finding classification rnodeIs which include explanatory elements. More specifically the proposed solution consists in merging a regularization path algorithm called DRSVM with a kernel approach called KERNEL BASIS. The first part of the thesis focuses on improving an algorithm called DRSVM from a reformulation of the thanks to the suh-differential theory. The second part of the thesis describes the extension of DRSVM afgorithm under a KERNEL BASIS framework via a dictionary approach. Finally, a series of experiments are conducted in order to validate the interpretable aspect of the rnodel
Brouard, Céline. "Inférence de réseaux d'interaction protéine-protéine par apprentissage statistique." Phd thesis, Université d'Evry-Val d'Essonne, 2013. http://tel.archives-ouvertes.fr/tel-00845692.
Full textSebti, Nadia. "Noyaux rationnels et automates d'arbres." Rouen, 2015. http://www.theses.fr/2015ROUES007.
Full textIn the case of words, a general scheme for computing rational kernels has been proposed. It is based on a general algorithm for composition of weighted transducers and a general algorithm computing smallest distance. Our goal is to generalize this computation scheme to the case of trees using tree automata. To do this we have established the following two main objectives : On the one hand,define tree automata for computing subtree kernel, subsettree kernel and tree factor kernel. On the other hand, from a regular tree expression, build a tree automaton to compute the various rational tree kernels described by regular tree expressions using the following scheme over two tree languages L1 and L2 : KE(L1;L2) = (AL1 \ AE \ AL1). We explored and proposed efficient algorithms for the conversion of a regular tree expression into tree automata. The first algorithm computes the Follow sets for a regular tree expression E of a size jEj and alphabetic width jjEjj in O(jjEjj jEj) time complexity. The second algorithm computes the equation tree automaton via the k-C-Continuations which is based on the acyclic minimization of Revuz. The algorithm is performed in an O(jQj jEj) time and space complexity, where jQj is the number of states of the produced automaton. Then we developed algorithms for the computation of subtree kernel, subsettree kernel and tree factor kernel. Our approach is based on the representation of all trees of the set S = fs1; : : : ; sng (resp. T = ft1; : : : ; tmg) by a particular weighted tree automaton called Root Weighted Tree Automaton the RWTA AS (resp. AT ) (equivalent to the prefix automaton in the case of words) such that jASj #Pn i=1 jsij = jSj (resp. JAT #Pm j=1 jtj j = jTj) ; then we compute the kernels between the two sets S and T. This amounts to compute the weight of the intersection automaton AS \ AT. We show that the computation of the kernel K(S; T) can be done in O(jASj jAT j) time and space complexity. Keywords: Finite Tree Automata, Rationnal Tree Languages, Regular Tree Expressions, Conversion of Regular Tree Expressions, Rational Kernels, Trees, Rational Tree Kernels
Alquier, Pierre. "Transductive and inductive adaptative inference for regression and density estimation." Paris 6, 2006. http://www.theses.fr/2006PA066436.
Full textSaide, Chafic. "Filtrage adaptatif à l’aide de méthodes à noyau : application au contrôle d’un palier magnétique actif." Thesis, Troyes, 2013. http://www.theses.fr/2013TROY0018/document.
Full textFunction approximation methods based on reproducing kernel Hilbert spaces are of great importance in kernel-based regression. However, the order of the model is equal to the number of observations, which makes this method inappropriate for online identification. To overcome this drawback, many sparsification methods have been proposed to control the order of the model. The coherence criterion is one of these sparsification methods. It has been shown possible to select a subset of the most relevant passed input vectors to form a dictionary to identify the model.A kernel function, once introduced into the dictionary, remains unchanged even if the non-stationarity of the system makes it less influent in estimating the output of the model. This observation leads to the idea of adapting the elements of the dictionary to obtain an improved one with an objective to minimize the resulting instantaneous mean square error and/or to control the order of the model.The first part deals with adaptive algorithms using the coherence criterion. The adaptation of the elements of the dictionary using a stochastic gradient method is presented for two types of kernel functions. Another topic is covered in this part which is the implementation of adaptive algorithms using the coherence criterion to identify Multiple-Outputs models.The second part introduces briefly the active magnetic bearing (AMB). A proposed method to control an AMB by an adaptive algorithm using kernel methods is presented to replace an existing method using neural networks
Boughorbel, Sabri. "Noyaux pour la classification d'images par les machines à vecteurs de support." Paris 11, 2005. http://www.theses.fr/2005PA112159.
Full textKernel methods are providing a ``plug and play'' approach to algorithm design. For instance, Support Vector Machines (SVMs) has been successfully applied to many tasks through simply substituting appropriate kernels. We design several kernels for different kinds of image representations. For global image representations, we propose a general form of the Histogram Intersection kernel by relaxing conditions of its positive definiteness. We propose also a compactly supported kernel (GCS kernel) based on a geometric approach. The compactness property enhances the computation of the SVM training using sparse linear algebra algorithms. For local representations, images are described by local image features such as local jets computed on interest points. We propose is the intermediate matching kernel which defines an alignement between the two sets of local features using a matching set. The latter is learned using local features from all training images. A crucial step after designing kernels is to carefully choose hyperparameters. In this research, we addressed also tuning kernel parameters. First we have been interested by generalization error estimation. We introduce a new estimator for the leave-two-out using the mean-field approach. We prove that the proposed estimator is statistically more robust than the leave-one-out ones. We propose also an optimization scheme (LCCP), based on the concave convex procedure, for tuning kernel parameters. The LCCP is more efficient than gradient descent technique since it insures that the optimization criterion decreases monotonically and converges to a local minimum without searching the size step
El, Dakdouki Aya. "Machine à vecteurs de support hyperbolique et ingénierie du noyau." Thesis, Lille 1, 2019. http://www.theses.fr/2019LIL1I046/document.
Full textStatistical learning theory is a field of inferential statistics whose foundations were laid by Vapnik at the end of the 1960s. It is considered a subdomain of artificial intelligence. In machine learning, support vector machines (SVM) are supervised learning models with associated learning algorithms that analyze data used for classification and regression analysis. In this thesis, our aim is to propose two new statistical learning problems : one on the conception and evaluation of a multi-class SVM extension and another on the design of a new kernel for support vectors machines. First, we introduced a new kernel machine for multi-class pattern recognition : the hyperbolic support vector machine. Geometrically, it is characterized by the fact that its decision boundaries in the feature space are defined by hyperbolic functions. We then established its main statistical properties. Among these properties we showed that the classes of component functions are uniform Glivenko-Cantelli, this by establishing an upper bound of the Rademacher complexity. Finally, we establish a guaranteed risk for our classifier. Second, we constructed a new kernel based on the Fourier transform of a Gaussian mixture model. We proceed in the following way: first, each class is fragmented into a number of relevant subclasses, then we consider the directions given by the vectors obtained by taking all pairs of subclass centers of the same class. Among these are excluded those allowing to connect two subclasses of two different classes. We can also see this as the search for translation invariance in each class. It successfully on several datasets in the context of machine learning using multiclass support vector machines
Mahé, Pierre. "Fonctions noyaux pour molécules et leur application au criblage virtuel par machines à vecteurs de support." Phd thesis, École Nationale Supérieure des Mines de Paris, 2006. http://pastel.archives-ouvertes.fr/pastel-00002191.
Full textSANTINI, DOMINIQUE. "Realisation du noyau logiciel d'une machine bases de donnees a hautes performances." Paris 6, 1992. http://www.theses.fr/1992PA066605.
Full textRamona, Mathieu. "Classification automatique de flux radiophoniques par Machines à Vecteurs de Support." Phd thesis, Télécom ParisTech, 2010. http://pastel.archives-ouvertes.fr/pastel-00529331.
Full textRamona, Mathieu. "Classification automatique de flux radiophoniques par machines à vecteurs de support." Phd thesis, Paris, Télécom ParisTech, 2010. https://pastel.hal.science/pastel-00529331.
Full textWe present here a system for speech/music audio classification, that relies on the excellent statistical properties of Support Vector Machines. This problems raises three questions : how can the SVM, by essence discriminative, be used effeciently on a problem involving more than two classes, how can an audio signal be characterized in a relevant way, and how can the temporel issue be adressed ? We propose a hybrid system for multi-class classification, based on a combination of One-vs-One and dendogram-based approaches, and allowing the estimation of posterior probabilities. The latter are used for the application of post-processing methods that take into account the neighboring frames' inter-dependancies. We thus propose a classification scheme based on the application of Hidden Markov Models on the posterior probabilities, along with an approach based on change detection between segments with "homogeneous" acoustic content. Concerning the audio signal characterization, since it involves a great amount of audio descriptors, we propose new algorithms for feature selection, based on the recent Kernel Alignement criterion. This criterion is also used for the kernel selection step in the classification process. The proposed algorithms are compared to the state-of-the-art, and constitute a relevant alternative in terms of computational cost and storage. The system built from these contributions has been used for a participation to the ESTER 2 evaluation campaign, that we present, along with our results
Mahe, Pierre. "Fonctions noyaux pour molécules et leur application au criblage virtuel par machines à vecteurs de support." Paris, ENMP, 2006. http://www.theses.fr/2006ENMP1381.
Full textMenneteau, François. "ParObj : un noyau de système parallèle à objets." Grenoble INPG, 1993. http://tel.archives-ouvertes.fr/tel-00005135.
Full textZwald, Laurent. "PERFORMANCES STATISTIQUES D'ALGORITHMES D'APPRENTISSAGE : ``KERNEL PROJECTION MACHINE'' ET ANALYSE EN COMPOSANTES PRINCIPALES A NOYAU." Phd thesis, Université Paris Sud - Paris XI, 2005. http://tel.archives-ouvertes.fr/tel-00012011.
Full textdes contributions à la communauté du machine learning en utilisant des
techniques de statistiques modernes basées sur des avancées dans l'étude
des processus empiriques. Dans une première partie, les propriétés statistiques de
l'analyse en composantes principales à noyau (KPCA) sont explorées. Le
comportement de l'erreur de reconstruction est étudié avec un point de vue
non-asymptotique et des inégalités de concentration des valeurs propres de la matrice de
Gram sont données. Tous ces résultats impliquent des vitesses de
convergence rapides. Des propriétés
non-asymptotiques concernant les espaces propres de la KPCA eux-mêmes sont également
proposées. Dans une deuxième partie, un nouvel
algorithme de classification a été
conçu : la Kernel Projection Machine (KPM).
Tout en s'inspirant des Support Vector Machines (SVM), il met en lumière que la sélection d'un espace vectoriel par une méthode de
réduction de la dimension telle que la KPCA régularise
convenablement. Le choix de l'espace vectoriel utilisé par la KPM est guidé par des études statistiques de sélection de modéle par minimisation pénalisée de la perte empirique. Ce
principe de régularisation est étroitement relié à la projection fini-dimensionnelle étudiée dans les travaux statistiques de
Birgé et Massart. Les performances de la KPM et de la SVM sont ensuite comparées sur différents jeux de données. Chaque thème abordé dans cette thèse soulève de nouvelles questions d'ordre théorique et pratique.
Zwald, Laurent. "Performances statistiques d'algorithmes d'apprentissage : "Kernel projection machine" et analyse en composantes principales à noyau." Paris 11, 2005. https://tel.archives-ouvertes.fr/tel-00012011.
Full textThis thesis takes place within the framework of statistical learning. It brings contributions to the machine learning community using modern statistical techniques based on progress in the study of empirical processes. The first part investigates the statistical properties of Kernel Principal Component Analysis (KPCA). The behavior of the reconstruction error is studied with a non-asymptotique point of view and concentration inequalities of the eigenvalues of the kernel matrix are provided. All these results correspond to fast convergence rates. Non-asymptotic results concerning the eigenspaces of KPCA themselves are also provided. A new algorithm of classification has been designed in the second part: the Kernel Projection Machine (KPM). It is inspired by the Support Vector Machines (SVM). Besides, it highlights that the selection of a vector space by a dimensionality reduction method such as KPCA regularizes suitably. The choice of the vector space involved in the KPM is guided by statistical studies of model selection using the penalized minimization of the empirical loss. This regularization procedure is intimately connected with the finite dimensional projections studied in the statistical work of Birge and Massart. The performances of KPM and SVM are then compared on some data sets. Each topic tackled in this thesis raises new questions
Villain, Jonathan. "Estimation de l'écotoxicité de substances chimiques par des méthodes à noyaux." Thesis, Lorient, 2016. http://www.theses.fr/2016LORIS404/document.
Full textIn chemistry and more particularly in chemoinformatics, QSAR models (Quantitative Structure Activity Relationship) are increasingly studied. They provide an in silico estimation of the properties of chemical compounds including ecotoxicological properties. These models are theoretically valid only for a class of compounds (validity domain) and are sensitive to the presence of outliers. This PhD thesis is focused on the construction of robust global models (including a maximum of compounds) to predict ecotoxicity of chemical compounds on algae P. subcapitata and to determine a validity domain in order to deduce the capacity of a model to predict the toxicity of a compound. These robust statistical models are based on quantile approach in linear regression and regression Support Vector Machine
Le, Van Luong. "Identification de systèmes dynamiques hybrides : géométrie, parcimonie et non-linéarités." Phd thesis, Université de Lorraine, 2013. http://tel.archives-ouvertes.fr/tel-00874283.
Full textYger, Florian. "Discrimination de signaux : contributions aux approches par noyaux et par projection." Rouen, 2013. http://www.theses.fr/2013ROUES016.
Full textKellner, Jérémie. "Gaussian models and kernel methods." Thesis, Lille 1, 2016. http://www.theses.fr/2016LIL10177/document.
Full textKernel methods have been extensively used to transform initial datasets by mapping them into a so-called kernel space or RKHS, before applying some statistical procedure onto transformed data. In particular, this kind of approach has been explored in the literature to try and make some prescribed probabilistic model more accurate in the RKHS, for instance Gaussian mixtures for classification or mere Gaussians for outlier detection. Therefore this thesis studies the relevancy of such models in kernel spaces.In a first time, we focus on a family of parameterized kernels - Gaussian RBF kernels - and study theoretically the distribution of an embedded random variable in a corresponding RKHS. We managed to prove that most marginals of such a distribution converge weakly to a so-called ''scale-mixture'' of Gaussians - basically a Gaussian with a random variance - when the parameter of the kernel tends to infinity. This result is used in practice to device a new method for outlier detection.In a second time, we present a one-sample test for normality in an RKHS based on the Maximum Mean Discrepancy. In particular, our test uses a fast parametric bootstrap procedure which circumvents the need for re-estimating Gaussian parameters for each bootstrap replication
Gkirtzou, Aikaterini. "Sparsity regularization and graph-based representation in medical imaging." Phd thesis, Ecole Centrale Paris, 2013. http://tel.archives-ouvertes.fr/tel-00960163.
Full textMialon, Grégoire. "On inductive biases for machine learning in data constrained settings." Electronic Thesis or Diss., Université Paris sciences et lettres, 2022. http://www.theses.fr/2022UPSLE011.
Full textLearning with limited data is one of the biggest problems of deep learning. Current, popular approaches to this issueconsist in training models on huge amounts of data, labelled or not, before re-training the model on a smaller dataset ofinterest from the same modality. Intuitively, this technique allows the model to learn a general representation for somekind of data first, such as images. Then, fewer data should be required to learn a specific task for this particular modality.While this approach coined as "transfer learning" is very effective in domains such as computer vision or natural languageprocessing, it does not solve common problems of deep learning such as model interpretability or the overall need fordata. This thesis explores a different answer to the problem of learning expressive models in data constrained settings.Instead of relying on big datasets to learn the parameters of a neural network, we will replace some of them by knownfunctions reflecting the structure of the data. Very often, these functions will be drawn from the rich literature of kernelmethods. Indeed, many kernels can be interpreted, and/or allow for learning with few data. Our approach falls under thehood of "inductive biases", which can be defined as hypothesis on the data at hand restricting the space of models toexplore during learning. In the first two chapters of the thesis, we demonstrate the effectiveness of this approach in thecontext of sequences, such as sentences in natural language or protein sequences, and graphs, such as molecules. Wealso highlight the relationship between our work and recent advances in deep learning. The last chapter of this thesisfocuses on convex machine learning models. Here, rather than proposing new models, we wonder which proportion ofthe samples in a dataset is really needed to learn a "good" model. More precisely, we study the problem of safe samplescreening, i.e, executing simple tests to discard uninformative samples from a dataset even before fitting a machinelearning model, without affecting the optimal model. Such techniques can be used to compress datasets or mine for raresamples
Zribi, Abir. "Apprentissage par noyaux multiples : application à la classification automatique des images biomédicales microscopiques." Thesis, Rouen, INSA, 2016. http://www.theses.fr/2016ISAM0001.
Full textThis thesis arises in the context of computer aided analysis for subcellular protein localization in microscopic images. The aim is the establishment of an automatic classification system allowing to identify the cellular compartment in which a protein of interest exerts its biological activity. In order to overcome the difficulties in attempting to discern the cellular compartments in microscopic images, the existing state-of-art systems use several descriptors to train an ensemble of classifiers. In this thesis, we propose a different classification scheme wich better cope with the requirement of genericity and flexibility to treat various image datasets. Aiming to provide an efficient image characterization of microscopic images, a new feature system combining local, frequency-domain, global, and region-based features is proposed. Then, we formulate the problem of heterogeneous feature fusion as a kernel selection problem. Using multiple kernel learning, the problems of optimal feature sets selection and classifier training are simultaneously resolved. The proposed combination scheme leads to a simple and a generic framework capable of providing a high performance for microscopy image classification. Extensive experiments were carried out using widely-used and best known datasets. When compared with the state-of-the-art systems, our framework is more generic and outperforms other classification systems. To further expand our study on multiple kernel learning, we introduce a new formalism for learning with multiple kernels performed in two steps. This contribution consists in proposing three regularized terms with in the minimization of kernels weights problem, formulated as a classification problem using Separators with Vast Margin on the space of pairs of data. The first term ensures that kernels selection leads to a sparse representation. While the second and the third terms introduce the concept of kernels similarity by using a correlation measure. Experiments on various biomedical image datasets show a promising performance of our method compared to states of art methods
Eid, Abdelrahman. "Stochastic simulations for graphs and machine learning." Thesis, Lille 1, 2020. http://www.theses.fr/2020LIL1I018.
Full textWhile it is impractical to study the population in many domains and applications, sampling is a necessary method allows to infer information. This thesis is dedicated to develop probability sampling algorithms to infer the whole population when it is too large or impossible to be obtained. Markov chain Monte Carlo (MCMC) techniques are one of the most important tools for sampling from probability distributions especially when these distributions haveintractable normalization constants.The work of this thesis is mainly interested in graph sampling techniques. Two methods in chapter 2 are presented to sample uniform subtrees from graphs using Metropolis-Hastings algorithms. The proposed methods aim to sample trees according to a distribution from a graph where the vertices are labelled. The efficiency of these methods is proved mathematically. Additionally, simulation studies were conducted and confirmed the theoretical convergence results to the equilibrium distribution.Continuing to the work on graph sampling, a method is presented in chapter 3 to sample sets of similar vertices in an arbitrary undirected graph using the properties of the Permanental Point processes PPP. Our algorithm to sample sets of k vertices is designed to overcome the problem of computational complexity when computing the permanent by sampling a joint distribution whose marginal distribution is a kPPP.Finally in chapter 4, we use the definitions of the MCMC methods and convergence speed to estimate the kernel bandwidth used for classification in supervised Machine learning. A simple and fast method called KBER is presented to estimate the bandwidth of the Radial basis function RBF kernel using the average Ricci curvature of graphs