Зміст
Добірка наукової літератури з теми "Apprentissage parcimonieux"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся зі списками актуальних статей, книг, дисертацій, тез та інших наукових джерел на тему "Apprentissage parcimonieux".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Статті в журналах з теми "Apprentissage parcimonieux"
BOCQUIER, F., N. DEBUS, A. LURETTE, C. MATON, G. VIUDES, C. H. MOULIN, and M. JOUVEN. "Elevage de précision en systèmes d’élevage peu intensifiés." INRAE Productions Animales 27, no. 2 (June 2, 2014): 101–12. http://dx.doi.org/10.20870/productions-animales.2014.27.2.3058.
Повний текст джерелаДисертації з теми "Apprentissage parcimonieux"
Avalos, Marta. "Modèles additifs parcimonieux." Phd thesis, Université de Technologie de Compiègne, 2004. http://tel.archives-ouvertes.fr/tel-00008802.
Повний текст джерелаAmate, Laure. "Apprentissage de modèles de formes parcimonieux basés sur des représentations splines." Phd thesis, Université de Nice Sophia-Antipolis, 2009. http://tel.archives-ouvertes.fr/tel-00456612.
Повний текст джерелаAmate, Laure. "Apprentissage de modèles de formes parcimonieux basés sur les représentations splines." Nice, 2009. http://www.theses.fr/2009NICE4117.
Повний текст джерелаIn many contexts it is important to be able to find compact representations of the collective morphological properties of a set of objects. This is the case of autonomous robotic platforms operating in natural environments that must use the perceptual properties of the objects present in their workspace to execute their mission. This thesis is a contribution to the definition of formalisms and methods for automatic identification of such models. The shapes we want to characterize are closed curves corresponding to contours of objects detected in the scene. We begin with the formal definition of the notion of shape as classes of equivalence with respect to groups of basic geometric operators, introducing two distinct approaches that have been used in the literature: discrete and continuous. The discrete theory, admitting the existence of a finite number of recognizable landmarks, provides in an obvious manner a compact representation but is sensible to their selection. The continuous theory of shapes provides a more fundamental approach, but leads to shape spaces of infinite dimension, lacking the parsimony of the discrete representation. We thus combine in our work the advantages of both approaches representing shapes of curves with splines: piece-wise continuous polynomials defined by sets of knots and control points. We first study the problem of fitting free-knots splines of varying complexity to a single observed curve. The trade-o_ between the parsimony of the representation and its fidelity to the observations is a well known characteristic of model identification using nested families of increasing dimension. After presenting an overview of methods previously proposed in the literature, we single out a two-step approach which is formally sound and matches our specific requirements. It splits the identification, simulating a reversible jump Markov chain to select the complexity of the model followed by a simulated annealing algorithm to estimate its parameters. We investigate the link between Kendall's shape space and spline representations when we take the spline control points as landmarks. We consider now the more complex problem of modeling a set of objects with similar morphological characteristics. We equate the problem to finding the statistical distribution of the parameters of the spline representation, modeling the knots and control points as unobserved variables. The identified distribution is the maximizer of a marginal likelihood criterion, and we propose a new Expectation-Maximization algorithm to optimize it. Because we may want to treat a large number of curves observed sequentially, we adapt an iterative (on-line) version of the EM algorithm recently proposed in the literature. For the choice of statistical distributions that we consider, both the expectation and the maximization steps must resort to numerical approximations, leading to a stochastic/on-line variant of the EM algorithm that, as far as we know, is implemented here for the first time
Huet, Romain. "Codage neural parcimonieux pour un système de vision." Thesis, Lorient, 2017. http://www.theses.fr/2017LORIS439/document.
Повний текст джерелаThe neural networks have gained a renewed interest through the deep learning paradigm. Whilethe so called optimised neural nets, by optimising the parameters necessary for learning, require massive computational resources, we focus here on neural nets designed as addressable content memories, or neural associative memories. The challenge consists in realising operations, traditionally obtained through computation, exclusively with neural memory in order to limit the need in computational resources. In this thesis, we study an associative memory based on cliques, whose sparse neural coding optimises the data diversity encoded in the network. This large diversity allows the clique based network to be more efficient in messages retrieval from its memory than other neural associative memories. The associative memories are known for their incapacity to identify without ambiguities the messages stored in a saturated memory. Indeed, depending of the information present in the network and its encoding, a memory can fail to retrieve a desired result. We are interested in tackle this issue and propose several contributions in order to reduce the ambiguities in the cliques based neural network. Besides, these cliques based nets are unable to retrieve an information within their memories if the message is unknown. We propose a solution to this problem through a new associative memory based on cliques which preserves the initial network's corrective ability while being able to hierarchise the information. The hierarchy relies on a surjective and bidirectional transition to generalise an unknown input with an approximation of learnt information. The associative memories' experimental validation is usually based on low dimension artificial dataset. In the computer vision context, we report here the results obtained with real datasets used in the state-of-the-art, such as MNIST, Yale or CIFAR
Belilovsky, Eugene. "Apprentissage de graphes structuré et parcimonieux dans des données de haute dimension avec applications à l’imagerie cérébrale." Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLC027.
Повний текст джерелаThis dissertation presents novel structured sparse learning methods on graphs that address commonly found problems in the analysis of neuroimaging data as well as other high dimensional data with few samples. The first part of the thesis proposes convex relaxations of discrete and combinatorial penalties involving sparsity and bounded total variation on a graph as well as bounded `2 norm. These are developed with the aim of learning an interpretable predictive linear model and we demonstrate their effectiveness on neuroimaging data as well as a sparse image recovery problem.The subsequent parts of the thesis considers structure discovery of undirected graphical models from few observational data. In particular we focus on invoking sparsity and other structured assumptions in Gaussian Graphical Models (GGMs). To this end we make two contributions. We show an approach to identify differences in Gaussian Graphical Models (GGMs) known to have similar structure. We derive the distribution of parameter differences under a joint penalty when parameters are known to be sparse in the difference. We then show how this approach can be used to obtain confidence intervals on edge differences in GGMs. We then introduce a novel learning based approach to the problem structure discovery of undirected graphical models from observational data. We demonstrate how neural networks can be used to learn effective estimators for this problem. This is empirically shown to be flexible and efficient alternatives to existing techniques
Mattei, Pierre-Alexandre. "Sélection de modèles parcimonieux pour l’apprentissage statistique en grande dimension." Thesis, Sorbonne Paris Cité, 2017. http://www.theses.fr/2017USPCB051/document.
Повний текст джерелаThe numerical surge that characterizes the modern scientific era led to the rise of new kinds of data united in one common immoderation: the simultaneous acquisition of a large number of measurable quantities. Whether coming from DNA microarrays, mass spectrometers, or nuclear magnetic resonance, these data, usually called high-dimensional, are now ubiquitous in scientific and technological worlds. Processing these data calls for an important renewal of the traditional statistical toolset, unfit for such frameworks that involve a large number of variables. Indeed, when the number of variables exceeds the number of observations, most traditional statistics becomes inefficient. First, we give a brief overview of the statistical issues that arise with high-dimensional data. Several popular solutions are presented, and we present some arguments in favor of the method utilized and advocated in this thesis: Bayesian model uncertainty. This chosen framework is the subject of a detailed review that insists on several recent developments. After these surveys come three original contributions to high-dimensional model selection. A new algorithm for high-dimensional sparse regression called SpinyReg is presented. It compares favorably to state-of-the-art methods on both real and synthetic data sets. A new data set for high-dimensional regression is also described: it involves predicting the number of visitors in the Orsay museum in Paris using bike-sharing data. We focus next on model selection for high-dimensional principal component analysis (PCA). Using a new theoretical result, we derive the first closed-form expression of the marginal likelihood of a PCA model. This allows us to propose two algorithms for model selection in PCA. A first one called globally sparse probabilistic PCA (GSPPCA) that allows to perform scalable variable selection, and a second one called normal-gamma probabilistic PCA (NGPPCA) that estimates the intrinsic dimensionality of a high-dimensional data set. Both methods are competitive with other popular approaches. In particular, using unlabeled DNA microarray data, GSPPCA is able to select genes that are more biologically relevant than several popular approaches
Hérault, Romain. "Vision et apprentissage statistique pour la reconnaissance d'items comportementaux." Compiègne, 2007. http://www.theses.fr/2007COMP1715.
Повний текст джерелаThis work consists in the detection of behavioral items in order to prevent driver drowsiness. Videos were shot from within a car, and each picture of the video was characterized by six behavioral items. Our work consists in the retrieval of these items, picture by picture. The study was decomposed into two phases: 1) A Head and facial action tracking. A framework to 3D head pose and facial action tracking with an adaptive appearance model based on a mixture model is proposed to deal with face occlusion ; 2) A recognition of the behavioral items based on data retrieved from the tracking. We propose a new criterion leading to an adaptation of maximum likelihood estimation. The model outputs proper conditional probabilities into a user-defined interval. This criterion is applied to MLPs and IOHMMs for the recognition of the behavioral items
Meghnoudj, Houssem. "Génération de caractéristiques à partir de séries temporelles physiologiques basée sur le contrôle optimal parcimonieux : application au diagnostic de maladies et de troubles humains." Electronic Thesis or Diss., Université Grenoble Alpes, 2024. http://www.theses.fr/2024GRALT003.
Повний текст джерелаIn this thesis, a novel methodology for features generation from physiological signals (EEG, ECG) has been proposed that is used for the diagnosis of a variety of brain and heart diseases. Based on sparse optimal control, the generation of Sparse Dynamical Features (SDFs) is inspired by the functioning of the brain. The method's fundamental concept revolves around sparsely decomposing the signal into dynamical modes that can be switched on and off at the appropriate time instants with the appropriate amplitudes. This decomposition provides a new point of view on the data which gives access to informative features that are faithful to the brain functioning. Nevertheless, the method remains generic and versatile as it can be applied to a wide range of signals. The methodology's performance was evaluated on three use cases using openly accessible real-world data: (1) Parkinson's Disease, (2) Schizophrenia, and (3) various cardiac diseases. For all three applications, the results are highly conclusive, achieving results that are comparable to the state-of-the-art methods while using only few features (one or two for brain applications) and a simple linear classifier supporting the significance and reliability of the findings. It's worth highlighting that special attention has been given to achieving significant and meaningful results with an underlying explainability
Laporte, Léa. "La sélection de variables en apprentissage d'ordonnancement pour la recherche d'information : vers une approche contextuelle." Toulouse 3, 2013. http://thesesups.ups-tlse.fr/2170/.
Повний текст джерелаLearning-to-rank aims at automatically optimizing a ranking function learned on training data by a machine learning algorithm. Existing approaches have two major drawbacks. Firstly, the ranking functions can use several thousands of features, which is an issue since algorithms have to deal with large scale data. This can also have a negative impact on the ranking quality. Secondly, algorithms learn an unique fonction for all queries. Then, nor the kind of user need neither the context of the query are taken into account in the ranking process. Our works focus on solving the large-scale issue and the context-aware issue by using feature selection methods dedicated to learning-to-rank. We propose five feature selection algorithms based on sparse Support Vector Machines (SVM). Three proceed to feature selection by reweighting the L2-norm, one solves a L1-regularized problem whereas the last algorithm consider nonconvex regularizations. Our methods are faster and sparser than state-of-the-art algorithms on benchmark datasets, while providing similar performances in terms of RI measures. We also evaluate our approches on a commercial dataset. Experimentations confirm the previous results. We propose in this context a relevance model based on users clicks, in the special case of multi-clickable documents. Finally, we propose an adaptative and query-dependent ranking system based on feature selection. This system considers several clusters of queries, each group defines a context. For each cluster, the system selects a group of features to learn a context-aware ranking function
Vezard, Laurent. "Réduction de dimension en apprentissage supervisé : applications à l’étude de l’activité cérébrale." Thesis, Bordeaux 1, 2013. http://www.theses.fr/2013BOR15005/document.
Повний текст джерелаThe aim of this work is to develop a method able to automatically determine the alertness state of humans. Such a task is relevant to diverse domains, where a person is expected or required to be in a particular state. For instance, pilots, security personnel or medical personnel are expected to be in a highly alert state, and this method could help to confirm this or detect possible problems. In this work, electroencephalographic data (EEG) of 58 subjects in two distinct vigilance states (state of high and low alertness) were collected via a cap with $58$ electrodes. Thus, a binary classification problem is considered. In order to use of this work on a real-world applications, it is necessary to build a prediction method that requires only a small number of sensors (electrodes) in order to minimize the time needed by the cap installation and the cap cost. During this thesis, several approaches have been developed. A first approach involves use of a pre-processing method for EEG signals based on the use of a discrete wavelet decomposition in order to extract the energy of each frequency in the signal. Then, a linear regression is performed on the energies of some of these frequencies and the slope of this regression is retained. A genetic algorithm (GA) is used to optimize the selection of frequencies on which the regression is performed. Moreover, the GA is used to select a single electrode .A second approach is based on the use of the Common Spatial Pattern method (CSP). This method allows to define linear combinations of the original variables to obtain useful synthetic signals for the task classification. In this work, a GA and a sequential search method have been proposed to select a subset of electrode which are keep in the CSP calculation.Finally, a sparse CSP algorithm, based on the use of existing work in the sparse principal component analysis, was developed.The results of the different approaches are detailed and compared. This work allows us to obtaining a reliable model to obtain fast prediction of the alertness of a new individual