Rozprawy doktorskie na temat „Modèle latent”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Sprawdź 50 najlepszych rozpraw doktorskich naukowych na temat „Modèle latent”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Przeglądaj rozprawy doktorskie z różnych dziedzin i twórz odpowiednie bibliografie.
Brault, Vincent. "Estimation et sélection de modèle pour le modèle des blocs latents". Thesis, Paris 11, 2014. http://www.theses.fr/2014PA112238/document.
Pełny tekst źródłaClassification aims at sharing data sets in homogeneous subsets; the observations in a class are more similar than the observations of other classes. The problem is compounded when the statistician wants to obtain a cross classification on the individuals and the variables. The latent block model uses a law for each crossing object class and class variables, and observations are assumed to be independent conditionally on the choice of these classes. However, factorizing the joint distribution of the labels is impossible, obstructing the calculation of the log-likelihood and the using of the EM algorithm. Several methods and criteria exist to find these partitions, some frequentist ones, some bayesian ones, some stochastic ones... In this thesis, we first proposed sufficient conditions to obtain the identifiability of the model. In a second step, we studied two proposed algorithms to counteract the problem of the EM algorithm: the VEM algorithm (Govaert and Nadif (2008)) and the SEM-Gibbs algorithm (Keribin, Celeux and Govaert (2010)). In particular, we analyzed the combination of both and highlighted why the algorithms degenerate (term used to say that it returns empty classes). By choosing priors wise, we then proposed a Bayesian adaptation to limit this phenomenon. In particular, we used a Gibbs sampler and we proposed a stopping criterion based on the statistics of Brooks-Gelman (1998). We also proposed an adaptation of the Largest Gaps algorithm (Channarond et al. (2012)). By taking their demonstrations, we have shown that the labels and parameters estimators obtained are consistent when the number of rows and columns tend to infinity. Furthermore, we proposed a method to select the number of classes in row and column, the estimation provided is also consistent when the number of row and column is very large. To estimate the number of classes, we studied the ICL criterion (Integrated Completed Likelihood) whose we proposed an exact shape. After studying the asymptotic approximation, we proposed a BIC criterion (Bayesian Information Criterion) and we conjecture that the two criteria select the same results and these estimates are consistent; conjecture supported by theoretical and empirical results. Finally, we compared the different combinations and proposed a methodology for co-clustering
Mura, Thibault. "Prévention des démences : analyse du déclin cognitif à l’aide d’un modèle longitudinal non linéaire à variable latente". Thesis, Montpellier 1, 2012. http://www.theses.fr/2012MON1T018/document.
Pełny tekst źródłaThe first aim of this doctoral work is to replace dementia in its public health context by estimating the number of dementia cases expected to occur in France and Europe over the next few decades until 2050. The sensitivity of these projections to hypotheses made on dementia incidence and mortality, demographic scenario used, and implementation of a prevention intervention, was also assessed. In this context of increasing number of future cases, the primary and secondary prevention of dementia will take a prominent place in the social management of this problem. Relevant research in the field of primary and secondary prevention requires an appropriate methodology and the use of relevant outcome. Cognitive decline seems to be an appropriate outcome, but a number of biases must be avoided. First, we illustrated the use of this criterion in the context of primary prevention using a nonlinear model with latent variable for longitudinal data to investigated the association between chronic use of benzodiazepines and cognitive decline. We showed the absence of association in a large population-based cohort. Secondly we used this model to describe and compare the metrological properties of a broad range of neuropsychological tests in a clinical cohort of patients with mild cognitive impairment (MCI). We also investigated the sensitivity of these tests to cognitive changes associated with prodromal Alzheimer's disease. Our work provides arguments for selecting neuropsychological tests which can be used in secondary prevention research, to identify and / or to follow patients with mild cognitive impairment (MCI) due to Alzheimer's disease
Samuth, Benjamin. "Ηybrid mοdels cοmbining deep neural representatiοns and nοn-parametric patch-based methοds fοr phοtοrealistic image generatiοn". Electronic Thesis or Diss., Normandie, 2024. http://www.theses.fr/2024NORMC249.
Pełny tekst źródłaImage generation has encountered great progress thanks to the quickevolution of deep neural models. Their reach went beyond thescientific domain and thus multiple legitimate concerns and questionshave been raised, in particular about how the training data aretreated. On the opposite, lightweight and explainable models wouldbe a fitting answer to these emerging problematics, but their qualityand range of applications are limited.This thesis strives to build “hybrid models”. They would efficientlycombine the qualities of lightweight or frugal methods with theperformance of deep networks. We first study the case of artisticstyle transfer with a multiscale and constrained patch-basedmethod. We qualitatively find out the potential of perceptual metricsin the process. Besides, we develop two hybrid models forphotorealistic face generation, each built around a pretrainedauto-encoder. The first model tackles the problem of few-shot facegeneration with the help of latent patches. Results shows a notablerobustness and convincing synthesis with a simple patch-basedsequential algorithm. The second model uses Gaussian mixtures modelsas a way to generalize the previous method to wider varieties offaces. In particular, we show that these models perform similarly toother neural methods, while removing a non-negligible number ofparameters and computing steps at the same time
Dantan, Etienne. "Modèles conjoints pour données longitudinales et données de survie incomplètes appliqués à l'étude du vieillissement cognitif". Thesis, Bordeaux 2, 2009. http://www.theses.fr/2009BOR21658/document.
Pełny tekst źródłaIn cognitive ageing study, older people are highly selected by a risk of death associated with poor cognitive performances. Modeling the natural history of cognitive decline is difficult in presence of incomplete longitudinal and survival data. Moreover, the non observed cognitive decline acceleration beginning before the dementia diagnosis is difficult to evaluate. Cognitive decline is highly heterogeneous, e.g. there are various patterns associated with different risks of survival event. The objective is to study joint models for incomplete longitudinal and survival data to describe the cognitive evolution in older people. Latent variable approaches were used to take into account the non-observed mechanisms, e.g. heterogeneity and decline acceleration. First, we compared two approaches to consider missing data in longitudinal data analysis. Second, we propose a joint model with a latent state to model cognitive evolution and its pre-dementia acceleration, dementia risk and death risk
Robert, Valérie. "Classification croisée pour l'analyse de bases de données de grandes dimensions de pharmacovigilance". Thesis, Université Paris-Saclay (ComUE), 2017. http://www.theses.fr/2017SACLS111/document.
Pełny tekst źródłaThis thesis gathers methodological contributions to the statistical analysis of large datasets in pharmacovigilance. The pharmacovigilance datasets produce sparse and large matrices and these two characteritics are the main statistical challenges for modelling them. The first part of the thesis is dedicated to the coclustering of the pharmacovigilance contingency table thanks to the normalized Poisson latent block model. The objective is on the one hand, to provide pharmacologists with some interesting and reduced areas to explore more precisely. On the other hand, this coclustering remains a useful background information for dealing with individual database. Within this framework, a parameter estimation procedure for this model is detailed and objective model selection criteria are developed to choose the best fit model. Datasets are so large that we propose a procedure to explore the model space in coclustering, in a non exhaustive way but a relevant one. Additionnally, to assess the performances of the methods, a convenient coclustering index is developed to compare partitions with high numbers of clusters. The developments of these statistical tools are not specific to pharmacovigilance and can be used for any coclustering issue. The second part of the thesis is devoted to the statistical analysis of the large individual data, which are more numerous but also provides even more valuable information. The aim is to produce individual clusters according their drug profiles and subgroups of drugs and adverse effects with possible links, which overcomes the coprescription and masking phenomenons, common contingency table issues in pharmacovigilance. Moreover, the interaction between several adverse effects is taken into account. For this purpose, we propose a new model, the multiple latent block model which enables to cocluster two binary tables by imposing the same row ranking. Assertions inherent to the model are discussed and sufficient identifiability conditions for the model are presented. Then a parameter estimation algorithm is studied and objective model selection criteria are developed. Moreover, a numeric simulation model of the individual data is proposed to compare existing methods and study its limits. Finally, the proposed methodology to deal with individual pharmacovigilance data is presented and applied to a sample of the French pharmacovigilance database between 2002 and 2010
Georgescu, Vera. "Classification de données multivariées multitypes basée sur des modèles de mélange : application à l'étude d'assemblages d'espèces en écologie". Phd thesis, Université d'Avignon, 2010. http://tel.archives-ouvertes.fr/tel-00624382.
Pełny tekst źródłaArzamendia, Lopez Juan Pablo. "Métholodogie de conception des matériaux architecturés pour le stockage latent dans le domaine du bâtiment". Thesis, Lyon, INSA, 2013. http://www.theses.fr/2013ISAL0060/document.
Pełny tekst źródłaThe use of energy storage systems that exploit latent heat represents a promising solution to erase the heating demand of residential buildings during periods of peak demand. Equipping a building with such components can contribute to the goal of peak shaving in terms of public electricity grid supply. Significant drawbacks, however, are the low thermal conductivity of Phase Change Materials (PCM) that typically constitute such systems,and the requirement for a high rate of discharge. Consequently, the use of so-called architectured materials has been put forward as a means to optimize the effective conductivity of storage materials. Our work is focused upon the development of a methodology to design optimal materials for such systems that meet the criteria of energy storage and energy output. A so-called “top-down metholodogy” was implemented for the present work. This approach includes three scales of interest: building (top), system and material (down). The aim of the building scale analysis is to formulate a set of general design requirements. These are complemented by performance indicators, which are defined at the scale of the system. Finally, at the scale of the material, the architecture of the identified material is elaborated. A numerical simulation tool was developed to determine performance indicators for a latent heat energy storage system comprising of an air/PCM heat exchanger. This model was tested against a benchmark analytical solution and validated though comparison to experimental data. The developed methodology is applied to the specific case of an air/PCM exchanger latent-heat energy storage system. The system is analysed through the study of dimensionless numbers, which provide a set of design indicators for the system. As a result of this stage, the optimal material and functional properties are thus identified. Finally, an architectured material is proposed that would satisfy the design requirements of the storage system. We demonstrate that an arrangement composed of a sandwich of planar layers with nails and PCM can offer the required material properties. Furthermore, in order to meet the desired functional properties, the system design is modified by the addition of fins at the exchange surfaces. With the addition of 20 fins of 3mm thickness attached to the exchange surface of the sandwich panel, the storage system eliminated the heating demand for 2 hours during the period of high daily demand in winter
Diallo, Alhassane. "Recherche de sous-groupes évolutifs et leur impact sur la survie : application à des patients atteints d'ataxie spinocérébelleuse (SCA) Body Mass Index Decline Is Related to Spinocerebellar Ataxia Disease Progression Survival in patients with spinocerebellar ataxia types 1, 2, 3, and 6 (EUROSCA): a longitudinal cohort study". Thesis, Sorbonne université, 2018. http://www.theses.fr/2018SORUS447.
Pełny tekst źródłaIn cohort studies, most often the models used assume that the study population follows an average pattern of evolution. However, in many cases, such as for spinocerebellar ataxias (SCA), it is not uncommon for heterogeneity to be suspected. This heterogeneity could also be inuenced by other intercurrent events for example joint evolution of a second phenotype or occurrence of an event such as dropout or death. In the first part of this thesis, we analyzed the evolution of the BMI of SCA patients and looked for different evolution profiles of BMI. We identified 3 subgroups of BMI evolution: decrease (23% of patients), increase (18%) and stable (59%) ; and we have shown that patients who lower their BMI are faster disease progression. In the second part, we studied the survival of SCA patients and developed a prognostic nomogram. We have shown that it is different according on the genotype. Survival is shorter in SCA1, intermediate in SCA2 and SCA3, and longer in SCA6. Finally, we assessed the long-term impact of ataxia progression on survival. We have shown that progression of ataxia is associated with shorter survival regardless of genotype. Only in SCA1 patients, we identified three subgroups of homogeneous patients in terms of disease progression and risk of death
Corneli, Marco. "Dynamic stochastic block models, clustering and segmentation in dynamic graphs". Thesis, Paris 1, 2017. http://www.theses.fr/2017PA01E012/document.
Pełny tekst źródłaThis thesis focuses on the statistical analysis of dynamic graphs, both defined in discrete or continuous time. We introduce a new extension of the stochastic block model (SBM) for dynamic graphs. The proposed approach, called dSBM, adopts non homogeneous Poisson processes to model the interaction times between pairs of nodes in dynamic graphs, either in discrete or continuous time. The intensity functions of the processes only depend on the node clusters, in a block modelling perspective. Moreover, all the intensity functions share some regularity properties on hidden time intervals that need to be estimated. A recent estimation algorithm for SBM, based on the greedy maximization of an exact criterion (exact ICL) is adopted for inference and model selection in dSBM. Moreover, an exact algorithm for change point detection in time series, the "pruned exact linear time" (PELT) method is extended to deal with dynamic graph data modelled via dSBM. The approach we propose can be used for change point analysis in graph data. Finally, a further extension of dSBM is developed to analyse dynamic net- works with textual edges (like social networks, for instance). In this context, the graph edges are associated with documents exchanged between the corresponding vertices. The textual content of the documents can provide additional information about the dynamic graph topological structure. The new model we propose is called "dynamic stochastic topic block model" (dSTBM).Graphs are mathematical structures very suitable to model interactions between objects or actors of interest. Several real networks such as communication networks, financial transaction networks, mobile telephone networks and social networks (Facebook, Linkedin, etc.) can be modelled via graphs. When observing a network, the time variable comes into play in two different ways: we can study the time dates at which the interactions occur and/or the interaction time spans. This thesis only focuses on the first time dimension and each interaction is assumed to be instantaneous, for simplicity. Hence, the network evolution is given by the interaction time dates only. In this framework, graphs can be used in two different ways to model networks. Discrete time […] Continuous time […]. In this thesis both these perspectives are adopted, alternatively. We consider new unsupervised methods to cluster the vertices of a graph into groups of homogeneous connection profiles. In this manuscript, the node groups are assumed to be time invariant to avoid possible identifiability issues. Moreover, the approaches that we propose aim to detect structural changes in the way the node clusters interact with each other. The building block of this thesis is the stochastic block model (SBM), a probabilistic approach initially used in social sciences. The standard SBM assumes that the nodes of a graph belong to hidden (disjoint) clusters and that the probability of observing an edge between two nodes only depends on their clusters. Since no further assumption is made on the connection probabilities, SBM is a very flexible model able to detect different network topologies (hubs, stars, communities, etc.)
Boucquemont, Julie. "Modèles statistiques pour l'étude de la progression de la maladie rénale chronique". Thesis, Bordeaux, 2014. http://www.theses.fr/2014BORD0411/document.
Pełny tekst źródłaThe objective of this thesis was to illustrate the benefit of using advanced statistical methods to study associations between risk factors and chrouic kidney disease (CKD) progression. In a first time, we conducted a literature review of statistical methods used to investigate risk factors of CKD progression, identified important methodological issues, and discussed solutions. In our sec ond work, we focused on survival analyses and issues with interval-censoring, which occurs when the event of interest is the progression to a specifie CKD stage, and competing risk with death. A comparison between standard survival models and the illness-death mode! for interval-censored data allowed us to illustrate the impact of modeling on the estimates of both the effects of risk factors and the probabilities of events, using data from the NephroTest cohort. Other works fo cused on analysis of longitudinal data on renal function. We illustrated the interest of linear mixed mode! in this context and presented its extension to account for sub-populations with different trajectories of renal function. We identified five classes, including one with a strong decline and one with an improvement of renal function over time. Severa! perspectives on predictions bind the two types of analyses presented in this thesis
Xiong, Hao. "Diversified Latent Variable Models". Thesis, The University of Sydney, 2018. http://hdl.handle.net/2123/18512.
Pełny tekst źródłaPodosinnikova, Anastasia. "Sur la méthode des moments pour l'estimation des modèles à variables latentes". Thesis, Paris Sciences et Lettres (ComUE), 2016. http://www.theses.fr/2016PSLEE050/document.
Pełny tekst źródłaLatent linear models are powerful probabilistic tools for extracting useful latent structure from otherwise unstructured data and have proved useful in numerous applications such as natural language processing and computer vision. However, the estimation and inference are often intractable for many latent linear models and one has to make use of approximate methods often with no recovery guarantees. An alternative approach, which has been popular lately, are methods based on the method of moments. These methods often have guarantees of exact recovery in the idealized setting of an infinite data sample and well specified models, but they also often come with theoretical guarantees in cases where this is not exactly satisfied. In this thesis, we focus on moment matchingbased estimation methods for different latent linear models. Using a close connection with independent component analysis, which is a well studied tool from the signal processing literature, we introduce several semiparametric models in the topic modeling context and for multi-view models and develop moment matching-based methods for the estimation in these models. These methods come with improved sample complexity results compared to the previously proposed methods. The models are supplemented with the identifiability guarantees, which is a necessary property to ensure their interpretability. This is opposed to some other widely used models, which are unidentifiable
Creagh-Osborne, Jane. "Latent variable generalized linear models". Thesis, University of Plymouth, 1998. http://hdl.handle.net/10026.1/1885.
Pełny tekst źródłaDallaire, Patrick. "Bayesian nonparametric latent variable models". Doctoral thesis, Université Laval, 2016. http://hdl.handle.net/20.500.11794/26848.
Pełny tekst źródłaOne of the important problems in machine learning is determining the complexity of the model to learn. Too much complexity leads to overfitting, which finds structures that do not actually exist in the data, while too low complexity leads to underfitting, which means that the expressiveness of the model is insufficient to capture all the structures present in the data. For some probabilistic models, the complexity depends on the introduction of one or more latent variables whose role is to explain the generative process of the data. There are various approaches to identify the appropriate number of latent variables of a model. This thesis covers various Bayesian nonparametric methods capable of determining the number of latent variables to be used and their dimensionality. The popularization of Bayesian nonparametric statistics in the machine learning community is fairly recent. Their main attraction is the fact that they offer highly flexible models and their complexity scales appropriately with the amount of available data. In recent years, research on Bayesian nonparametric learning methods have focused on three main aspects: the construction of new models, the development of inference algorithms and new applications. This thesis presents our contributions to these three topics of research in the context of learning latent variables models. Firstly, we introduce the Pitman-Yor process mixture of Gaussians, a model for learning infinite mixtures of Gaussians. We also present an inference algorithm to discover the latent components of the model and we evaluate it on two practical robotics applications. Our results demonstrate that the proposed approach outperforms, both in performance and flexibility, the traditional learning approaches. Secondly, we propose the extended cascading Indian buffet process, a Bayesian nonparametric probability distribution on the space of directed acyclic graphs. In the context of Bayesian networks, this prior is used to identify the presence of latent variables and the network structure among them. A Markov Chain Monte Carlo inference algorithm is presented and evaluated on structure identification problems and as well as density estimation problems. Lastly, we propose the Indian chefs process, a model more general than the extended cascading Indian buffet process for learning graphs and orders. The advantage of the new model is that it accepts connections among observable variables and it takes into account the order of the variables. We also present a reversible jump Markov Chain Monte Carlo inference algorithm which jointly learns graphs and orders. Experiments are conducted on density estimation problems and testing independence hypotheses. This model is the first Bayesian nonparametric model capable of learning Bayesian learning networks with completely arbitrary graph structures.
Wegelin, Jacob A. "Latent models for cross-covariance /". Thesis, Connect to this title online; UW restricted, 2001. http://hdl.handle.net/1773/8982.
Pełny tekst źródłaMena-Chavez, Ramses H. "Stationary models using latent structures". Thesis, University of Bath, 2003. https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.425643.
Pełny tekst źródłaAmoualian, Hesam. "Modélisation et apprentissage de dépendances á l’aide de copules dans les modéles probabilistes latents". Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAM078/document.
Pełny tekst źródłaThis thesis focuses on scaling latent topic models for big data collections, especiallywhen document streams. Although the main goal of probabilistic modeling is to find word topics, an equally interesting objective is to examine topic evolutions and transitions. To accomplish this task, we propose in Chapter 3, three new models for modeling topic and word-topic dependencies between consecutive documents in document streams. The first model is a direct extension of Latent Dirichlet Allocation model (LDA) and makes use of a Dirichlet distribution to balance the influence of the LDA prior parameters with respect to topic and word-topic distributions of the previous document. The second extension makes use of copulas, which constitute a generic tool to model dependencies between random variables. We rely here on Archimedean copulas, and more precisely on Franck copula, as they are symmetric and associative and are thus appropriate for exchangeable random variables. Lastly, the third model is a non-parametric extension of the second one through the integration of copulas in the stick-breaking construction of Hierarchical Dirichlet Processes (HDP). Our experiments, conducted on five standard collections that have been used in several studies on topic modeling, show that our proposals outperform previous ones, as dynamic topic models, temporal LDA and the Evolving Hierarchical Processes,both in terms of perplexity and for tracking similar topics in document streams. Compared to previous proposals, our models have extra flexibility and can adapt to situations where there are no dependencies between the documents.On the other hand, the "Exchangeability" assumption in topic models like LDA oftenresults in inferring inconsistent topics for the words of text spans like noun-phrases, which are usually expected to be topically coherent. In Chapter 4, we propose copulaLDA (copLDA), that extends LDA by integrating part of the text structure to the model and relaxes the conditional independence assumption between the word-specific latent topics given the per-document topic distributions. To this end, we assume that the words of text spans like noun-phrases are topically bound and we model this dependence with copulas. We demonstrate empirically the effectiveness of copLDA on both intrinsic and extrinsic evaluation tasks on several publicly available corpora. To complete the previous model (copLDA), Chapter 5 presents an LDA-based model that generates topically coherent segments within documents by jointly segmenting documents and assigning topics to their words. The coherence between topics is ensured through a copula, binding the topics associated to the words of a segment. In addition, this model relies on both document and segment specific topic distributions so as to capture fine-grained differences in topic assignments. We show that the proposed model naturally encompasses other state-of-the-art LDA-based models designed for similar tasks. Furthermore, our experiments, conducted on six different publicly available datasets, show the effectiveness of our model in terms of perplexity, Normalized Pointwise Mutual Information, which captures the coherence between the generated topics, and the Micro F1 measure for text classification
Dupuy, Christophe. "Inference and applications for topic models". Thesis, Paris Sciences et Lettres (ComUE), 2017. http://www.theses.fr/2017PSLEE055/document.
Pełny tekst źródłaMost of current recommendation systems are based on ratings (i.e. numbers between 0 and 5) and try to suggest a content (movie, restaurant...) to a user. These systems usually allow users to provide a text review for this content in addition to ratings. It is hard to extract useful information from raw text while a rating does not contain much information on the content and the user. In this thesis, we tackle the problem of suggesting personalized readable text to users to help them make a quick decision about a content. More specifically, we first build a topic model that predicts personalized movie description from text reviews. Our model extracts distinct qualitative (i.e., which convey opinion) and descriptive topics by combining text reviews and movie ratings in a joint probabilistic model. We evaluate our model on an IMDB dataset and illustrate its performance through comparison of topics. We then study parameter inference in large-scale latent variable models, that include most topic models. We propose a unified treatment of online inference for latent variable models from a non-canonical exponential family, and draw explicit links between several previously proposed frequentist or Bayesian methods. We also propose a novel inference method for the frequentist estimation of parameters, that adapts MCMC methods to online inference of latent variable models with the proper use of local Gibbs sampling.~For the specific latent Dirichlet allocation topic model, we provide an extensive set of experiments and comparisons with existing work, where our new approach outperforms all previously proposed methods. Finally, we propose a new class of determinantal point processes (DPPs) which can be manipulated for inference and parameter learning in potentially sublinear time in the number of items. This class, based on a specific low-rank factorization of the marginal kernel, is particularly suited to a subclass of continuous DPPs and DPPs defined on exponentially many items. We apply this new class to modelling text documents as sampling a DPP of sentences, and propose a conditional maximum likelihood formulation to model topic proportions, which is made possible with no approximation for our class of DPPs. We present an application to document summarization with a DPP on 2 to the power 500 items, where the summaries are composed of readable sentences
Sagara, Issaka. "Méthodes d'analyse statistique pour données répétées dans les essais cliniques : intérêts et applications au paludisme". Thesis, Aix-Marseille, 2014. http://www.theses.fr/2014AIXM5081/document.
Pełny tekst źródłaNumerous clinical studies or control interventions were done or are ongoing in Africa for malaria control. For an efficient control of this disease, the strategies should be closer to the reality of the field and the data should be analyzed appropriately. In endemic areas, malaria is a recurrent disease. Repeated malaria episodes are common in African. However, the literature review indicates a limited application of appropriate statistical tools for the analysis of recurrent malaria data. We implemented appropriate statistical methods for the analysis of these data We have also studied the repeated measurements of hemoglobin during malaria treatments follow-up in order to assess the safety of the study drugs by pooling data from 13 clinical trials.For the analysis of the number of malaria episodes, the negative binomial regression has been implemented. To model the recurrence of malaria episodes, four models were used: i) the generalized estimating equations (GEE) using the Poisson distribution; and three models that are an extension of the Cox model: ii) Andersen-Gill counting process (AG-CP), iii) Prentice-Williams-Peterson counting process (PWP-CP); and (iv) the shared gamma frailty model. For the safety analysis, i.e. the assessment of the impact of malaria treatment on hemoglobin levels or the onset of anemia, the generalized linear and latent mixed models (GLLAMM) has been implemented. We have shown how to properly apply the existing statistical tools in the analysis of these data. The prospects of this work remain in the development of guides on good practices on the methodology of the preparation and analysis and storage network for malaria data
Christmas, Jacqueline. "Robust spatio-temporal latent variable models". Thesis, University of Exeter, 2011. http://hdl.handle.net/10036/3051.
Pełny tekst źródłaChen, George H. "Latent source models for nonparametric inference". Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/99774.
Pełny tekst źródłaThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 95-101).
Nearest-neighbor inference methods have been widely and successfully used in numerous applications such as forecasting which news topics will go viral, recommending products to people in online stores, and delineating objects in images by looking at image patches. However, there is little theoretical understanding of when, why, and how well these nonparametric inference methods work in terms of key problem-specific quantities relevant to practitioners. This thesis bridges the gap between theory and practice for these methods in the three specific case studies of time series classification, online collaborative filtering, and patch-based image segmentation. To do so, for each of these problems, we prescribe a probabilistic model in which the data appear generated from unknown "latent sources" that capture salient structure in the problem. These latent source models naturally lead to nearest-neighbor or nearest-neighbor-like inference methods similar to ones already used in practice. We derive theoretical performance guarantees for these methods, relating inference quality to the amount of training data available and problems-specific structure modeled by the latent sources.
by George H. Chen.
Ph. D.
Wanigasekara, Prashan. "Latent state space models for prediction". Thesis, Massachusetts Institute of Technology, 2016. http://hdl.handle.net/1721.1/106269.
Pełny tekst źródłaCataloged from PDF version of thesis.
Includes bibliographical references (pages 95-98).
In this thesis, I explore a novel algorithm to model the joint behavior of multiple correlated signals. Our chosen example is the ECG (Electrocardiogram) and ABP (Arterial Blood Pressure) signals from patients in the ICU (Intensive Care Unit). I then use the generated models to predict blood pressure levels of ICU patients based on their historical ECG and ABP signals. The algorithm used is a variant of a Hidden Markov model. The new extension is termed as the Latent State Space Copula Model. In the novel Latent State Space Copula Modelthe ECG, ABP signals are considered to be correlated and are modeled using a bivariate Gaussian copula with Weibull marginals generated by a hidden state. We assume that there are hidden patient "states" that transition from one hidden state to another driving a joint ECG-ABP behavior. We estimate the parameters of the model using a novel Gibbs sampling approach. Using this model, we generate predictors that are the state probabilities at any given time step and use them to predict a patient's future health condition. The predictions made by the model are binary and detects whether the Mean arterial pressure(MAP) is going to be above or below a certain threshold at a future time step. Towards the end of the thesis I do a comparison between the new Latent State Space Copula Model and a state of the art Classical Discrete HMM. The Latent State Space Copula Model achieves an Area Under the ROC (AUROC) curve of .7917 for 5 states while the Classical Discrete HMM achieves an AUROC of .7609 for 5 states.
by Prashan Wanigasekara.
S.M. in Engineering and Management
Paquet, Ulrich. "Bayesian inference for latent variable models". Thesis, University of Cambridge, 2007. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.613111.
Pełny tekst źródłaO'Sullivan, Aidan Michael. "Bayesian latent variable models with applications". Thesis, Imperial College London, 2013. http://hdl.handle.net/10044/1/19191.
Pełny tekst źródłaZhang, Cheng. "Structured Representation Using Latent Variable Models". Doctoral thesis, KTH, Datorseende och robotik, CVAP, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-191455.
Pełny tekst źródłaQC 20160905
Surian, Didi. "Novel Applications Using Latent Variable Models". Thesis, The University of Sydney, 2015. http://hdl.handle.net/2123/14014.
Pełny tekst źródłaParsons, S. "Approximation methods for latent variable models". Thesis, University College London (University of London), 2016. http://discovery.ucl.ac.uk/1513250/.
Pełny tekst źródłaOldmeadow, Christopher. "Latent variable models in statistical genetics". Thesis, Queensland University of Technology, 2009. https://eprints.qut.edu.au/31995/1/Christopher_Oldmeadow_Thesis.pdf.
Pełny tekst źródłaLaclau, Charlotte. "Hard and fuzzy block clustering algorithms for high dimensional data". Thesis, Sorbonne Paris Cité, 2016. http://www.theses.fr/2016USPCB014.
Pełny tekst źródłaWith the increasing number of data available, unsupervised learning has become an important tool used to discover underlying patterns without the need to label instances manually. Among different approaches proposed to tackle this problem, clustering is arguably the most popular one. Clustering is usually based on the assumption that each group, also called cluster, is distributed around a center defined in terms of all features while in some real-world applications dealing with high-dimensional data, this assumption may be false. To this end, co-clustering algorithms were proposed to describe clusters by subsets of features that are the most relevant to them. The obtained latent structure of data is composed of blocks usually called co-clusters. In first two chapters, we describe two co-clustering methods that proceed by differentiating the relevance of features calculated with respect to their capability of revealing the latent structure of the data in both probabilistic and distance-based framework. The probabilistic approach uses the mixture model framework where the irrelevant features are assumed to have a different probability distribution that is independent of the co-clustering structure. On the other hand, the distance-based (also called metric-based) approach relied on the adaptive metric where each variable is assigned with its weight that defines its contribution in the resulting co-clustering. From the theoretical point of view, we show the global convergence of the proposed algorithms using Zangwill convergence theorem. In the last two chapters, we consider a special case of co-clustering where contrary to the original setting, each subset of instances is described by a unique subset of features resulting in a diagonal structure of the initial data matrix. Same as for the two first contributions, we consider both probabilistic and metric-based approaches. The main idea of the proposed contributions is to impose two different kinds of constraints: (1) we fix the number of row clusters to the number of column clusters; (2) we seek a structure of the original data matrix that has the maximum values on its diagonal (for instance for binary data, we look for diagonal blocks composed of ones with zeros outside the main diagonal). The proposed approaches enjoy the convergence guarantees derived from the results of the previous chapters. Finally, we present both hard and fuzzy versions of the proposed algorithms. We evaluate our contributions on a wide variety of synthetic and real-world benchmark binary and continuous data sets related to text mining applications and analyze advantages and inconvenients of each approach. To conclude, we believe that this thesis covers explicitly a vast majority of possible scenarios arising in hard and fuzzy co-clustering and can be seen as a generalization of some popular biclustering approaches
Martino, Sara. "Approximate Bayesian Inference for Latent Gaussian Models". Doctoral thesis, Norwegian University of Science and Technology, Department of Mathematical Sciences, 2007. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-1949.
Pełny tekst źródłaThis thesis consists of five papers, presented in chronological order. Their content is summarised in this section.
Paper I introduces the approximation tool for latent GMRF models and discusses, in particular, the approximation for the posterior of the hyperparameters θ in equation (1). It is shown that this approximation is indeed very accurate, as even long MCMC runs cannot detect any error in it. A Gaussian approximation to the density of χi|θ, y is also discussed. This appears to give reasonable results and it is very fast to compute. However, slight errors are detected when comparing the approximation with long MCMC runs. These are mostly due to the fact that a possible - skewed density is approximated via a symmetric one. Paper I presents also some details about sparse matrices algorithms.
The core of the thesis is presented in Paper II. Here most of the remaining issues present in Paper I are solved. Three different approximation for χi|θ, y with different degrees of accuracy and computational costs are described. Moreover, ways to assess the approximation error and considerations about the asymptotical behaviour of the approximations are also discussed. Through a series of examples covering a wide range of commonly used latent GMRF models, the approximations are shown to give extremely accurate results in a fraction of the computing time used by MCMC algorithms.
Paper III applies the same ideas as Paper II to generalised linear mixed models where χ represents a latent variable at n spatial sites on a two dimensional domain. Out of these n sites k, with n >> k , are observed through data. The n sites are assumed to be on a regular grid and wrapped on a torus. For the class of models described in Paper III the computations are based on discrete Fourier transform instead of sparse matrices. Paper III illustrates also how marginal likelihood π (y) can be approximated, provides approximate strategies for Bayesian outlier detection and perform approximate evaluation of spatial experimental design.
Paper IV presents yet another application of the ideas in Paper II. Here approximate techniques are used to do inference on multivariate stochastic volatility models, a class of models widely used in financial applications. Paper IV discusses also problems deriving from the increased dimension of the parameter vector θ, a condition which makes all numerical integration more computationally intensive. Different approximations for the posterior marginals of the parameters θ, π(θi)|y), are also introduced. Approximations to the marginal likelihood π(y) are used in order to perform model comparison.
Finally, Paper V is a manual for a program, named inla which implements all approximations described in Paper II. A large series of worked out examples, covering many well known models, illustrate the use and the performance of the inla program. This program is a valuable instrument since it makes most of the Bayesian inference techniques described in this thesis easily available for everyone.
Dominicus, Annica. "Latent variable models for longitudinal twin data". Doctoral thesis, Stockholm : Mathematical statistics, Dept. of mathematics, Stockholm university, 2006. http://urn.kb.se/resolve?urn=urn:nbn:se:su:diva-848.
Pełny tekst źródłaSaba, Laura M. "Latent pattern mixture models for binary outcomes /". Connect to full text via ProQuest. Limited to UCD Anschutz Medical Campus, 2007.
Znajdź pełny tekst źródłaTypescript. Includes bibliographical references (leaves 70-71). Free to UCD affiliates. Online version available via ProQuest Digital Dissertations;
Jung, Sunho. "Regularized structural equation models with latent variables". Thesis, McGill University, 2009. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=66858.
Pełny tekst źródłaDans les modèles d'équations structurales avec des variables latentes, l'estimation demaximum devraisemblance est la méthode d'estimation la plus utilisée. Par contre, la méthode de maximum devraisemblance souvent ne réussit pas á fournir des solutions exactes, par exemple lorsque les échantillons sont petits, les données ne sont pas normale, ou lorsque le modèle est mal specifié. L'estimation des moindres carrés á deux-phases est asymptotiquement sans distribution et robuste contre mauvaises spécifications, mais elle manque de robustesse quand les chantillons sont petits. Afin de surmonter les trois difficultés mentionnés ci-dessus et d'obtenir une estimation plus exacte, des extensions régularisées des moindres carrés á deux phases sont proposé á qui incorporent directement un type de régularisation dans les modèles d'équations structurales avec des variables latentes. Deux études de simulation et deux applications empiriques démontrent que la méthode propose est une alternative prometteuse aux méthodes de maximum vraisemblance et de l'estimation des moindres carrés á deux-phases. Un paramètre de régularisation valeur optimale a été trouvé par la technique de validation croisé d'ordre K. Une méthode non-paramétrique Bootstrap est utilisée afin d'évaluer la stabilité des solutions. Une mesure d'adéquation est utilisée pour estimer l'adéquation globale.
Moustaki, Irini. "Latent variable models for mixed manifest variables". Thesis, London School of Economics and Political Science (University of London), 1996. http://etheses.lse.ac.uk/78/.
Pełny tekst źródłaWhite, S. A. "Latent structure models for repeated measurements experiments". Thesis, University of Nottingham, 1986. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.376163.
Pełny tekst źródłaBurridge, C. Y. "Latent variable models for genotype-environment interaction". Thesis, University of Reading, 1988. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.383469.
Pełny tekst źródłaChallis, E. A. L. "Variational approximate inference in latent linear models". Thesis, University College London (University of London), 2013. http://discovery.ucl.ac.uk/1414228/.
Pełny tekst źródłaAlbanese, Maria Teresinha. "Latent variable models for binary response data". Thesis, London School of Economics and Political Science (University of London), 1990. http://etheses.lse.ac.uk/1220/.
Pełny tekst źródłaBasbug, Mehmet Emin. "Integrating Exponential Dispersion Models to Latent Structures". Thesis, Princeton University, 2017. http://pqdtopen.proquest.com/#viewpdf?dispub=10254057.
Pełny tekst źródłaLatent variable models have two basic components: a latent structure encoding a hypothesized complex pattern and an observation model capturing the data distribution. With the advancements in machine learning and increasing availability of resources, we are able to perform inference in deeper and more sophisticated latent variable models. In most cases, these models are designed with a particular application in mind; hence, they tend to have restrictive observation models. The challenge, surfaced with the increasing diversity of data sets, is to generalize these latent models to work with different data types. We aim to address this problem by utilizing exponential dispersion models (EDMs) and proposing mechanisms for incorporating them into latent structures. (Abstract shortened by ProQuest.)
Wenzel, Florian. "Scalable Inference in Latent Gaussian Process Models". Doctoral thesis, Humboldt-Universität zu Berlin, 2020. http://dx.doi.org/10.18452/20926.
Pełny tekst źródłaLatent Gaussian process (GP) models help scientists to uncover hidden structure in data, express domain knowledge and form predictions about the future. These models have been successfully applied in many domains including robotics, geology, genetics and medicine. A GP defines a distribution over functions and can be used as a flexible building block to develop expressive probabilistic models. The main computational challenge of these models is to make inference about the unobserved latent random variables, that is, computing the posterior distribution given the data. Currently, most interesting Gaussian process models have limited applicability to big data. This thesis develops a new efficient inference approach for latent GP models. Our new inference framework, which we call augmented variational inference, is based on the idea of considering an augmented version of the intractable GP model that renders the model conditionally conjugate. We show that inference in the augmented model is more efficient and, unlike in previous approaches, all updates can be computed in closed form. The ideas around our inference framework facilitate novel latent GP models that lead to new results in language modeling, genetic association studies and uncertainty quantification in classification tasks.
Chen, Tao. "Search-based learning of latent tree models /". View abstract or full-text, 2009. http://library.ust.hk/cgi/db/thesis.pl?CSED%202009%20CHEN.
Pełny tekst źródłaFusi, Nicolo. "Probabilistic latent variable models in statistical genomics". Thesis, University of Sheffield, 2015. http://etheses.whiterose.ac.uk/8326/.
Pełny tekst źródłaRidall, Peter Gareth. "Bayesian Latent Variable Models for Biostatistical Applications". Thesis, Queensland University of Technology, 2004. https://eprints.qut.edu.au/16164/1/Peter_Ridall_Thesis.pdf.
Pełny tekst źródłaRidall, Peter Gareth. "Bayesian Latent Variable Models for Biostatistical Applications". Queensland University of Technology, 2004. http://eprints.qut.edu.au/16164/.
Pełny tekst źródłaPegoraro, Fulvio <1974>. "Discrete time pricing: models with latent variables". Doctoral thesis, Università Ca' Foscari Venezia, 2004. http://hdl.handle.net/10579/197.
Pełny tekst źródłaGao, Sheng. "Latent factor models for link prediction problems". Paris 6, 2012. http://www.theses.fr/2012PA066056.
Pełny tekst źródłaWith the rising of Internet as well as modern social media, relational data has become ubiquitous, which consists of those kinds of data where the objects are linked to each other with various relation types. Accordingly, various relational learning techniques have been studied in a large variety of applications with relational data, such as recommender systems, social network analysis, Web mining or bioinformatic. Among a wide range of tasks encompassed by relational learning, we address the problem of link prediction in this thesis. Link prediction has arisen as a fundamental task in relational learning, which considers to predict the presence or absence of links between objects in the relational data based on the topological structure of the network and/or the attributes of objects. However, the complexity and sparsity of network structure make this a great challenging problem. In this thesis, we propose solutions to reduce the difficulties in learning and fit various models into corresponding applications. Basically, in Chapter 3 we present a unified framework of latent factor models to address the generic link prediction problem, in which we specifically discuss various configurations in the models from computational perspective and probabilistic view. Then, according to the applications addressed in this dissertation, we propose different latentfactor models for two classes of link prediction problems: (i) structural link prediction. (ii) temporal link prediction. In terms of structural link prediction problem, in Chapter 4 we define a new task called Link Pattern Prediction (LPP) in multi-relational networks. By introducing a specific latent factor for different relation types in addition to using latent feature factors to characterize objects, we develop a computational tensor factorization model, and the probabilistic version with its Bayesian treatment to reveal the intrinsic causality of interaction patterns in multi-relational networks. Moreover, considering the complex structural patterns in relational data, in Chapter 5 we propose a novel model that simultaneously incorporates the effect of latent feature factors and the impact from the latent cluster structures in the network, and also develop an optimization transfer algorithm to facilitate the model learning procedure. In terms of temporal link prediction problem in time-evolving networks, in Chapter 6 we propose a unified latent factor model which integrates multiple information sources in the network, including the global network structure, the content of objects and the graph proximity information from the network to capture the time-evolving patterns of links. This joint model is constructed based on matrix factorization and graph regularization technique. Each model proposed in this thesis achieves state-of-the-art performances, extensive experiments are conducted on real world datasets to demonstrate their significant improvements over baseline methods. Almost all of themhave been published in international or national peer-reviewed conference proceedings
Pasquiou, Alexandre. "Deciphering the neural bases of language comprehension using latent linguistic representations". Electronic Thesis or Diss., université Paris-Saclay, 2023. http://www.theses.fr/2023UPASG041.
Pełny tekst źródłaIn the last decades, language models (LMs) have reached human level performance on several tasks. They can generate rich representations (features) that capture various linguistic properties such has semantics or syntax. Following these improvements, neuroscientists have increasingly used them to explore the neural bases of language comprehension. Specifically, LM's features computed from a story are used to fit the brain data of humans listening to the same story, allowing the examination of multiple levels of language processing in the brain. If LM's features closely align with a specific brain region, then it suggests that both the model and the region are encoding the same information. LM-brain comparisons can then teach us about language processing in the brain. Using the fMRI brain data of fifty US participants listening to "The Little Prince" story, this thesis 1) investigates the reasons why LMs' features fit brain activity and 2) examines the limitations of such comparisons. The comparison of several pre-trained and custom-trained LMs (GloVe, LSTM, GPT-2 and BERT) revealed that Transformers better fit fMRI brain data than LSTM and GloVe. Yet, none are able to explain all the fMRI signal, suggesting either limitations related to the encoding paradigm or to the LMs. Focusing specifically on Transformers, we found that no brain region is better fitted by specific attentional head or layer. Our results caution that the nature and the amount of training data greatly affects the outcome, indicating that using off-the-shelf models trained on small datasets is not effective in capturing brain activations. We showed that LMs' training influences their ability to fit fMRI brain data, and that perplexity was not a good predictor of brain score. Still, training LMs particularly improves their fitting performance in core semantic regions, irrespective of the architecture and training data. Moreover, we showed a partial convergence between brain's and LM's representations.Specifically, they first converge during model training before diverging from one another. This thesis further investigates the neural bases of syntax, semantics and context-sensitivity by developing a method that can probe specific linguistic dimensions. This method makes use of "information-restricted LMs", that are customized LMs architectures trained on feature spaces containing a specific type of information, in order to fit brain data. First, training LMs on semantic and syntactic features revealed a good fitting performance in a widespread network, albeit with varying relative degrees. The quantification of this relative sensitivity to syntax and semantics showed that brain regions most attuned to syntax tend to be more localized, while semantic processing remain widely distributed over the cortex. One notable finding from this analysis was that the extent of semantic and syntactic sensitive brain regions was similar across hemispheres. However, the left hemisphere had a greater tendency to distinguish between syntactic and semantic processing compared to the right hemisphere. In a last set of experiments we designed "masked-attention generation", a method that controls the attention mechanisms in transformers, in order to generate latent representations that leverage fixed-size context. This approach provides evidence of context-sensitivity across most of the cortex. Moreover, this analysis found that the left and right hemispheres tend to process shorter and longer contextual information respectively
Cuesta, Ramirez Jhouben Janyk. "Optimization of a computationally expensive simulator with quantitative and qualitative inputs". Thesis, Lyon, 2022. http://www.theses.fr/2022LYSEM010.
Pełny tekst źródłaIn this thesis, costly mixed problems are approached through gaussian processes where the discrete variables are relaxed into continuous latent variables. the continuous space is more easily harvested by classical bayesian optimization techniques than a mixed space would. discrete variables are recovered either subsequently to the continuous optimization, or simultaneously with an additional continuous-discrete compatibility constraint that is handled with augmented lagrangians. several possible implementations of such bayesian mixed optimizers are compared. in particular, the reformulation of the problem with continuous latent variables is put in competition with searches working directly in the mixed space. among the algorithms involving latent variables and an augmented lagrangian, a particular attention is devoted to the lagrange multipliers for which a local and a global estimation techniques are studied. the comparisons are based on the repeated optimization of three analytical functions and a mechanical application regarding a beam design. an additional study for applying a proposed mixed optimization strategy in the field of mixed self-calibration is made. this analysis was inspired in an application in radionuclide quantification, which defined an specific inverse function that required the study of its multiple properties in the continuous scenario. a proposition of different deterministic and bayesian strategies was made towards a complete definition in a mixed variable setup
Ödling, David, i Arvid Österlund. "Factorisation of Latent Variables in Word Space Models : Studying redistribution of weight on latent variables". Thesis, KTH, Skolan för teknikvetenskap (SCI), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-153776.
Pełny tekst źródłaMålet med alla semantiska fördelningsmodeller (DSMs) är en skalbaroch precis representation av semantiska relationer. Nya rön från Bullinaria & Levy (2012) och Caron (2001) indikerar att man kan förbättra prestandan avsevärt genom att omfördela vikten ifrån principalkomponenterna med störst varians mot de lägre. Varför metoden fungerar är dock fortfarande oklart, delvis på grund av höga beräkningskostnader för PCA men även på grund av att resultaten strider mot tidigare praxis. Vi börjar med att replikera resultaten i Bullinaria & Levy (2012) för att sedan fördjupa oss i resultaten, både kvantitativt och kvalitativt, genom att använda oss av BLESS testet. Huvudresultaten av denna studie är verifiering av 100% på TOEFL testet och ett nytt resultat på en paradigmatisk variant av BLESStestet på 91.5%. Våra resultat tyder på att en omfördelning av vikten ifrån de första principalkomponenterna leder till en förändring i fördelningensins emellan de semantiska relationerna vilket delvis förklarar förbättringen i TOEFL resultaten. Vidare finner vi i enlighet med tidigare resultat ingen signifikant relation mellan ordfrekvenser och viktomfördelning. Utifrån dessa resultat föreslår vi en rad experiment som kan ge vidare insikt till dessa intressanta resultat.
PENNONI, FULVIA. "Issues on the Estimation of Latent Variable and Latent Class Models with Social Science Applications". Doctoral thesis, Università degli Studi di Firenze, 2004. http://hdl.handle.net/10281/46004.
Pełny tekst źródła