Dissertations / Theses on the topic 'Gaussian process mixture model'

To see the other types of publications on this topic, follow the link: Gaussian process mixture model.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Gaussian process mixture model.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Zhang, Lin. "Semiparametric Bayesian Kernel Survival Model for Highly Correlated High-Dimensional Data." Diss., Virginia Tech, 2018. http://hdl.handle.net/10919/95040.

Full text
Abstract:
We are living in an era in which many mysteries related to science, technologies and design can be answered by "learning" the huge amount of data accumulated over the past few decades. In the processes of those endeavors, highly-correlated high-dimensional data are frequently observed in many areas including predicting shelf life, controlling manufacturing processes, and identifying important pathways related with diseases. We define a "set" as a group of highly-correlated high-dimensional (HCHD) variables that possess a certain practical meaning or control a certain process, and define an "element" as one of the HCHD variables within a certain set. Such an elements-within-a-set structure is very complicated because: (i) the dimensions of elements in different sets can vary dramatically, ranging from two to hundreds or even thousands; (ii) the true relationships, include element-wise associations, set-wise interactions, and element-set interactions, are unknown; (iii) and the sample size (n) is usually much smaller than the dimension of the elements (p). The goal of this dissertation is to provide a systematic way to identify both the set effects and the element effects associated with survival outcomes from heterogeneous populations using Bayesian survival kernel models. By connecting kernel machines with semiparametric Bayesian hierarchical models, the proposed unified model frameworks can identify significant elements as well as sets regardless of mis-specifications of distributions or kernels. The proposed methods can potentially be applied to a vast range of fields to solve real-world problems.
PHD
APA, Harvard, Vancouver, ISO, and other styles
2

Xu, Li. "Statistical Methods for Variability Management in High-Performance Computing." Diss., Virginia Tech, 2021. http://hdl.handle.net/10919/104184.

Full text
Abstract:
High-performance computing (HPC) variability management is an important topic in computer science. Research topics include experimental designs for efficient data collection, surrogate models for predicting the performance variability, and system configuration optimization. Due to the complex architecture of HPC systems, a comprehensive study of HPC variability needs large-scale datasets, and experimental design techniques are useful for improved data collection. Surrogate models are essential to understand the variability as a function of system parameters, which can be obtained by mathematical and statistical models. After predicting the variability, optimization tools are needed for future system designs. This dissertation focuses on HPC input/output (I/O) variability through three main chapters. After the general introduction in Chapter 1, Chapter 2 focuses on the prediction models for the scalar description of I/O variability. A comprehensive comparison study is conducted, and major surrogate models for computer experiments are investigated. In addition, a tool is developed for system configuration optimization based on the chosen surrogate model. Chapter 3 conducts a detailed study for the multimodal phenomena in I/O throughput distribution and proposes an uncertainty estimation method for the optimal number of runs for future experiments. Mixture models are used to identify the number of modes for throughput distributions at different configurations. This chapter also addresses the uncertainty in parameter estimation and derives a formula for sample size calculation. The developed method is then applied to HPC variability data. Chapter 4 focuses on the prediction of functional outcomes with both qualitative and quantitative factors. Instead of a scalar description of I/O variability, the distribution of I/O throughput provides a comprehensive description of I/O variability. We develop a modified Gaussian process for functional prediction and apply the developed method to the large-scale HPC I/O variability data. Chapter 5 contains some general conclusions and areas for future work.
Doctor of Philosophy
This dissertation focuses on three projects that are all related to statistical methods in performance variability management in high-performance computing (HPC). HPC systems are computer systems that create high performance by aggregating a large number of computing units. The performance of HPC is measured by the throughput of a benchmark called the IOZone Filesystem Benchmark. The performance variability is the variation among throughputs when the system configuration is fixed. Variability management involves studying the relationship between performance variability and the system configuration. In Chapter 2, we use several existing prediction models to predict the standard deviation of throughputs given different system configurations and compare the accuracy of predictions. We also conduct HPC system optimization using the chosen prediction model as the objective function. In Chapter 3, we use the mixture model to determine the number of modes in the distribution of throughput under different system configurations. In addition, we develop a model to determine the number of additional runs for future benchmark experiments. In Chapter 4, we develop a statistical model that can predict the throughout distributions given the system configurations. We also compare the prediction of summary statistics of the throughput distributions with existing prediction models.
APA, Harvard, Vancouver, ISO, and other styles
3

Erich, Roger Alan. "Regression Modeling of Time to Event Data Using the Ornstein-Uhlenbeck Process." The Ohio State University, 2012. http://rave.ohiolink.edu/etdc/view?acc_num=osu1342796812.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Bjarnason, Brynjar Smári. "Clustering metagenome contigs using coverage with CONCOCT." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-208944.

Full text
Abstract:
Metagenomics allows studying genetic potentials of microorganisms without prior cultivation. Since metagenome assembly results in fragmented genomes, a key challenge is to cluster the genome fragments (contigs) into more or less complete genomes. The goal of this project was to investigate how well CONCOCT bins assembled contigs into taxonomically relevant clusters using the abundance profiles of the contigs over multiple samples. This was done by studying the effects of different parameter settings for CONCOCT on the clustering results when clustering metagenome contigs from in silico model communities generated by mixing data from isolate genomes. These parameters control how the model that CONCOCT trains is tuned and then how the model fits contigs to their cluster. Each parameter was tested in isolation while others were kept at their default values. For each of the data set used, the number of clusters was kept constant at the known number of species and strains in their respective data set. The resulting configuration was to use a tied covariance model, using principal components explaining 90% of the variance, and filtering out contigs shorter than 3000 bp. It also suggested that all available samples should be used for the abundance profiles. Using these parameters for CONCOCT, it was executed to have it estimate the number of clusters automatically. This gave poor results which lead to the conclusion that the process for selecting the number of clusters that was implemented in CONCOCT, “Bayesian Information Criterion”, was not good enough. That led to the testing of another similar mathematical model, “Dirichlet Process Gaussian Mixture Model”, that uses a different algorithm to estimate number of clusters. This new model gave much better results and CONCOCT has adapted a similar model in later versions.
Metagenomik möjliggör analys av arvsmassor i mikrobiella floror utan att först behöva odla mikroorgansimerna. Metoden innebär att man läser korta DNA-snuttar som sedan pusslas ihop till längre genomfragment (kontiger). Genom att gruppera kontiger som härstammar från samma organism kan man sedan återskapa mer eller mindre fullständiga genom, men detta är en svår bioinformatisk utmaning. Målsättningen med det här projektet var att utvärdera precisionen med vilken mjukvaran CONCOCT, som vi nyligen utvecklat, grupperar kontiger som härstammar från samma organism baserat på information om kontigernas sekvenskomposition och abundansprofil över olika prover. Vi testade hur olika parametrar påverkade klustringen av kontiger i artificiella metagenomdataset av olika komplexitet som vi skapade in silico genom att blanda data från tidigare sekvenserade genom. Parametrarna som testades rörde indata såväl som den statistiska modell som CONCOCT använder för att utföra klustringen. Parametrarna varierades en i taget medan de andra parametrarna hölls konstanta. Antalet kluster hölls också konstant och motsvarade antalet olika organismer i flororna. Bäst resultat erhölls då vi använde en låst kovariansmodell och använde principalkomponenter som förklarade 90% av variansen, samt filtrerade bort kontiger som var kortare än 3000 baspar. Vi fick också bäst resultat då vi använde alla tillgängliga prover. Därefter använde vi dessa parameterinställningar och lät CONCOCT själv bestämma lämpligt antal kluster i dataseten med “Bayesian Information Criterion” - metoden som då var implementerad i CONCOCT. Detta gav otillfredsställande resultat med i regel för få och för stora kluster. Därför testade vi en alternativ metod, “Dirichlet Process Gaussian Mixture Model”, för att uppskatta antal kluster. Denna metod gav avsevärt bättre resultat och i senare versioner av CONCOCT har en liknande metod implementerats.
APA, Harvard, Vancouver, ISO, and other styles
5

Tang, Man. "Statistical methods for variant discovery and functional genomic analysis using next-generation sequencing data." Diss., Virginia Tech, 2020. http://hdl.handle.net/10919/104039.

Full text
Abstract:
The development of high-throughput next-generation sequencing (NGS) techniques produces massive amount of data, allowing the identification of biomarkers in early disease diagnosis and driving the transformation of most disciplines in biology and medicine. A greater concentration is needed in developing novel, powerful, and efficient tools for NGS data analysis. This dissertation focuses on modeling ``omics'' data in various NGS applications with a primary goal of developing novel statistical methods to identify sequence variants, find transcription factor (TF) binding patterns, and decode the relationship between TF and gene expression levels. Accurate and reliable identification of sequence variants, including single nucleotide polymorphisms (SNPs) and insertion-deletion polymorphisms (INDELs), plays a fundamental role in NGS applications. Existing methods for calling these variants often make simplified assumption of positional independence and fail to leverage the dependence of genotypes at nearby loci induced by linkage disequilibrium. We propose vi-HMM, a hidden Markov model (HMM)-based method for calling SNPs and INDELs in mapped short read data. Simulation experiments show that, under various sequencing depths, vi-HMM outperforms existing methods in terms of sensitivity and F1 score. When applied to the human whole genome sequencing data, vi-HMM demonstrates higher accuracy in calling SNPs and INDELs. One important NGS application is chromatin immunoprecipitation followed by sequencing (ChIP-seq), which characterizes protein-DNA relations through genome-wide mapping of TF binding sites. Multiple TFs, binding to DNA sequences, often show complex binding patterns, which indicate how TFs with similar functionalities work together to regulate the expression of target genes. To help uncover the transcriptional regulation mechanism, we propose a novel nonparametric Bayesian method to detect the clustering pattern of multiple-TF bindings from ChIP-seq datasets. Simulation study demonstrates that our method performs best with regard to precision, recall, and F1 score, in comparison to traditional methods. We also apply the method on real data and observe several TF clusters that have been recognized previously in mouse embryonic stem cells. Recent advances in ChIP-seq and RNA sequencing (RNA-Seq) technologies provides more reliable and accurate characterization of TF binding sites and gene expression measurements, which serves as a basis to study the regulatory functions of TFs on gene expression. We propose a log Gaussian cox process with wavelet-based functional model to quantify the relationship between TF binding site locations and gene expression levels. Through the simulation study, we demonstrate that our method performs well, especially with large sample size and small variance. It also shows a remarkable ability to distinguish real local feature in the function estimates.
Doctor of Philosophy
The development of high-throughput next-generation sequencing (NGS) techniques produces massive amount of data and bring out innovations in biology and medicine. A greater concentration is needed in developing novel, powerful, and efficient tools for NGS data analysis. In this dissertation, we mainly focus on three problems closely related to NGS and its applications: (1) how to improve variant calling accuracy, (2) how to model transcription factor (TF) binding patterns, and (3) how to quantify of the contribution of TF binding on gene expression. We develop novel statistical methods to identify sequence variants, find TF binding patterns, and explore the relationship between TF binding and gene expressions. We expect our findings will be helpful in promoting a better understanding of disease causality and facilitating the design of personalized treatments.
APA, Harvard, Vancouver, ISO, and other styles
6

Fang, Zaili. "Some Advanced Model Selection Topics for Nonparametric/Semiparametric Models with High-Dimensional Data." Diss., Virginia Tech, 2012. http://hdl.handle.net/10919/40090.

Full text
Abstract:
Model and variable selection have attracted considerable attention in areas of application where datasets usually contain thousands of variables. Variable selection is a critical step to reduce the dimension of high dimensional data by eliminating irrelevant variables. The general objective of variable selection is not only to obtain a set of cost-effective predictors selected but also to improve prediction and prediction variance. We have made several contributions to this issue through a range of advanced topics: providing a graphical view of Bayesian Variable Selection (BVS), recovering sparsity in multivariate nonparametric models and proposing a testing procedure for evaluating nonlinear interaction effect in a semiparametric model. To address the first topic, we propose a new Bayesian variable selection approach via the graphical model and the Ising model, which we refer to the ``Bayesian Ising Graphical Model'' (BIGM). There are several advantages of our BIGM: it is easy to (1) employ the single-site updating and cluster updating algorithm, both of which are suitable for problems with small sample sizes and a larger number of variables, (2) extend this approach to nonparametric regression models, and (3) incorporate graphical prior information. In the second topic, we propose a Nonnegative Garrote on a Kernel machine (NGK) to recover sparsity of input variables in smoothing functions. We model the smoothing function by a least squares kernel machine and construct a nonnegative garrote on the kernel model as the function of the similarity matrix. An efficient coordinate descent/backfitting algorithm is developed. The third topic involves a specific genetic pathway dataset in which the pathways interact with the environmental variables. We propose a semiparametric method to model the pathway-environment interaction. We then employ a restricted likelihood ratio test and a score test to evaluate the main pathway effect and the pathway-environment interaction.
Ph. D.
APA, Harvard, Vancouver, ISO, and other styles
7

Chu, Shuyu. "Change Detection and Analysis of Data with Heterogeneous Structures." Diss., Virginia Tech, 2017. http://hdl.handle.net/10919/78613.

Full text
Abstract:
Heterogeneous data with different characteristics are ubiquitous in the modern digital world. For example, the observations collected from a process may change on its mean or variance. In numerous applications, data are often of mixed types including both discrete and continuous variables. Heterogeneity also commonly arises in data when underlying models vary across different segments. Besides, the underlying pattern of data may change in different dimensions, such as in time and space. The diversity of heterogeneous data structures makes statistical modeling and analysis challenging. Detection of change-points in heterogeneous data has attracted great attention from a variety of application areas, such as quality control in manufacturing, protest event detection in social science, purchase likelihood prediction in business analytics, and organ state change in the biomedical engineering. However, due to the extraordinary diversity of the heterogeneous data structures and complexity of the underlying dynamic patterns, the change-detection and analysis of such data is quite challenging. This dissertation aims to develop novel statistical modeling methodologies to analyze four types of heterogeneous data and to find change-points efficiently. The proposed approaches have been applied to solve real-world problems and can be potentially applied to a broad range of areas.
Ph. D.
APA, Harvard, Vancouver, ISO, and other styles
8

Lan, Jing. "Gaussian mixture model based system identification and control." [Gainesville, Fla.] : University of Florida, 2006. http://purl.fcla.edu/fcla/etd/UFE0014640.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Vakil, Sam. "Gaussian mixture model based coding of speech and audio." Thesis, McGill University, 2004. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=81575.

Full text
Abstract:
The transmission of speech and audio over communication channels has always required speech and audio coders with reasonable search and computational complexity and good performance relative to the corresponding distortion measure.
This work introduces a coding scheme which works in a perceptual auditory domain. The input high dimensional frames of audio and speech are transformed to power spectral domain, using either DFT or MDCT. The log spectral vectors are then transformed to the excitation domain. In the quantizer section the vectors are DCT transformed and decorrelated. This operation gives the possibility of using diagonal covariances in modelling the data. Finally, a GMM based VQ is performed on the vectors.
In the decoder part the inverse operations are done. However, in order to prevent negative power spectrum elements due to inverse perceptual transformation in the decoder, instead of direct inversion, a Nonnegative Least Squares Algorithm has been used to switch back to frequency domain. For the sake of comparison, a reference subband based "Excitation Distortion coder" is implemented and comparing the resulting coded files showed a better performance for the proposed GMM based coder.
APA, Harvard, Vancouver, ISO, and other styles
10

Sadarangani, Nikhil 1979. "An improved Gaussian mixture model algorithm for background subtraction." Thesis, Massachusetts Institute of Technology, 2002. http://hdl.handle.net/1721.1/87293.

Full text
Abstract:
Thesis (M.Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2002.
Includes bibliographical references (leaves 71-72).
by Nikhil Sadarangani.
M.Eng.
APA, Harvard, Vancouver, ISO, and other styles
11

Stuttle, Matthew Nicholas. "A gaussian mixture model spectral representation for speech recognition." Thesis, University of Cambridge, 2003. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.620077.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Wang, Juan. "Estimation of individual treatment effect via Gaussian mixture model." HKBU Institutional Repository, 2020. https://repository.hkbu.edu.hk/etd_oa/839.

Full text
Abstract:
In this thesis, we investigate the estimation problem of treatment effect from Bayesian perspective through which one can first obtain the posterior distribution of unobserved potential outcome from observed data, and then obtain the posterior distribution of treatment effect. We mainly consider how to represent a joint distribution of two potential outcomes - one from treated group and another from control group, which can give us an indirect impression of correlation, since the estimation of treatment effect depends on correlation between two potential outcomes. The first part of this thesis illustrates the effectiveness of adapting Gaussian mixture models in solving the treatment effect problem. We apply the mixture models - Gaussian Mixture Regression (GMR) and Gaussian Mixture Linear Regression (GMLR)- as a potentially simple and powerful tool to investigate the joint distribution of two potential outcomes. For GMR, we consider a joint distribution of the covariate and two potential outcomes. For GMLR, we consider a joint distribution of two potential outcomes, which linearly depend on covariate. Through developing an EM algorithm for GMLR, we find that GMR and GMLR are effective in estimating means and variances, but they are not effective in capturing correlation between two potential outcomes. In the second part of this thesis, GMLR is modified to capture unobserved covariance structure (correlation between outcomes) that can be explained by latent variables introduced through making an important model assumption. We propose a much more efficient Pre-Post EM Algorithm to implement our proposed GMLR model with unobserved covariance structure in practice. Simulation studies show that Pre-Post EM Algorithm performs well not only in estimating means and variances, but also in estimating covariance.
APA, Harvard, Vancouver, ISO, and other styles
13

Lu, Liang. "Subspace Gaussian mixture models for automatic speech recognition." Thesis, University of Edinburgh, 2013. http://hdl.handle.net/1842/8065.

Full text
Abstract:
In most of state-of-the-art speech recognition systems, Gaussian mixture models (GMMs) are used to model the density of the emitting states in the hidden Markov models (HMMs). In a conventional system, the model parameters of each GMM are estimated directly and independently given the alignment. This results a large number of model parameters to be estimated, and consequently, a large amount of training data is required to fit the model. In addition, different sources of acoustic variability that impact the accuracy of a recogniser such as pronunciation variation, accent, speaker factor and environmental noise are only weakly modelled and factorized by adaptation techniques such as maximum likelihood linear regression (MLLR), maximum a posteriori adaptation (MAP) and vocal tract length normalisation (VTLN). In this thesis, we will discuss an alternative acoustic modelling approach — the subspace Gaussian mixture model (SGMM), which is expected to deal with these two issues better. In an SGMM, the model parameters are derived from low-dimensional model and speaker subspaces that can capture phonetic and speaker correlations. Given these subspaces, only a small number of state-dependent parameters are required to derive the corresponding GMMs. Hence, the total number of model parameters can be reduced, which allows acoustic modelling with a limited amount of training data. In addition, the SGMM-based acoustic model factorizes the phonetic and speaker factors and within this framework, other source of acoustic variability may also be explored. In this thesis, we propose a regularised model estimation for SGMMs, which avoids overtraining in case that the training data is sparse. We will also take advantage of the structure of SGMMs to explore cross-lingual acoustic modelling for low-resource speech recognition. Here, the model subspace is estimated from out-domain data and ported to the target language system. In this case, only the state-dependent parameters need to be estimated which relaxes the requirement of the amount of training data. To improve the robustness of SGMMs against environmental noise, we propose to apply the joint uncertainty decoding (JUD) technique that is shown to be efficient and effective. We will report experimental results on the Wall Street Journal (WSJ) database and GlobalPhone corpora to evaluate the regularisation and cross-lingual modelling of SGMMs. Noise compensation using JUD for SGMM acoustic models is evaluated on the Aurora 4 database.
APA, Harvard, Vancouver, ISO, and other styles
14

Delport, Marion. "A spatial variant of the Gaussian mixture of regressions model." Diss., University of Pretoria, 2017. http://hdl.handle.net/2263/65883.

Full text
Abstract:
In this study the nite mixture of multivariate Gaussian distributions is discussed in detail including the derivation of maximum likelihood estimators, a discussion on identi ability of mixture components as well as a discussion on the singularities typically occurring during the estimation process. Examples demonstrate the application of the nite mixture of univariate and bivariate Gaussian distributions. The nite mixture of multivariate Gaussian regressions is discussed including the derivation of maximum likelihood estimators. An example is used to demonstrate the application of the mixture of regressions model. Two methods of calculating the coe cient of determination for measuring model performance are introduced. The application of nite mixtures of Gaussian distributions and regressions to image segmentation problems is examined. The traditional nite mixture models however, have a shortcoming in that commonality of location of observations (pixels) is not taken into account when clustering the data. In literature, this shortcoming is addressed by including a Markov random eld prior for the mixing probabilities and the present study discusses this theoretical development. The resulting nite spatial variant mixture of Gaussian regressions model is de ned and its application is demonstrated in a simulated example. It was found that the spatial variant mixture of Gaussian regressions delivered accurate spatial clustering results and simultaneously accurately estimated the component model parameters. This study contributes an application of the spatial variant mixture of Gaussian regressions model in the agricultural context: maize yields in the Free State are modelled as a function of precipitation, type of maize and season; GPS coordinates linked to the observations provide the location information. A simple linear regression and traditional mixture of Gaussian regressions model were tted for comparative purposes and the latter identi ed three distinct clusters without accounting for location information. It was found that the application of the spatial variant mixture of regressions model resulted in spatially distinct and informative clusters, especially with respect to the type of maize covariate. However, the estimated component regression models for this data set were quite similar. The investigated data set was not perfectly suited for the spatial variant mixture of regressions model application and possible solutions were proposed to improve the model results in future studies. A key learning from the present study is that the e ectiveness of the spatial variant mixture of regressions model is dependent on the clear and distinguishable spatial dependencies in the underlying data set when it is applied to map-type data.
Dissertation (MSc)--University of Pretoria, 2017.
Statistics
MSc
Unrestricted
APA, Harvard, Vancouver, ISO, and other styles
15

Srinivasan, Balaji Vasan. "Gaussian process regression for model estimation." College Park, Md.: University of Maryland, 2008. http://hdl.handle.net/1903/8962.

Full text
Abstract:
Thesis (M.S.) -- University of Maryland, College Park, 2008.
Thesis research directed by: Dept. of Electrical and Computer Engineering E. Title from t.p. of PDF. Includes bibliographical references. Published by UMI Dissertation Services, Ann Arbor, Mich. Also available in paper.
APA, Harvard, Vancouver, ISO, and other styles
16

Cuesta, Ramirez Jhouben Janyk. "Optimization of a computationally expensive simulator with quantitative and qualitative inputs." Thesis, Lyon, 2022. http://www.theses.fr/2022LYSEM010.

Full text
Abstract:
Dans cette thèse, les problèmes mixtes couteux sont abordés par le biais de processus gaussiens où les variables discrètes sont relaxées en variables latentes continues. L'espace continu est plus facilement exploité par les techniques classiques d'optimisation bayésienne que ne le serait un espace mixte. Les variables discrètes sont récupérées soit après l'optimisation continue, soit simultanément avec une contrainte supplémentaire de compatibilité continue-discrète qui est traitée avec des lagrangiens augmentés. Plusieurs implémentations possibles de ces optimiseurs mixtes bayésiens sont comparées. En particulier, la reformulation du problème avec des variables latentes continues est mise en concurrence avec des recherches travaillant directement dans l'espace mixte. Parmi les algorithmes impliquant des variables latentes et un lagrangien augmenté, une attention particulière est consacrée aux multiplicateurs de lagrange pour lesquels des techniques d'estimation locale et globale sont étudiées. Les comparaisons sont basées sur l'optimisation répétée de trois fonctions analytiques et sur une application mécanique concernant la conception d'une poutre. Une étude supplémentaire pour l'application d'une stratégie d'optimisation mixte proposée dans le domaine de l'auto-calibrage mixte est faite. Cette analyse s'inspire d'une application de quantification des radionucléides, qui définit une fonction inverse spécifique nécessitant l'étude de ses multiples propriétés dans le scenario continu. une proposition de différentes stratégies déterministes et bayésiennes a été faite en vue d'une définition complète dans un contexte de variables mixtes
In this thesis, costly mixed problems are approached through gaussian processes where the discrete variables are relaxed into continuous latent variables. the continuous space is more easily harvested by classical bayesian optimization techniques than a mixed space would. discrete variables are recovered either subsequently to the continuous optimization, or simultaneously with an additional continuous-discrete compatibility constraint that is handled with augmented lagrangians. several possible implementations of such bayesian mixed optimizers are compared. in particular, the reformulation of the problem with continuous latent variables is put in competition with searches working directly in the mixed space. among the algorithms involving latent variables and an augmented lagrangian, a particular attention is devoted to the lagrange multipliers for which a local and a global estimation techniques are studied. the comparisons are based on the repeated optimization of three analytical functions and a mechanical application regarding a beam design. an additional study for applying a proposed mixed optimization strategy in the field of mixed self-calibration is made. this analysis was inspired in an application in radionuclide quantification, which defined an specific inverse function that required the study of its multiple properties in the continuous scenario. a proposition of different deterministic and bayesian strategies was made towards a complete definition in a mixed variable setup
APA, Harvard, Vancouver, ISO, and other styles
17

Tran, Denis. "A study of bit allocation for Gaussian mixture model quantizers and image coders /." Thesis, McGill University, 2005. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=83937.

Full text
Abstract:
This thesis describes different bit allocation schemes and their performances when applied on coding line spectral frequencies (LSF) using the GMM-based coder designed by Subramaniam and a simple image transform coder. The new algorithms are compared to the original bit allocation formula; the Pruning algorithm used by Subramaniam, Segall's method and the Greedy bit allocation algorithm using the Log Spectral Distortion and the Mean-Square Error for the LSF quantizer and the Peak Signal-to-Noise Ratio for the image coder.
First, a Greedy level allocation algorithm is developed based on the philosophy of the Greedy algorithin but, it does so level by level, considering the best benefit and bit cost yielded by an allocation. The Greedy level allocation algorithm is computationally intensive in general, thus we discuss combining it with other algorithms to obtain lower costs.
Second, another algorithm solving problems of negative bit allocations and integer level is proposed. The level allocations are to keep a certain ratio with respect to each other throughout the algorithm in order to remain closest to the condition for lowest distortion. Moreover, the original formula assumes a 6dB gain for each added bit, which is not generally true. The algorithm presents a new parameter k, which controls the benefit of adding one bit, usually set at 0.5 in the high-rate optimal bit allocation formula for MSE calling the new algorithm, the Two-Stage Iterative Bit Allocation (TSIBA) algorithm. Simulations show that modifying the bit allocation formula effectively brings about some gains over the previous methods.
The formula containing the new parameter is generalized into a, formula introducing a new parameter which weights not only the variances but also the dimensions, training the new parameter on their distribution function. The TSIBA was an a-posteriori decision algorithm, where the decision on which value of k to select for lowest distortion was decided after computing all distortions. The Generalized TSIBA (GTSIBA), on the other hand, uses a training procedure to estimate which weighting factor to set for each dimension at a certain bit rate. Simulation results show yet another improvement when using the Generalized TSIBA over all previous methods.
APA, Harvard, Vancouver, ISO, and other styles
18

Seidu, Mohammed Nazib. "Predicting Bankruptcy Risk: A Gaussian Process Classifciation Model." Thesis, Linköpings universitet, Institutionen för datavetenskap, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-119120.

Full text
Abstract:
This thesis develops a Gaussian processes model for bankruptcy risk classification and prediction in a Bayesian framework. Gaussian processes and linear logistic models are discriminative methods used for classification and prediction purposes. The Gaussian processes model is a much more flexible model than the linear logistic model with smoothness encoded in the kernel with the potential to improve the modeling of the highly nonlinear relationships between accounting ratios and bankruptcy risk. We compare the linear logistic regression with the Gaussian process classification model in the context of bankruptcy prediction. The posterior distributions of the GPs are non-Gaussian, and we investigate the effectiveness of the Laplace approximation and the expectation propagation approximation across several different kernels for the Gaussian process. The approximate methods are compared to the gold standard of Markov Chain Monte Carlo (MCMC) sampling from the posterior. The dataset is an unbalanced panel consisting of 21846 yearly observations for about 2000 corporate firms in Sweden recorded between 1991−2008. We used 5000 observations to train the models and the rest for evaluating the predictions. We find that the choice of covariance kernel affects the GP model’s performance and we find support for the squared exponential covariance function (SEXP) as an optimal kernel. The empirical evidence suggests that a multivariate Gaussian processes classifier with squared exponential kernel can effectively improve bankruptcy risk prediction with high accuracy (90.19 percent) compared to the linear logistic model (83.25 percent).
APA, Harvard, Vancouver, ISO, and other styles
19

Adamou, Maria. "Bayesian optimal designs for the Gaussian Process Model." Thesis, University of Southampton, 2014. https://eprints.soton.ac.uk/373881/.

Full text
Abstract:
This thesis is concerned with methodology for finding Bayesian optimal designs for the Gaussian process model when the aim is precise prediction at unobserved points. The fundamental problem addressed is that the design selection criterion obtained from the Bayesian decision theoretic approach is often, in practice, computationally infeasible to apply. We propose an approximation to the objective function in the criterion and develop this approximation for spatial and spatio-temporal studies, and for computer experiments. We provide empirical evidence and theoretical insights to support the approximation. For spatial studies, we use the approximation to find optimal designs for the general sensor placement problem, and also to find the best sensors to remove from an existing monitoring network. We assess the performance of the criterion using a prospective study and also from a retrospective study based on an air pollution dataset. We investigate the robustness of designs to misspecification of the mean function and correlation function in the model through a factorial sensitivity study that compares the performance of optimal designs for the sensor placement problem under different assumptions. In computer experiments, using a Gaussian process model as a surrogate for the output from a computer model, we find optimal designs for prediction using the proposed approximation. A comparison is made of optimal designs obtained from commonly used model-free methods such as the maximin criterion and Latin hypercube sampling via both the space-filling and prediction properties of the designs. For spatio-temporal studies, we extend our proposed approximation to include both space and time dependency and investigate the approximation for a particular choice of separable spatio-temporal correlation function. Two cases are considered: (i) the temporal design is fixed and an optimal spatial design is found; (ii) both optimal temporal and spatial designs are found. For all three of the application areas, we found that the choice of optimal design depends on the degree and the range of the correlation in the Gaussian process model.
APA, Harvard, Vancouver, ISO, and other styles
20

Shashidhar, Sanda, and Amirisetti Sravya. "Online Handwritten Signature Verification System : using Gaussian Mixture Model and Longest Common Sub-Sequences." Thesis, Blekinge Tekniska Högskola, Institutionen för tillämpad signalbehandling, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-15807.

Full text
APA, Harvard, Vancouver, ISO, and other styles
21

Moradiannejad, Ghazaleh. "People Tracking Under Occlusion Using Gaussian Mixture Model and Fast Level Set Energy Minimization." Thèse, Université d'Ottawa / University of Ottawa, 2013. http://hdl.handle.net/10393/24304.

Full text
Abstract:
Tracking multiple articulated objects (such as a human body) and handling occlusion between them is a challenging problem in automated video analysis. This work proposes a new approach for accurately and steadily visual tracking people, which should function even if the system encounters occlusion in video sequences. In this approach, targets are represented with a Gaussian mixture, which are adapted to regions of the target automatically using an EM-model algorithm. Field speeds are defined for changed pixels in each frame based on the probability of their belonging to a particular person's blobs. Pixels are matched to the models using a fast numerical level set method. Since each target is tracked with its blob's information, the system is capable of handling partial or full occlusion during tracking. Experimental results on a number of challenging sequences that were collected in non-experimental environments demonstrate the effectiveness of the approach.
APA, Harvard, Vancouver, ISO, and other styles
22

Webb, Grayson. "A Gaussian Mixture Model based Level Set Method for Volume Segmentation in Medical Images." Thesis, Linköpings universitet, Beräkningsmatematik, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-148548.

Full text
Abstract:
This thesis proposes a probabilistic level set method to be used in segmentation of tumors with heterogeneous intensities. It models the intensities of the tumor and surrounding tissue using Gaussian mixture models. Through a contour based initialization procedure samples are gathered to be used in expectation maximization of the mixture model parameters. The proposed method is compared against a threshold-based segmentation method using MRI images retrieved from The Cancer Imaging Archive. The cases are manually segmented and an automated testing procedure is used to find optimal parameters for the proposed method and then it is tested against the threshold-based method. Segmentation times, dice coefficients, and volume errors are compared. The evaluation reveals that the proposed method has a comparable mean segmentation time to the threshold-based method, and performs faster in cases where the volume error does not exceed 40%. The mean dice coefficient and volume error are also improved while achieving lower deviation.
APA, Harvard, Vancouver, ISO, and other styles
23

Lindström, Kevin. "Fault Clustering With Unsupervised Learning Using a Modified Gaussian Mixture Model and Expectation Maximization." Thesis, Linköpings universitet, Fordonssystem, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-176535.

Full text
Abstract:
When a fault is detected in the engine, the check engine light will come on. After that, it is often up to the mechanic to diagnose the engine fault. Manual fault classification by a mechanic can be time-consuming and expensive. Recent technological advancements have granted us immense computing power, which can be utilized to diagnose faults using data-driven classifiers. Data-driven classifiers generally require a lot of training data to be able to accurately diagnose system faults by comparing sensor data to training data because labeled training data is required for a wide variety of different realizations of the same faults. In this study an algorithm is proposed that does not rely on labeled training data, instead the proposed algorithm clusters similar fault data together by combining an engine model and unsupervised learning in the form of a modified Gaussian mixture model using Expectation Maximization. If one or more of the fault scenarios in a cluster is later diagnosed, the rest of the data in the same cluster is likely to have the same diagnosis. The modified Gaussian mixture model proposed in this study takes into account that residual data, in some cases including the case in this study when the data is from an internal combustion engine, seem to diverge from the nominal case (data points near the origin) along a linear trajectory as the fault size increases. This is taken into account by modeling the clusters as Gaussian distributions around fault vectors that each represent the trajectories the data moves along as the fault size increases for each cluster or fault mode. The algorithm also takes into account that data from one scenario are likely to belong to the same fault class i.e. it is not necessary to classify each data point separately, instead the data can be clustered as batches. This study also evaluates the proposed model as a semi-supervised learner, where some data is known. In this case, the algorithm can also be used to estimate the fault sizes of unknown faults by using the acquired fault vectors, given that there are known fault sizes for other data in the same cluster. The algorithm is evaluated with data collected from an engine test bench using a commercial Volvo engine and shows promising results as most fault scenarios can be correctly clustered. However, results show that there are clustering ambiguities for data from small faults, as they are more similar to the nominal case and overlap more with data from other fault modes.
APA, Harvard, Vancouver, ISO, and other styles
24

Malsiner-Walli, Gertraud, Sylvia Frühwirth-Schnatter, and Bettina Grün. "Model-based clustering based on sparse finite Gaussian mixtures." Springer, 2016. http://dx.doi.org/10.1007/s11222-014-9500-2.

Full text
Abstract:
In the framework of Bayesian model-based clustering based on a finite mixture of Gaussian distributions, we present a joint approach to estimate the number of mixture components and identify cluster-relevant variables simultaneously as well as to obtain an identified model. Our approach consists in specifying sparse hierarchical priors on the mixture weights and component means. In a deliberately overfitting mixture model the sparse prior on the weights empties superfluous components during MCMC. A straightforward estimator for the true number of components is given by the most frequent number of non-empty components visited during MCMC sampling. Specifying a shrinkage prior, namely the normal gamma prior, on the component means leads to improved parameter estimates as well as identification of cluster-relevant variables. After estimating the mixture model using MCMC methods based on data augmentation and Gibbs sampling, an identified model is obtained by relabeling the MCMC output in the point process representation of the draws. This is performed using K-centroids cluster analysis based on the Mahalanobis distance. We evaluate our proposed strategy in a simulation setup with artificial data and by applying it to benchmark data sets. (authors' abstract)
APA, Harvard, Vancouver, ISO, and other styles
25

Fry, James Thomas. "Hierarchical Gaussian Processes for Spatially Dependent Model Selection." Diss., Virginia Tech, 2018. http://hdl.handle.net/10919/84161.

Full text
Abstract:
In this dissertation, we develop a model selection and estimation methodology for nonstationary spatial fields. Large, spatially correlated data often cover a vast geographical area. However, local spatial regions may have different mean and covariance structures. Our methodology accomplishes three goals: (1) cluster locations into small regions with distinct, stationary models, (2) perform Bayesian model selection within each cluster, and (3) correlate the model selection and estimation in nearby clusters. We utilize the Conditional Autoregressive (CAR) model and Ising distribution to provide intra-cluster correlation on the linear effects and model inclusion indicators, while modeling inter-cluster correlation with separate Gaussian processes. We apply our model selection methodology to a dataset involving the prediction of Brook trout presence in subwatersheds across Pennsylvania. We find that our methodology outperforms the stationary spatial model and that different regions in Pennsylvania are governed by separate Gaussian process regression models.
Ph. D.
APA, Harvard, Vancouver, ISO, and other styles
26

Cheng, Nan. "Bayesian Nonparametric Reliability Analysis Using Dirichlet Process Mixture Model." Ohio University / OhioLINK, 2011. http://rave.ohiolink.edu/etdc/view?acc_num=ohiou1307988021.

Full text
APA, Harvard, Vancouver, ISO, and other styles
27

Fergie, Martin Paul. "Discriminative pose estimation using mixtures of Gaussian processes." Thesis, University of Manchester, 2013. https://www.research.manchester.ac.uk/portal/en/theses/discriminative-pose-estimation-using-mixtures-of-gaussian-processes(462e72b8-ab98-4517-9649-4c2b9bec4b04).html.

Full text
Abstract:
This thesis proposes novel algorithms for using Gaussian processes for Discriminative pose estimation. We overcome the traditional limitations of Gaussian processes, their cubic training complexity and their uni-modal predictive distribution by assembling them in a mixture of experts formulation. Our First contribution shows that by creating a large number of Fixed size Gaussian process experts, we can build a model that is able to scale to large data sets and accurately learn the multi-modal and non- linear mapping between image features and the subject’s pose. We demonstrate that this model gives state of the art performance compared to other discriminative pose estimation techniques.We then extend the model to automatically learn the size and location of each expert. Gaussian processes are able to accurately model non-linear functional regression problems where the output is given as a function of the input. However, when an individual Gaussian process is trained on data which contains multi-modalities, or varying levels of ambiguity, the Gaussian process is unable to accurately model the data. We propose a novel algorithm for learning the size and location of each expert in our mixture of Gaussian processes model to ensure that the training data of each expert matches the assumptions of a Gaussian process. We show that this model is able to out perform our previous mixture of Gaussian processes model.Our final contribution is a dynamics framework for inferring a smooth sequence of pose estimates from a sequence of independent predictive distributions. Discriminative pose estimation infers the pose of each frame independently, leading to jittery tracking results. Our novel algorithm uses a model of human dynamics to infer a smooth path through a sequence of Gaussian mixture models as given by our mixture of Gaussian processes model. We show that our algorithm is able to smooth and correct some mis- takes made by the appearance model alone, and outperform a baseline linear dynamical system.
APA, Harvard, Vancouver, ISO, and other styles
28

Yang, Xiaoke. "Fault-tolerant predictive control : a Gaussian process model based approach." Thesis, University of Cambridge, 2015. https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.708784.

Full text
APA, Harvard, Vancouver, ISO, and other styles
29

Kawaguchi, Nobuo, Katsuhiko Kaji, Susumu Fujita, 信夫 河口, 克彦 梶, and 迪. 藤田. "Gaussian Mixture Model を用いた無線LAN位置推定手法." 一般社団法人情報処理学会, 2010. http://hdl.handle.net/2237/15430.

Full text
APA, Harvard, Vancouver, ISO, and other styles
30

Kawaguchi, Nobuo, Katsuhiko Kaji, Susumu Fujita, 信夫 河口, 克彦 梶, and 迪. 藤田. "Gaussian Mixture Model を用いた無線LAN位置推定手法." 一般社団法人情報処理学会, 2011. http://hdl.handle.net/2237/15440.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Dahlqwist, Elisabeth. "Birthweight-specific neonatal health : With application on data from a tertiaryhospital in Tanzania." Thesis, Uppsala universitet, Statistiska institutionen, 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-227531.

Full text
Abstract:
The following study analyzes birthweight-specific neonatal health using a combination of a mixture model and logistic regression: the extended Parametric Mixture of Logistic Regression. The data are collected from the Obstetric database at Muhimbili National Hospital in Dar es Salaam, Tanzania and the years 2009 -2013 are used in the analysis. Due to rounding in the birthweight data a novel method to adjust for rounding when estimating a mixture model is applied. The influence of rounding on the estimates is then investigated. A three-component model is selected. The variables used in the analysis of neonatal health are early neonatal mortality, if the mother has HIV, anaemia, is a private patient and if the neonate is born after 36 completed weeks of gestation. It can be concluded that the mortality rates are high especially for low birthweights (2000 or less) in the estimated first and second components. However, due to wide confidence bounds it is hard to draw conclusions from the data.
APA, Harvard, Vancouver, ISO, and other styles
32

Zhang, Huaiye. "Bayesian Approach Dealing with Mixture Model Problems." Diss., Virginia Tech, 2012. http://hdl.handle.net/10919/37681.

Full text
Abstract:
In this dissertation, we focus on two research topics related to mixture models. The first topic is Adaptive Rejection Metropolis Simulated Annealing for Detecting Global Maximum Regions, and the second topic is Bayesian Model Selection for Nonlinear Mixed Effects Model. In the first topic, we consider a finite mixture model, which is used to fit the data from heterogeneous populations for many applications. An Expectation Maximization (EM) algorithm and Markov Chain Monte Carlo (MCMC) are two popular methods to estimate parameters in a finite mixture model. However, both of the methods may converge to local maximum regions rather than the global maximum when multiple local maxima exist. In this dissertation, we propose a new approach, Adaptive Rejection Metropolis Simulated Annealing (ARMS annealing), to improve the EM algorithm and MCMC methods. Combining simulated annealing (SA) and adaptive rejection metropolis sampling (ARMS), ARMS annealing generate a set of proper starting points which help to reach all possible modes. ARMS uses a piecewise linear envelope function for a proposal distribution. Under the SA framework, we start with a set of proposal distributions, which are constructed by ARMS, and this method finds a set of proper starting points, which help to detect separate modes. We refer to this approach as ARMS annealing. By combining together ARMS annealing with the EM algorithm and with the Bayesian approach, respectively, we have proposed two approaches: an EM ARMS annealing algorithm and a Bayesian ARMS annealing approach. EM ARMS annealing implement the EM algorithm by using a set of starting points proposed by ARMS annealing. ARMS annealing also helps MCMC approaches determine starting points. Both approaches capture the global maximum region and estimate the parameters accurately. An illustrative example uses a survey data on the number of charitable donations. The second topic is related to the nonlinear mixed effects model (NLME). Typically a parametric NLME model requires strong assumptions which make the model less flexible and often are not satisfied in real applications. To allow the NLME model to have more flexible assumptions, we present three semiparametric Bayesian NLME models, constructed with Dirichlet process (DP) priors. Dirichlet process models often refer to an infinite mixture model. We propose a unified approach, the penalized posterior Bayes factor, for the purpose of model comparison. Using simulation studies, we compare the performance of two of the three semiparametric hierarchical Bayesian approaches with that of the parametric Bayesian approach. Simulation results suggest that our penalized posterior Bayes factor is a robust method for comparing hierarchical parametric and semiparametric models. An application to gastric emptying studies is used to demonstrate the advantage of our estimation and evaluation approaches.
Ph. D.
APA, Harvard, Vancouver, ISO, and other styles
33

Su, Weiji. "Flexible Joint Hierarchical Gaussian Process Model for Longitudinal and Recurrent Event Data." University of Cincinnati / OhioLINK, 2020. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1595850414934069.

Full text
APA, Harvard, Vancouver, ISO, and other styles
34

Rezvani, Arany Roushan. "Gaussian Process Model Predictive Control for Autonomous Driving in Safety-Critical Scenarios." Thesis, Linköpings universitet, Reglerteknik, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-161430.

Full text
Abstract:
This thesis is concerned with model predictive control (MPC) within the field of autonomous driving. MPC requires a model of the system to be controlled. Since a vehicle is expected to handle a wide range of driving conditions, it is crucial that the model of the vehicle dynamics is able to account for this. Differences in road grip caused by snowy, icy or muddy roads change the driving dynamics and relying on a single model, based on ideal conditions, could possibly lead to dangerous behaviour. This work investigates the use of Gaussian processes for learning a model that can account for varying road friction coefficients. This model is incorporated as an extension to a nominal vehicle model. A double lane change scenario is considered and the aim is to learn a GP model of the disturbance based on previous driving experiences with a road friction coefficient of 0.4 and 0.6 performed with a regular MPC controller. The data is then used to train a GP model. The GPMPC controller is then compared with the regular MPC controller in the case of trajectory tracking. The results show that the obtained GP models in most cases correctly predict the model error in one prediction step. For multi-step predictions, the results vary more with some cases showing an improved prediction with a GP model compared to the nominal model. In all cases, the GPMPC controller gives a better trajectory tracking than the MPC controller while using less control input.
APA, Harvard, Vancouver, ISO, and other styles
35

Tomashenko, Natalia. "Speaker adaptation of deep neural network acoustic models using Gaussian mixture model framework in automatic speech recognition systems." Thesis, Le Mans, 2017. http://www.theses.fr/2017LEMA1040/document.

Full text
Abstract:
Les différences entre conditions d'apprentissage et conditions de test peuvent considérablement dégrader la qualité des transcriptions produites par un système de reconnaissance automatique de la parole (RAP). L'adaptation est un moyen efficace pour réduire l'inadéquation entre les modèles du système et les données liées à un locuteur ou un canal acoustique particulier. Il existe deux types dominants de modèles acoustiques utilisés en RAP : les modèles de mélanges gaussiens (GMM) et les réseaux de neurones profonds (DNN). L'approche par modèles de Markov cachés (HMM) combinés à des GMM (GMM-HMM) a été l'une des techniques les plus utilisées dans les systèmes de RAP pendant de nombreuses décennies. Plusieurs techniques d'adaptation ont été développées pour ce type de modèles. Les modèles acoustiques combinant HMM et DNN (DNN-HMM) ont récemment permis de grandes avancées et surpassé les modèles GMM-HMM pour diverses tâches de RAP, mais l'adaptation au locuteur reste très difficile pour les modèles DNN-HMM. L'objectif principal de cette thèse est de développer une méthode de transfert efficace des algorithmes d'adaptation des modèles GMM aux modèles DNN. Une nouvelle approche pour l'adaptation au locuteur des modèles acoustiques de type DNN est proposée et étudiée : elle s'appuie sur l'utilisation de fonctions dérivées de GMM comme entrée d'un DNN. La technique proposée fournit un cadre général pour le transfert des algorithmes d'adaptation développés pour les GMM à l'adaptation des DNN. Elle est étudiée pour différents systèmes de RAP à l'état de l'art et s'avère efficace par rapport à d'autres techniques d'adaptation au locuteur, ainsi que complémentaire
Differences between training and testing conditions may significantly degrade recognition accuracy in automatic speech recognition (ASR) systems. Adaptation is an efficient way to reduce the mismatch between models and data from a particular speaker or channel. There are two dominant types of acoustic models (AMs) used in ASR: Gaussian mixture models (GMMs) and deep neural networks (DNNs). The GMM hidden Markov model (GMM-HMM) approach has been one of the most common technique in ASR systems for many decades. Speaker adaptation is very effective for these AMs and various adaptation techniques have been developed for them. On the other hand, DNN-HMM AMs have recently achieved big advances and outperformed GMM-HMM models for various ASR tasks. However, speaker adaptation is still very challenging for these AMs. Many adaptation algorithms that work well for GMMs systems cannot be easily applied to DNNs because of the different nature of these models. The main purpose of this thesis is to develop a method for efficient transfer of adaptation algorithms from the GMM framework to DNN models. A novel approach for speaker adaptation of DNN AMs is proposed and investigated. The idea of this approach is based on using so-called GMM-derived features as input to a DNN. The proposed technique provides a general framework for transferring adaptation algorithms, developed for GMMs, to DNN adaptation. It is explored for various state-of-the-art ASR systems and is shown to be effective in comparison with other speaker adaptation techniques and complementary to them
APA, Harvard, Vancouver, ISO, and other styles
36

Kullmann, Emelie. "Speech to Text for Swedish using KALDI." Thesis, KTH, Optimeringslära och systemteori, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-189890.

Full text
Abstract:
The field of speech recognition has during the last decade left the re- search stage and found its way in to the public market. Most computers and mobile phones sold today support dictation and transcription in a number of chosen languages.  Swedish is often not one of them. In this thesis, which is executed on behalf of the Swedish Radio, an Automatic Speech Recognition model for Swedish is trained and the performance evaluated. The model is built using the open source toolkit Kaldi.  Two approaches of training the acoustic part of the model is investigated. Firstly, using Hidden Markov Model and Gaussian Mixture Models and secondly, using Hidden Markov Models and Deep Neural Networks. The later approach using deep neural networks is found to achieve a better performance in terms of Word Error Rate.
De senaste åren har olika tillämpningar inom människa-dator interaktion och främst taligenkänning hittat sig ut på den allmänna marknaden. Många system och tekniska produkter stöder idag tjänsterna att transkribera tal och diktera text. Detta gäller dock främst de större språken och sällan finns samma stöd för mindre språk som exempelvis svenskan. I detta examensprojekt har en modell för taligenkänning på svenska ut- vecklas. Det är genomfört på uppdrag av Sveriges Radio som skulle ha stor nytta av en fungerande taligenkänningsmodell på svenska. Modellen är utvecklad i ramverket Kaldi. Två tillvägagångssätt för den akustiska träningen av modellen är implementerade och prestandan för dessa två är evaluerade och jämförda. Först tränas en modell med användningen av Hidden Markov Models och Gaussian Mixture Models och slutligen en modell där Hidden Markov Models och Deep Neural Networks an- vänds, det visar sig att den senare uppnår ett bättre resultat i form av måttet Word Error Rate.
APA, Harvard, Vancouver, ISO, and other styles
37

Slifko, Matthew D. "The Cauchy-Net Mixture Model for Clustering with Anomalous Data." Diss., Virginia Tech, 2019. http://hdl.handle.net/10919/93576.

Full text
Abstract:
We live in the data explosion era. The unprecedented amount of data offers a potential wealth of knowledge but also brings about concerns regarding ethical collection and usage. Mistakes stemming from anomalous data have the potential for severe, real-world consequences, such as when building prediction models for housing prices. To combat anomalies, we develop the Cauchy-Net Mixture Model (CNMM). The CNMM is a flexible Bayesian nonparametric tool that employs a mixture between a Dirichlet Process Mixture Model (DPMM) and a Cauchy distributed component, which we call the Cauchy-Net (CN). Each portion of the model offers benefits, as the DPMM eliminates the limitation of requiring a fixed number of a components and the CN captures observations that do not belong to the well-defined components by leveraging its heavy tails. Through isolating the anomalous observations in a single component, we simultaneously identify the observations in the net as warranting further inspection and prevent them from interfering with the formation of the remaining components. The result is a framework that allows for simultaneously clustering observations and making predictions in the face of the anomalous data. We demonstrate the usefulness of the CNMM in a variety of experimental situations and apply the model for predicting housing prices in Fairfax County, Virginia.
Doctor of Philosophy
We live in the data explosion era. The unprecedented amount of data offers a potential wealth of knowledge but also brings about concerns regarding ethical collection and usage. Mistakes stemming from anomalous data have the potential for severe, real-world consequences, such as when building prediction models for housing prices. To combat anomalies, we develop the Cauchy-Net Mixture Model (CNMM). The CNMM is a flexible tool for identifying and isolating the anomalies, while simultaneously discovering cluster structure and making predictions among the nonanomalous observations. The result is a framework that allows for simultaneously clustering and predicting in the face of the anomalous data. We demonstrate the usefulness of the CNMM in a variety of experimental situations and apply the model for predicting housing prices in Fairfax County, Virginia.
APA, Harvard, Vancouver, ISO, and other styles
38

Minh, Tuan Pham, Tomohiro Yoshikawa, Takeshi Furuhashi, and Kaita Tachibana. "Robust feature extractions from geometric data using geometric algebra." IEEE, 2009. http://hdl.handle.net/2237/13896.

Full text
APA, Harvard, Vancouver, ISO, and other styles
39

Yang, Chenguang. "Security in Voice Authentication." Digital WPI, 2014. https://digitalcommons.wpi.edu/etd-dissertations/79.

Full text
Abstract:
We evaluate the security of human voice password databases from an information theoretical point of view. More specifically, we provide a theoretical estimation on the amount of entropy in human voice when processed using the conventional GMM-UBM technologies and the MFCCs as the acoustic features. The theoretical estimation gives rise to a methodology for analyzing the security level in a corpus of human voice. That is, given a database containing speech signals, we provide a method for estimating the relative entropy (Kullback-Leibler divergence) of the database thereby establishing the security level of the speaker verification system. To demonstrate this, we analyze the YOHO database, a corpus of voice samples collected from 138 speakers and show that the amount of entropy extracted is less than 14-bits. We also present a practical attack that succeeds in impersonating the voice of any speaker within the corpus with a 98% success probability with as little as 9 trials. The attack will still succeed with a rate of 62.50% if 4 attempts are permitted. Further, based on the same attack rationale, we mount an attack on the ALIZE speaker verification system. We show through experimentation that the attacker can impersonate any user in the database of 69 people with about 25% success rate with only 5 trials. The success rate can achieve more than 50% by increasing the allowed authentication attempts to 20. Finally, when the practical attack is cast in terms of an entropy metric, we find that the theoretical entropy estimate almost perfectly predicts the success rate of the practical attack, giving further credence to the theoretical model and the associated entropy estimation technique.
APA, Harvard, Vancouver, ISO, and other styles
40

Bekli, Zeid, and William Ouda. "A performance measurement of a Speaker Verification system based on a variance in data collection for Gaussian Mixture Model and Universal Background Model." Thesis, Malmö universitet, Fakulteten för teknik och samhälle (TS), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-20122.

Full text
Abstract:
Voice recognition has become a more focused and researched field in the last century,and new techniques to identify speech has been introduced. A part of voice recognition isspeaker verification which is divided into Front-end and Back-end. The first componentis the front-end or feature extraction where techniques such as Mel-Frequency CepstrumCoefficients (MFCC) is used to extract the speaker specific features of a speech signal,MFCC is mostly used because it is based on the known variations of the humans ear’scritical frequency bandwidth. The second component is the back-end and handles thespeaker modeling. The back-end is based on the Gaussian Mixture Model (GMM) andGaussian Mixture Model-Universal Background Model (GMM-UBM) methods forenrollment and verification of the specific speaker. In addition, normalization techniquessuch as Cepstral Means Subtraction (CMS) and feature warping is also used forrobustness against noise and distortion. In this paper, we are going to build a speakerverification system and experiment with a variance in the amount of training data for thetrue speaker model, and to evaluate the system performance. And further investigate thearea of security in a speaker verification system then two methods are compared (GMMand GMM-UBM) to experiment on which is more secure depending on the amount oftraining data available.This research will therefore give a contribution to how much data is really necessary fora secure system where the False Positive is as close to zero as possible, how will theamount of training data affect the False Negative (FN), and how does this differ betweenGMM and GMM-UBM.The result shows that an increase in speaker specific training data will increase theperformance of the system. However, too much training data has been proven to beunnecessary because the performance of the system will eventually reach its highest point and in this case it was around 48 min of data, and the results also show that the GMMUBM model containing 48- to 60 minutes outperformed the GMM models.
APA, Harvard, Vancouver, ISO, and other styles
41

Zhao, David Yuheng. "Model Based Speech Enhancement and Coding." Doctoral thesis, Stockholm : Kungliga Tekniska högskolan, 2007. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-4412.

Full text
APA, Harvard, Vancouver, ISO, and other styles
42

Bakhtiari, Koohsorkhi Alireza. "Analysis of the Dirichlet Process Mixture Model with Application to Dialogue Act Classification." Thesis, Université Laval, 2011. http://www.theses.ulaval.ca/2011/28503/28503.pdf.

Full text
Abstract:
La reconnaissance des intentions de l’utilisateur est l’un des problèmes les plus difficiles dans la conception des systèmes de dialogues. Ces intentions sont généralement codés en termes d’actes de dialogue, où un rôle fonctionnel est attribué à chaque énoncé d’une conversation. L’annotation manuelle des actes de dialogue est généralement coûteuse et prends du temps, il y a donc un grand intérêt à plutôt annoter automatiquement des corpus de dialogue. Dans ce mémoire, nous proposons une approche non paramétrique bayésienne pour la classification automatique des actes de dialogue. Nous utilisons les mélanges par processus de Dirichlet (DPMM), dans lesquels chacune des composantes est déterminée par une distribution de Dirichlet-multinomial. Deux nouvelles approches pour l’estimation des hyperparamètres dans ces distributions sont introduites. Les résultats de l’application de ce modèle au corpus DIHANA montre que la DPMM peut récupérer le nombre réel d’étiquettes en haute précision.
Recognition of user intentions is one of the most challenging problems in the design of dialogue systems. These intentions are usually coded in terms of Dialogue Acts (Following Austin’s work on speech act theory), where a functional role is assigned to each utterance of a conversation. Manual annotation of dialogue acts is both time consuming and expensive, therefore there is a huge interest in systems which are able to automatically annotate dialogue corpora. In this thesis, we propose a nonparametric Bayesian approach for the automatic classification of dialogue acts. We make use of the Dirichlet Process Mixture Model (DPMM), within which each of the components is governed by a Dirichlet-Multinomial distribution. Two novel approaches for hyperparameter estimation in these distributions are also introduced. Results of the application of this model to the DIHANA corpus shows that the DPMM can successfully recover the true number of DA labels with high precision
APA, Harvard, Vancouver, ISO, and other styles
43

Liu, Xi. "Semi-parametric Bayesian Inference of Accelerated Life Test Using Dirichlet Process Mixture Model." Ohio University / OhioLINK, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=ohiou1447193154.

Full text
APA, Harvard, Vancouver, ISO, and other styles
44

Kim, Sinae. "Bayesian variable selection in clustering via dirichlet process mixture models." Texas A&M University, 2003. http://hdl.handle.net/1969.1/5888.

Full text
Abstract:
The increased collection of high-dimensional data in various fields has raised a strong interest in clustering algorithms and variable selection procedures. In this disserta- tion, I propose a model-based method that addresses the two problems simultane- ously. I use Dirichlet process mixture models to define the cluster structure and to introduce in the model a latent binary vector to identify discriminating variables. I update the variable selection index using a Metropolis algorithm and obtain inference on the cluster structure via a split-merge Markov chain Monte Carlo technique. I evaluate the method on simulated data and illustrate an application with a DNA microarray study. I also show that the methodology can be adapted to the problem of clustering functional high-dimensional data. There I employ wavelet thresholding methods in order to reduce the dimension of the data and to remove noise from the observed curves. I then apply variable selection and sample clustering methods in the wavelet domain. Thus my methodology is wavelet-based and aims at clustering the curves while identifying wavelet coefficients describing discriminating local features. I exemplify the method on high-dimensional and high-frequency tidal volume traces measured under an induced panic attack model in normal humans.
APA, Harvard, Vancouver, ISO, and other styles
45

"ADAPTIVE LEARNING OF NEURAL ACTIVITY DURING DEEP BRAIN STIMULATION." Master's thesis, 2015. http://hdl.handle.net/2286/R.I.29727.

Full text
Abstract:
abstract: Parkinson's disease is a neurodegenerative condition diagnosed on patients with clinical history and motor signs of tremor, rigidity and bradykinesia, and the estimated number of patients living with Parkinson's disease around the world is seven to ten million. Deep brain stimulation (DBS) provides substantial relief of the motor signs of Parkinson's disease patients. It is an advanced surgical technique that is used when drug therapy is no longer sufficient for Parkinson's disease patients. DBS alleviates the motor symptoms of Parkinson's disease by targeting the subthalamic nucleus using high-frequency electrical stimulation. This work proposes a behavior recognition model for patients with Parkinson's disease. In particular, an adaptive learning method is proposed to classify behavioral tasks of Parkinson's disease patients using local field potential and electrocorticography signals that are collected during DBS implantation surgeries. Unique patterns exhibited between these signals in a matched feature space would lead to distinction between motor and language behavioral tasks. Unique features are first extracted from deep brain signals in the time-frequency space using the matching pursuit decomposition algorithm. The Dirichlet process Gaussian mixture model uses the extracted features to cluster the different behavioral signal patterns, without training or any prior information. The performance of the method is then compared with other machine learning methods and the advantages of each method is discussed under different conditions.
Dissertation/Thesis
Masters Thesis Electrical Engineering 2015
APA, Harvard, Vancouver, ISO, and other styles
46

(11073474), Bin Zhang. "Data-driven Uncertainty Analysis in Neural Networks with Applications to Manufacturing Process Monitoring." Thesis, 2021.

Find full text
Abstract:

Artificial neural networks, including deep neural networks, play a central role in data-driven science due to their superior learning capacity and adaptability to different tasks and data structures. However, although quantitative uncertainty analysis is essential for training and deploying reliable data-driven models, the uncertainties in neural networks are often overlooked or underestimated in many studies, mainly due to the lack of a high-fidelity and computationally efficient uncertainty quantification approach. In this work, a novel uncertainty analysis scheme is developed. The Gaussian mixture model is used to characterize the probability distributions of uncertainties in arbitrary forms, which yields higher fidelity than the presumed distribution forms, like Gaussian, when the underlying uncertainty is multimodal, and is more compact and efficient than large-scale Monte Carlo sampling. The fidelity of the Gaussian mixture is refined through adaptive scheduling of the width of each Gaussian component based on the active assessment of the factors that could deteriorate the uncertainty representation quality, such as the nonlinearity of activation functions in the neural network.

Following this idea, an adaptive Gaussian mixture scheme of nonlinear uncertainty propagation is proposed to effectively propagate the probability distributions of uncertainties through layers in deep neural networks or through time in recurrent neural networks. An adaptive Gaussian mixture filter (AGMF) is then designed based on this uncertainty propagation scheme. By approximating the dynamics of a highly nonlinear system with a feedforward neural network, the adaptive Gaussian mixture refinement is applied at both the state prediction and Bayesian update steps to closely track the distribution of unmeasurable states. As a result, this new AGMF exhibits state-of-the-art accuracy with a reasonable computational cost on highly nonlinear state estimation problems subject to high magnitudes of uncertainties. Next, a probabilistic neural network with Gaussian-mixture-distributed parameters (GM-PNN) is developed. The adaptive Gaussian mixture scheme is extended to refine intermediate layer states and ensure the fidelity of both linear and nonlinear transformations within the network so that the predictive distribution of output target can be inferred directly without sampling or approximation of integration. The derivatives of the loss function with respect to all the probabilistic parameters in this network are derived explicitly, and therefore, the GM-PNN can be easily trained with any backpropagation method to address practical data-driven problems subject to uncertainties.

The GM-PNN is applied to two data-driven condition monitoring schemes of manufacturing processes. For tool wear monitoring in the turning process, a systematic feature normalization and selection scheme is proposed for the engineering of optimal feature sets extracted from sensor signals. The predictive tool wear models are established using two methods, one is a type-2 fuzzy network for interval-type uncertainty quantification and the other is the GM-PNN for probabilistic uncertainty quantification. For porosity monitoring in laser additive manufacturing processes, convolutional neural network (CNN) is used to directly learn patterns from melt-pool patterns to predict porosity. The classical CNN models without consideration of uncertainty are compared with the CNN models in which GM-PNN is embedded as an uncertainty quantification module. For both monitoring schemes, experimental results show that the GM-PNN not only achieves higher prediction accuracies of process conditions than the classical models but also provides more effective uncertainty quantification to facilitate the process-level decision-making in the manufacturing environment.

Based on the developed uncertainty analysis methods and their proven successes in practical applications, some directions for future studies are suggested. Closed-loop control systems may be synthesized by combining the AGMF with data-driven controller design. The AGMF can also be extended from a state estimator to the parameter estimation problems in data-driven models. In addition, the GM-PNN scheme may be expanded to directly build more complicated models like convolutional or recurrent neural networks.

APA, Harvard, Vancouver, ISO, and other styles
47

Lai, Chu-Shiuan, and 賴竹煖. "Gaussian Mixture of Background and Shadow Model." Thesis, 2010. http://ndltd.ncl.edu.tw/handle/38760050190895130218.

Full text
Abstract:
碩士
國立臺灣師範大學
資訊工程研究所
98
In this paper, we integrate shadow information into the background model of a scene in an attempt to detect both shadows and foreground objects at a time. Since shadows accompanying foreground objects are viewed as parts of the foreground objects, shadows will be extracted as well during foreground object detection. Shadows can distort object shapes and may connect multiple objects into one object. On the other hand, shadows tell the directions of light sources. In other words, shadows can be advantageous as well as disadvantageous. To begin, we use an adaptive Gaussian mixture model to describe the background of a scene. Based on this preliminary background model, we extract foreground objects and their accompanying shadows. Shadows are next separated from foreground objects through a series of intensity and color analyses. The characteristics of shadows are finally determined with the principal component analysis method and are embedded as an additional Gaussian in the background model. Experimental results demonstrated the feasibility of the proposed background model. Keywords: Dynamic scene, Adaptive Gaussian Mixture Model, Foreground detection, Shadow detection
APA, Harvard, Vancouver, ISO, and other styles
48

Sue, Yung-Chun, and 蘇詠鈞. "Specified Gestures Identification using Gaussian Mixture Model." Thesis, 2012. http://ndltd.ncl.edu.tw/handle/15219540833691164661.

Full text
Abstract:
碩士
清雲科技大學
電子工程所
100
Sign language recognition technique is composed by the hand images detection and the hand gestures recognition. Hand images detection is locating the sign language select, sign language capture, the palm and fingers part from the sensed image, and rotating them to the appropriate hand posture, both are the important pre-processing for sign language identification and recognition. This paper first introduced sequentially throughout the study practices, as well as the process of image pre-processing instructions. The major work in the hand gestures recognition is to identify the variance of the fingers. In this paper the creation of sign language image of slash encoding, Department of the advantages of slash encoding the difference between your fingers the number of changes, and the Gaussian mixture model (GMM) to establish the model of sign language and identification. Such as poor recognition rate is adjusted probability distribution of weight values to improve the recognition rate. The entire the paper Shushing is the Gaussian mixture model (GMM), slash code, adjust the probability distribution of the weight value. Finally, after adjusting the probability distribution of weight values, we learned from the conclusion that the overall recognition results rose to 98.33%from 92.66% of the original, so changing the probability distribution of the weight value can effectively improve the recognition rate.
APA, Harvard, Vancouver, ISO, and other styles
49

莊清乾. "Automatic Bird Songs Recognition using Gaussian Mixture Model." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/09774268339453426682.

Full text
Abstract:
碩士
中華大學
資訊工程學系(所)
96
In this paper, Gaussian mixture models (GMM) were applied to identify bird species from their sounds. First, each syllable corresponding to a piece of vocalization is manually segmented. Two-dimension MFCC (TDMFCC), dynamic two-dimension MFCC (DTDMFCC), and normalized audio spectrum envelope (NASE) modulation coefficients are calculated for each syllable and regarded as the vocalization features of each syllable. Principal component analysis (PCA) is used to reduce the feature space dimension of the original input features vector space. GMM is used to cluster the feature vectors from the same bird species into several groups with each group represented by a Gaussian distribution. The self-splitting Gaussian mixture learning (SGML) algorithm is then employed to find an appropriate number of Gaussian components for each GMM. In addition, a model selection algorithm based on the Bayesian information criterion (BIC) is applied to select the optimal model between GMM and extended VQ (EVQ) according to the amount of training data available. Linear discriminant analysis (LDA) is finally exploited to increase the classification accuracy at a lower dimensional feature vector space. In our experiments, the combination of TDMFCC, DTDMFCC, and NASE modulation coefficients achieve the average classification accuracy of 83.9% for the classification of 28 bird species.
APA, Harvard, Vancouver, ISO, and other styles
50

LIN, YU-JUNG, and 林昱融. "Modified Gaussian Mixture Model Applied to Speaker Verification." Thesis, 2016. http://ndltd.ncl.edu.tw/handle/33cbau.

Full text
Abstract:
碩士
中央警察大學
刑事警察研究所
104
Gaussian mixture model (GMM) is a combination of a plurality of Gaussian probability density function, it can be smoothly approximate the probability density distribution of any arbitrary shape. In various areas of pattern recognition, it has a good recognition results. However, during building the speaker model process, we must determine the parameters of each Gaussian probability density function through constantly iterative calculation, the calculation process is quite complex. This paper presents modified Gaussian mixture model, each characteristic for recognition has its own independent Gaussian probability density function. Since the process without iteration, it can significantly reduce the amount of calculation. And the speaker verification results show that it can still maintain a good recognition results. In this paper, we use Mel frequency cepstral coefficients(MFCCs) as the voice characteristic for speaker verification. The average error rate for speaker verification on Gaussian mixture model is 0.5901%, while it on modified Gaussian mixture model is 1.6700%, the gap between them was 1.0799%. The error rate of two methods during the speaker verification has less difference. But in the speaker model build process, modified Gaussian mixture model does not need to go through an iterative calculation. The calculation ways and time are more simple and faster than Gaussian mixture model. It can also be another consideration of algorithm for more speed and convenience.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography