Дисертації з теми "POSTERIORI ALGORITHM"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся з топ-39 дисертацій для дослідження на тему "POSTERIORI ALGORITHM".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Переглядайте дисертації для різних дисциплін та оформлюйте правильно вашу бібліографію.
Ghoumari, Asmaa. "Métaheuristiques adaptatives d'optimisation continue basées sur des méthodes d'apprentissage." Thesis, Paris Est, 2018. http://www.theses.fr/2018PESC1114/document.
Повний текст джерелаThe problems of continuous optimization are numerous, in economics, in signal processing, in neural networks, and so on. One of the best-known and most widely used solutions is the evolutionary algorithm, a metaheuristic algorithm based on evolutionary theories that borrows stochastic mechanisms and has shown good performance in solving problems of continuous optimization. The use of this family of algorithms is very popular, despite the many difficulties that can be encountered in their design. Indeed, these algorithms have several parameters to adjust and a lot of operators to set according to the problems to solve. In the literature, we find a plethora of operators described, and it becomes complicated for the user to know which one to select in order to have the best possible result. In this context, this thesis has the main objective to propose methods to solve the problems raised without deteriorating the performance of these algorithms. Thus we propose two algorithms:- a method based on the maximum a posteriori that uses diversity probabilities for the operators to apply, and which puts this choice regularly in play,- a method based on a dynamic graph of operators representing the probabilities of transitions between operators, and relying on a model of the objective function built by a neural network to regularly update these probabilities. These two methods are detailed, as well as analyzed via a continuous optimization benchmark
Moon, Kyoung-Sook. "Adaptive Algorithms for Deterministic and Stochastic Differential Equations." Doctoral thesis, KTH, Numerical Analysis and Computer Science, NADA, 2003. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-3586.
Повний текст джерелаBekkouche, Fatiha. "Étude théorique et numérique des équations non-linéaires de Sobolev." Thesis, Valenciennes, 2018. http://www.theses.fr/2018VALE0018/document.
Повний текст джерелаThe purpose of this work is the mathematical study and the numerical analysis of the nonlinear Sobolev problem. A first chapter is devoted to the a priori analysis for the Sobolev problem, where we use an explicit semidiscretization in time. A priori error estimates were obtained ensuring that the used numerical schemes converge when the time step discretization and the spatial step discretization tend to zero. In a second chapter, we are interested in the singularly perturbed Sobolev problem. For the stability of numerical schemes, we used in this part implicit semidiscretizations in time (the Euler method and the Crank-Nicolson method). Our estimates of Chapters 1 and 2 are confirmed in the third chapter by some numerical experiments. In the last chapter, we consider a Sobolev equation and we derive a posteriori error estimates for the discretization of this equation by a conforming finite element method in space and an implicit Euler scheme in time. The upper bound is global in space and time and allows effective control of the global error. At the end of the chapter, we propose an adaptive algorithm which ensures the control of the total error with respect to a user-defined relative precision by refining the meshes adaptively, equilibrating the time and space contributions of the error. We also present numerical experiments
Giacomini, Matteo. "Quantitative a posteriori error estimators in Finite Element-based shape optimization." Thesis, Université Paris-Saclay (ComUE), 2016. http://www.theses.fr/2016SACLX070/document.
Повний текст джерелаGradient-based shape optimization strategies rely on the computation of the so-called shape gradient. In many applications, the objective functional depends both on the shape of the domain and on the solution of a PDE which can only be solved approximately (e.g. via the Finite Element Method). Hence, the direction computed using the discretized shape gradient may not be a genuine descent direction for the objective functional. This Ph.D. thesis is devoted to the construction of a certification procedure to validate the descent direction in gradient-based shape optimization methods using a posteriori estimators of the error due to the Finite Element approximation of the shape gradient.By means of a goal-oriented procedure, we derive a fully computable certified upper bound of the aforementioned error. The resulting Certified Descent Algorithm (CDA) for shape optimization is able to identify a genuine descent direction at each iteration and features a reliable stopping criterion basedon the norm of the shape gradient.Two main applications are tackled in the thesis. First, we consider the scalar inverse identification problem of Electrical Impedance Tomography and we investigate several a posteriori estimators. A first procedure is inspired by the complementary energy principle and involves the solution of additionalglobal problems. In order to reduce the computational cost of the certification step, an estimator which depends solely on local quantities is derived via an equilibrated fluxes approach. The estimators are validated for a two-dimensional case and some numerical simulations are presented to test the discussed methods. A second application focuses on the vectorial problem of optimal design of elastic structures. Within this framework, we derive the volumetric expression of the shape gradient of the compliance using both H 1 -based and dual mixed variational formulations of the linear elasticity equation. Some preliminary numerical tests are performed to minimize the compliance under a volume constraint in 2D using the Boundary Variation Algorithm and an a posteriori estimator of the error in the shape gradient is obtained via the complementary energy principle
Chalhoub, Nancy. "Estimations a posteriori pour l'équation de convection-diffusion-réaction instationnaire et applications aux volumes finis." Phd thesis, Université Paris-Est, 2012. http://pastel.archives-ouvertes.fr/pastel-00794392.
Повний текст джерелаMoon, Kyoung-Sook. "Convergence rates of adaptive algorithms for deterministic and stochastic differential equations." Licentiate thesis, KTH, Numerical Analysis and Computer Science, NADA, 2001. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-1382.
Повний текст джерелаSánchez, Góez Sebastián. "Algoritmo de reconstrucción analítico para el escáner basado en cristales monolíticos MINDView." Doctoral thesis, Universitat Politècnica de València, 2021. http://hdl.handle.net/10251/159259.
Повний текст джерела[CAT] La tomografia per emissió de positrons és una tècnica de medicina nuclear en la qual es genera una imatge a partir de la detecció de raigs gamma en coincidència. Aquests raigs són produïts dins d'un pacient a què se li injecta una radiotraçador emissor de positrons, els quals s'aniquilen amb electrons de l'medi circumdant. El procés de adquición d'esdeveniments d'interacció, té com a unitat central el detector de l'escàner PET, el qual es compon al seu torn d'un vidre de centelleig, encarregat de transformar els raigs gamma incidents en fotons òptics dins el vidre. La finalitat és llavors, determinar les coordenades d'impacte dins el vidre de centelleig amb la major precisió possible, perquè, a partir d'aquests punts, es pugui reconstruir una imatge. Al llarg de la història, els detectors basats en cristalls pixelats han representat l'elecció per excellència per a la la fabricació d'escàners PET. En aquesta tesi s'avalua l'impacte en la resolució espacial de l'escàner PET MINDView, desenvolupat dins el setè programa Marc de la Unió Europea No 603.002, el qual es basa en l'ús de vidres monolítics. L'ús de vidres monolítics, facilita la determinació de la profunditat d'interacció dels raigs gamma incidents, augmenta la precisió en les coordenades d'impacte determinades, i disminueix l'error de parallaxi que s'indueix en cristalls pixelats, a causa de la dificultat per determinar la DOI. En aquesta tesi, hem aconseguit dos objectius principals relacionats amb el mesurament de la resolució espacial de l'escàner MINDView: l'adaptació de l'un algoritme de STIR de Retroprojecció Filtrada en 3D a un escàner basat en cristalls monolítics i la implementació d'un algoritme de Retroprojecció i filtrat a posteriori. Pel que fa a l'adaptació de l'algoritme FBP3DRP, les resolucions espacials obtingudes varien en els intervals [2 mm, 3,4 mm], [2,3 mm, 3,3 mm] i [2,2 mm, 2,3 mm] per les direccions radial, tangencial i axial, respectivament, en el primer prototip de l'escàner MINDView dedicat a cervell. D'altra banda, en la implementació de l'algoritme de tipus BPF, es va realitzar una adquisició d'un maniquí de derenzo i es va comparar la resolució obtinguda amb l'algorisme de FBP3DRP i una implementació de l'algoritme de subconjunts ordenats en mode llista (LMOS - de l'anglès List Mode Ordered Subset). Mitjançant l'algoritme de tipus BPF es van obtenir valors pic-vall de 2.4 al llarg dels cilindres de l'maniquí de 1.6 mm de diàmetre, en contrast amb les mesures obtingudes de 1.34 i 1.44 per als algoritmes de FBP3DRP i LMOS, respectivament. L'anterior es tradueix en que, mitjançant l'algoritme de tipus BPF, s'aconsegueix millorar la resolució per obtenir-se un valor mitjà 1.6 mm.
[EN] Positron Emission Tomography (PET) is a medical imaging technique, in which an image is generated from the detection of gamma rays in coincidence. These rays are produced within a patient, who is injected with a positron emmiter radiotracer, from which positrons are annihilated with electrons in the media. The event acquisition process is focused on the scanner detector. The detector is in turn composed of a scintillation crystal, which transform the incident ray gamma into optical photons within the crystal. The purpose is then to determine the impact coordinates within the scintillation crystal with the greatest possible precision, so that, from these points, an image can be reconstructed. Throughout history, detectors based on pixelated crystals have represented the quintessential choice for PET scanners manufacture. This thesis evaluates the impact on the spatial resolution of the MINDView PET scanner, developed in the seventh Framework program of the European Union No. 603002, which detectors are based on monolithic crystals. The use of monolithic crystals facilitates the determination of the depth of interaction (DOI - Depth Of Interaction) of the incident gamma rays, increases the precision in the determined impact coordinates, and reduces the parallax error induces in pixelated crystals, due to the difficulties in determining DOI. In this thesis, we have achieved two main goals related to the measurement of the spatial resolution of the MINDView PET scanner: the adaptation of an STIR algorithm for Filtered BackProjection 3D Reproyected (FBP3DRP) to a scanner based on monolithic crystals, and the implementation of a BackProjection then Filtered algorithm (BPF). Regarding the FBP algorithm adaptation, we achieved resolutions ranging in the intervals [2 mm, 3.4 mm], [2.3 mm, 3.3 mm] and [2.2 mm, 2.3 mm] for the radial, tangential and axial directions, respectively. On the an acquisition of a derenzo phantom was performed to measure the spacial resolution, which was obtained using three reconstruction algorithms: the BPF-type algorithm, the FBP3DRP algorithm and an implementation of the list-mode ordered subsets algorithm (LMOS). Regarding the BPF-type algorithm, a peak-to-valley value of 2.4 were obtain along rod of 1.6 mm, in contrast to the measurements of 1.34 and 1.44 obtained for the FBP3DRP and LMOS algorithms, respectively. This means that, by means of the BPF-type algorithm, it is possible to improve the resolution to obtain an average value of 1.6 mm.
Sánchez Góez, S. (2020). Algoritmo de reconstrucción analítico para el escáner basado en cristales monolíticos MINDView [Tesis doctoral]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/159259
TESIS
Hu, Ying. "Maximum a posteriori estimation algorithms for image segmentation and restoration." Thesis, University of Essex, 1992. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.317698.
Повний текст джерелаRenaud, Gabriel. "Bayesian maximum a posteriori algorithms for modern and ancient DNA." Doctoral thesis, Universitätsbibliothek Leipzig, 2016. http://nbn-resolving.de/urn:nbn:de:bsz:15-qucosa-195705.
Повний текст джерелаGrosman, Sergey. "Adaptivity in anisotropic finite element calculations." Doctoral thesis, Universitätsbibliothek Chemnitz, 2006. http://nbn-resolving.de/urn:nbn:de:swb:ch1-200600815.
Повний текст джерелаZoubert-Ousseni, Kersane. "Algorithmes de géolocalisation à l’intérieur d’un bâtiment en temps différé." Thesis, Rennes 1, 2018. http://www.theses.fr/2018REN1S015/document.
Повний текст джерелаReal time indoor geolocalization has recently been widely studied, and has many applications. Off-line (post-processing) trajectory estimation also presents some interest. Off-line indoor geolocalization makes it possible for instance to develop crowdsourcing approaches that take advantage of a large number of users to collect a large number of measurements: knowing the trajectory of a smartphone user makes it possible for instance to feed an attendance map. Estimating this trajectory does not need to be performed in real-time and can be performed off-line, two main benefits. Firstly, the real-time approach estimates a current position using present and past measurements only, when the off-line approach has access to the whole measurements, and makes it possible to obtain an estimated trajectory that is smoother and more accurate than with a real-time approach. Secondly, this estimation can be done on a server and does not need to be implemented in the smartphone as it is the case in the real-time approach, with the consequence that more computing power and size memory are available. The objective of this PhD is to provide an off-line estimation of the trajectory of a smartphone user receiving signal strength (RSS) of wifi or bluetooth measurements and collecting inertial measurements (IMU). In the beginning, without the floorplan of the building, a parametric model is proposed, based on an adaptive pathloss model for RSS measurements and on a piecewise parametrization for the inertial trajectory, obtained with IMU measurements. Results are an average error of 6.2mfor the off-line estimation against 12.5m for the real-time estimation. Then, information on displacement constraints induced by the walls is considered, that makes it possible to adjust the estimated trajectory by using a particle technique as often done in the state-of-the-art. With this second approach we developped a particle smoother and a maximum a posteriori estimator using the Viterbi algorithm. Other numerical heuristics have been introduced. A first heuristic makes use of the parametric model developed without the floorplan to adjust the state model of the user which was originally based on IMUalone. A second heuristic proposes to performseveral realization of a particle filter and to define two score functions based on RSS and on the continuity of the estimated trajectory. The scores are then used to select the best realization of the particle filter as the estimated trajectory. A global algorithm, which uses all of the aforementioned approaches, leads to an error of 3.6m against 5.8m in real-time. Lastly, a statistical machine learning model produced with random forests makes it possible to distinguish the correct estimated trajectories by only using few variables to be used in a crowdsourcing framework
Riedlbeck, Rita. "Adaptive algorithms for poromechanics and poroplasticity." Thesis, Montpellier, 2017. http://www.theses.fr/2017MONTS055/document.
Повний текст джерелаIn this Ph.D. thesis we develop equilibrated flux a posteriori error estimates for poro-mechanical and poro-plasticity problems.Based on these estimations we propose adaptive algorithms for the numerical solution of problems in soil mechanics.The first chapter deals with linear poro-elasticity problems.Using equilibrated $H({rm div})$-conforming flux reconstructions of the Darcy velocity and the mechanical stress tensor, we obtain a guaranteed upper bound on the error.We apply this estimate in an adaptive algorithm balancing the space and time discretisation error components in simulations in two space dimensions.The main contribution of this chapter is the symmetric reconstruction of the stress tensor.In the second chapter we propose another reconstruction technique for the stress tensor, while considering nonlinear elasticity problems.By imposing the symmetry of the tensor only weakly, we reduce computation time and simplify the implementation.We prove that the estimate obtained using this stress reconstuction is locally and globally efficient for a wide range of hyperelasticity problems.We add a linearization error estimator, enabling us to introduce adaptive stopping criteria for the linearization solver.The third chapter adresses the industrial application of the obtained results.We apply an adaptive algorithm to three-dimensional poro-mechanical problems involving elasto-plastic mechanical behavior laws
Dakroub, Jad. "Analyse a posteriori d'algorithmes itératifs pour des problèmes non linéaires." Thesis, Paris 6, 2014. http://www.theses.fr/2014PA066259/document.
Повний текст джерелаThe numerical resolution of any discretization of nonlinear PDEs most often requires an iterative algorithm. In general, the discretization of partial differential equations leads to large systems. As the resolution of large systems is very costly in terms of computation time, an important question arises. To obtain an approximate solution of good quality, when is it necessary to stop the iteration in order to avoid unnecessary iterations? A posteriori error indicators have been studied in recent years owing to their remarkable capacity to enhance both speed and accuracy in computing. This thesis deals with a posteriori error estimation for the finite element discretization of nonlinear problems. Our purpose is to apply a new method that allows us to reduce the number of iterations of the resolution system while keeping a good accuracy of the numerical method. In other words, our goal is to apply a new method that provides a remarkable gain in computation time. For a given nonlinear equation we propose a finite element discretization relying on the Galerkin method. We solve the discrete problem using two iterative methods involving some kind of linearization. For each of them, there are actually two sources of error, namely discretization and linearization. Balancing these two errors can be very important, since it avoids performing an excessive number of iterations. Our results lead to the construction of computable upper indicators for the full error. Similarly, we apply this approach to the Navier-Stokes equations. Several numerical tests are provided to evaluate the efficiency of our indicators
Zhang, Li. "Automated segmentation and analysis of layers and structures of human posterior eye." Diss., University of Iowa, 2015. https://ir.uiowa.edu/etd/2030.
Повний текст джерелаNasr, Imen. "Algorithmes et Bornes minimales pour la Synchronisation Temporelle à Haute Performance : Application à l’internet des objets corporels." Thesis, Université Paris-Saclay (ComUE), 2017. http://www.theses.fr/2017SACLY007/document.
Повний текст джерелаTime synchronization is the first function performed by the demodulator. It ensures that the samples transmitted to the demodulation processes allow to achieve the lowest bit error rate.In this thesis we propose the study of innovative algorithms for high performance time synchronization.First, we propose algorithms exploiting the soft information from the decoder in addition to the received signal to improve the blind estimation of the time delay. Next, we develop an original algorithm based on low complexity smoothing synchronization techniques. This step consisted in proposing a technique operating in an off-line context, making it possible to estimate a random delay that varies over time on several iterations via Forward- Backward loops. The performance of such estimators exceeds that of traditional algorithms. In order to evaluate the relevance of all the proposed estimators, for deterministic and random delays, we evaluated and compared their performance to Cramer-Rao bounds that we developed within these frameworks. We, finally, evaluated the proposed algorithms on WBAN signals
Renaud, Gabriel [Verfasser], Janet [Akademischer Betreuer] Kelso, and Peter [Gutachter] Stadler. "Bayesian maximum a posteriori algorithms for modern and ancient DNA / Gabriel Renaud ; Gutachter: Peter Stadler ; Betreuer: Janet Kelso." Leipzig : Universitätsbibliothek Leipzig, 2016. http://d-nb.info/1240317271/34.
Повний текст джерелаNazarov, Murtazo. "Adaptive Algorithms and High Order Stabilization for Finite Element Computation of Turbulent Compressible Flow." Doctoral thesis, KTH, Numerisk analys, NA, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-34532.
Повний текст джерелаQC 20110627
Johansson, August. "Duality-based adaptive finite element methods with application to time-dependent problems." Doctoral thesis, Umeå : Institutionen för matematik och matematisk statistik, Umeå universitet, 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-33872.
Повний текст джерелаMurphy, Steven. "Methods for solving discontinuous-Galerkin finite element equations with application to neutron transport." Phd thesis, Toulouse, INPT, 2015. http://oatao.univ-toulouse.fr/14650/1/murphy.pdf.
Повний текст джерелаPereira, Filipe de Oliveira. "Separação cega de misturas com não-linearidade posterior utilizando estruturas monotônicas e algoritmos bio-inspirados de otimização." [s.n.], 2010. http://repositorio.unicamp.br/jspui/handle/REPOSIP/259842.
Повний текст джерелаDissertação (mestrado) - Universidade Estadual de Campinas, Faculdade de Engenharia Elétrica e de Computação
Made available in DSpace on 2018-08-16T19:27:38Z (GMT). No. of bitstreams: 1 Pereira_FilipedeOliveira_M.pdf: 3292959 bytes, checksum: b07b4141d2a1f443eb3ab766909a099c (MD5) Previous issue date: 2010
Resumo: O presente trabalho se propõe a desenvolver métodos de Separação Cega de Fontes (BSS) para modelos de mistura com Não-Linearidade Posterior (PNL). Neste caso particular, a despeito da não-linearidade do modelo, ainda é possível recuperar as fontes através de técnicas de Análise de Componentes Independentes (ICA). No entanto, há duas dificuldades maiores no emprego da ICA em modelos PNL. A primeira delas diz respeito a uma restrição sobre as funções não-lineares presentes no modelo PNL: elas devem ser monotônicas por construção. O segundo problema se encontra no ajuste do sistema separador com base em funções custo associadas à ICA: pode haver mínimos locais sub-ótimos. De modo a contornar o primeiro problema, investigamos a adequabilidade de três tipos distintos de estruturas não-lineares monotônicas. Para lidar com a presença de mínimos sub-ótimos no ajuste do sistema separador, empregamos algoritmos bio-inspirados com significativa capacidade de busca global. Finalmente, buscamos, através de experimentos em diversos cenários representativos, identificar dentre as estratégias estudadas qual a melhor configuração, tanto em termos de qualidade da estimação das fontes quanto em termos de complexidade
Abstract: This work aims at the development of Blind Source Separation (BSS) methods for Post-NonLinear (PNL) mixing models. In this particular case, despite the presence of nonlinear elements in the mixing model, it is still possible to recover the sources through Independent Component Analysis (ICA) methods. However, there are two major problems in the application of ICA techniques to PNL models. The first one concerns a restriction on the nonlinear functions present in the PNL model: they must be monotonic functions by construction. The second one is related to the adjustment of the PNL separating system via ICA-based cost functions: there may be sub-optimal local minima. To cope with the first problem, we investigate three types of monotonic nonlinear structures. Moreover, to circumvent the problem related to the presence of sub-optimal minima, we consider bio-inspired algorithms that have a significant global search potential. Finally, we perform a set of experiments in representative scenarios in order to identify, among the considered strategies, the best ones in terms of quality of the retrieved sources and overall complexity
Mestrado
Mestre em Engenharia Elétrica
Harroue, Benjamin. "Approche bayésienne pour la sélection de modèles : application à la restauration d’image." Thesis, Bordeaux, 2020. http://www.theses.fr/2020BORD0127.
Повний текст джерелаInversing main goal is about reconstructing objects from data. Here, we focus on the special case of image restauration in convolution problems. The data are acquired through a altering observation system and additionnaly distorted by errors. The problem becomes ill-posed due to the loss of information. One way to tackle it is to exploit Bayesian approach in order to regularize the problem. Introducing prior information about the unknown quantities osset the loss, and it relies on stochastic models. We have to test all the candidate models, in order to select the best one. But some questions remain : how do you choose the best model? Which features or quantities should we rely on ? In this work, we propose a method to automatically compare and choose the model, based on Bayesion decision theory : objectively compare the models based on their posterior probabilities. These probabilities directly depend on the marginal likelihood or “evidence” of the models. The evidence comes from the marginalization of the jointe law according to the unknow image and the unknow hyperparameters. This a difficult integral calculation because of the complex dependancies between the quantities and the high dimension of the image. That way, we have to work with computationnal methods and approximations. There are several methods on the test stand as Harmonic Mean, Laplace method, discrete integration, Chib from Gibbs approximation or the power posteriors. Comparing is those methods is significative step to determine which ones are the most competent in image restauration. As a first lead of research, we focus on the family of Gaussian models with circulant covariance matrices to lower some difficulties
Fontana, Ilaria. "Interface problems for dam modeling." Thesis, Université de Montpellier (2022-….), 2022. http://www.theses.fr/2022UMONS020.
Повний текст джерелаEngineering teams often use finite element numerical simulations for the design, study and analysis of the behavior of large hydraulic structures. For concrete structures, models of increasing complexity must be able to take into account the nonlinear behavior of discontinuities at the various interfaces located in the foundation, in the body of the dam or at the interface between structure and foundation. Besides representing the nonlinear mechanical behavior of these interfaces (rupture, sliding, contact), one should also be able to take into account the hydraulic flow through these openings.In this thesis, we first focus on the topic of interface behavior modeling, which we address through the Cohesive Zone Model (CZM). This model was introduced in various finite element codes (with the joint elements), and it is a relevant approach to describe the physics of cracking and friction problems at the geometrical discontinuities level. Although initially the CZM was introduced to take into account the phenomenon of rupture, we show in this thesis that it can be extended to sliding problems by possibly relying on the elasto-plastic formalism coupled to the damage. In addition, nonlinear hydro-mechanical constitutive relations can be introduced to model the notion of crack opening and the coupling with the laws of fluid flow. At the mechanical level, we work in the Standard Generalized Materials (SGM) framework, which provides a class of models automatically satisfying some thermodynamical principles, while having good mathematical and numerical properties that are useful for robust numerical modeling. We adapt the formalism of volumetric SGM to the interface zones description. In this first part of the thesis, we present our developpements under the hypothesis of SGM adapted to CZM, capable of reproducing the physical phenomena observed experimentally: rupture, friction, adhesion.In practice, nonlinearities of behavior of interface zones are dominated by the presence of contact, which generates significant numerical difficulties for the convergence of finite element computations. The development of efficient numerical methods for the contact problem is thus a key stage for achieving the goal of robust industrial numerical simulators. Recently, the weak enforcement of contact conditions à la Nitsche has been proposed as a mean to reduce numerical complexity. This technique displays several advantages, among which the most important for our work are: 1) it can handle a wide range of conditions (slip with or without friction, no interpenetration, etc.); 2) it lends itself for a rigorous a posteriori error analysis. This scheme based on the weak contact conditions represents in this work the starting point for the a posteriori error estimation via equilibrated stress reconstruction. This analysis is then used to estimate the different error components (e.g., spatial, nonlinear), and to develop an adaptive resolution algorithm, as well as stopping criteria for iterative solvers and the automatic tuning of possible numerical parameters.The main goal of this thesis is thus to make the finite element numerical simulation of structures with geometrical discontinuities robust. We address this question from two angles: on one side, we revisit the existing methods for the crack representation working on the mechanical constitutive relation for joints; on the other, we introduce a new a posteriori method for the contact problem and we propose its adaptation for the generic interface models
Cooper, Jonathan Paul. "Automatic validation and optimisation of biological models." Thesis, University of Oxford, 2009. http://ora.ox.ac.uk/objects/uuid:24b96d62-b47c-458d-9dff-79b27dbdc9f2.
Повний текст джерелаBéchet, Clémentine. "Commande optimale rapide pour l'optique adaptative des futurs télescopes hectométriques." Phd thesis, Ecole Centrale de Lyon, 2008. http://tel.archives-ouvertes.fr/tel-00363940.
Повний текст джерелаYousef, Soleiman. "Etude d'estimations d'erreur a posteriori et d'adaptivité basée sur des critères d'arrêt et raffinement de maillages pour des problèmes d'écoulements multiphasiques et thermiques. Application aux procédés de récupération assistée d'huile." Phd thesis, Université Pierre et Marie Curie - Paris VI, 2013. http://tel.archives-ouvertes.fr/tel-00918782.
Повний текст джерелаTremblay, Marie. "Estimation des paramètres des modèles de culture : application au modèle STICS Tournesol." Toulouse 3, 2004. http://www.theses.fr/2004TOU30020.
Повний текст джерелаCeleste, Francis. "Cartes incertaines et planification optimale pour la localisation d'un engin autonome." Phd thesis, Université Européenne de Bretagne, 2010. http://tel.archives-ouvertes.fr/tel-00483272.
Повний текст джерелаPankajakshan, Praveen. "Déconvolution Aveugle en Imagerie de Microscopie Confocale À Balayage Laser." Phd thesis, Université de Nice Sophia-Antipolis, 2009. http://tel.archives-ouvertes.fr/tel-00474264.
Повний текст джерелаOzerov, Alexey. "Adaptation de modèles statistiques pour la séparation de sources mono-capteur : application à la séparation voix / musique dans les chansons." Phd thesis, Rennes 1, 2006. https://tel.archives-ouvertes.fr/tel-00564866.
Повний текст джерелаBontemps, Dominique. "Statistiques discrètes et Statistiques bayésiennes en grande dimension." Phd thesis, Université Paris Sud - Paris XI, 2010. http://tel.archives-ouvertes.fr/tel-00561749.
Повний текст джерелаKao, Chih-wei, and 高志偉. "Circuit Design of Maximum a Posteriori Algorithm for Turbo Code Decoder." Thesis, 2010. http://ndltd.ncl.edu.tw/handle/17949025353187667484.
Повний текст джерелаChen, Jiunn-Yann, and 陳俊諺. "Low-Complexity Symmetry-Constrained Maximum-A-Posteriori Probability Algorithm for Adaptive Blind Beamforming." Thesis, 2013. http://ndltd.ncl.edu.tw/handle/09482189312002402759.
Повний текст джерела國立中央大學
通訊工程學系
101
In this paper, we propose a real-valued SC-MAP (RSC-MAP) algorithm for concurrent adaptive filter (CAF) applied to beamforming. We first contribute to deriving a closed-form optimal weight expression for blind MAP algorithm. A conjugate symmetric property associated with optimal blind MAP weights is further acquired. Then, we use the conjugate symmetric constraint to guide the proposed RSC-MAP algorithms to follow the optimal blind MAP expression form during adapting procedure. In the simulations, we show that the proposed RSC-MAP algorithms have better performance than the classic ones. Compared with SC-MAP, the RSC-MAP with less computational complexity has the same bit-error rate performance.
CHANDRA, MUDITA. "PERFORMANCE ANALYSIS OF TURBO CODES IN FADING ENVIRONMENTS." Thesis, 2016. http://dspace.dtu.ac.in:8080/jspui/handle/repository/14972.
Повний текст джерелаChen, Chih-Cheng, and 陳志權. "Application and Complexity Reduction of "Adaptive Maximum A Posteriori Symbol Detection Algorithm" Under Multipath Fading Channel." Thesis, 1998. http://ndltd.ncl.edu.tw/handle/67862919205169266594.
Повний текст джерелаHossain, Md Nurtaj. "Adaptive reduced order modeling of dynamical systems through novel a posteriori error estimators : Application to uncertainty quantification." Thesis, 2021. https://etd.iisc.ac.in/handle/2005/5218.
Повний текст джерелаSamarasinghe, Devanarayanage Pradeepa. "Efficient methodologies for real-time image restoration." Phd thesis, 2011. http://hdl.handle.net/1885/9859.
Повний текст джерелаVicente, Sergio. "Apprentissage statistique avec le processus ponctuel déterminantal." Thesis, 2021. http://hdl.handle.net/1866/25249.
Повний текст джерелаThis thesis presents the determinantal point process, a probabilistic model that captures repulsion between points of a certain space. This repulsion is encompassed by a similarity matrix, the kernel matrix, which selects which points are more similar and then less likely to appear in the same subset. This point process gives more weight to subsets characterized by a larger diversity of its elements, which is not the case with the traditional uniform random sampling. Diversity has become a key concept in domains such as medicine, sociology, forensic sciences and behavioral sciences. The determinantal point process is considered a promising alternative to traditional sampling methods, since it takes into account the diversity of selected elements. It is already actively used in machine learning as a subset selection method. Its application in statistics is illustrated with three papers. The first paper presents the consensus clustering, which consists in running a clustering algorithm on the same data, a large number of times. To sample the initials points of the algorithm, we propose the determinantal point process as a sampling method instead of a uniform random sampling and show that the former option produces better clustering results. The second paper extends the methodology developed in the first paper to large-data. Such datasets impose a computational burden since sampling with the determinantal point process is based on the spectral decomposition of the large kernel matrix. We introduce two methods to deal with this issue. These methods also produce better clustering results than consensus clustering based on a uniform sampling of initial points. The third paper addresses the problem of variable selection for the linear model and the logistic regression, when the number of predictors is large. A Bayesian approach is adopted, using Markov Chain Monte Carlo methods with Metropolis-Hasting algorithm. We show that setting the determinantal point process as the prior distribution for the model space selects a better final model than the model selected by a uniform prior on the model space.
Liu, Hui. "On Regularized Newton-type Algorithms and A Posteriori Error Estimates for Solving Ill-posed Inverse Problems." 2015. http://scholarworks.gsu.edu/math_diss/25.
Повний текст джерелаSingh, Inderjeet 1978. "Risk-averse periodic preventive maintenance optimization." Thesis, 2011. http://hdl.handle.net/2152/ETD-UT-2011-08-4203.
Повний текст джерелаtext