Dissertations / Theses on the topic 'Rare event probability'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 19 dissertations / theses for your research on the topic 'Rare event probability.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Drozdenko, Myroslav. "Weak Convergence of First-Rare-Event Times for Semi-Markov Processes." Doctoral thesis, Västerås : Mälardalen University, 2007. http://urn.kb.se/resolve?urn=urn:nbn:se:mdh:diva-394.
Razaaly, Nassim. "Rare Event Estimation and Robust Optimization Methods with Application to ORC Turbine Cascade." Thesis, Université Paris-Saclay (ComUE), 2019. http://www.theses.fr/2019SACLX027.
This thesis aims to formulate innovative Uncertainty Quantification (UQ) methods in both Robust Optimization (RO) and Reliability-Based Design Optimization (RBDO) problems. The targeted application is the optimization of supersonic turbines used in Organic Rankine Cycle (ORC) power systems.Typical energy sources for ORC power systems feature variable heat load and turbine inlet/outlet thermodynamic conditions. The use of organic compounds with a heavy molecular weight typically leads to supersonic turbine configurations featuring supersonic flows and shocks, which grow in relevance in the aforementioned off-design conditions; these features also depend strongly on the local blade shape, which can be influenced by the geometric tolerances of the blade manufacturing. A consensus exists about the necessity to include these uncertainties in the design process, so requiring fast UQ methods and a comprehensive tool for performing shape optimization efficiently.This work is decomposed in two main parts. The first one addresses the problem of rare events estimation, proposing two original methods for failure probability (metaAL-OIS and eAK-MCS) and one for quantile computation (QeAK-MCS). The three methods rely on surrogate-based (Kriging) adaptive strategies, aiming at refining the so-called Limit-State Surface (LSS) directly, unlike Subset Simulation (SS) derived methods. Indeed, the latter consider intermediate threshold associated with intermediate LSSs to be refined. This direct refinement property is of crucial importance since it enables the adaptability of the developed methods for RBDO algorithms. Note that the proposed algorithms are not subject to restrictive assumptions on the LSS (unlike the well-known FORM/SORM), such as the number of failure modes, however need to be formulated in the Standard Space. The eAK-MCS and QeAK-MCS methods are derived from the AK-MCS method and inherit a parallel adaptive sampling based on weighed K-Means. MetaAL-OIS features a more elaborate sequential refinement strategy based on MCMC samples drawn from a quasi-optimal ISD. It additionally proposes the construction of a Gaussian mixture ISD, permitting the accurate estimation of small failure probabilities when a large number of evaluations (several millions) is tractable, as an alternative to SS. The three methods are shown to perform very well for 2D to 8D analytical examples popular in structural reliability literature, some featuring several failure modes, all subject to very small failure probability/quantile level. Accurate estimations are performed in the cases considered using a reasonable number of calls to the performance function.The second part of this work tackles original Robust Optimization (RO) methods applied to the Shape Design of a supersonic ORC Turbine cascade. A comprehensive Uncertainty Quantification (UQ) analysis accounting for operational, fluid parameters and geometric (aleatoric) uncertainties is illustrated, permitting to provide a general overview over the impact of multiple effects and constitutes a preliminary study necessary for RO. Then, several mono-objective RO formulations under a probabilistic constraint are considered in this work, including the minimization of the mean or a high quantile of the Objective Function. A critical assessment of the (Robust) Optimal designs is finally investigated
Sinks, Shuxian. "Response Adaptive Design using Auxiliary and Primary Outcomes." VCU Scholars Compass, 2013. http://scholarscompass.vcu.edu/etd/572.
Chabridon, Vincent. "Analyse de sensibilité fiabiliste avec prise en compte d'incertitudes sur le modèle probabiliste - Application aux systèmes aérospatiaux." Thesis, Université Clermont Auvergne (2017-2020), 2018. http://www.theses.fr/2018CLFAC054/document.
Aerospace systems are complex engineering systems for which reliability has to be guaranteed at an early design phase, especially regarding the potential tremendous damage and costs that could be induced by any failure. Moreover, the management of various sources of uncertainties, either impacting the behavior of systems (“aleatory” uncertainty due to natural variability of physical phenomena) and/or their modeling and simulation (“epistemic” uncertainty due to lack of knowledge and modeling choices) is a cornerstone for reliability assessment of those systems. Thus, uncertainty quantification and its underlying methodology consists in several phases. Firstly, one needs to model and propagate uncertainties through the computer model which is considered as a “black-box”. Secondly, a relevant quantity of interest regarding the goal of the study, e.g., a failure probability here, has to be estimated. For highly-safe systems, the failure probability which is sought is very low and may be costly-to-estimate. Thirdly, a sensitivity analysis of the quantity of interest can be set up in order to better identify and rank the influential sources of uncertainties in input. Therefore, the probabilistic modeling of input variables (epistemic uncertainty) might strongly influence the value of the failure probability estimate obtained during the reliability analysis. A deeper investigation about the robustness of the probability estimate regarding such a type of uncertainty has to be conducted. This thesis addresses the problem of taking probabilistic modeling uncertainty of the stochastic inputs into account. Within the probabilistic framework, a “bi-level” input uncertainty has to be modeled and propagated all along the different steps of the uncertainty quantification methodology. In this thesis, the uncertainties are modeled within a Bayesian framework in which the lack of knowledge about the distribution parameters is characterized by the choice of a prior probability density function. During a first phase, after the propagation of the bi-level input uncertainty, the predictive failure probability is estimated and used as the current reliability measure instead of the standard failure probability. Then, during a second phase, a local reliability-oriented sensitivity analysis based on the use of score functions is achieved to study the impact of hyper-parameterization of the prior on the predictive failure probability estimate. Finally, in a last step, a global reliability-oriented sensitivity analysis based on Sobol indices on the indicator function adapted to the bi-level input uncertainty is proposed. All the proposed methodologies are tested and challenged on a representative industrial aerospace test-case simulating the fallout of an expendable space launcher
Krauth, Timothé. "Modèle génératif profond pour l'estimation de probabilité de collision en vol." Electronic Thesis or Diss., Toulouse, ISAE, 2024. http://www.theses.fr/2024ESAE0018.
It is essential to calculate the probability of aircraft collisions to optimise air traffic while maintaining high safety standards. This need became more pronounced in the 1960s with the increase in transatlantic commercial air traffic. Initially, analytical models such as those of Reich and Anderson-Hsu were benchmarks for assessing in-flight collision risks, but they proved to be less suited for the complex airspace around airports.Data-driven methods, especially Monte Carlo simulations, have become a promising alternative for collision risk assessment. They offer significant flexibility through simplified assumptions, making them adaptable to various contexts. However, traditional Monte Carlo simulations are inefficient for estimating rare event probabilities, requiring a large number of aircraft trajectories and substantial computational resources. This thesis proposes a collision risk model based on Monte Carlo simulations, using a trajectory generation model to overcome these limitations associated with rare events. These generative methods faithfully reproduce observed trajectory distributions while incorporating uncertainties from external factors. Three main research areas are defined: (i) developing a trajectory generation method, (ii) constructing a Monte Carlo-based collision risk model using synthetic trajectories, and (iii) improving the interpretability of collision risk estimates.Generating synthetic samples involves estimating the distribution of observed data to ensure identical distribution in new samples. This is particularly important for aircraft trajectories, where the model must reflect uncertainty sources causing deviations from standard trajectories. We initially use traditional statistical learning methods to estimate complex two-dimensional aircraft trajectories. Despite reducing the problem's dimensionality, conventional methods struggle with high-dimensional distribution estimation. We then explore the use of variational autoencoders for more refined probability density estimation. Suitably adapted for multivariate time-series applications, variational autoencoders prove effective for estimating the distribution of complex aircraft trajectories.Using the developed generation method, we estimate the risk of loss of separation induced by the departure and approach procedures of Paris-Orly Airport using Monte Carlo simulations. The use of a trajectory generation method proves promising, allowing the creation of the equivalent of 20 years of air traffic trajectories from only two months of observations. However, this direct method has limitations for estimating extremely low collision probabilities, requiring the use of one variational autoencoder per flight procedure considered in the studied scenario. The processes of trajectory generation and collision risk evaluation are distinctly separated. Consequently, the inherent constraints of classical Monte Carlo methods are not truly overcome but merely postponed by the production of a set of arbitrarily large trajectories.The thesis's final work unifies the frameworks of variational autoencoders and uncertainty quantification. It demonstrates how variational autoencoders can build suitable input distributions for uncertainty quantification algorithms, enhancing the reliability of Monte Carlo simulations through subset simulation and the explainability of mid-air collision probability estimation through sensitivity analysis. More broadly, we show that the variational autoencoder represents a promising tool to be associated with uncertainty quantification problems
Jacquemart, Damien. "Contributions aux méthodes de branchement multi-niveaux pour les évènements rares, et applications au trafic aérien." Thesis, Rennes 1, 2014. http://www.theses.fr/2014REN1S186/document.
The thesis deals with the design and mathematical analysis of reliable and accurate Monte Carlo methods in order to estimate the (very small) probability that a Markov process reaches a critical region of the state space before a deterministic final time. The underlying idea behind the multilevel splitting methods studied here is to design an embedded sequence of intermediate more and more critical regions, in such a way that reaching an intermediate region, given that the previous intermediate region has already been reached, is not so rare. In practice, trajectories are propagated, selected and replicated as soon as the next intermediate region is reached, and it is easy to accurately estimate the transition probability between two successive intermediate regions. The bias due to time discretization of the Markov process trajectories is corrected using perturbed intermediate regions as proposed by Gobet and Menozzi. An adaptive version would consist in the automatic design of the intermediate regions, using empirical quantiles. However, it is often difficult if not impossible to remember where (in which state) and when (at which time instant) did each successful trajectory reach the empirically defined intermediate region. The contribution of the thesis consists in using a first population of pilot trajectories to define the next threshold, in using a second population of trajectories to estimate the probability of exceeding this empirically defined threshold, and in iterating these two steps (definition of the next threshold, and evaluation of the transition probability) until the critical region is reached. The convergence of this adaptive two-step algorithm is studied in the asymptotic framework of a large number of trajectories. Ideally, the intermediate regions should be defined in terms of the spatial and temporal variables jointly (for example, as the set of states and times for which a scalar function of the state exceeds a time-dependent threshold). The alternate point of view proposed in the thesis is to keep intermediate regions as simple as possible, defined in terms of the spatial variable only, and to make sure that trajectories that manage to exceed a threshold at an early time instant are more replicated than trajectories that exceed the same threshold at a later time instant. The resulting algorithm combines importance sampling and multilevel splitting. Its preformance is evaluated in the asymptotic framework of a large number of trajectories, and in particular a central limit theorem is obtained for the relative approximation error
Yu, Xiaomin. "Simulation Study of Sequential Probability Ratio Test (SPRT) in Monitoring an Event Rate." University of Cincinnati / OhioLINK, 2009. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1244562576.
DHAMODARAN, RAMYA. "EFFICIENT ANALYSIS OF RARE EVENTS ASSOCIATED WITH INDIVIDUAL BUFFERS IN A TANDEM JACKSON NETWORK." University of Cincinnati / OhioLINK, 2004. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1099073321.
Estecahandy, Maïder. "Méthodes accélérées de Monte-Carlo pour la simulation d'événements rares. Applications aux Réseaux de Petri." Thesis, Pau, 2016. http://www.theses.fr/2016PAUU3008/document.
The dependability analysis of safety instrumented systems is an important industrial concern. To be able to carry out such safety studies, TOTAL develops since the eighties the dependability software GRIF. To take into account the increasing complexity of the operating context of its safety equipment, TOTAL is more frequently led to use the engine MOCA-RP of the GRIF Simulation package. Indeed, MOCA-RP allows to estimate quantities associated with complex aging systems modeled in Petri nets thanks to the standard Monte Carlo (MC) simulation. Nevertheless, deriving accurate estimators, such as the system unavailability, on very reliable systems involves rare event simulation, which requires very long computing times with MC. In order to address this issue, the common fast Monte Carlo methods do not seem to be appropriate. Many of them are originally defined to improve only the estimate of the unreliability and/or well-suited for Markovian processes. Therefore, the work accomplished in this thesis pertains to the development of acceleration methods adapted to the problematic of performing safety studies modeled in Petri nets and estimating in particular the unavailability. More specifically, we propose the Extension of the "Méthode de Conditionnement Temporel" to accelerate the individual failure of the components, and we introduce the Dissociation Method as well as the Truncated Fixed Effort Method to increase the occurrence of their simultaneous failures. Then, we combine the first technique with the two other ones, and we also associate them with the Randomized Quasi-Monte Carlo method. Through different sensitivities studies and benchmark experiments, we assess the performance of the acceleration methods and observe a significant improvement of the results compared with MC. Furthermore, we discuss the choice of the confidence interval method to be used when considering rare event simulation, which is an unfamiliar topic in the field of dependability. Last, an application to an industrial case permits the illustration of the potential of our solution methodology
Mattrand, Cécile. "Approche probabiliste de la tolérance aux dommages." Phd thesis, Université Blaise Pascal - Clermont-Ferrand II, 2011. http://tel.archives-ouvertes.fr/tel-00738947.
Lelièvre, Nicolas. "Développement des méthodes AK pour l'analyse de fiabilité. Focus sur les évènements rares et la grande dimension." Thesis, Université Clermont Auvergne (2017-2020), 2018. http://www.theses.fr/2018CLFAC045/document.
Engineers increasingly use numerical model to replace the experimentations during the design of new products. With the increase of computer performance and numerical power, these models are more and more complex and time-consuming for a better representation of reality. In practice, optimization is very challenging when considering real mechanical problems since they exhibit uncertainties. Reliability is an interesting metric of the failure risks of design products due to uncertainties. The estimation of this metric, the failure probability, requires a high number of evaluations of the time-consuming model and thus becomes intractable in practice. To deal with this problem, surrogate modeling is used here and more specifically AK-based methods to enable the approximation of the physical model with much fewer time-consuming evaluations. The first objective of this thesis work is to discuss the mathematical formulations of design problems under uncertainties. This formulation has a considerable impact on the solution identified by the optimization during design process of new products. A definition of both concepts of reliability and robustness is also proposed. These works are presented in a publication in the international journal: Structural and Multidisciplinary Optimization (Lelièvre, et al. 2016). The second objective of this thesis is to propose a new AK-based method to estimate failure probabilities associated with rare events. This new method, named AK-MCSi, presents three enhancements of AK-MCS: (i) sequential Monte Carlo simulations to reduce the time associated with the evaluation of the surrogate model, (ii) a new stricter stopping criterion on learning evaluations to ensure the good classification of the Monte Carlo population and (iii) a multipoints enrichment permitting the parallelization of the evaluation of the time-consuming model. This work has been published in Structural Safety (Lelièvre, et al. 2018). The last objective of this thesis is to propose new AK-based methods to estimate the failure probability of a high-dimensional reliability problem, i.e. a problem defined by both a time-consuming model and a high number of input random variables. Two new methods, AK-HDMR1 and AK-PCA, are proposed to deal with this problem based on respectively a functional decomposition and a dimensional reduction technique. AK-HDMR1 has been submitted to Reliability Enginnering and Structural Safety on 1st October 2018
Shao, Jun. "Calcul de probabilités d'événements rares liés aux maxima en horizon fini de processus stochastiques." Thesis, Clermont-Ferrand 2, 2016. http://www.theses.fr/2016CLF22771/document.
Initiated within the framework of an ANR project (the MODNAT project) targeted on the stochastic modeling of natural hazards and the probabilistic quantification of their dynamic effects on mechanical and structural systems, this thesis aims at the calculation of probabilities of rare events related to the maxima of stochastic processes over a finite time interval, taking into account the following four constraints : (1) the set of considered processes must contain the four main categories of processes encountered in random dynamics, namely stationary Gaussian, non-stationary Gaussian, stationary non-Gaussian and non-stationary non-Gaussian ones ; (2) these processes can be either described by their distributions, or functions of processes described by their distributions, or solutions of stochastic differential equations, or solutions of stochastic differential inclusions ; (3) the events in question are crossings of high thresholds by the maxima of the considered processes over finite time intervals and these events are of very weak occurrence, hence of very small probability, due to the high size of thresholds ; and finally (4) the use of a Monte Carlo approach to perform this type of calculation must be proscribed because it is too time-consuming given the above constraints. To solve such a problem, whose field of interest extends well beyond probabilistic mechanics and structural reliability (it is found in all scientific domains in connection with the extreme values theory, such as financial mathematics or economical sciences), an innovative method is proposed, whose main idea emerged from the analysis of the results of a large-scale statistical study carried out within the MODNAT project. This study, which focuses on analyzing the behavior of the extreme values of elements of a large set of processes, has indeed revealed two germ functions explicitly related to the target probability (the first directly related, the second indirectly via a conditional auxiliary probability which itself depend on the target probability) which possess remarkable and recurring regularity properties for all the processes of the database, and the method is based on the joint exploitation of these properties and a "low level approximation-high level extrapolation" principle. Two versions of this method are first proposed, which are distinguished by the choice of the germ function and in each of which the latter is approximated by a polynomial. A third version has also been developed. It is based on the formalism of the second version but which uses as germ function an approximation of "Pareto survival function" type. The numerous presented numerical results attest to the remarkable effectiveness of the first two versions. They also show that they are of comparable precision. The third version, slightly less efficient than the first two, presents the interest of establishing a direct link with the extreme values theory. In each of its three versions, the proposed method is clearly an improvement compared to current methods dedicated to this type of problem. Thanks to its structure, it also offers the advantage of remaining operational in industrial context
Wilson, James Adams. "A New Volcanic Event Recurrence Rate Model and Code For Estimating Uncertainty in Recurrence Rate and Volume Flux Through Time With Selected Examples." Scholar Commons, 2016. http://scholarcommons.usf.edu/etd/6435.
Sallin, Mathieu. "Approche probabiliste du diagnostic de l'état de santé des véhicules militaires terrestres en environnement incertain." Thesis, Université Clermont Auvergne (2017-2020), 2018. http://www.theses.fr/2018CLFAC099.
This thesis is a contribution to the structural health analysis of the body of ground military vehicles. Belonging to the 20 - 30 tons range, such vehicles are deployed in a variety of operational contexts where driving conditions are severe and difficult to characterize. In addition, due to a growing industrial competition, the mobility function of vehicles is acquired from suppliers and is no longer developed by Nexter Systems. As a result, the complete definition of this function is unknown. Based on this context, the main objective of this thesis is to analyze the health of the vehicle body using a probabilistic approach in order to control the calculation techniques allowing to take into account the random nature of loads related to the use of ground military vehicles. In particular, the most relevant strategies for propagating uncertainties due to the terrain within a vehicle dynamics model are defined. This work describes how it is possible to manage an observation data measured in the vehicle for the purpose of assessing the reliability with respect to a given damage criterion. An application on a demonstrator entirely designed by Nexter Systems illustrates the proposed approach
Pham, Quang Khoai. "Estimation non paramétrique adaptative dans la théorie des valeurs extrêmes : application en environnement." Thesis, Lorient, 2015. http://www.theses.fr/2015LORIS361/document.
The objective of this PhD thesis is to develop statistical methods based on the theory of extreme values to estimate the probabilities of rare events and conditional extreme quantiles. We consider independent random variables $X_{t_1},…,X_{t_n}$ associated to a sequence of times $0 ≤t_1 <… < t_n ≤ T_{\max}$ where $X_{t_i}$ has distribution function $F_{t_i}$ and $F_t$ is the conditional distribution of $X$ given $T = t \in [0,T_{\max}]$. For each $ t \in [0, T {\max}]$, we propose a nonparametric adaptive estimator for extreme quantiles of $F_t$. The idea of our approach is to adjust the tail of the distribution function $F_t$ with a Pareto distribution of parameter $\theta {t,\tau}$ starting from a threshold $\tau$. The parameter $\theta {t,\tau}$ is estimated using a nonparametric kernel estimator of bandwidth $h$ based on the observations larger than $\tau$. We propose a sequence testing based procedure for the choice of the threshold $\tau$ and we determine the bandwidth $h$ by two methods: cross validation and an adaptive procedure. Under some regularity assumptions, we prove that the adaptive estimator of $\theta {t, \tau}$ is consistent and we determine its rate of convergence. We also propose a method to choose simultaneously the threshold $\tau$ and the bandwidth $h$. Finally, we study the proposed procedures by simulation and on real data set to contribute to the survey of aquatic systems
Dubourg, Vincent. "Méta-modèles adaptatifs pour l'analyse de fiabilité et l'optimisation sous contrainte fiabiliste." Phd thesis, Université Blaise Pascal - Clermont-Ferrand II, 2011. http://tel.archives-ouvertes.fr/tel-00697026.
Lagnoux, Agnès. "Analyse des modeles de branchement avec duplication des trajectoires pour l'étude des événements rares." Toulouse 3, 2006. http://www.theses.fr/2006TOU30231.
This thesis deals with the splitting method first introduced in rare event analysis in order to speed-up simulation. In this technique, the sample paths are split into R multiple copies at various stages during the simulation. Given the cost, the optimization of the algorithm suggests to take the transition probabilities between stages equal to some constant and to resample the inverse of that constant subtrials, which may be non-integer and even unknown but estimated. First, we study the sensitivity of the relative error between the probability of interest P(A) and its estimator depending on the strategy that makes the resampling numbers integers. Then, since in practice the transition probabilities are generally unknown (and so the optimal resampling umbers), we propose a two-steps algorithm to face that problem. Several numerical applications and comparisons with other models are proposed
Lejarraga, Tomás. "Decisions from experience: Time delays, complexity and illusions of control." Doctoral thesis, Universitat Pompeu Fabra, 2009. http://hdl.handle.net/10803/7395.
This thesis includes three chapters that study different aspects of the distinction between decisions from description and decisions from experience. Chapter 1 studies choice when decision makers have both information from description and information from experience. Results suggest that experience is disregarded in the face of description. Individual differences with respect to rational ability are also explored. Participants with higher rational ability draw larger samples than participants with lower rational ability. Chapter 2 examines situations in which information from experience is a better source of information than information from description. Complex scenarios and delayed judgmental tasks favor experience over description as source of information. Moreover, there were no individual differences due to numerical/rational abilities. Additional evidence was found that relates higher rational ability to larger samples.Finally,chapter 3 explores how illusion of control interacts with the source of information in a lottery task.
Ben, Issaid Chaouki. "Effcient Monte Carlo Simulations for the Estimation of Rare Events Probabilities in Wireless Communication Systems." Diss., 2019. http://hdl.handle.net/10754/660001.