Dissertations / Theses on the topic 'Functional expansion methods'

To see the other types of publications on this topic, follow the link: Functional expansion methods.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 18 dissertations / theses for your research on the topic 'Functional expansion methods.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Correia, Fagner Cintra [UNESP]. "The standard model effective field theory: integrating UV models via functional methods." Universidade Estadual Paulista (UNESP), 2017. http://hdl.handle.net/11449/151703.

Full text
Abstract:
Submitted by FAGNER CINTRA CORREIA null (ccorreia@ift.unesp.br) on 2017-09-24T14:11:35Z No. of bitstreams: 1 Correia_TeseIFT.pdf: 861574 bytes, checksum: 1829fcb0903e20303312d37d7c1e0ffc (MD5)
Approved for entry into archive by Monique Sasaki (sayumi_sasaki@hotmail.com) on 2017-09-27T19:37:47Z (GMT) No. of bitstreams: 1 correia_fc_dr_ift.pdf: 861574 bytes, checksum: 1829fcb0903e20303312d37d7c1e0ffc (MD5)
Made available in DSpace on 2017-09-27T19:37:47Z (GMT). No. of bitstreams: 1 correia_fc_dr_ift.pdf: 861574 bytes, checksum: 1829fcb0903e20303312d37d7c1e0ffc (MD5) Previous issue date: 2017-07-27
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
O Modelo Padrão Efetivo é apresentado como um método consistente de parametrizar Física Nova. Os conceitos de Matching e Power Counting são tratados, assim como a Expansão em Derivadas Covariantes introduzida como alternativa à construção do conjunto de operadores efetivos resultante de um modelo UV particular. A técnica de integração funcional é aplicada em casos que incluem o MP com Tripleto de Escalares e diferentes setores do modelo 3-3-1 na presença de Leptons pesados. Finalmente, o coeficiente de Wilson de dimensão-6 gerado a partir da integração de um quark-J pesado é limitado pelos valores recentes do parâmetro obliquo Y.
It will be presented the principles behind the use of the Standard Model Effective Field Theory as a consistent method to parametrize New Physics. The concepts of Matching and Power Counting are covered and a Covariant Derivative Expansion introduced to the construction of the operators set coming from the particular integrated UV model. The technique is applied in examples including the SM with a new Scalar Triplet and for different sectors of the 3-3-1 model in the presence of Heavy Leptons. Finally, the Wilson coefficient for a dimension-6 operator generated from the integration of a heavy J-quark is then compared with the measurements of the oblique Y parameter.
CNPq: 142492/2013-2
CAPES: 88881.132498/2016-01
APA, Harvard, Vancouver, ISO, and other styles
2

Correia, Fagner Cintra. "The standard model effective field theory : integrating UV models via functional methods /." São Paulo, 2017. http://hdl.handle.net/11449/151703.

Full text
Abstract:
Orientador: Vicente Pleitez
Resumo: O Modelo Padrão Efetivo é apresentado como um método consistente de parametrizar FísicaNova. Os conceitos de Matching e Power Counting são tratados, assim como a Expansão emDerivadas Covariantes introduzida como alternativa à construção do conjunto de operadoresefetivos resultante de um modelo UV particular. A técnica de integração funcional é aplicadaem casos que incluem o MP com Tripleto de Escalares e diferentes setores do modelo 3-3-1 napresença de Leptons pesados. Finalmente, o coeficiente de Wilson de dimensão-6 gerado a partirda integração de um quark-J pesado é limitado pelos valores recentes do parâmetro obliquo Y.
Doutor
APA, Harvard, Vancouver, ISO, and other styles
3

Rau, Christian, and rau@maths anu edu au. "Curve Estimation and Signal Discrimination in Spatial Problems." The Australian National University. School of Mathematical Sciences, 2003. http://thesis.anu.edu.au./public/adt-ANU20031215.163519.

Full text
Abstract:
In many instances arising prominently, but not exclusively, in imaging problems, it is important to condense the salient information so as to obtain a low-dimensional approximant of the data. This thesis is concerned with two basic situations which call for such a dimension reduction. The first of these is the statistical recovery of smooth edges in regression and density surfaces. The edges are understood to be contiguous curves, although they are allowed to meander almost arbitrarily through the plane, and may even split at a finite number of points to yield an edge graph. A novel locally-parametric nonparametric method is proposed which enjoys the benefit of being relatively easy to implement via a `tracking' approach. These topics are discussed in Chapters 2 and 3, with pertaining background material being given in the Appendix. In Chapter 4 we construct concomitant confidence bands for this estimator, which have asymptotically correct coverage probability. The construction can be likened to only a few existing approaches, and may thus be considered as our main contribution. ¶ Chapter 5 discusses numerical issues pertaining to the edge and confidence band estimators of Chapters 2-4. Connections are drawn to popular topics which originated in the fields of computer vision and signal processing, and which surround edge detection. These connections are exploited so as to obtain greater robustness of the likelihood estimator, such as with the presence of sharp corners. ¶ Chapter 6 addresses a dimension reduction problem for spatial data where the ultimate objective of the analysis is the discrimination of these data into one of a few pre-specified groups. In the dimension reduction step, an instrumental role is played by the recently developed methodology of functional data analysis. Relatively standar non-linear image processing techniques, as well as wavelet shrinkage, are used prior to this step. A case study for remotely-sensed navigation radar data exemplifies the methodology of Chapter 6.
APA, Harvard, Vancouver, ISO, and other styles
4

Rustaey, Abid 1961. "A comparison of conventional acceleration schemes to the method of residual expansion functions." Thesis, The University of Arizona, 1989. http://hdl.handle.net/10150/277176.

Full text
Abstract:
The algebraic equations resulting from a finite difference approximation may be solved numerically. A new scheme that appears quite promising is the method of residual expansion functions. In addition to speedy convergence, it is also independent of the number of algebraic equations under consideration, hence enabling us to analyze larger systems with higher accuracies. A factor which plays an important role in convergence of some numerical schemes is the concept of diagonal dominance. Matrices that converge at high rates are indeed the ones that possess a high degree of diagonal dominance. Another attractive feature of the method of residual expansion functions is its accurate convergence with minimal degree of diagonal dominance. Methods such as simultaneous and successive displacements, Chebyshev and projection are also discussed, but unlike the method of residual expansion functions, their convergence rates are strongly dependent on the degree of diagonal dominance.
APA, Harvard, Vancouver, ISO, and other styles
5

Zipperer, Travis Jonathan. "Pulse height tally response expansion method for application in detector problems." Thesis, Georgia Institute of Technology, 2011. http://hdl.handle.net/1853/44816.

Full text
Abstract:
A pulse height tally response expansion (PHRE) method is developed for detectors. By expanding the incident flux at the detector window/surface, a set of response functions is constructed via Monte Carlo estimators for pulse height tallies. B-spline functions are selected to perform the expansion of the response functions as well as for the expansion of the incident flux in photon energy. The method is verified for several incident flux spectra on a CsI(Na) detector. Results are compared to the solutions generated using direct Monte Carlo calculations. It is found that the method is several orders faster than MCNP5 while maintaining paralleled accuracy.
APA, Harvard, Vancouver, ISO, and other styles
6

Lladser, Manuel Eugenio. "Asymptotic enumeration via singularity analysis." Connect to this title online, 2003. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=osu1060976912.

Full text
Abstract:
Thesis (Ph. D.)--Ohio State University, 2003.
Title from first page of PDF file. Document formatted into pages; contains x, 227 p.; also includes graphics Includes bibliographical references (p. 224-227). Available online via OhioLINK's ETD Center
APA, Harvard, Vancouver, ISO, and other styles
7

Calatayud, Gregori Julia. "Computational methods for random differential equations: probability density function and estimation of the parameters." Doctoral thesis, Universitat Politècnica de València, 2020. http://hdl.handle.net/10251/138396.

Full text
Abstract:
[EN] Mathematical models based on deterministic differential equations do not take into account the inherent uncertainty of the physical phenomenon (in a wide sense) under study. In addition, inaccuracies in the collected data often arise due to errors in the measurements. It thus becomes necessary to treat the input parameters of the model as random quantities, in the form of random variables or stochastic processes. This gives rise to the study of random ordinary and partial differential equations. The computation of the probability density function of the stochastic solution is important for uncertainty quantification of the model output. Although such computation is a difficult objective in general, certain stochastic expansions for the model coefficients allow faithful representations for the stochastic solution, which permits approximating its density function. In this regard, Karhunen-Loève and generalized polynomial chaos expansions become powerful tools for the density approximation. Also, methods based on discretizations from finite difference numerical schemes permit approximating the stochastic solution, therefore its probability density function. The main part of this dissertation aims at approximating the probability density function of important mathematical models with uncertainties in their formulation. Specifically, in this thesis we study, in the stochastic sense, the following models that arise in different scientific areas: in Physics, the model for the damped pendulum; in Biology and Epidemiology, the models for logistic growth and Bertalanffy, as well as epidemiological models; and in Thermodynamics, the heat partial differential equation. We rely on Karhunen-Loève and generalized polynomial chaos expansions and on finite difference schemes for the density approximation of the solution. These techniques are only applicable when we have a forward model in which the input parameters have certain probability distributions already set. When the model coefficients are estimated from collected data, we have an inverse problem. The Bayesian inference approach allows estimating the probability distribution of the model parameters from their prior probability distribution and the likelihood of the data. Uncertainty quantification for the model output is then carried out using the posterior predictive distribution. In this regard, the last part of the thesis shows the estimation of the distributions of the model parameters from experimental data on bacteria growth. To do so, a hybrid method that combines Bayesian parameter estimation and generalized polynomial chaos expansions is used.
[ES] Los modelos matemáticos basados en ecuaciones diferenciales deterministas no tienen en cuenta la incertidumbre inherente del fenómeno físico (en un sentido amplio) bajo estudio. Además, a menudo se producen inexactitudes en los datos recopilados debido a errores en las mediciones. Por lo tanto, es necesario tratar los parámetros de entrada del modelo como cantidades aleatorias, en forma de variables aleatorias o procesos estocásticos. Esto da lugar al estudio de las ecuaciones diferenciales aleatorias. El cálculo de la función de densidad de probabilidad de la solución estocástica es importante en la cuantificación de la incertidumbre de la respuesta del modelo. Aunque dicho cálculo es un objetivo difícil en general, ciertas expansiones estocásticas para los coeficientes del modelo dan lugar a representaciones fieles de la solución estocástica, lo que permite aproximar su función de densidad. En este sentido, las expansiones de Karhunen-Loève y de caos polinomial generalizado constituyen herramientas para dicha aproximación de la densidad. Además, los métodos basados en discretizaciones de esquemas numéricos de diferencias finitas permiten aproximar la solución estocástica, por lo tanto, su función de densidad de probabilidad. La parte principal de esta disertación tiene como objetivo aproximar la función de densidad de probabilidad de modelos matemáticos importantes con incertidumbre en su formulación. Concretamente, en esta memoria se estudian, en un sentido estocástico, los siguientes modelos que aparecen en diferentes áreas científicas: en Física, el modelo del péndulo amortiguado; en Biología y Epidemiología, los modelos de crecimiento logístico y de Bertalanffy, así como modelos de tipo epidemiológico; y en Termodinámica, la ecuación en derivadas parciales del calor. Utilizamos expansiones de Karhunen-Loève y de caos polinomial generalizado y esquemas de diferencias finitas para la aproximación de la densidad de la solución. Estas técnicas solo son aplicables cuando tenemos un modelo directo en el que los parámetros de entrada ya tienen determinadas distribuciones de probabilidad establecidas. Cuando los coeficientes del modelo se estiman a partir de los datos recopilados, tenemos un problema inverso. El enfoque de inferencia Bayesiana permite estimar la distribución de probabilidad de los parámetros del modelo a partir de su distribución de probabilidad previa y la verosimilitud de los datos. La cuantificación de la incertidumbre para la respuesta del modelo se lleva a cabo utilizando la distribución predictiva a posteriori. En este sentido, la última parte de la tesis muestra la estimación de las distribuciones de los parámetros del modelo a partir de datos experimentales sobre el crecimiento de bacterias. Para hacerlo, se utiliza un método híbrido que combina la estimación de parámetros Bayesianos y los desarrollos de caos polinomial generalizado.
[CAT] Els models matemàtics basats en equacions diferencials deterministes no tenen en compte la incertesa inherent al fenomen físic (en un sentit ampli) sota estudi. A més a més, sovint es produeixen inexactituds en les dades recollides a causa d'errors de mesurament. Es fa així necessari tractar els paràmetres d'entrada del model com a quantitats aleatòries, en forma de variables aleatòries o processos estocàstics. Açò dóna lloc a l'estudi de les equacions diferencials aleatòries. El càlcul de la funció de densitat de probabilitat de la solució estocàstica és important per a quantificar la incertesa de la sortida del model. Tot i que, en general, aquest càlcul és un objectiu difícil d'assolir, certes expansions estocàstiques dels coeficients del model donen lloc a representacions fidels de la solució estocàstica, el que permet aproximar la seua funció de densitat. En aquest sentit, les expansions de Karhunen-Loève i de caos polinomial generalitzat esdevenen eines per a l'esmentada aproximació de la densitat. A més a més, els mètodes basats en discretitzacions mitjançant esquemes numèrics de diferències finites permeten aproximar la solució estocàstica, per tant la seua funció de densitat de probabilitat. La part principal d'aquesta dissertació té com a objectiu aproximar la funció de densitat de probabilitat d'importants models matemàtics amb incerteses en la seua formulació. Concretament, en aquesta memòria s'estudien, en un sentit estocàstic, els següents models que apareixen en diferents àrees científiques: en Física, el model del pèndol amortit; en Biologia i Epidemiologia, els models de creixement logístic i de Bertalanffy, així com models de tipus epidemiològic; i en Termodinàmica, l'equació en derivades parcials de la calor. Per a l'aproximació de la densitat de la solució, ens basem en expansions de Karhunen-Loève i de caos polinomial generalitzat i en esquemes de diferències finites. Aquestes tècniques només són aplicables quan tenim un model cap avant en què els paràmetres d'entrada tenen ja determinades distribucions de probabilitat. Quan els coeficients del model s'estimen a partir de les dades recollides, tenim un problema invers. L'enfocament de la inferència Bayesiana permet estimar la distribució de probabilitat dels paràmetres del model a partir de la seua distribució de probabilitat prèvia i la versemblança de les dades. La quantificació de la incertesa per a la resposta del model es fa mitjançant la distribució predictiva a posteriori. En aquest sentit, l'última part de la tesi mostra l'estimació de les distribucions dels paràmetres del model a partir de dades experimentals sobre el creixement de bacteris. Per a fer-ho, s'utilitza un mètode híbrid que combina l'estimació de paràmetres Bayesiana i els desenvolupaments de caos polinomial generalitzat.
This work has been supported by the Spanish Ministerio de Econom´ıa y Competitividad grant MTM2017–89664–P.
Calatayud Gregori, J. (2020). Computational methods for random differential equations: probability density function and estimation of the parameters [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/138396
TESIS
Premiado
APA, Harvard, Vancouver, ISO, and other styles
8

Jornet, Sanz Marc. "Mean square solutions of random linear models and computation of their probability density function." Doctoral thesis, Universitat Politècnica de València, 2020. http://hdl.handle.net/10251/138394.

Full text
Abstract:
[EN] This thesis concerns the analysis of differential equations with uncertain input parameters, in the form of random variables or stochastic processes with any type of probability distributions. In modeling, the input coefficients are set from experimental data, which often involve uncertainties from measurement errors. Moreover, the behavior of the physical phenomenon under study does not follow strict deterministic laws. It is thus more realistic to consider mathematical models with randomness in their formulation. The solution, considered in the sample-path or the mean square sense, is a smooth stochastic process, whose uncertainty has to be quantified. Uncertainty quantification is usually performed by computing the main statistics (expectation and variance) and, if possible, the probability density function. In this dissertation, we study random linear models, based on ordinary differential equations with and without delay and on partial differential equations. The linear structure of the models makes it possible to seek for certain probabilistic solutions and even approximate their probability density functions, which is a difficult goal in general. A very important part of the dissertation is devoted to random second-order linear differential equations, where the coefficients of the equation are stochastic processes and the initial conditions are random variables. The study of this class of differential equations in the random setting is mainly motivated because of their important role in Mathematical Physics. We start by solving the randomized Legendre differential equation in the mean square sense, which allows the approximation of the expectation and the variance of the stochastic solution. The methodology is extended to general random second-order linear differential equations with analytic (expressible as random power series) coefficients, by means of the so-called Fröbenius method. A comparative case study is performed with spectral methods based on polynomial chaos expansions. On the other hand, the Fröbenius method together with Monte Carlo simulation are used to approximate the probability density function of the solution. Several variance reduction methods based on quadrature rules and multilevel strategies are proposed to speed up the Monte Carlo procedure. The last part on random second-order linear differential equations is devoted to a random diffusion-reaction Poisson-type problem, where the probability density function is approximated using a finite difference numerical scheme. The thesis also studies random ordinary differential equations with discrete constant delay. We study the linear autonomous case, when the coefficient of the non-delay component and the parameter of the delay term are both random variables while the initial condition is a stochastic process. It is proved that the deterministic solution constructed with the method of steps that involves the delayed exponential function is a probabilistic solution in the Lebesgue sense. Finally, the last chapter is devoted to the linear advection partial differential equation, subject to stochastic velocity field and initial condition. We solve the equation in the mean square sense and provide new expressions for the probability density function of the solution, even in the non-Gaussian velocity case.
[ES] Esta tesis trata el análisis de ecuaciones diferenciales con parámetros de entrada aleatorios, en la forma de variables aleatorias o procesos estocásticos con cualquier tipo de distribución de probabilidad. En modelización, los coeficientes de entrada se fijan a partir de datos experimentales, los cuales suelen acarrear incertidumbre por los errores de medición. Además, el comportamiento del fenómeno físico bajo estudio no sigue patrones estrictamente deterministas. Es por tanto más realista trabajar con modelos matemáticos con aleatoriedad en su formulación. La solución, considerada en el sentido de caminos aleatorios o en el sentido de media cuadrática, es un proceso estocástico suave, cuya incertidumbre se tiene que cuantificar. La cuantificación de la incertidumbre es a menudo llevada a cabo calculando los principales estadísticos (esperanza y varianza) y, si es posible, la función de densidad de probabilidad. En este trabajo, estudiamos modelos aleatorios lineales, basados en ecuaciones diferenciales ordinarias con y sin retardo, y en ecuaciones en derivadas parciales. La estructura lineal de los modelos nos permite buscar ciertas soluciones probabilísticas e incluso aproximar su función de densidad de probabilidad, lo cual es un objetivo complicado en general. Una parte muy importante de la disertación se dedica a las ecuaciones diferenciales lineales de segundo orden aleatorias, donde los coeficientes de la ecuación son procesos estocásticos y las condiciones iniciales son variables aleatorias. El estudio de esta clase de ecuaciones diferenciales en el contexto aleatorio está motivado principalmente por su importante papel en la Física Matemática. Empezamos resolviendo la ecuación diferencial de Legendre aleatorizada en el sentido de media cuadrática, lo que permite la aproximación de la esperanza y la varianza de la solución estocástica. La metodología se extiende al caso general de ecuaciones diferenciales lineales de segundo orden aleatorias con coeficientes analíticos (expresables como series de potencias), mediante el conocido método de Fröbenius. Se lleva a cabo un estudio comparativo con métodos espectrales basados en expansiones de caos polinomial. Por otro lado, el método de Fröbenius junto con la simulación de Monte Carlo se utilizan para aproximar la función de densidad de probabilidad de la solución. Para acelerar el procedimiento de Monte Carlo, se proponen varios métodos de reducción de la varianza basados en reglas de cuadratura y estrategias multinivel. La última parte sobre ecuaciones diferenciales lineales de segundo orden aleatorias estudia un problema aleatorio de tipo Poisson de difusión-reacción, en el que la función de densidad de probabilidad es aproximada mediante un esquema numérico de diferencias finitas. En la tesis también se tratan ecuaciones diferenciales ordinarias aleatorias con retardo discreto y constante. Estudiamos el caso lineal y autónomo, cuando el coeficiente de la componente no retardada i el parámetro del término retardado son ambos variables aleatorias mientras que la condición inicial es un proceso estocástico. Se demuestra que la solución determinista construida con el método de los pasos y que involucra la función exponencial retardada es una solución probabilística en el sentido de Lebesgue. Finalmente, el último capítulo lo dedicamos a la ecuación en derivadas parciales lineal de advección, sujeta a velocidad y condición inicial estocásticas. Resolvemos la ecuación en el sentido de media cuadrática y damos nuevas expresiones para la función de densidad de probabilidad de la solución, incluso en el caso de velocidad no Gaussiana.
[CAT] Aquesta tesi tracta l'anàlisi d'equacions diferencials amb paràmetres d'entrada aleatoris, en la forma de variables aleatòries o processos estocàstics amb qualsevol mena de distribució de probabilitat. En modelització, els coeficients d'entrada són fixats a partir de dades experimentals, les quals solen comportar incertesa pels errors de mesurament. A més a més, el comportament del fenomen físic sota estudi no segueix patrons estrictament deterministes. És per tant més realista treballar amb models matemàtics amb aleatorietat en la seua formulació. La solució, considerada en el sentit de camins aleatoris o en el sentit de mitjana quadràtica, és un procés estocàstic suau, la incertesa del qual s'ha de quantificar. La quantificació de la incertesa és sovint duta a terme calculant els principals estadístics (esperança i variància) i, si es pot, la funció de densitat de probabilitat. En aquest treball, estudiem models aleatoris lineals, basats en equacions diferencials ordinàries amb retard i sense, i en equacions en derivades parcials. L'estructura lineal dels models ens fa possible cercar certes solucions probabilístiques i inclús aproximar la seua funció de densitat de probabilitat, el qual és un objectiu complicat en general. Una part molt important de la dissertació es dedica a les equacions diferencials lineals de segon ordre aleatòries, on els coeficients de l'equació són processos estocàstics i les condicions inicials són variables aleatòries. L'estudi d'aquesta classe d'equacions diferencials en el context aleatori està motivat principalment pel seu important paper en Física Matemàtica. Comencem resolent l'equació diferencial de Legendre aleatoritzada en el sentit de mitjana quadràtica, el que permet l'aproximació de l'esperança i la variància de la solució estocàstica. La metodologia s'estén al cas general d'equacions diferencials lineals de segon ordre aleatòries amb coeficients analítics (expressables com a sèries de potències), per mitjà del conegut mètode de Fröbenius. Es duu a terme un estudi comparatiu amb mètodes espectrals basats en expansions de caos polinomial. Per altra banda, el mètode de Fröbenius juntament amb la simulació de Monte Carlo són emprats per a aproximar la funció de densitat de probabilitat de la solució. Per a accelerar el procediment de Monte Carlo, es proposen diversos mètodes de reducció de la variància basats en regles de quadratura i estratègies multinivell. L'última part sobre equacions diferencials lineals de segon ordre aleatòries estudia un problema aleatori de tipus Poisson de difusió-reacció, en què la funció de densitat de probabilitat és aproximada mitjançant un esquema numèric de diferències finites. En la tesi també es tracten equacions diferencials ordinàries aleatòries amb retard discret i constant. Estudiem el cas lineal i autònom, quan el coeficient del component no retardat i el paràmetre del terme retardat són ambdós variables aleatòries mentre que la condició inicial és un procés estocàstic. Es prova que la solució determinista construïda amb el mètode dels passos i que involucra la funció exponencial retardada és una solució probabilística en el sentit de Lebesgue. Finalment, el darrer capítol el dediquem a l'equació en derivades parcials lineal d'advecció, subjecta a velocitat i condició inicial estocàstiques. Resolem l'equació en el sentit de mitjana quadràtica i donem noves expressions per a la funció de densitat de probabilitat de la solució, inclús en el cas de velocitat no Gaussiana.
This work has been supported by the Spanish Ministerio de Economía y Competitividad grant MTM2017–89664–P. I acknowledge the doctorate scholarship granted by Programa de Ayudas de Investigación y Desarrollo (PAID), Universitat Politècnica de València.
Jornet Sanz, M. (2020). Mean square solutions of random linear models and computation of their probability density function [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/138394
TESIS
APA, Harvard, Vancouver, ISO, and other styles
9

Starkloff, Hans-Jörg, and Ralf Wunderlich. "Stationary solutions of linear ODEs with a randomly perturbed system matrix and additive noise." Universitätsbibliothek Chemnitz, 2005. http://nbn-resolving.de/urn:nbn:de:swb:ch1-200501335.

Full text
Abstract:
The paper considers systems of linear first-order ODEs with a randomly perturbed system matrix and stationary additive noise. For the description of the long-term behavior of such systems it is necessary to study their stationary solutions. We deal with conditions for the existence of stationary solutions as well as with their representations and the computation of their moment functions. Assuming small perturbations of the system matrix we apply perturbation techniques to find series representations of the stationary solutions and give asymptotic expansions for their first- and second-order moment functions. We illustrate the findings with a numerical example of a scalar ODE, for which the moment functions of the stationary solution still can be computed explicitly. This allows the assessment of the goodness of the approximations found from the derived asymptotic expansions.
APA, Harvard, Vancouver, ISO, and other styles
10

Cao, Liang. "Numerical analysis and multi-precision computational methods applied to the extant problems of Asian option pricing and simulating stable distributions and unit root densities." Thesis, University of St Andrews, 2014. http://hdl.handle.net/10023/6539.

Full text
Abstract:
This thesis considers new methods that exploit recent developments in computer technology to address three extant problems in the area of Finance and Econometrics. The problem of Asian option pricing has endured for the last two decades in spite of many attempts to find a robust solution across all parameter values. All recently proposed methods are shown to fail when computations are conducted using standard machine precision because as more and more accuracy is forced upon the problem, round-off error begins to propagate. Using recent methods from numerical analysis based on multi-precision arithmetic, we show using the Mathematica platform that all extant methods have efficacy when computations use sufficient arithmetic precision. This creates the proper framework to compare and contrast the methods based on criteria such as computational speed for a given accuracy. Numerical methods based on a deformation of the Bromwich contour in the Geman-Yor Laplace transform are found to perform best provided the normalized strike price is above a given threshold; otherwise methods based on Euler approximation are preferred. The same methods are applied in two other contexts: the simulation of stable distributions and the computation of unit root densities in Econometrics. The stable densities are all nested in a general function called a Fox H function. The same computational difficulties as above apply when using only double-precision arithmetic but are again solved using higher arithmetic precision. We also consider simulating the densities of infinitely divisible distributions associated with hyperbolic functions. Finally, our methods are applied to unit root densities. Focusing on the two fundamental densities, we show our methods perform favorably against the extant methods of Monte Carlo simulation, the Imhof algorithm and some analytical expressions derived principally by Abadir. Using Mathematica, the main two-dimensional Laplace transform in this context is reduced to a one-dimensional problem.
APA, Harvard, Vancouver, ISO, and other styles
11

Svoboda, Petr. "Popis rozložení napětí v okolí ostrého vrubu." Master's thesis, Vysoké učení technické v Brně. Fakulta strojního inženýrství, 2018. http://www.nusl.cz/ntk/nusl-382552.

Full text
Abstract:
The presented diploma thesis deals with the problem of determining the stress singularity exponent of the V-notch. This task can be divided into two parts. The first deals with the theoretical background, that means the basic relations of mechanics and the basic concepts of fracture mechanics. The second part deals with the elaboration of the Williams method and the creation of a program for calculating the stress singularity exponent.
APA, Harvard, Vancouver, ISO, and other styles
12

Rau, Christian. "Curve Estimation and Signal Discrimination in Spatial Problems." Phd thesis, 2003. http://hdl.handle.net/1885/48023.

Full text
Abstract:
In many instances arising prominently, but not exclusively, in imaging problems, it is important to condense the salient information so as to obtain a low-dimensional approximant of the data. This thesis is concerned with two basic situations which call for such a dimension reduction. The first of these is the statistical recovery of smooth edges in regression and density surfaces. The edges are understood to be contiguous curves, although they are allowed to meander almost arbitrarily through the plane, and may even split at a finite number of points to yield an edge graph. A novel locally-parametric nonparametric method is proposed which enjoys the benefit of being relatively easy to implement via a `tracking' approach. These topics are discussed in Chapters 2 and 3, with pertaining background material being given in the Appendix. In Chapter 4 we construct concomitant confidence bands for this estimator, which have asymptotically correct coverage probability. The construction can be likened to only a few existing approaches, and may thus be considered as our main contribution. ¶ Chapter 5 discusses numerical issues pertaining to the edge and confidence band estimators of Chapters 2-4. Connections are drawn to popular topics which originated in the fields of computer vision and signal processing, and which surround edge detection. These connections are exploited so as to obtain greater robustness of the likelihood estimator, such as with the presence of sharp corners. ¶ Chapter 6 addresses a dimension reduction problem for spatial data where the ultimate objective of the analysis is the discrimination of these data into one of a few pre-specified groups. In the dimension reduction step, an instrumental role is played by the recently developed methodology of functional data analysis. Relatively standar non-linear image processing techniques, as well as wavelet shrinkage, are used prior to this step. A case study for remotely-sensed navigation radar data exemplifies the methodology of Chapter 6.
APA, Harvard, Vancouver, ISO, and other styles
13

LIN, JUN-LIANG, and 林俊良. "Missile specification design by using model expansion method and covariance analysis describing function technique." Thesis, 1986. http://ndltd.ncl.edu.tw/handle/58758985588816498724.

Full text
APA, Harvard, Vancouver, ISO, and other styles
14

Chan, Chih-Yuan, and 詹智淵. "Analysis on Eigenstates in Two Dimensional System by Expansion method: Using Sinc Function as Basis." Thesis, 2011. http://ndltd.ncl.edu.tw/handle/60769415135518945230.

Full text
Abstract:
碩士
國立交通大學
理學院應用科技學程
99
Eigenvalue problem has been widely existed in the nature. Unfortunately, due to the complexities of the physical system, analytical solutions to eigenvalue problem can be obtained only in few cases. Some numerical methods must be used to solve the eigenvalue problem. In this paper, we used expansion method based on Sinc function as basis to analyze the famous eigenvalue problem in physical system: 2D Helmholtz equation. To consider the feasibility of numerical method, we numerically calculated the eigenstates and eigenenergy in rectangular and circular membrane and compared the numerical results with the analytical solutions. We also solved the two dimensional (2D) membranes problems with arbitrary shapes of Kinmen and violin by using the numerical method. We further used the numerical method to simulate Chladni Nodal line pattern and investigate the influence of perturbation on Chladni Nodal line pattern. It can be seen that there is a good agreement between numerical results and experiment results.
APA, Harvard, Vancouver, ISO, and other styles
15

Huang, Chiao-Wen, and 黃巧文. "The Effect of Different Lung Expansion Training Methods on Pulmonary Complications and Pulmonary Function in Patients Receiving Lobectomy." Thesis, 2004. http://ndltd.ncl.edu.tw/handle/10806481782598008335.

Full text
Abstract:
碩士
國立臺灣大學
護理學研究所
92
Lung expansion is an important clinical intervention. It plays an important role in the prevention of postoperative pulmonary complication and the recovery of pulmonary function. Few researches focused on the long-term effect of the lung expansion on the prevention of pulmonary complication and the recovery of pulmonary function. This study was to investigate the effect of the different lung expansion training methods on pulmonary complication and pulmonary function in patients receiving lobectomy. In this study, a repeated-measure longitudinal quai- experimental study was employed. The research took place in the medical-surgical units of a medical center in Taipei. All elective probable lobectomy patients were divided into three groups, each receiving preoperative bedside testing of pulmonary function and instruction in the use of three randomly assigned lung expansion methods (deep breath exercise, TrifloII, or Couch2). Repeated lung function tests were provided preoperatively and postoperatively at days 2, 4, 6, two weeks, 1.5 months, and 3 months. And six-minute walk was tested the same time as lung function except postoperatively at days 2, 4, 6. All the patients’ postoperative pulmonary complications, lung function recovery were recorded. Fifty-four patients met the criteria were included. There were no statistical significant differences between three groups in postoperative pulmonary complications, except more Chest X-ray abnormal finding of the TrifloII group postoperatively at days 7 to 3 months. There were no statistical significant differences between three groups in postoperative pulmonary function, except higher dyspnea scale score of the TrifloII group after six-minute walk test. Furthermore, different lung expansion training methods, gender, diagnosis, past cardiopulmonary disease history, duration of anesthesia, blood loss in operation, surgical methods, length of the surgical site, religious, retention time of chest tube, duration of ventilator use, body mass index, length of ICU stay, preoperative pulmonary function(%FVCpred, %PFpred, and %FEV1pred ), frequency of the lung expansion training, and the effort of the lung expansion training would be the influential factors to the recovery of pulmonary function in statistics. The changes SpO2 and heart rate after six-minute walk test would be the predictive indicators of the recovery of pulmonary function. The study showed deep breath exercise group was the most cost-effective lung expansion training in lobectomy patients. Otherwise, some of the lobectomy patients lack the correct deep breath exercise skill after training or surgery. We suggest deep breath exercise training for all preoperative and postoperative lobectomy patients and incentive spirometry as the alternative if lacking the correct deep breath exercise skill.
APA, Harvard, Vancouver, ISO, and other styles
16

Stephens, John Adam. "Simulation tools for predicting the atomic configuration of bimetallic catalytic surfaces." 2012. http://hdl.handle.net/2152/22175.

Full text
Abstract:
Transition metal alloys are an important class of materials in heterogeneous catalysis due in no small part to the often greatly enhanced activity and selectivity they exhibit compared to their monometallic constituents. A host of experimental and theoretical studies have demonstrated that, in many cases, these synergistic effects can be attributed to atomic-scale features of the catalyst surface. Realizing the goal of designing -- rather than serendipitously discovering -- new alloy catalysts thus depends on our ability to predict their atomic configuration under technologically relevant conditions. This dissertation presents original research into the development and use of computational tools to accomplish this objective. These tools are all based on a similar strategy: For each of the alloy systems examined, cluster expansion (CE) Hamiltonians were constructed from the results of density functional theory (DFT) calculations, and then used in Metropolis Monte Carlo (MC) simulations to predict properties of interest. Following a detailed description of the DFT+CE+MC simulation scheme, results for the AuPd/Pd(111) and AuPt/Pt(111) surface alloys are presented. These two systems exhibit considerably different trends in their atomic arrangement, which are explicable in terms of their interatomic interactions. In AuPd, a preference for heteronuclear, Au-Pd interactions results in the preferential formation of Pd monomers and other small ensembles, while in AuPt, a preference for homonuclear interactions results in the opposite. AuPd/Pd(100) and AuPt/Pt(100) were similarly examined, revealing not only the effects of the same heteronuclear/homonuclear preferences in this facet, but also a propensity for the formation of second nearest-neighbor pairs of Pd monomers, in close agreement with experiment. Subsequent simulations of the AuPd/Pd(100) surface suggest the application of biaxial compressive strain as a means increasing the population of this catalytically important ensemble of atoms. A method to incorporate the effects of subsurface atomic configuration is also presented, using AuPd as an example. This method represents several improvements over others previously reported in the literature, especially in terms of its simplicity. Finally, we introduce the dimensionless scaled pair interaction, whereby the finite-temperature atomic configuration of any bimetallic surface alloy may be predicted from a small number of relatively inexpensive calculations.
text
APA, Harvard, Vancouver, ISO, and other styles
17

Tiegel, Alexander Clemens. "Finite-temperature dynamics of low-dimensional quantum systems with DMRG methods." Doctoral thesis, 2016. http://hdl.handle.net/11858/00-1735-0000-0028-8801-A.

Full text
APA, Harvard, Vancouver, ISO, and other styles
18

Kotchoni, Rachidi. "Efficient estimation using the characteristic function : theory and applications with high frequency data." Thèse, 2010. http://hdl.handle.net/1866/4392.

Full text
Abstract:
Nous abordons deux sujets distincts dans cette thèse: l'estimation de la volatilité des prix d'actifs financiers à partir des données à haute fréquence, et l'estimation des paramétres d'un processus aléatoire à partir de sa fonction caractéristique. Le chapitre 1 s'intéresse à l'estimation de la volatilité des prix d'actifs. Nous supposons que les données à haute fréquence disponibles sont entachées de bruit de microstructure. Les propriétés que l'on prête au bruit sont déterminantes dans le choix de l'estimateur de la volatilité. Dans ce chapitre, nous spécifions un nouveau modèle dynamique pour le bruit de microstructure qui intègre trois propriétés importantes: (i) le bruit peut être autocorrélé, (ii) le retard maximal au delà duquel l'autocorrélation est nulle peut être une fonction croissante de la fréquence journalière d'observations; (iii) le bruit peut avoir une composante correlée avec le rendement efficient. Cette dernière composante est alors dite endogène. Ce modèle se différencie de ceux existant en ceci qu'il implique que l'autocorrélation d'ordre 1 du bruit converge vers 1 lorsque la fréquence journalière d'observation tend vers l'infini. Nous utilisons le cadre semi-paramétrique ainsi défini pour dériver un nouvel estimateur de la volatilité intégrée baptisée "estimateur shrinkage". Cet estimateur se présente sous la forme d'une combinaison linéaire optimale de deux estimateurs aux propriétés différentes, l'optimalité étant défini en termes de minimisation de la variance. Les simulations indiquent que l'estimateur shrinkage a une variance plus petite que le meilleur des deux estimateurs initiaux. Des estimateurs sont également proposés pour les paramètres du modèle de microstructure. Nous clôturons ce chapitre par une application empirique basée sur des actifs du Dow Jones Industrials. Les résultats indiquent qu'il est pertinent de tenir compte de la dépendance temporelle du bruit de microstructure dans le processus d'estimation de la volatilité. Les chapitres 2, 3 et 4 s'inscrivent dans la littérature économétrique qui traite de la méthode des moments généralisés. En effet, on rencontre en finance des modèles dont la fonction de vraisemblance n'est pas connue. On peut citer en guise d'exemple la loi stable ainsi que les modèles de diffusion observés en temps discrets. Les méthodes d'inférence basées sur la fonction caractéristique peuvent être envisagées dans ces cas. Typiquement, on spécifie une condition de moment basée sur la différence entre la fonction caractéristique (conditionnelle) théorique et sa contrepartie empirique. Le défit ici est d'exploiter au mieux le continuum de conditions de moment ainsi spécifié pour atteindre la même efficacité que le maximum de vraisemblance dans les inférences. Ce défit a été relevé par Carrasco et Florens (2000) qui ont proposé la procédure CGMM (continuum GMM). La fonction objectif que ces auteurs proposent est une forme quadratique hilbertienne qui fait intervenir l'opérateur inverse de covariance associé au continuum de condition de moments. Cet opérateur inverse est régularisé à la Tikhonov pour en assurer l'existence globale et la continuité. Carrasco et Florens (2000) ont montré que l'estimateur obtenu en minimisant cette forme quadratique est asymptotiquement aussi efficace que l'estimateur du maximum de vraisemblance si le paramètre de régularisation (α) tend vers zéro lorsque la taille de l'échatillon tend vers l'infini. La nature de la fonction objectif du CGMM soulève deux questions importantes. La première est celle de la calibration de α en pratique, et la seconde est liée à la présence d'intégrales multiples dans l'expression de la fonction objectif. C'est à ces deux problématiques qu'essayent de répondent les trois derniers chapitres de la présente thèse. Dans le chapitre 2, nous proposons une méthode de calibration de α basée sur la minimisation de l'erreur quadratique moyenne (EQM) de l'estimateur. Nous suivons une approche similaire à celle de Newey et Smith (2004) pour calculer un développement d'ordre supérieur de l'EQM de l'estimateur CGMM de sorte à pouvoir examiner sa dépendance en α en échantillon fini. Nous proposons ensuite deux méthodes pour choisir α en pratique. La première se base sur le développement de l'EQM, et la seconde se base sur des simulations Monte Carlo. Nous montrons que la méthode Monte Carlo délivre un estimateur convergent de α optimal. Nos simulations confirment la pertinence de la calibration de α en pratique. Le chapitre 3 essaye de vulgariser la théorie du chapitre 2 pour les modèles univariés ou bivariés. Nous commençons par passer en revue les propriétés de convergence et de normalité asymptotique de l'estimateur CGMM. Nous proposons ensuite des recettes numériques pour l'implémentation. Enfin, nous conduisons des simulations Monte Carlo basée sur la loi stable. Ces simulations démontrent que le CGMM est une méthode fiable d'inférence. En guise d'application empirique, nous estimons par CGMM un modèle de variance autorégressif Gamma. Les résultats d'estimation confirment un résultat bien connu en finance: le rendement est positivement corrélé au risque espéré et négativement corrélé au choc sur la volatilité. Lorsqu'on implémente le CGMM, une difficulté majeure réside dans l'évaluation numérique itérative des intégrales multiples présentes dans la fonction objectif. Les méthodes de quadrature sont en principe parmi les plus précises que l'on puisse utiliser dans le présent contexte. Malheureusement, le nombre de points de quadrature augmente exponentiellement en fonction de la dimensionalité (d) des intégrales. L'utilisation du CGMM devient pratiquement impossible dans les modèles multivariés et non markoviens où d≥3. Dans le chapitre 4, nous proposons une procédure alternative baptisée "reéchantillonnage dans le domaine fréquentielle" qui consiste à fabriquer des échantillons univariés en prenant une combinaison linéaire des éléments du vecteur initial, les poids de la combinaison linéaire étant tirés aléatoirement dans un sous-espace normalisé de ℝ^{d}. Chaque échantillon ainsi généré est utilisé pour produire un estimateur du paramètre d'intérêt. L'estimateur final que nous proposons est une combinaison linéaire optimale de tous les estimateurs ainsi obtenus. Finalement, nous proposons une étude par simulation et une application empirique basées sur des modèles autorégressifs Gamma. Dans l'ensemble, nous faisons une utilisation intensive du bootstrap, une technique selon laquelle les propriétés statistiques d'une distribution inconnue peuvent être estimées à partir d'un estimé de cette distribution. Nos résultats empiriques peuvent donc en principe être améliorés en faisant appel aux connaissances les plus récentes dans le domaine du bootstrap.
In estimating the integrated volatility of financial assets using noisy high frequency data, the time series properties assumed for the microstructure noise determines the proper choice of the volatility estimator. In the first chapter of the current thesis, we propose a new model for the microstructure noise with three important features. First of all, our model assumes that the noise is L-dependent. Secondly, the memory lag L is allowed to increase with the sampling frequency. And thirdly, the noise may include an endogenous part, that is, a piece that is correlated with the latent returns. The main difference between this microstructure model and existing ones is that it implies a first order autocorrelation that converges to 1 as the sampling frequency goes to infinity. We use this semi-parametric model to derive a new shrinkage estimator for the integrated volatility. The proposed estimator makes an optimal signal-to-noise trade-off by combining a consistent estimators with an inconsistent one. Simulation results show that the shrinkage estimator behaves better than the best of the two combined ones. We also propose some estimators for the parameters of the noise model. An empirical study based on stocks listed in the Dow Jones Industrials shows the relevance of accounting for possible time dependence in the noise process. Chapters 2, 3 and 4 pertain to the generalized method of moments based on the characteristic function. In fact, the likelihood functions of many financial econometrics models are not known in close form. For example, this is the case for the stable distribution and a discretely observed continuous time model. In these cases, one may estimate the parameter of interest by specifying a moment condition based on the difference between the theoretical (conditional) characteristic function and its empirical counterpart. The challenge is then to exploit the whole continuum of moment conditions hence defined to achieve the maximum likelihood efficiency. This problem has been solved in Carrasco and Florens (2000) who propose the CGMM procedure. The objective function of the CGMM is a quadrqtic form on the Hilbert space defined by the moment function. That objective function depends on a Tikhonov-type regularized inverse of the covariance operator associated with the moment function. Carrasco and Florens (2000) have shown that the estimator obtained by minimizing the proposed objective function is asymptotically as efficient as the maximum likelihood estimator provided that the regularization parameter (α) converges to zero as the sample size goes to infinity. However, the nature of this objective function raises two important questions. First of all, how do we select α in practice? And secondly, how do we implement the CGMM when the multiplicity (d) of the integrals embedded in the objective-function d is large. These questions are tackled in the last three chapters of the thesis. In Chapter 2, we propose to choose α by minimizing the approximate mean square error (MSE) of the estimator. Following an approach similar to Newey and Smith (2004), we derive a higher-order expansion of the estimator from which we characterize the finite sample dependence of the MSE on α. We provide two data-driven methods for selecting the regularization parameter in practice. The first one relies on the higher-order expansion of the MSE whereas the second one uses only simulations. We show that our simulation technique delivers a consistent estimator of α. Our Monte Carlo simulations confirm the importance of the optimal selection of α. The goal of Chapter 3 is to illustrate how to efficiently implement the CGMM for d≤2. To start with, we review the consistency and asymptotic normality properties of the CGMM estimator. Next we suggest some numerical recipes for its implementation. Finally, we carry out a simulation study with the stable distribution that confirms the accuracy of the CGMM as an inference method. An empirical application based on the autoregressive variance Gamma model led to a well-known conclusion: investors require a positive premium for bearing the expected risk while a negative premium is attached to the unexpected risk. In implementing the characteristic function based CGMM, a major difficulty lies in the evaluation of the multiple integrals embedded in the objective function. Numerical quadratures are among the most accurate methods that can be used in the present context. Unfortunately, the number of quadrature points grows exponentially with d. When the data generating process is Markov or dependent, the accurate implementation of the CGMM becomes roughly unfeasible when d≥3. In Chapter 4, we propose a strategy that consists in creating univariate samples by taking a linear combination of the elements of the original vector process. The weights of the linear combinations are drawn from a normalized set of ℝ^{d}. Each univariate index generated in this way is called a frequency domain bootstrap sample that can be used to compute an estimator of the parameter of interest. Finally, all the possible estimators obtained in this fashion can be aggregated to obtain the final estimator. The optimal aggregation rule is discussed in the paper. The overall method is illustrated by a simulation study and an empirical application based on autoregressive Gamma models. This thesis makes an extensive use of the bootstrap, a technique according to which the statistical properties of an unknown distribution can be estimated from an estimate of that distribution. It is thus possible to improve our simulations and empirical results by using the state-of-the-art refinements of the bootstrap methodology.
The attached file is created with Scientific Workplace Latex
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography