Academic literature on the topic 'Generalised regression estimators'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Generalised regression estimators.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Generalised regression estimators"

1

Zhang, Li-Chun. "Generalised Regression Estimation Given Imperfectly Matched Auxiliary Data." Journal of Official Statistics 37, no. 1 (March 1, 2021): 239–55. http://dx.doi.org/10.2478/jos-2021-0010.

Full text
Abstract:
Abstract Generalised regression estimation allows one to make use of available auxiliary information in survey sampling. We develop three types of generalised regression estimator when the auxiliary data cannot be matched perfectly to the sample units, so that the standard estimator is inapplicable. The inference remains design-based. Consistency of the proposed estimators is either given by construction or else can be tested given the observed sample and links. Mean square errors can be estimated. A simulation study is used to explore the potentials of the proposed estimators.
APA, Harvard, Vancouver, ISO, and other styles
2

Wada, Kazumi, Keiichiro Sakashita, and Hiroe Tsubaki. "Robust Estimation for a Generalised Ratio Model." Austrian Journal of Statistics 50, no. 1 (February 3, 2021): 74–87. http://dx.doi.org/10.17713/ajs.v50i1.994.

Full text
Abstract:
It is known that data such as business sales and household income need data transformation prior to regression estimate as the data has a homoscedastic error. However, data transformations make the estimation of mean and total unstable. Therefore, the ratio model is often used for imputation in the field of official statistics to avoid the problem. Our study aims to robustify the estimator following the ratio model by means of M-estimation. Reformulation of the conventional ratio model with homoscedastic quasi-error term provides quasi-residuals which can be used as a measure of outlyingness as same as a linear regression model. A generalisation of the model, which accommodates varied error terms with different heteroscedasticity, is also proposed. Functions for robustified estimators of the generalised ratio model are implemented by the iterative re-weighted least squares algorithm in R environment and illustrated using random datasets. Monte Carlo simulation confirms accuracy of the proposed estimators, as well as their computational efficiency. A comparison of the scale parameters between the average absolute deviation (AAD) and median absolute deviation (MAD) is made regarding Tukey's biweight function. The results with Huber's weight function are also provided for reference. The proposed robust estimator of the generalised ratio model is used for imputation of major corporate accounting items of the 2016 Economic Census for Business Activity in Japan.
APA, Harvard, Vancouver, ISO, and other styles
3

Mohammed, M. A., Huda M. Alshanbari, and Abdal-Aziz H. El-Bagoury. "Application of the LINEX Loss Function with a Fundamental Derivation of Liu Estimator." Computational Intelligence and Neuroscience 2022 (March 14, 2022): 1–9. http://dx.doi.org/10.1155/2022/2307911.

Full text
Abstract:
For a variety of well-known approaches, optimum predictors and estimators are determined in relation to the asymmetrical LINEX loss function. The applications of an iteratively practicable lowest mean squared error estimation of the regression disturbance variation with the LINEX loss function are discussed in this research. This loss is a symmetrical generalisation of the quadratic loss function. Whenever the LINEX loss function is applied, we additionally look at the risk performance of the feasible virtually unbiased generalised Liu estimator and practicable generalised Liu estimator. Whenever the variation σ 2 is specified, we get all acceptable linear estimation in the class of linear estimation techniques, and when σ 2 is undetermined, we get all acceptable linear estimation in the class of linear estimation techniques. During position transformations, the proposed Liu estimators are stable. The estimators’ biases and hazards are calculated and evaluated. We utilize an asymmetrical loss function, the LINEX loss function, to calculate the actual hazards of several error variation estimators. The employment of δ P σ , which is easy to use and maximin, is recommended in the conclusions.
APA, Harvard, Vancouver, ISO, and other styles
4

Laroussi, Ilhem. "A generalised censored least squares and smoothing spline estimators of regression function." International Journal of Mathematics in Operational Research 20, no. 4 (2021): 506. http://dx.doi.org/10.1504/ijmor.2021.120102.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Sutradhar, B. "Miscellanea. On the efficiency of regression estimators in generalised linear models for longitudinal data." Biometrika 86, no. 2 (June 1, 1999): 459–65. http://dx.doi.org/10.1093/biomet/86.2.459.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Khare, B. B., and Sanjay Kumar. "Generalised chain ratio-in-regression estimators for population mean using two-phase sampling in the presence of non-response." Journal of Information and Optimization Sciences 36, no. 4 (June 9, 2015): 317–38. http://dx.doi.org/10.1080/02522667.2014.926706.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Slaoui, Y., and A. Jmaei. "Recursive and non-recursive regression estimators using Bernstein polynomials." Theory of Stochastic Processes 26(42), no. 1 (December 27, 2022): 60–95. http://dx.doi.org/10.37863/tsp-2899660400-77.

Full text
Abstract:
If a regression function has a bounded support, the kernel estimates often exceed the boundaries and are therefore biased on and near these limits. In this paper, we focus on mitigating this boundary problem. We apply Bernstein polynomials and the Robbins-Monro algorithm to construct a non-recursive and recursive regression estimator. We study the asymptotic properties of these estimators, and we compare them with those of the Nadaraya-Watson estimator and the generalized Révész estimator introduced by [21]. In addition, through some simulation studies, we show that our non-recursive estimator has the lowest integrated root mean square error (ISE) in most of the considered cases. Finally, using a set of real data, we demonstrate how our non-recursive and recursive regression estimators can lead to very satisfactory estimates, especially near the boundaries.
APA, Harvard, Vancouver, ISO, and other styles
8

DEVITA, HANY, I. KOMANG GDE SUKARSA, and I. PUTU EKA N. KENCANA. "KINERJA JACKKNIFE RIDGE REGRESSION DALAM MENGATASI MULTIKOLINEARITAS." E-Jurnal Matematika 3, no. 4 (November 28, 2014): 146. http://dx.doi.org/10.24843/mtk.2014.v03.i04.p077.

Full text
Abstract:
Ordinary least square is a parameter estimations for minimizing residual sum of squares. If the multicollinearity was found in the data, unbias estimator with minimum variance could not be reached. Multicollinearity is a linear correlation between independent variabels in model. Jackknife Ridge Regression(JRR) as an extension of Generalized Ridge Regression (GRR) for solving multicollinearity. Generalized Ridge Regression is used to overcome the bias of estimators caused of presents multicollinearity by adding different bias parameter for each independent variabel in least square equation after transforming the data into an orthoghonal form. Beside that, JRR can reduce the bias of the ridge estimator. The result showed that JRR model out performs GRR model.
APA, Harvard, Vancouver, ISO, and other styles
9

Shaheen, Nazia, Muhammad Nouman Qureshi, Osama Abdulaziz Alamri, and Muhammad Hanif. "Optimized inferences of finite population mean using robust parameters in systematic sampling." PLOS ONE 18, no. 1 (January 23, 2023): e0278619. http://dx.doi.org/10.1371/journal.pone.0278619.

Full text
Abstract:
In this article, we have proposed a generalized estimator for mean estimation by combining the ratio and regression methods of estimation in the presence of auxiliary information using systematic sampling. We incorporated some robust parameters of the auxiliary variable to obtain precise estimates of the proposed estimator. The mathematical expressions for bias and mean square error of proposed the estimator are derived under large sample approximation. Many other generalized ratio and product-type estimators are obtained from the proposed estimator using different choices of scalar constants. Some special cases are also discussed in which the proposed generalized estimator reduces to the usual mean, classical ratio, product, and regression type estimators. Mathematical conditions are obtained for which the proposed estimator will perform more precisely than the challenging estimators mentioned in this article. The efficiency of the proposed estimator is evaluated using four populations. Results showed that the proposed estimator is efficient and useful for survey sampling in comparison to the other existing estimators.
APA, Harvard, Vancouver, ISO, and other styles
10

SÖKÜT AÇAR, Tuğba. "Kibria-Lukman Estimator for General Linear Regression Model with AR(2) Errors: A Comparative Study with Monte Carlo Simulation." Journal of New Theory, no. 41 (December 31, 2022): 1–17. http://dx.doi.org/10.53570/jnt.1139885.

Full text
Abstract:
The sensitivity of the least-squares estimation in a regression model is impacted by multicollinearity and autocorrelation problems. To deal with the multicollinearity, Ridge, Liu, and Ridge-type biased estimators have been presented in the statistical literature. The recently proposed Kibria-Lukman estimator is one of the Ridge-type estimators. The literature has compared the Kibria-Lukman estimator with the others using the mean square error criterion for the linear regression model. It was achieved in a study conducted on the Kibria-Lukman estimator's performance under the first-order autoregressive erroneous autocorrelation. When there is an autocorrelation problem with the second-order, evaluating the performance of the Kibria-Lukman estimator according to the mean square error criterion makes this paper original. The scalar mean square error of the Kibria-Lukman estimator under the second-order autoregressive error structure was evaluated using a Monte Carlo simulation and two real examples, and compared with the Generalized Least-squares, Ridge, and Liu estimators. The findings revealed that when the variance of the model was small, the mean square error of the Kibria-Lukman estimator gave very close values with the popular biased estimators. As the model variance grew, Kibria-Lukman did not give fairly similar values with popular biased estimators as in the model with small variance. However, according to the mean square error criterion the Kibria-Lukman estimator outperformed the Generalized Least-Squares estimator in all possible cases.
APA, Harvard, Vancouver, ISO, and other styles

Dissertations / Theses on the topic "Generalised regression estimators"

1

Bae, Deok Hwan. "Models for target detection times." Thesis, Monterey, California. Naval Postgraduate School, 1989. http://hdl.handle.net/10945/27141.

Full text
Abstract:
Approved for public release; distribution in unlimited.
Some battlefield models have a component in them which models the time it takes for an observer to detect a target. Different observers may have different mean detection times due to various factors such as the type of sensor used, environmental conditions, fatigue of the observer, etc. Two parametric models for the distribution of time to target detection are considered which can incorporate these factors. Maximum likelihood estimation procedures for the parameters are described. Results of simulation experiments to study the small sample behavior of the estimators are presented.
http://archive.org/details/modelsfortargetd00baed
Major, Korean Air Force
APA, Harvard, Vancouver, ISO, and other styles
2

Santos, Helton Saulo Bezerra dos. "Essays on Birnbaum-Saunders models." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2013. http://hdl.handle.net/10183/87375.

Full text
Abstract:
Nessa tese apresentamos três diferentes aplicações dos modelos Birnbaum-Saunders. No capítulo 2 introduzimos um novo método por função-núcleo não-paramétrico para a estimação de densidades assimétricas, baseado nas distribuições Birnbaum-Saunders generalizadas assimétricas. Funções-núcleo baseadas nessas distribuições têm a vantagem de fornecer flexibilidade nos níveis de assimetria e curtose. Em adição, os estimadores da densidade por função-núcleo Birnbaum-Saunders gene-ralizadas assimétricas são livres de viés na fronteira e alcançam a taxa ótima de convergência para o erro quadrático integrado médio dos estimadores por função-núcleo-assimétricas-não-negativos da densidade. Realizamos uma análise de dados consistindo de duas partes. Primeiro, conduzimos uma simulação de Monte Carlo para avaliar o desempenho do método proposto. Segundo, usamos esse método para estimar a densidade de três dados reais da concentração de poluentes atmosféricos. Os resultados numéricos favorecem os estimadores não-paramétricos propostos. No capítulo 3 propomos uma nova família de modelos autorregressivos de duração condicional baseados nas distribuições misturas de escala Birnbaum-Saunders (SBS). A distribuição Birnbaum-Saunders (BS) é um modelo que tem recebido considerável atenção recentemente devido às suas boas propriedades. Uma extensão dessa distribuição é a classe de distribuições SBS, a qual (i) herda várias das boas propriedades da distribuição BS, (ii) permite a estimação de máxima verossimilhança em uma forma eficiente usando o algoritmo EM, e (iii) possibilita a obtenção de um procedimento de estimação robusta, entre outras propriedades. O modelo autorregressivo de duração condicional é a família primária de modelos para analisar dados de duração de transações de alta frequência. A metodologia estudada aqui inclui estimação dos parâmetros pelo algoritmo EM, inferência para esses parâmetros, modelo preditivo e uma análise residual. Realizamos simulações de Monte Carlo para avaliar o desempenho da metodologia proposta. Ainda, avalia-mos a utilidade prática dessa metodologia usando dados reais de transações financeiras da bolsa de valores de Nova Iorque. O capítulo 4 trata de índices de capacidade do processo (PCIs), os quais são ferramentas utilizadas pelas empresas para determinar a qualidade de um produto e avaliar o desempenho de seus processos de produção. Estes índices foram desenvolvidos para processos cuja característica de qualidade tem uma distribuição normal. Na prática, muitas destas ca-racterísticas não seguem esta distribuição. Nesse caso, os PCIs devem ser modificados considerando a não-normalidade. O uso de PCIs não-modificados podemlevar a resultados inadequados. De maneira a estabelecer políticas de qualidade para resolver essa inadequação, transformação dos dados tem sido proposta, bem como o uso de quantis de distribuições não-normais. Um distribuição não-normal assimétrica o qual tem tornado muito popular em tempos recentes é a distribuição Birnbaum-Saunders (BS). Propomos, desenvolvemos, implementamos e aplicamos uma metodologia baseada em PCIs para a distribuição BS. Além disso, realizamos um estudo de simulação para avaliar o desempenho da metodologia proposta. Essa metodologia foi implementada usando o software estatístico chamado R. Aplicamos essa metodologia para um conjunto de dados reais de maneira a ilustrar a sua flexibilidade e potencialidade.
In this thesis, we present three different applications of Birnbaum-Saunders models. In Chapter 2, we introduce a new nonparametric kernel method for estimating asymmetric densities based on generalized skew-Birnbaum-Saunders distributions. Kernels based on these distributions have the advantage of providing flexibility in the asymmetry and kurtosis levels. In addition, the generalized skew-Birnbaum-Saunders kernel density estimators are boundary bias free and achieve the optimal rate of convergence for the mean integrated squared error of the nonnegative asymmetric kernel density estimators. We carry out a data analysis consisting of two parts. First, we conduct a Monte Carlo simulation study for evaluating the performance of the proposed method. Second, we use this method for estimating the density of three real air pollutant concentration data sets, whose numerical results favor the proposed nonparametric estimators. In Chapter 3, we propose a new family of autoregressive conditional duration models based on scale-mixture Birnbaum-Saunders (SBS) distributions. The Birnbaum-Saunders (BS) distribution is a model that has received considerable attention recently due to its good properties. An extension of this distribution is the class of SBS distributions, which allows (i) several of its good properties to be inherited; (ii) maximum likelihood estimation to be efficiently formulated via the EM algorithm; (iii) a robust estimation procedure to be obtained; among other properties. The autoregressive conditional duration model is the primary family of models to analyze high-frequency financial transaction data. This methodology includes parameter estimation by the EM algorithm, inference for these parameters, the predictive model and a residual analysis. We carry out a Monte Carlo simulation study to evaluate the performance of the proposed methodology. In addition, we assess the practical usefulness of this methodology by using real data of financial transactions from the New York stock exchange. Chapter 4 deals with process capability indices (PCIs), which are tools widely used by companies to determine the quality of a product and the performance of their production processes. These indices were developed for processes whose quality characteristic has a normal distribution. In practice, many of these characteristics do not follow this distribution. In such a case, the PCIs must be modified considering the non-normality. The use of unmodified PCIs can lead to inadequacy results. In order to establish quality policies to solve this inadequacy, data transformation has been proposed, as well as the use of quantiles from non-normal distributions. An asymmetric non-normal distribution which has become very popular in recent times is the Birnbaum-Saunders (BS) distribution. We propose, develop, implement and apply a methodology based on PCIs for the BS distribution. Furthermore, we carry out a simulation study to evaluate the performance of the proposed methodology. This methodology has been implemented in a noncommercial and open source statistical software called R. We apply this methodology to a real data set to illustrate its flexibility and potentiality.
APA, Harvard, Vancouver, ISO, and other styles
3

Tencaliec, Patricia. "Developments in statistics applied to hydrometeorology : imputation of streamflow data and semiparametric precipitation modeling." Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAM006/document.

Full text
Abstract:
Les précipitations et les débits des cours d'eau constituent les deux variables hydrométéorologiques les plus importantes pour l'analyse des bassins versants. Ils fournissent des informations fondamentales pour la gestion intégrée des ressources en eau, telles que l’approvisionnement en eau potable, l'hydroélectricité, les prévisions d'inondations ou de sécheresses ou les systèmes d'irrigation.Dans cette thèse de doctorat sont abordés deux problèmes distincts. Le premier prend sa source dans l’étude des débits des cours d’eau. Dans le but de bien caractériser le comportement global d'un bassin versant, de longues séries temporelles de débit couvrant plusieurs dizaines d'années sont nécessaires. Cependant les données manquantes constatées dans les séries représentent une perte d'information et de fiabilité, et peuvent entraîner une interprétation erronée des caractéristiques statistiques des données. La méthode que nous proposons pour aborder le problème de l'imputation des débits se base sur des modèles de régression dynamique (DRM), plus spécifiquement, une régression linéaire multiple couplée à une modélisation des résidus de type ARIMA. Contrairement aux études antérieures portant sur l'inclusion de variables explicatives multiples ou la modélisation des résidus à partir d'une régression linéaire simple, l'utilisation des DRMs permet de prendre en compte les deux aspects. Nous appliquons cette méthode pour reconstruire les données journalières de débit à huit stations situées dans le bassin versant de la Durance (France), sur une période de 107 ans. En appliquant la méthode proposée, nous parvenons à reconstituer les débits sans utiliser d'autres variables explicatives. Nous comparons les résultats de notre modèle avec ceux obtenus à partir d'un modèle complexe basé sur les analogues et la modélisation hydrologique et d'une approche basée sur le plus proche voisin. Dans la majorité des cas, les DRMs montrent une meilleure performance lors de la reconstitution de périodes de données manquantes de tailles différentes, dans certains cas pouvant allant jusqu'à 20 ans.Le deuxième problème que nous considérons dans cette thèse concerne la modélisation statistique des quantités de précipitations. La recherche dans ce domaine est actuellement très active car la distribution des précipitations exhibe une queue supérieure lourde et, au début de cette thèse, il n'existait aucune méthode satisfaisante permettant de modéliser toute la gamme des précipitations. Récemment, une nouvelle classe de distribution paramétrique, appelée distribution généralisée de Pareto étendue (EGPD), a été développée dans ce but. Cette distribution exhibe une meilleure performance, mais elle manque de flexibilité pour modéliser la partie centrale de la distribution. Dans le but d’améliorer la flexibilité, nous développons, deux nouveaux modèles reposant sur des méthodes semiparamétriques.Le premier estimateur développé transforme d'abord les données avec la distribution cumulative EGPD puis estime la densité des données transformées en appliquant un estimateur nonparamétrique par noyau. Nous comparons les résultats de la méthode proposée avec ceux obtenus en appliquant la distribution EGPD paramétrique sur plusieurs simulations, ainsi que sur deux séries de précipitations au sud-est de la France. Les résultats montrent que la méthode proposée se comporte mieux que l'EGPD, l’erreur absolue moyenne intégrée (MIAE) de la densité étant dans tous les cas presque deux fois inférieure.Le deuxième modèle considère une distribution EGPD semiparamétrique basée sur les polynômes de Bernstein. Plus précisément, nous utilisons un mélange creuse de densités béta. De même, nous comparons nos résultats avec ceux obtenus par la distribution EGPD paramétrique sur des jeux de données simulés et réels. Comme précédemment, le MIAE de la densité est considérablement réduit, cet effet étant encore plus évident à mesure que la taille de l'échantillon augmente
Precipitation and streamflow are the two most important meteorological and hydrological variables when analyzing river watersheds. They provide fundamental insights for water resources management, design, or planning, such as urban water supplies, hydropower, forecast of flood or droughts events, or irrigation systems for agriculture.In this PhD thesis we approach two different problems. The first one originates from the study of observed streamflow data. In order to properly characterize the overall behavior of a watershed, long datasets spanning tens of years are needed. However, the quality of the measurement dataset decreases the further we go back in time, and blocks of data of different lengths are missing from the dataset. These missing intervals represent a loss of information and can cause erroneous summary data interpretation or unreliable scientific analysis.The method that we propose for approaching the problem of streamflow imputation is based on dynamic regression models (DRMs), more specifically, a multiple linear regression with ARIMA residual modeling. Unlike previous studies that address either the inclusion of multiple explanatory variables or the modeling of the residuals from a simple linear regression, the use of DRMs allows to take into account both aspects. We apply this method for reconstructing the data of eight stations situated in the Durance watershed in the south-east of France, each containing daily streamflow measurements over a period of 107 years. By applying the proposed method, we manage to reconstruct the data without making use of additional variables, like other models require. We compare the results of our model with the ones obtained from a complex approach based on analogs coupled to a hydrological model and a nearest-neighbor approach, respectively. In the majority of cases, DRMs show an increased performance when reconstructing missing values blocks of various lengths, in some of the cases ranging up to 20 years.The second problem that we approach in this PhD thesis addresses the statistical modeling of precipitation amounts. The research area regarding this topic is currently very active as the distribution of precipitation is a heavy-tailed one, and at the moment, there is no general method for modeling the entire range of data with high performance. Recently, in order to propose a method that models the full-range precipitation amounts, a new class of distribution called extended generalized Pareto distribution (EGPD) was introduced, specifically with focus on the EGPD models based on parametric families. These models provide an improved performance when compared to previously proposed distributions, however, they lack flexibility in modeling the bulk of the distribution. We want to improve, through, this aspect by proposing in the second part of the thesis, two new models relying on semiparametric methods.The first method that we develop is the transformed kernel estimator based on the EGPD transformation. That is, we propose an estimator obtained by, first, transforming the data with the EGPD cdf, and then, estimating the density of the transformed data by applying a nonparametric kernel density estimator. We compare the results of the proposed method with the ones obtained by applying EGPD on several simulated scenarios, as well as on two precipitation datasets from south-east of France. The results show that the proposed method behaves better than parametric EGPD, the MIAE of the density being in all the cases almost twice as small.A second approach consists of a new model from the general EGPD class, i.e., we consider a semiparametric EGPD based on Bernstein polynomials, more specifically, we use a sparse mixture of beta densities. Once again, we compare our results with the ones obtained by EGPD on both simulated and real datasets. As before, the MIAE of the density is considerably reduced, this effect being even more obvious as the sample size increases
APA, Harvard, Vancouver, ISO, and other styles
4

Ahmed, Mohamed Salem. "Contribution à la statistique spatiale et l'analyse de données fonctionnelles." Thesis, Lille 3, 2017. http://www.theses.fr/2017LIL30047/document.

Full text
Abstract:
Ce mémoire de thèse porte sur la statistique inférentielle des données spatiales et/ou fonctionnelles. En effet, nous nous sommes intéressés à l’estimation de paramètres inconnus de certains modèles à partir d’échantillons obtenus par un processus d’échantillonnage aléatoire ou non (stratifié), composés de variables indépendantes ou spatialement dépendantes.La spécificité des méthodes proposées réside dans le fait qu’elles tiennent compte de la nature de l’échantillon étudié (échantillon stratifié ou composé de données spatiales dépendantes).Tout d’abord, nous étudions des données à valeurs dans un espace de dimension infinie ou dites ”données fonctionnelles”. Dans un premier temps, nous étudions les modèles de choix binaires fonctionnels dans un contexte d’échantillonnage par stratification endogène (échantillonnage Cas-Témoin ou échantillonnage basé sur le choix). La spécificité de cette étude réside sur le fait que la méthode proposée prend en considération le schéma d’échantillonnage. Nous décrivons une fonction de vraisemblance conditionnelle sous l’échantillonnage considérée et une stratégie de réduction de dimension afin d’introduire une estimation du modèle par vraisemblance conditionnelle. Nous étudions les propriétés asymptotiques des estimateurs proposées ainsi que leurs applications à des données simulées et réelles. Nous nous sommes ensuite intéressés à un modèle linéaire fonctionnel spatial auto-régressif. La particularité du modèle réside dans la nature fonctionnelle de la variable explicative et la structure de la dépendance spatiale des variables de l’échantillon considéré. La procédure d’estimation que nous proposons consiste à réduire la dimension infinie de la variable explicative fonctionnelle et à maximiser une quasi-vraisemblance associée au modèle. Nous établissons la consistance, la normalité asymptotique et les performances numériques des estimateurs proposés.Dans la deuxième partie du mémoire, nous abordons des problèmes de régression et prédiction de variables dépendantes à valeurs réelles. Nous commençons par généraliser la méthode de k-plus proches voisins (k-nearest neighbors; k-NN) afin de prédire un processus spatial en des sites non-observés, en présence de co-variables spatiaux. La spécificité du prédicteur proposé est qu’il tient compte d’une hétérogénéité au niveau de la co-variable utilisée. Nous établissons la convergence presque complète avec vitesse du prédicteur et donnons des résultats numériques à l’aide de données simulées et environnementales.Nous généralisons ensuite le modèle probit partiellement linéaire pour données indépendantes à des données spatiales. Nous utilisons un processus spatial linéaire pour modéliser les perturbations du processus considéré, permettant ainsi plus de flexibilité et d’englober plusieurs types de dépendances spatiales. Nous proposons une approche d’estimation semi paramétrique basée sur une vraisemblance pondérée et la méthode des moments généralisées et en étudions les propriétés asymptotiques et performances numériques. Une étude sur la détection des facteurs de risque de cancer VADS (voies aéro-digestives supérieures)dans la région Nord de France à l’aide de modèles spatiaux à choix binaire termine notre contribution
This thesis is about statistical inference for spatial and/or functional data. Indeed, weare interested in estimation of unknown parameters of some models from random or nonrandom(stratified) samples composed of independent or spatially dependent variables.The specificity of the proposed methods lies in the fact that they take into considerationthe considered sample nature (stratified or spatial sample).We begin by studying data valued in a space of infinite dimension or so-called ”functionaldata”. First, we study a functional binary choice model explored in a case-controlor choice-based sample design context. The specificity of this study is that the proposedmethod takes into account the sampling scheme. We describe a conditional likelihoodfunction under the sampling distribution and a reduction of dimension strategy to definea feasible conditional maximum likelihood estimator of the model. Asymptotic propertiesof the proposed estimates as well as their application to simulated and real data are given.Secondly, we explore a functional linear autoregressive spatial model whose particularityis on the functional nature of the explanatory variable and the structure of the spatialdependence. The estimation procedure consists of reducing the infinite dimension of thefunctional variable and maximizing a quasi-likelihood function. We establish the consistencyand asymptotic normality of the estimator. The usefulness of the methodology isillustrated via simulations and an application to some real data.In the second part of the thesis, we address some estimation and prediction problemsof real random spatial variables. We start by generalizing the k-nearest neighbors method,namely k-NN, to predict a spatial process at non-observed locations using some covariates.The specificity of the proposed k-NN predictor lies in the fact that it is flexible and allowsa number of heterogeneity in the covariate. We establish the almost complete convergencewith rates of the spatial predictor whose performance is ensured by an application oversimulated and environmental data. In addition, we generalize the partially linear probitmodel of independent data to the spatial case. We use a linear process for disturbancesallowing various spatial dependencies and propose a semiparametric estimation approachbased on weighted likelihood and generalized method of moments methods. We establishthe consistency and asymptotic distribution of the proposed estimators and investigate thefinite sample performance of the estimators on simulated data. We end by an applicationof spatial binary choice models to identify UADT (Upper aerodigestive tract) cancer riskfactors in the north region of France which displays the highest rates of such cancerincidence and mortality of the country
APA, Harvard, Vancouver, ISO, and other styles
5

Tam, Siu Ming. "Estimation in finite population sampling : robustness and optimality." Phd thesis, 1988. http://hdl.handle.net/1885/10575.

Full text
Abstract:
Considerable progress has been made in the past two decades on research into survey sampling methodologies. A significant trend in this development is that the classical survey sampling techniques are formulated using an inference framework, which is made possible by assuming that the given finite population is a realization of a superpopulation model with unknown parameters. One attraction of this method of formulation of survey sampling problems is that not only the classical tools of inference can usefully be brought to bear on the problem, but more importantly, the optimal properties of certain familiar estimators in classical survey sampling are more and more better understood. A central issue in the use of superpopulation models for designing sampling strategies is robustness of the procedures. Under the design-based approach, it is claimed in the literature that robustness is provided by requiring that the estimators are design-unbiased, a property which is independent of the assumed superpopulation model. Chapter 1 examines the claim in detail and provide sufficient conditions for the sampling strategy to attain the Godambe-Joshi lower bound to the expected variance (Godambe & Joshi , 1965) . Chapter 2 and Chapter 3 examine the issues of reconciling design-based estimators with model-based predictors. The approach adopted in Chapter 2 is to retain the asymptotic design unbiasedness of the predictors, but modify the form of the estimators for the superpopulation model parameters in constructing the predictors of.population totals. On the other hand, the approach adopted in Chapter 3 is to retain the model based properties of the predictors but modify the selection probailites for selecting the sample. The discussion for Chapters 1 to 3 is restricted to the design and analysis of ad hoc surveys, that is, surveys which are conducted once on a given finite population. Chapter 4 and Chapter 5 examine issues related to the analysis of repeated surveys of the finite population over time. Chapter 6 looks at specification of superpopulation models for survey sampling from the point of view of performing regression analysis on the survey data from ad hoc surveys. It is argued that under certain circumstances, the uniform intra-class correlation assumed in the covariance matrix of the superpopulation model is unnecessary.
APA, Harvard, Vancouver, ISO, and other styles
6

Yang, Szu-peng, and 楊思芃. "A class of generalized ridge estimator for high-dimensional linear regression." Thesis, 2014. http://ndltd.ncl.edu.tw/handle/7aw844.

Full text
Abstract:
碩士
國立中央大學
統計研究所
102
In multiple linear regression, the least square estimator is inappropriate for high-dimensional regressors, especially for p≥n. Consider the linear regression model. The generalized ridge estimator has been considered by many authors under the usual p
APA, Harvard, Vancouver, ISO, and other styles
7

Yeh, Chi-Kuang. "Optimal regression design under second-order least squares estimator: theory, algorithm and applications." Thesis, 2018. https://dspace.library.uvic.ca//handle/1828/9765.

Full text
Abstract:
In this thesis, we first review the current development of optimal regression designs under the second-order least squares estimator in the literature. The criteria include A- and D-optimality. We then introduce a new formulation of A-optimality criterion so the result can be extended to c-optimality which has not been studied before. Following Kiefer's equivalence results, we derive the optimality conditions for A-, c- and D-optimal designs under the second-order least squares estimator. In addition, we study the number of support points for various regression models including Peleg models, trigonometric models, regular and fractional polynomial models. A generalized scale invariance property for D-optimal designs is also explored. Furthermore, we discuss one computing algorithm to find optimal designs numerically. Several interesting applications are presented and related MATLAB code are provided in the thesis.
Graduate
APA, Harvard, Vancouver, ISO, and other styles
8

Yzenbrandt, Kai. "Minimax D-optimal designs for regression models with heteroscedastic errors." Thesis, 2021. http://hdl.handle.net/1828/12863.

Full text
Abstract:
Minimax D-optimal designs for regression models with heteroscedastic errors are studied and constructed. These designs are robust against possible misspecification of the error variance in the model. We propose a flexible assumption for the error variance and use a minimax approach to define robust designs. As usual it is hard to find robust designs analytically, since the associated design problem is not a convex optimization problem. However, the minimax D-optimal design problem has an objective function as a difference of two convex functions. An effective algorithm is developed to compute minimax D-optimal designs under the least squares estimator and generalized least squares estimator. The algorithm can be applied to construct minimax D-optimal designs for any linear or nonlinear regression model with heteroscedastic errors. In addition, several theoretical results are obtained for the minimax D-optimal designs.
Graduate
APA, Harvard, Vancouver, ISO, and other styles
9

Klička, Petr. "Kalibrační odhady ve výběrových šetřeních." Master's thesis, 2018. http://www.nusl.cz/ntk/nusl-386957.

Full text
Abstract:
V této práci se zabýváme odhady populačního úhrnu s využitím pomoc- ných informací. V práci je popsán obecný regresní odhad a předpoklady, za kterých je splněna asymptotická normalita tohoto odhadu. Dále jsou zde po- psány kalibrační odhady a zmínka o jejich asymptotické ekvivalenci s obec- ným regresním odhadem. Odvozené závěry aplikujeme na data z RADIO- PROJEKTu a porovnáme je s výsledky získanými společnostmi, které tento projekt realizovali. Na závěr pomocí simulací porovnáme skutečné pravdě- podobnosti pokrytí interval· spolehlivosti pro populační úhrn spočítané na základě teorie uvedené v této práci a na základě metod společností realizu- jících RADIOPROJEKT. 1
APA, Harvard, Vancouver, ISO, and other styles
10

Mundhenk, Philip Henrich. "Integrating remotely sensed data into forest resource inventories." Doctoral thesis, 2014. http://hdl.handle.net/11858/00-1735-0000-0022-5FE6-3.

Full text
Abstract:
Die letzten zwanzig Jahre haben gezeigt, dass die Integration luftgestützter Lasertechnologien (Light Detection and Ranging; LiDAR) in die Erfassung von Waldressourcen dazu beitragen kann, die Genauigkeit von Schätzungen zu erhöhen. Um diese zu ermöglichen, müssen Feldaten mit LiDAR-Daten kombiniert werden. Diverse Techniken der Modellierung bieten die Möglichkeit, diese Verbindung statistisch zu beschreiben. Während die Wahl der Methode in der Regel nur geringen Einfluss auf Punktschätzer hat, liefert sie unterschiedliche Schätzungen der Genauigkeit. In der vorliegenden Studie wurde der Einfluss verschiedener Modellierungstechniken und Variablenauswahl auf die Genauigkeit von Schätzungen untersucht. Der Schwerpunkt der Arbeit liegt hierbei auf LiDAR Anwendungen im Rahmen von Waldinventuren. Die Methoden der Variablenauswahl, welche in dieser Studie berücksichtigt wurden, waren das Akaike Informationskriterium (AIC), das korrigierte Akaike Informationskriterium (AICc), und das bayesianische (oder Schwarz) Informationskriterium. Zudem wurden Variablen anhand der Konditionsnummer und des Varianzinflationsfaktors ausgewählt. Weitere Methoden, die in dieser Studie Berücksichtigung fanden, umfassen Ridge Regression, der least absolute shrinkage and selection operator (Lasso), und der Random Forest Algorithmus. Die Methoden der schrittweisen Variablenauswahl wurden sowohl im Rahmen der Modell-assistierten als auch der Modell-basierten Inferenz untersucht. Die übrigen Methoden wurden nur im Rahmen der Modell-assistierten Inferenz untersucht. In einer umfangreichen Simulationsstudie wurden die Einflüsse der Art der Modellierungsmethode und Art der Variablenauswahl auf die Genauigkeit der Schätzung von Populationsparametern (oberirdische Biomasse in Megagramm pro Hektar) ermittelt. Hierzu wurden fünf unterschiedliche Populationen genutzt. Drei künstliche Populationen wurden simuliert, zwei weitere basierten auf in Kanada und Norwegen erhobenen Waldinveturdaten. Canonical vine copulas wurden genutzt um synthetische Populationen aus diesen Waldinventurdaten zu generieren. Aus den Populationen wurden wiederholt einfache Zufallsstichproben gezogen und für jede Stichprobe wurden der Mittelwert und die Genauigkeit der Mittelwertschätzung geschäzt. Während für das Modell-basierte Verfahren nur ein Varianzschätzer untersucht wurde, wurden für den Modell-assistierten Ansatz drei unterschiedliche Schätzer untersucht. Die Ergebnisse der Simulationsstudie zeigten, dass das einfache Anwenden von schrittweisen Methoden zur Variablenauswahl generell zur Überschätzung der Genauigkeiten in LiDAR unterstützten Waldinventuren führt. Die verzerrte Schätzung der Genauigkeiten war vor allem für kleine Stichproben (n = 40 und n = 50) von Bedeutung. Für Stichproben von größerem Umfang (n = 400), war die Überschätzung der Genauigkeit vernachlässigbar. Gute Ergebnisse, im Hinblick auf Deckungsraten und empirischem Standardfehler, zeigten Ridge Regression, Lasso und der Random Forest Algorithmus. Aus den Ergebnissen dieser Studie kann abgeleitet werden, dass die zuletzt genannten Methoden in zukünftige LiDAR unterstützten Waldinventuren Berücksichtigung finden sollten.
APA, Harvard, Vancouver, ISO, and other styles

Books on the topic "Generalised regression estimators"

1

Linton, Oliver. Edgeworth approximation for generalised minimum contrast estimators in semiparametric regression models. Oxford: Nuffield College, 1992.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
2

Flynn, Robert H. Generalized estimates from streamflow data of annual and seasonal ground-water-recharge rates for drainage basins in New Hampshire. Pembroke, N.H: U.S. Dept. of the Interior, U.S. Geological Survey, 2004.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
3

Flynn, Robert H. Generalized estimates from streamflow data of annual and seasonal ground-water-recharge rates for drainage basins in New Hampshire. Pembroke, N.H: U.S. Dept. of the Interior, U.S. Geological Survey, 2004.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
4

Flynn, Robert H. Generalized estimates from streamflow data of annual and seasonal ground-water-recharge rates for drainage basins in New Hampshire. Pembroke, N.H: U.S. Dept. of the Interior, U.S. Geological Survey, 2004.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
5

Flynn, Robert H. Generalized estimates from streamflow data of annual and seasonal ground-water-recharge rates for drainage basins in New Hampshire. Pembroke, N.H: U.S. Dept. of the Interior, U.S. Geological Survey, 2004.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
6

Cardot, Hervé, and Pascal Sarda. Functional Linear Regression. Edited by Frédéric Ferraty and Yves Romain. Oxford University Press, 2018. http://dx.doi.org/10.1093/oxfordhb/9780199568444.013.2.

Full text
Abstract:
This article presents a selected bibliography on functional linear regression (FLR) and highlights the key contributions from both applied and theoretical points of view. It first defines FLR in the case of a scalar response and shows how its modelization can also be extended to the case of a functional response. It then considers two kinds of estimation procedures for this slope parameter: projection-based estimators in which regularization is performed through dimension reduction, such as functional principal component regression, and penalized least squares estimators that take into account a penalized least squares minimization problem. The article proceeds by discussing the main asymptotic properties separating results on mean square prediction error and results on L2 estimation error. It also describes some related models, including generalized functional linear models and FLR on quantiles, and concludes with a complementary bibliography and some open problems.
APA, Harvard, Vancouver, ISO, and other styles
7

D, Tasker Gary, New Hampshire. Dept. of Environmental Services., and Geological Survey (U.S.), eds. Generalized estimates from streamflow data of annual and seasonal ground-water-recharge rates for drainage basins in New Hampshire. Pembroke, N.H: U.S. Dept. of the Interior, U.S. Geological Survey, 2004.

Find full text
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "Generalised regression estimators"

1

Bartels, Robert, and Denzil G. Fiebig. "Efficiency of Alternative Estimators in Generalized Seemingly Unrelated Regression Models." In Contributions to Consumer Demand and Econometrics, 125–39. London: Palgrave Macmillan UK, 1992. http://dx.doi.org/10.1007/978-1-349-12221-9_7.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Taylan, Pakize, and Gerhard Wilhelm Weber. "C-LASSO Estimator for Generalized Additive Logistic Regression Based on B-Spline." In Data Science and Digital Business, 173–90. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-319-95651-0_10.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Sengupta, Ramprasad. "Crime, Inequality, and Poverty." In Entropy Law, Sustainability, and Third Industrial Revolution, 70–100. Oxford University Press, 2020. http://dx.doi.org/10.1093/oso/9780190121143.003.0004.

Full text
Abstract:
Chapter 3 enquires into the issues of the social sustainability aspect of development by analyzing the inter-relationships between crime and deprivation, and the resulting social tension in the Indian context. It estimates the relationship of violent crimes such as homicide and property-related crime—dacoity, burglary and robbery, riots and left-wing extremism—with economic inequality or poverty, or social tension induced by either of these, along with other developmental variables such as state of education, infrastructural development, urbanization, and so on. It has used panel data for both simple correlation analysis as well as multivariate regression analysis (generalised moments method) in its different dimensions of use and treatment. The analysis of riots also considered the role of share of minority population, share of SC/ST population, as well as religious polarization as important social explanatory variables.
APA, Harvard, Vancouver, ISO, and other styles
4

Kiziltan, Mustafa. "The Effects of Population Aging and Life Expectancy on Economic Growth." In Advances in Human Services and Public Health, 97–118. IGI Global, 2021. http://dx.doi.org/10.4018/978-1-7998-7327-3.ch007.

Full text
Abstract:
This study examines the effects of gross saving rates, gross capital formation, population aging, and life expectancy on income growth between 1985 and 2018 for selected emerging market and middle-income economies. The estimates by feasible generalized least squares (FGLS), fully modified ordinary least squares (FMOLS), dynamic ordinary least squares (DOLS), canonical cointegrating regressions (CCR), and Driscoll-Kraay methods show the impact of population aging on income growth is positive. However, life expectancy gives positive results for gross domestic product per capita growth and negative results for gross domestic product growth. The results confirm that there is still an increasing population for emerging economies in general. In this respect, they also point to the neutral view approach for these countries. This situation highlights the importance of considering the effects of population aging on sustainable economic growth in emerging countries. For this reason, decision makers in these economies must monitor population structures closely and carefully.
APA, Harvard, Vancouver, ISO, and other styles
5

Andrews, Martyn, Alastair R. Hall, Rabeya Khatoon, and James Lincoln. "Info-metric Methods for the Estimation of Models with Group-Specific Moment Conditions." In Advances in Info-Metrics, 349–84. Oxford University Press, 2020. http://dx.doi.org/10.1093/oso/9780190636685.003.0013.

Full text
Abstract:
Motivated by empirical analyses in economics using repeated cross-sectional data, we propose info-metric methods (IM) for estimation of the parameters of statistical models based on the information in population moment conditions that hold at group level. The info-metric estimation can be viewed as the primary approach to a constrained optimization. The estimators can also be obtained via the dual approach to this optimization, known as generalized empirical likelihood (GEL). In Andrews, Hall, Khatoon and Lincoln (2019), we provide a comprehensive framework for inference based on GEL with the grouped-specific moment conditions. In this chapter, we compare the computational requirements of the primary and dual approaches. We also describe the IM/GEL inference framework in the context of a linear regression model that is estimated using the information that the mean of the error is zero for each group. For the latter setting, we use analytical arguments and a small simulation study to compare the properties of IM/GEL-based inferences to those of inferences based on certain extant methods. The IM/GEL methods are illustrated through an application to estimation of the returns to education in which the groups are defined via information on family background.
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Generalised regression estimators"

1

Kolansky, Jeremy, and Corina Sandu. "Generalized Polynomial Chaos-Based Extended Kalman Filter: Improvement and Expansion." In ASME 2013 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. American Society of Mechanical Engineers, 2013. http://dx.doi.org/10.1115/detc2013-12082.

Full text
Abstract:
The generalized polynomial chaos (gPC) method for propagating uncertain parameters through dynamical systems (previously developed at Virginia Tech) has been shown to be very computationally efficient. This method seems also to be ideal for real-time parameter estimation when merged with the Extended Kalman Filter (EKF). The resulting technique is shown in the present paper for systems in state-space representations, and then expanded to systems in regressions formulations. Due to the way the filter interacts with the polynomial chaos expansions, the covariance matrix is forced to zero in finite time. This problem shows itself as an inability to perform state estimations and causes the parameters to converge to incorrect values for state space systems. In order to address this issue, improvements to the method are implemented and the updated method is applied to both state space and regression systems. The resultant technique shows high accuracy of both state and parameter estimations.
APA, Harvard, Vancouver, ISO, and other styles
2

Hage, Ilige S., Charbel Y. Seif, Ré-Mi Hage, and Ramsey F. Hamade. "A Verified Non-Linear Regression Model for Elastic Stiffness Estimates of Finite Composite Domains Considering Combined Effects of Volume Fractions, Shapes, Orientations, Locations, and Number of Multiple Inclusions." In ASME 2018 International Mechanical Engineering Congress and Exposition. American Society of Mechanical Engineers, 2018. http://dx.doi.org/10.1115/imece2018-86231.

Full text
Abstract:
A non-linear regression model using SAS/STAT (JMP® software; Proc regression module) is developed for estimating the elastic stiffness of finite composite domains considering the combined effects of volume fractions, shapes, orientations, inclusion locations, and number of multiple inclusions. These estimates are compared to numerical solutions that utilized another developed homogenization methodology by the authors (dubbed the generalized stiffness formulation, GSF) to numerically determine the elastic stiffness tensor of a composite domain having multiple inclusions with various combinations of geometric attributes. For each inclusion, these considered variables represent the inclusions’ combined attributes of volume fraction, aspect ratio, orientation, number of inclusions, and their locations. The GSF methodology’s solutions were compared against literature-reported solutions of simple cases according to such well-known techniques as Mori-Tanaka and generalized self-consistent type methods. In these test cases, the effect of only one variable was considered at a time: volume fraction, aspect ratio, or orientation (omitting the number and locations of inclusions). For experimental corroboration of the numerical solutions, testing (uniaxial compression) was performed on test cases of 3D printed test cubes. The regression equation returns estimates of the composite’s ratio of normalized longitudinal modulus (E11) to that of the matrix modulus (Em) or E11/Em when considering any combination of all of the aforementioned inclusions’ variables. All parameters were statistically analyzed with the parameters retained are only those deemed statistically significant (p-values less than 0.05). Values returned by the regression stiffness formulation solutions were compared against values returned by the GSF formulation numerical and against the experimentally found stiffness values. Results show good agreement between the regression model estimates as compared with both numerical and experimental results.
APA, Harvard, Vancouver, ISO, and other styles
3

Vanem, Erik, and Sam-Erik Walker. "Time Series Analysis of Significant Wave Height Data for Identification of Trends in the Ocean Wave Climate." In ASME 2013 32nd International Conference on Ocean, Offshore and Arctic Engineering. American Society of Mechanical Engineers, 2013. http://dx.doi.org/10.1115/omae2013-10024.

Full text
Abstract:
Reliable return period estimates of sea state parameters such as the significant wave height is of great importance in marine structural design and ocean engineering. Hence, time series of significant wave height have been extensively studied in recent years. However, with the possibility of an ongoing change in the global climate, this might influence the ocean wave climate as well and it would be of great interest to analyze long time series to see if any long-term trends can be detected. In this paper, long time series of significant wave height stemming from the ERA-40 reanalysis project, containing 6-hourly data over a period of more than 44 years are investigated with the purpose of identifying long term trends. Different time series analysis methods are employed, i.e. seasonal ARIMA, multiple linear regression, the Theil-Sen estimator and generalized additive models, and the results are discussed. These results are then compared to previous studies; in particular results are compared to a recent study where a spatio-temporal stochastic model was applied to the same data. However, in the current analysis, the spatial dimension has been reduced and spatial minima, mean and maxima have been analysed for temporal trends. Overall, increasing trends in the wave climate have been identified by most of the modelling approaches explored in the paper, although some of the trends are not statistically significant at the 95% level. Based on the results presented in this paper, it may be argued that there is evidence of a roughening trend in the recent ocean wave climate, and more detailed analyses of individual months and seasons indicate that these trends might be mostly due to trends during the winter months.
APA, Harvard, Vancouver, ISO, and other styles
4

Zhou, Peng, Ligang Lu, Huiyan Sang, and Birol Dindoruk. "Application of Machine Learning Methods to Well Completion Optimization: Problems with Groups of Interactive Inputs." In SPE Annual Technical Conference and Exhibition. SPE, 2021. http://dx.doi.org/10.2118/206174-ms.

Full text
Abstract:
Abstract In unconventional reservoirs, optimal completion controls are essential to improving well productivity and reducing costs. In this article, we propose a statistical model to investigate associations between shale oil production and completion parameters (e.g., completion lateral length, total proppant, number of hydraulic fracturing stages), while accounting for the influence of spatially heterogeneous geological conditions on hydrocarbon production. We develop a non-parametric regression method that combines a generalized additive model with a fused LASSO regularization for geological homogeneity pursuit. We present an alternating augmented Lagrangian method for model parameter estimations. The novelty and advantages of our method over the published ones are a) it can control or remove the heterogeneous non-completion effects; 2) it can account for and analyze the interactions among the completion parameters. We apply our method to the analysis of a real case from a Permian Basin US onshore field and show how our model can account for the interaction between the completion parameters. Our results provide key findings on how completion parameters affect oil production in that can lead to optimal well completion designs.
APA, Harvard, Vancouver, ISO, and other styles
5

Banerjee, Ashis Gopal, Walter Yund, Dan Yang, Peter Koudal, John Carbone, and Joseph Salvo. "A Hybrid Statistical Method for Accurate Prediction of Supplier Delivery Times of Aircraft Engine Parts." In ASME 2015 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. American Society of Mechanical Engineers, 2015. http://dx.doi.org/10.1115/detc2015-47605.

Full text
Abstract:
Aircraft engine assembly operations require thousands of parts provided by several geographically distributed suppliers. A majority of the operation steps are sequential, necessitating the availability of all the parts at appropriate times for these steps to be completed successfully. Thus, being able to accurately predict the availabilities of parts based on supplier deliveries is critical to minimizing the delays in meeting the customer demands. However, such accurate prediction is challenging due to the large lead times of these parts, limited knowledge of supplier capacities and capabilities, macroeconomic trends affecting material procurement and transportation times, and unreliable delivery date estimates provided by the suppliers themselves. We address these challenges by developing a statistical method that learns a hybrid stepwise regression — generalized multivariate gamma distribution model from historical transactional data on closed part purchase orders and is able to infer part delivery dates sufficiently before the supplier-promised delivery dates for open purchase orders. The hybrid form of the model makes it robust to data quality and short-term temporal effects as well as biased toward overestimating rather than underestimating the part delivery dates. Test results on real-world purchase orders demonstrate effective performance with low prediction errors and constantly high ratios of true positive to false positive predictions.
APA, Harvard, Vancouver, ISO, and other styles
6

Randell, David, Elena Zanini, Michael Vogel, Kevin Ewans, and Philip Jonathan. "Omnidirectional Return Values for Storm Severity From Directional Extreme Value Models: The Effect of Physical Environment and Sample Size." In ASME 2014 33rd International Conference on Ocean, Offshore and Arctic Engineering. American Society of Mechanical Engineers, 2014. http://dx.doi.org/10.1115/omae2014-23156.

Full text
Abstract:
Ewans and Jonathan [2008] shows that characteristics of extreme storm severity in the northern North Sea vary with storm direction. Jonathan et al. [2008] demonstrates, when directional effects are present, that omnidirectional return values should be estimated using a directional extreme value model. Omnidirectional return values so calculated are different in general to those estimated using a model which incorrectly assumes stationarity with respect to direction. The extent of directional variability of extreme storm severity depends on a number of physical factors, including fetch variability. Our ability to assess directional variability of extreme value parameters and return values also improves with increasing sample size in general. In this work, we estimate directional extreme value models for samples of hindcast storm peak significant wave height from locations in ocean basins worldwide, for a range of physical environments, sample sizes and periods of observation. At each location, we compare distributions of omnidirectional 100-year return values estimated using a directional model, to those (incorrectly) estimated assuming stationarity. The directional model for peaks over threshold of storm peak significant wave height is estimated using a non-homogeneous point process model as outlined in Randell et al. [2013]. Directional models for extreme value threshold (using quantile regression), rate of occurrence of threshold exceedances (using a Poisson model), and size of exceedances (using a generalised Pareto model) are estimated. Model parameters are described as smooth functions of direction using periodic B-splines. Parameter estimation is performed using maximum likelihood estimation penalised for parameter roughness. A bootstrap re-sampling procedure, encompassing all inference steps, quantifies uncertainties in, and dependence structure of, parameter estimates and omnidirectional return values.
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Generalised regression estimators"

1

Over, Thomas, Riki Saito, Andrea Veilleux, Padraic O’Shea, Jennifer Sharpe, David Soong, and Audrey Ishii. Estimation of Peak Discharge Quantiles for Selected Annual Exceedance Probabilities in Northeastern Illinois. Illinois Center for Transportation, June 2016. http://dx.doi.org/10.36501/0197-9191/16-014.

Full text
Abstract:
This report provides two sets of equations for estimating peak discharge quantiles at annual exceedance probabilities (AEPs) of 0.50, 0.20, 0.10, 0.04, 0.02, 0.01, 0.005, and 0.002 (recurrence intervals of 2, 5, 10, 25, 50, 100, 200, and 500 years, respectively) for watersheds in Illinois based on annual maximum peak discharge data from 117 watersheds in and near northeastern Illinois. One set of equations was developed through a temporal analysis with a two-step least squares-quantile regression technique that measures the average effect of changes in the urbanization of the watersheds used in the study. The resulting equations can be used to adjust rural peak discharge quantiles for the effect of urbanization, and in this study the equations also were used to adjust the annual maximum peak discharges from the study watersheds to 2010 urbanization conditions. The other set of equations was developed by a spatial analysis. This analysis used generalized least-squares regression to fit the peak discharge quantiles computed from the urbanization-adjusted annual maximum peak discharges from the study watersheds to drainage-basin characteristics. The peak discharge quantiles were computed by using the Expected Moments Algorithm following the removal of potentially influential low floods defined by a multiple Grubbs-Beck test. To improve the quantile estimates, regional skew coefficients were obtained from a newly developed regional skew model in which the skew increases with the urbanized land use fraction. The skew coefficient values for each streamgage were then computed as the variance-weighted average of at-site and regional skew coefficients. The drainage-basin characteristics used as explanatory variables in the spatial analysis include drainage area, the fraction of developed land, the fraction of land with poorly drained soils or likely water, and the basin slope estimated as the ratio of the basin relief to basin perimeter. This report also provides: (1) examples to illustrate the use of the spatial and urbanization-adjustment equations for estimating peak discharge quantiles at ungaged sites and to improve flood-quantile estimates at and near a gaged site; (2) the urbanization-adjusted annual maximum peak discharges and peak discharge quantile estimates at streamgages from 181 watersheds including the 117 study watersheds and 64 additional watersheds in the study region that were originally considered for use in the study but later deemed to be redundant. The urbanization-adjustment equations, spatial regression equations, and peak discharge quantile estimates developed in this study will be made available in the web-based application StreamStats, which provides automated regression-equation solutions for user-selected stream locations. Figures and tables comparing the observed and urbanization-adjusted peak discharge records by streamgage are provided at http://dx.doi.org/10.3133/sir20165050 for download.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!