Literatura científica selecionada sobre o tema "Robbins-Monro algorithm"
Crie uma referência precisa em APA, MLA, Chicago, Harvard, e outros estilos
Consulte a lista de atuais artigos, livros, teses, anais de congressos e outras fontes científicas relevantes para o tema "Robbins-Monro algorithm".
Ao lado de cada fonte na lista de referências, há um botão "Adicionar à bibliografia". Clique e geraremos automaticamente a citação bibliográfica do trabalho escolhido no estilo de citação de que você precisa: APA, MLA, Harvard, Chicago, Vancouver, etc.
Você também pode baixar o texto completo da publicação científica em formato .pdf e ler o resumo do trabalho online se estiver presente nos metadados.
Artigos de revistas sobre o assunto "Robbins-Monro algorithm"
Moler, José A., Fernando Plo e Miguel San Miguel. "Adaptive designs and Robbins–Monro algorithm". Journal of Statistical Planning and Inference 131, n.º 1 (abril de 2005): 161–74. http://dx.doi.org/10.1016/j.jspi.2003.12.018.
Texto completo da fonteWardi, Y. "On a proof of a Robbins-Monro algorithm". Journal of Optimization Theory and Applications 64, n.º 1 (janeiro de 1990): 217. http://dx.doi.org/10.1007/bf00940033.
Texto completo da fonteLin, Siming, e Jennie Si. "Weight-Value Convergence of the SOM Algorithm for Discrete Input". Neural Computation 10, n.º 4 (1 de maio de 1998): 807–14. http://dx.doi.org/10.1162/089976698300017485.
Texto completo da fonteMoser, Barry Kurt, e Melinda H. McCann. "Algorithm AS 316: A Robbins-Monro-based Sequential Procedure". Journal of the Royal Statistical Society: Series C (Applied Statistics) 46, n.º 3 (1997): 388–99. http://dx.doi.org/10.1111/1467-9876.00078.
Texto completo da fonteEl Moumen, AbdelKader, Salim Benslimane e Samir Rahmani. "Robbins–Monro Algorithm with $$\boldsymbol{\psi}$$-Mixing Random Errors". Mathematical Methods of Statistics 31, n.º 3 (setembro de 2022): 105–19. http://dx.doi.org/10.3103/s1066530722030024.
Texto completo da fonteXU, ZI, YINGYING LI e XINGFANG ZHAO. "SIMULATION-BASED OPTIMIZATION BY NEW STOCHASTIC APPROXIMATION ALGORITHM". Asia-Pacific Journal of Operational Research 31, n.º 04 (agosto de 2014): 1450026. http://dx.doi.org/10.1142/s0217595914500262.
Texto completo da fonteCai, Li. "Metropolis-Hastings Robbins-Monro Algorithm for Confirmatory Item Factor Analysis". Journal of Educational and Behavioral Statistics 35, n.º 3 (junho de 2010): 307–35. http://dx.doi.org/10.3102/1076998609353115.
Texto completo da fonteChen, Han-Fu. "Stochastic approximation with non-additive measurement noise". Journal of Applied Probability 35, n.º 2 (junho de 1998): 407–17. http://dx.doi.org/10.1239/jap/1032192856.
Texto completo da fonteChen, Han-Fu. "Stochastic approximation with non-additive measurement noise". Journal of Applied Probability 35, n.º 02 (junho de 1998): 407–17. http://dx.doi.org/10.1017/s0021900200015035.
Texto completo da fonteBuckland, S. T., e P. H. Garthwaite. "Algorithm AS 259: Estimating Confidence Intervals by the Robbins-Monro Search Process". Applied Statistics 39, n.º 3 (1990): 413. http://dx.doi.org/10.2307/2347401.
Texto completo da fonteTeses / dissertações sobre o assunto "Robbins-Monro algorithm"
Lu, Wei. "Μéthοdes stοchastiques du secοnd οrdre pοur le traitement séquentiel de dοnnées massives". Electronic Thesis or Diss., Normandie, 2024. http://www.theses.fr/2024NORMIR13.
Texto completo da fonteWith the rapid development of technologies and the acquisition of big data, methods capable of processing data sequentially (online) have become indispensable. Among these methods, stochastic gradient algorithms have been established for estimating the minimizer of a function expressed as the expectation of a random function. Although they have become essential, these algorithms encounter difficulties when the problem is ill-conditioned. In this thesis, we focus on second-order stochastic algorithms, such as those of the Newton type, and their applications to various statistical and optimization problems. After establishing theoretical foundations and exposing the motivations that lead us to explore stochastic Newton algorithms, we develop the various contributions of this thesis. The first contribution concerns the study and development of stochastic Newton algorithms for ridge linear regression and ridge logistic regression. These algorithms are based on the Riccati formula (Sherman-Morrison) to recursively estimate the inverse of the Hessian. As the acquisition of big data is generally accompanied by a contamination of the latter, in a second contribution, we focus on the online estimation of the geometric median, which is a robust indicator, i.e., not very sensitive to the presence of atypical data. More specifically, we propose a new stochastic Newton estimator to estimate the geometric median. In the first two contributions, the estimators of the Hessians' inverses are constructed using the Riccati formula, but this is only possible for certain functions. Thus, our third contribution introduces a new Robbins-Monro type method for online estimation of the Hessian's inverse, allowing us then to develop universal stochastic Newton algorithms. Finally, our last contribution focuses on Full Adagrad type algorithms, where the difficulty lies in the fact that there is an adaptive step based on the square root of the inverse of the gradient's covariance. We thus propose a Robbins-Monro type algorithm to estimate this matrix, allowing us to propose a recursive approach for Full AdaGrad and its streaming version, with reduced computational costs. For all the new estimators we propose, we establish their convergence rates as well as their asymptotic efficiency. Moreover, we illustrate the efficiency of these algorithms using numerical simulations and by applying them to real data
Stanley, Leanne M. "Flexible Multidimensional Item Response Theory Models Incorporating Response Styles". The Ohio State University, 2017. http://rave.ohiolink.edu/etdc/view?acc_num=osu1494316298549437.
Texto completo da fonteArouna, Bouhari. "Méthodes de Monté Carlo et algorithmes stochastiques". Marne-la-vallée, ENPC, 2004. https://pastel.archives-ouvertes.fr/pastel-00001269.
Texto completo da fonteHajji, Kaouther. "Accélération de la méthode de Monte Carlo pour des processus de diffusions et applications en Finance". Thesis, Paris 13, 2014. http://www.theses.fr/2014PA132054/document.
Texto completo da fonteIn this thesis, we are interested in studying the combination of variance reduction methods and complexity improvement of the Monte Carlo method. In the first part of this thesis,we consider a continuous diffusion model for which we construct an adaptive algorithm by applying importance sampling to Statistical Romberg method. Then, we prove a central limit theorem of Lindeberg-Feller type for this algorithm. In the same setting and in the same spirit, we apply the importance sampling to the Multilevel Monte Carlo method. We also prove a central limit theorem for the obtained adaptive algorithm. In the second part of this thesis, we develop the same type of adaptive algorithm for a discontinuous model namely the Lévy processes and we prove the associated central limit theorem. Numerical simulations are processed for the different obtained algorithms in both settings with and without jumps
Trabalhos de conferências sobre o assunto "Robbins-Monro algorithm"
Ram, S. Sundhar, V. V. Veeravalli e A. Nedic. "Incremental Robbins-Monro Gradient Algorithm for Regression in Sensor Networks". In 2007 2nd IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing. IEEE, 2007. http://dx.doi.org/10.1109/camsap.2007.4498027.
Texto completo da fonteIooss, Bertrand, e Jérôme Lonchampt. "Robust Tuning of Robbins-Monro Algorithm for Quantile Estimation -- Application to Wind-Farm Asset Management". In Proceedings of the 31st European Safety and Reliability Conference. Singapore: Research Publishing Services, 2021. http://dx.doi.org/10.3850/978-981-18-2016-8_084-cd.
Texto completo da fonte