Um die anderen Arten von Veröffentlichungen zu diesem Thema anzuzeigen, folgen Sie diesem Link: Convergence of Markov processes.

Zeitschriftenartikel zum Thema „Convergence of Markov processes“

Geben Sie eine Quelle nach APA, MLA, Chicago, Harvard und anderen Zitierweisen an

Wählen Sie eine Art der Quelle aus:

Machen Sie sich mit Top-50 Zeitschriftenartikel für die Forschung zum Thema "Convergence of Markov processes" bekannt.

Neben jedem Werk im Literaturverzeichnis ist die Option "Zur Bibliographie hinzufügen" verfügbar. Nutzen Sie sie, wird Ihre bibliographische Angabe des gewählten Werkes nach der nötigen Zitierweise (APA, MLA, Harvard, Chicago, Vancouver usw.) automatisch gestaltet.

Sie können auch den vollen Text der wissenschaftlichen Publikation im PDF-Format herunterladen und eine Online-Annotation der Arbeit lesen, wenn die relevanten Parameter in den Metadaten verfügbar sind.

Sehen Sie die Zeitschriftenartikel für verschiedene Spezialgebieten durch und erstellen Sie Ihre Bibliographie auf korrekte Weise.

1

Abakuks, A., S. N. Ethier und T. G. Kurtz. „Markov Processes: Characterization and Convergence.“ Biometrics 43, Nr. 2 (Juni 1987): 484. http://dx.doi.org/10.2307/2531839.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
2

Perkins, Edwin, S. N. Ethier und T. G. Kurtz. „Markov Processes, Characterization and Convergence.“ Journal of the Royal Statistical Society. Series A (Statistics in Society) 151, Nr. 2 (1988): 367. http://dx.doi.org/10.2307/2982773.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
3

Franz, Uwe, Volkmar Liebscher und Stefan Zeiser. „Piecewise-Deterministic Markov Processes as Limits of Markov Jump Processes“. Advances in Applied Probability 44, Nr. 3 (September 2012): 729–48. http://dx.doi.org/10.1239/aap/1346955262.

Der volle Inhalt der Quelle
Annotation:
A classical result about Markov jump processes states that a certain class of dynamical systems given by ordinary differential equations are obtained as the limit of a sequence of scaled Markov jump processes. This approach fails if the scaling cannot be carried out equally across all entities. In the present paper we present a convergence theorem for such an unequal scaling. In contrast to an equal scaling the limit process is not purely deterministic but still possesses randomness. We show that these processes constitute a rich subclass of piecewise-deterministic processes. Such processes apply in molecular biology where entities often occur in different scales of numbers.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
4

Franz, Uwe, Volkmar Liebscher und Stefan Zeiser. „Piecewise-Deterministic Markov Processes as Limits of Markov Jump Processes“. Advances in Applied Probability 44, Nr. 03 (September 2012): 729–48. http://dx.doi.org/10.1017/s0001867800005851.

Der volle Inhalt der Quelle
Annotation:
A classical result about Markov jump processes states that a certain class of dynamical systems given by ordinary differential equations are obtained as the limit of a sequence of scaled Markov jump processes. This approach fails if the scaling cannot be carried out equally across all entities. In the present paper we present a convergence theorem for such an unequal scaling. In contrast to an equal scaling the limit process is not purely deterministic but still possesses randomness. We show that these processes constitute a rich subclass of piecewise-deterministic processes. Such processes apply in molecular biology where entities often occur in different scales of numbers.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
5

HWANG, CHII-RUEY. „ACCELERATING MONTE CARLO MARKOV PROCESSES“. COSMOS 01, Nr. 01 (Mai 2005): 87–94. http://dx.doi.org/10.1142/s0219607705000085.

Der volle Inhalt der Quelle
Annotation:
Let π be a probability density proportional to exp - U(x) in S. A convergent Markov process to π(x) may be regarded as a "conceptual" algorithm. Assume that S is a finite set. Let X0,X1,…,Xn,… be a Markov chain with transition matrix P and invariant probability π. Under suitable condition on P, it is known that [Formula: see text] converges to π(f) and the corresponding asymptotic variance v(f, P) depends only on f and P. It is natural to consider criteria vw(P) and va(P), defined respectively by maximizing and averaging v(f, P) over f. Two families of transition matrices are considered. There are four problems to be investigated. Some results and conjectures are given. As for the continuum case, to accelerate the convergence a family of diffusions with drift ∇U(x) + C(x) with div(C(x)exp - U(x)) = 0 is considered.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
6

Aldous, David J. „Book Review: Markov processes: Characterization and convergence“. Bulletin of the American Mathematical Society 16, Nr. 2 (01.04.1987): 315–19. http://dx.doi.org/10.1090/s0273-0979-1987-15533-9.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
7

Swishchuk, Anatoliy, und M. Shafiqul Islam. „Diffusion Approximations of the Geometric Markov Renewal Processes and Option Price Formulas“. International Journal of Stochastic Analysis 2010 (19.12.2010): 1–21. http://dx.doi.org/10.1155/2010/347105.

Der volle Inhalt der Quelle
Annotation:
We consider the geometric Markov renewal processes as a model for a security market and study this processes in a diffusion approximation scheme. Weak convergence analysis and rates of convergence of ergodic geometric Markov renewal processes in diffusion scheme are presented. We present European call option pricing formulas in the case of ergodic, double-averaged, and merged diffusion geometric Markov renewal processes.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
8

Crank, Keith N., und Prem S. Puri. „A method of approximating Markov jump processes“. Advances in Applied Probability 20, Nr. 1 (März 1988): 33–58. http://dx.doi.org/10.2307/1427269.

Der volle Inhalt der Quelle
Annotation:
We present a method of approximating Markov jump processes which was used by Fuhrmann [7] in a special case. We generalize the method and prove weak convergence results under mild assumptions. In addition we obtain bounds on the rates of convergence of the probabilities at arbitrary fixed times. The technique is demonstrated using a state-dependent branching process as an example.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
9

Crank, Keith N., und Prem S. Puri. „A method of approximating Markov jump processes“. Advances in Applied Probability 20, Nr. 01 (März 1988): 33–58. http://dx.doi.org/10.1017/s0001867800017936.

Der volle Inhalt der Quelle
Annotation:
We present a method of approximating Markov jump processes which was used by Fuhrmann [7] in a special case. We generalize the method and prove weak convergence results under mild assumptions. In addition we obtain bounds on the rates of convergence of the probabilities at arbitrary fixed times. The technique is demonstrated using a state-dependent branching process as an example.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
10

Deng, Chang-Song, René L. Schilling und Yan-Hong Song. „Subgeometric rates of convergence for Markov processes under subordination“. Advances in Applied Probability 49, Nr. 1 (März 2017): 162–81. http://dx.doi.org/10.1017/apr.2016.83.

Der volle Inhalt der Quelle
Annotation:
Abstract We are interested in the rate of convergence of a subordinate Markov process to its invariant measure. Given a subordinator and the corresponding Bernstein function (Laplace exponent), we characterize the convergence rate of the subordinate Markov process; the key ingredients are the rate of convergence of the original process and the (inverse of the) Bernstein function. At a technical level, the crucial point is to bound three types of moment (subexponential, algebraic, and logarithmic) for subordinators as time t tends to ∞. We also discuss some concrete models and we show that subordination can dramatically change the speed of convergence to equilibrium.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
11

Champagnat, Nicolas, und Denis Villemonais. „Uniform convergence of penalized time-inhomogeneous Markov processes“. ESAIM: Probability and Statistics 22 (2018): 129–62. http://dx.doi.org/10.1051/ps/2017022.

Der volle Inhalt der Quelle
Annotation:
We provide a general criterion ensuring the exponential contraction of Feynman–Kac semi-groups of penalized processes. This criterion applies to time-inhomogeneous Markov processes with absorption and killing through penalization. We also give the asymptotic behavior of the expected penalization and provide results of convergence in total variation of the process penalized up to infinite time. For exponential convergence of penalized semi-groups with bounded penalization, a converse result is obtained, showing that our criterion is sharp in this case. Several cases are studied: we first show how our criterion can be simply checked for processes with bounded penalization, and we then study in detail more delicate examples, including one-dimensional diffusion processes conditioned not to hit 0 and penalized birth and death processes evolving in a quenched random environment.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
12

Xia, Aihua. „Weak Convergence of Markov Processes with Extended Generators“. Annals of Probability 22, Nr. 4 (Oktober 1994): 2183–202. http://dx.doi.org/10.1214/aop/1176988499.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
13

Mao, Yong-Hua. „Convergence rates in strong ergodicity for Markov processes“. Stochastic Processes and their Applications 116, Nr. 12 (Dezember 2006): 1964–76. http://dx.doi.org/10.1016/j.spa.2006.05.008.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
14

Turner, Amanda G. „Convergence of Markov processes near saddle fixed points“. Annals of Probability 35, Nr. 3 (Mai 2007): 1141–71. http://dx.doi.org/10.1214/009117906000000836.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
15

Malyk, Igor V. „Compensating Operator and Weak Convergence of Semi-Markov Process to the Diffusion Process without Balance Condition“. Journal of Applied Mathematics 2015 (2015): 1–7. http://dx.doi.org/10.1155/2015/563060.

Der volle Inhalt der Quelle
Annotation:
Weak convergence of semi-Markov processes in the diffusive approximation scheme is studied in the paper. This problem is not new and it is studied in many papers, using convergence of random processes. Unlike other studies, we used in this paper concept of the compensating operator. It enables getting sufficient conditions of weak convergence under the conditions on the local characteristics of output semi-Markov process.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
16

Swishchuk, Anatoliy, und Nikolaos Limnios. „Controlled Discrete-Time Semi-Markov Random Evolutions and Their Applications“. Mathematics 9, Nr. 2 (13.01.2021): 158. http://dx.doi.org/10.3390/math9020158.

Der volle Inhalt der Quelle
Annotation:
In this paper, we introduced controlled discrete-time semi-Markov random evolutions. These processes are random evolutions of discrete-time semi-Markov processes where we consider a control. applied to the values of random evolution. The main results concern time-rescaled weak convergence limit theorems in a Banach space of the above stochastic systems as averaging and diffusion approximation. The applications are given to the controlled additive functionals, controlled geometric Markov renewal processes, and controlled dynamical systems. We provide dynamical principles for discrete-time dynamical systems such as controlled additive functionals and controlled geometric Markov renewal processes. We also produce dynamic programming equations (Hamilton–Jacobi–Bellman equations) for the limiting processes in diffusion approximation such as controlled additive functionals, controlled geometric Markov renewal processes and controlled dynamical systems. As an example, we consider the solution of portfolio optimization problem by Merton for the limiting controlled geometric Markov renewal processes in diffusion approximation scheme. The rates of convergence in the limit theorems are also presented.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
17

Liu, Yuanyuan, und Zhenting Hou. „Several Types of Ergodicity for M/G/1-Type Markov Chains and Markov Processes“. Journal of Applied Probability 43, Nr. 1 (März 2006): 141–58. http://dx.doi.org/10.1239/jap/1143936249.

Der volle Inhalt der Quelle
Annotation:
In this paper we study polynomial and geometric (exponential) ergodicity for M/G/1-type Markov chains and Markov processes. First, practical criteria for M/G/1-type Markov chains are obtained by analyzing the generating function of the first return probability to level 0. Then the corresponding criteria for M/G/1-type Markov processes are given, using their h-approximation chains. Our method yields the radius of convergence of the generating function of the first return probability, which is very important in obtaining explicit bounds on geometric (exponential) convergence rates. Our results are illustrated, in the final section, in some examples.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
18

Liu, Yuanyuan, und Zhenting Hou. „Several Types of Ergodicity for M/G/1-Type Markov Chains and Markov Processes“. Journal of Applied Probability 43, Nr. 01 (März 2006): 141–58. http://dx.doi.org/10.1017/s002190020000142x.

Der volle Inhalt der Quelle
Annotation:
In this paper we study polynomial and geometric (exponential) ergodicity for M/G/1-type Markov chains and Markov processes. First, practical criteria for M/G/1-type Markov chains are obtained by analyzing the generating function of the first return probability to level 0. Then the corresponding criteria for M/G/1-type Markov processes are given, using their h-approximation chains. Our method yields the radius of convergence of the generating function of the first return probability, which is very important in obtaining explicit bounds on geometric (exponential) convergence rates. Our results are illustrated, in the final section, in some examples.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
19

Macci, Claudio. „Continuous-time Markov additive processes: Composition of large deviations principles and comparison between exponential rates of convergence“. Journal of Applied Probability 38, Nr. 4 (Dezember 2001): 917–31. http://dx.doi.org/10.1239/jap/1011994182.

Der volle Inhalt der Quelle
Annotation:
We consider a continuous-time Markov additive process (Jt,St) with (Jt) an irreducible Markov chain on E = {1,…,s}; it is known that (St/t) satisfies the large deviations principle as t → ∞. In this paper we present a variational formula H for the rate function κ∗ and, in some sense, we have a composition of two large deviations principles. Moreover, under suitable hypotheses, we can consider two other continuous-time Markov additive processes derived from (Jt,St): the averaged parameters model (Jt,St(A)) and the fluid model (Jt,St(F)). Then some results of convergence are presented and the variational formula H can be employed to show that, in some sense, the convergences for (Jt,St(A)) and (Jt,St(F)) are faster than the corresponding convergences for (Jt,St).
APA, Harvard, Vancouver, ISO und andere Zitierweisen
20

Cipra, Tomáš. „Autoregressive processes in optimization“. Journal of Applied Probability 25, Nr. 2 (Juni 1988): 302–12. http://dx.doi.org/10.2307/3214438.

Der volle Inhalt der Quelle
Annotation:
Vector autoregressive processes of the first order are considered which are non-negative and optimize a linear objective function. These processes may be used in stochastic linear programming with a dynamic structure. By using Tweedie's results from the theory of Markov chains, conditions for geometric rates of convergence to stationarity (i.e. so-called geometric ergodicity) and for existence and geometric convergence of moments of these processes are obtained.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
21

Cipra, Tomáš. „Autoregressive processes in optimization“. Journal of Applied Probability 25, Nr. 02 (Juni 1988): 302–12. http://dx.doi.org/10.1017/s0021900200040948.

Der volle Inhalt der Quelle
Annotation:
Vector autoregressive processes of the first order are considered which are non-negative and optimize a linear objective function. These processes may be used in stochastic linear programming with a dynamic structure. By using Tweedie's results from the theory of Markov chains, conditions for geometric rates of convergence to stationarity (i.e. so-called geometric ergodicity) and for existence and geometric convergence of moments of these processes are obtained.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
22

Haas, Peter J., und Gerald S. Shedler. „Stochastic Petri Nets: Modeling Power and Limit Theorems“. Probability in the Engineering and Informational Sciences 5, Nr. 4 (Oktober 1991): 477–98. http://dx.doi.org/10.1017/s0269964800002242.

Der volle Inhalt der Quelle
Annotation:
Generalized semi-Markov processes and stochastic Petri nets provide building blocks for specification of discrete event system simulations on a finite or countable state space. The two formal systems differ, however, in the event scheduling (clock-setting) mechanism, the state transition mechanism, and the form of the state space. We have shown previously that stochastic Petri nets have at least the modeling power of generalized semi-Markov processes. In this paper we show that stochastic Petri nets and generalized semi-Markov processes, in fact, have the same modeling power. Combining this result with known results for generalized semi-Markov processes, we also obtain conditions for time-average convergence and convergence in distribution along with a central limit theorem for the marking process of a stochastic Petri net.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
23

Macci, Claudio. „Continuous-time Markov additive processes: Composition of large deviations principles and comparison between exponential rates of convergence“. Journal of Applied Probability 38, Nr. 04 (Dezember 2001): 917–31. http://dx.doi.org/10.1017/s0021900200019136.

Der volle Inhalt der Quelle
Annotation:
We consider a continuous-time Markov additive process (J t ,S t ) with (J t ) an irreducible Markov chain on E = {1,…,s}; it is known that (S t /t) satisfies the large deviations principle as t → ∞. In this paper we present a variational formula H for the rate function κ∗ and, in some sense, we have a composition of two large deviations principles. Moreover, under suitable hypotheses, we can consider two other continuous-time Markov additive processes derived from (J t ,S t ): the averaged parameters model (J t ,S t (A)) and the fluid model (J t ,S t (F)). Then some results of convergence are presented and the variational formula H can be employed to show that, in some sense, the convergences for (J t ,S t (A)) and (J t ,S t (F)) are faster than the corresponding convergences for (J t ,S t ).
APA, Harvard, Vancouver, ISO und andere Zitierweisen
24

Kalpazidou, Sophia. „On the weak convergence of sequences of circuit processes: a probabilistic approach“. Journal of Applied Probability 29, Nr. 2 (Juni 1992): 374–83. http://dx.doi.org/10.2307/3214574.

Der volle Inhalt der Quelle
Annotation:
The asymptotic behaviour of sequences of Markov processes whose finite distributions depend upon the sample paths ω of a positive recurrent Markov chain ξ is studied. The existence of such sequences depends upon the existence of a unique class of directed weighted circuits having a probabilistic interpretation in terms of the directed circuits occurring along the sample paths of ξ. An application to multiple Markov chains is given.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
25

Kalpazidou, Sophia. „On the weak convergence of sequences of circuit processes: a probabilistic approach“. Journal of Applied Probability 29, Nr. 02 (Juni 1992): 374–83. http://dx.doi.org/10.1017/s0021900200043126.

Der volle Inhalt der Quelle
Annotation:
The asymptotic behaviour of sequences of Markov processes whose finite distributions depend upon the sample paths ω of a positive recurrent Markov chain ξ is studied. The existence of such sequences depends upon the existence of a unique class of directed weighted circuits having a probabilistic interpretation in terms of the directed circuits occurring along the sample paths of ξ. An application to multiple Markov chains is given.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
26

Szehr, Oleg, David Reeb und Michael M. Wolf. „Spectral Convergence Bounds for Classical and Quantum Markov Processes“. Communications in Mathematical Physics 333, Nr. 2 (12.10.2014): 565–95. http://dx.doi.org/10.1007/s00220-014-2188-5.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
27

Lund, Robert B., Sean P. Meyn und Richard L. Tweedie. „Computable exponential convergence rates for stochastically ordered Markov processes“. Annals of Applied Probability 6, Nr. 1 (Februar 1996): 218–37. http://dx.doi.org/10.1214/aoap/1034968072.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
28

Park, Y. S., J. C. Bean und R. L. Smith. „Optimal Average Value Convergence in Nonhomogeneous Markov Decision Processes“. Journal of Mathematical Analysis and Applications 179, Nr. 2 (November 1993): 525–36. http://dx.doi.org/10.1006/jmaa.1993.1367.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
29

Cooper, William L., Shane G. Henderson und Mark E. Lewis. „CONVERGENCE OF SIMULATION-BASED POLICY ITERATION“. Probability in the Engineering and Informational Sciences 17, Nr. 2 (27.02.2003): 213–34. http://dx.doi.org/10.1017/s0269964803172051.

Der volle Inhalt der Quelle
Annotation:
Simulation-based policy iteration (SBPI) is a modification of the policy iteration algorithm for computing optimal policies for Markov decision processes. At each iteration, rather than solving the average evaluation equations, SBPI employs simulation to estimate a solution to these equations. For recurrent average-reward Markov decision processes with finite state and action spaces, we provide easily verifiable conditions that ensure that simulation-based policy iteration almost-surely eventually never leaves the set of optimal decision rules. We analyze three simulation estimators for solutions to the average evaluation equations. Using our general results, we derive simple conditions on the simulation run lengths that guarantee the almost-sure convergence of the algorithm.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
30

Gravereaux, Jean-Bernard, und James Ledoux. „Poisson approximation for some point processes in reliability“. Advances in Applied Probability 36, Nr. 2 (Juni 2004): 455–70. http://dx.doi.org/10.1239/aap/1086957581.

Der volle Inhalt der Quelle
Annotation:
In this paper, we consider a failure point process related to the Markovian arrival process defined by Neuts. We show that it converges in distribution to a homogeneous Poisson process. This convergence takes place in the context of rare occurrences of failures. We also provide a convergence rate of the convergence in total variation of this point process using an approach developed by Kabanov, Liptser and Shiryaev for the doubly stochastic Poisson process driven by a finite Markov process.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
31

Gravereaux, Jean-Bernard, und James Ledoux. „Poisson approximation for some point processes in reliability“. Advances in Applied Probability 36, Nr. 02 (Juni 2004): 455–70. http://dx.doi.org/10.1017/s0001867800013562.

Der volle Inhalt der Quelle
Annotation:
In this paper, we consider a failure point process related to the Markovian arrival process defined by Neuts. We show that it converges in distribution to a homogeneous Poisson process. This convergence takes place in the context of rare occurrences of failures. We also provide a convergence rate of the convergence in total variation of this point process using an approach developed by Kabanov, Liptser and Shiryaev for the doubly stochastic Poisson process driven by a finite Markov process.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
32

Ying, Donghao, Mengzi Amy Guo, Yuhao Ding, Javad Lavaei und Zuo-Jun Shen. „Policy-Based Primal-Dual Methods for Convex Constrained Markov Decision Processes“. Proceedings of the AAAI Conference on Artificial Intelligence 37, Nr. 9 (26.06.2023): 10963–71. http://dx.doi.org/10.1609/aaai.v37i9.26299.

Der volle Inhalt der Quelle
Annotation:
We study convex Constrained Markov Decision Processes (CMDPs) in which the objective is concave and the constraints are convex in the state-action occupancy measure. We propose a policy-based primal-dual algorithm that updates the primal variable via policy gradient ascent and updates the dual variable via projected sub-gradient descent. Despite the loss of additivity structure and the nonconvex nature, we establish the global convergence of the proposed algorithm by leveraging a hidden convexity in the problem, and prove the O(T^-1/3) convergence rate in terms of both optimality gap and constraint violation. When the objective is strongly concave in the occupancy measure, we prove an improved convergence rate of O(T^-1/2). By introducing a pessimistic term to the constraint, we further show that a zero constraint violation can be achieved while preserving the same convergence rate for the optimality gap. This work is the first one in the literature that establishes non-asymptotic convergence guarantees for policy-based primal-dual methods for solving infinite-horizon discounted convex CMDPs.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
33

LIU, YUANYUAN, HANJUN ZHANG und YIQIANG ZHAO. „COMPUTABLE STRONGLY ERGODIC RATES OF CONVERGENCE FOR CONTINUOUS-TIME MARKOV CHAINS“. ANZIAM Journal 49, Nr. 4 (April 2008): 463–78. http://dx.doi.org/10.1017/s1446181108000114.

Der volle Inhalt der Quelle
Annotation:
AbstractIn this paper, we investigate computable lower bounds for the best strongly ergodic rate of convergence of the transient probability distribution to the stationary distribution for stochastically monotone continuous-time Markov chains and reversible continuous-time Markov chains, using a drift function and the expectation of the first hitting time on some state. We apply these results to birth–death processes, branching processes and population processes.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
34

Böttcher, Björn. „Embedded Markov chain approximations in Skorokhod topologies“. Probability and Mathematical Statistics 39, Nr. 2 (19.12.2019): 259–77. http://dx.doi.org/10.19195/0208-4147.39.2.2.

Der volle Inhalt der Quelle
Annotation:
We prove a J1-tightness condition for embedded Markov chains and discuss four Skorokhod topologies in a unified manner. To approximate a continuous time stochastic process by discrete time Markov chains, one has several options to embed the Markov chains into continuous time processes. On the one hand, there is a Markov embedding which uses exponential waiting times. On the other hand, each Skorokhod topology naturally suggests a certain embedding. These are the step function embedding for J1, the linear interpolation embedding forM1, the multistep embedding for J2 and a more general embedding for M2. We show that the convergence of the step function embedding in J1 implies the convergence of the other embeddings in the corresponding topologies. For the converse statement, a J1-tightness condition for embedded time-homogeneous Markov chains is given.Additionally, it is shown that J1 convergence is equivalent to the joint convergence in M1 and J2.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
35

Giroux, Gaston. „Asymptotic results for non-linear processes of the McKean tagged-molecule type“. Journal of Applied Probability 23, Nr. 1 (März 1986): 42–51. http://dx.doi.org/10.2307/3214115.

Der volle Inhalt der Quelle
Annotation:
McKean's tagged-molecule process is a non-linear homogeneous two-state Markov chain in continuous time, constructed with the aid of a binary branching process. For each of a large class of branching processes we construct a similar process. The construction is carefully done and the weak homogeneity is deduced. A simple probability argument permits us to show convergence to the equidistribution (½, ½) and to note that this limit is a strong equilibrium. A non-homogeneous Markov chain result is also used to establish the geometric rate of convergence. A proof of a Boltzmann H-theorem is also established.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
36

Giroux, Gaston. „Asymptotic results for non-linear processes of the McKean tagged-molecule type“. Journal of Applied Probability 23, Nr. 01 (März 1986): 42–51. http://dx.doi.org/10.1017/s0021900200106266.

Der volle Inhalt der Quelle
Annotation:
McKean's tagged-molecule process is a non-linear homogeneous two-state Markov chain in continuous time, constructed with the aid of a binary branching process. For each of a large class of branching processes we construct a similar process. The construction is carefully done and the weak homogeneity is deduced. A simple probability argument permits us to show convergence to the equidistribution (½, ½) and to note that this limit is a strong equilibrium. A non-homogeneous Markov chain result is also used to establish the geometric rate of convergence. A proof of a Boltzmann H-theorem is also established.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
37

Pakes, Anthony G. „Convergence Rates and Limit Theorems for the Dual Markov Branching Process“. Journal of Probability and Statistics 2017 (2017): 1–13. http://dx.doi.org/10.1155/2017/1410507.

Der volle Inhalt der Quelle
Annotation:
This paper studies aspects of the Siegmund dual of the Markov branching process. The principal results are optimal convergence rates of its transition function and limit theorems in the case that it is not positive recurrent. Additional discussion is given about specifications of the Markov branching process and its dual. The dualising Markov branching processes need not be regular or even conservative.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
38

Jacka, S. D., und G. O. Roberts. „Weak convergence of conditioned processes on a countable state space“. Journal of Applied Probability 32, Nr. 4 (Dezember 1995): 902–16. http://dx.doi.org/10.2307/3215203.

Der volle Inhalt der Quelle
Annotation:
We consider the problem of conditioning a continuous-time Markov chain (on a countably infinite state space) not to hit an absorbing barrier before time T; and the weak convergence of this conditional process as T → ∞. We prove a characterization of convergence in terms of the distribution of the process at some arbitrary positive time, t, introduce a decay parameter for the time to absorption, give an example where weak convergence fails, and give sufficient conditions for weak convergence in terms of the existence of a quasi-stationary limit, and a recurrence property of the original process.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
39

Jacka, S. D., und G. O. Roberts. „Weak convergence of conditioned processes on a countable state space“. Journal of Applied Probability 32, Nr. 04 (Dezember 1995): 902–16. http://dx.doi.org/10.1017/s0021900200103377.

Der volle Inhalt der Quelle
Annotation:
We consider the problem of conditioning a continuous-time Markov chain (on a countably infinite state space) not to hit an absorbing barrier before time T; and the weak convergence of this conditional process as T → ∞. We prove a characterization of convergence in terms of the distribution of the process at some arbitrary positive time, t, introduce a decay parameter for the time to absorption, give an example where weak convergence fails, and give sufficient conditions for weak convergence in terms of the existence of a quasi-stationary limit, and a recurrence property of the original process.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
40

Deng, C. S., R. L. Schilling und Y. H. Song. „Subgeometric rates of convergence for Markov processes under subordination - Correction“. Advances in Applied Probability 50, Nr. 3 (September 2018): 1005. http://dx.doi.org/10.1017/apr.2018.44.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
41

Gaudio, Julia, Saurabh Amin und Patrick Jaillet. „Exponential convergence rates for stochastically ordered Markov processes under perturbation“. Systems & Control Letters 133 (November 2019): 104515. http://dx.doi.org/10.1016/j.sysconle.2019.104515.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
42

Douc, Randal, Gersende Fort und Arnaud Guillin. „Subgeometric rates of convergence of f-ergodic strong Markov processes“. Stochastic Processes and their Applications 119, Nr. 3 (März 2009): 897–923. http://dx.doi.org/10.1016/j.spa.2008.03.007.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
43

Huang, Gang, Michel Mandjes und Peter Spreij. „Weak convergence of Markov-modulated diffusion processes with rapid switching“. Statistics & Probability Letters 86 (März 2014): 74–79. http://dx.doi.org/10.1016/j.spl.2013.12.013.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
44

Mao, Yong-Hua, Liping Xu, Ming Zhang und Yu-Hui Zhang. „Convergence in total variation distance for (in)homogeneous Markov processes“. Statistics & Probability Letters 137 (Juni 2018): 54–62. http://dx.doi.org/10.1016/j.spl.2018.01.011.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
45

Wang, Fengyu. „Coupling, convergence rates of Markov processes and weak Poincaré inequalities“. Science in China Series A: Mathematics 45, Nr. 8 (August 2002): 975–83. http://dx.doi.org/10.1007/bf02879980.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
46

Alvarez-Mena, Jorge, und Onésimo Hernández-Lerma. „Convergence of the optimal values of constrained Markov control processes“. Mathematical Methods of Operations Research 55, Nr. 3 (Juni 2002): 461–84. http://dx.doi.org/10.1007/s001860200209.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
47

Hart, Andrew G., und Richard L. Tweedie. „Convergence of Invariant Measures of Truncation Approximations to Markov Processes“. Applied Mathematics 03, Nr. 12 (2012): 2205–15. http://dx.doi.org/10.4236/am.2012.312a301.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
48

Mufa, Chen. „ExponentialL 2-convergence andL 2-spectral gap for Markov processes“. Acta Mathematica Sinica 7, Nr. 1 (März 1991): 19–37. http://dx.doi.org/10.1007/bf02582989.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
49

White, D. J., und W. T. Scherer. „The Convergence of Value Iteration in Discounted Markov Decision Processes“. Journal of Mathematical Analysis and Applications 182, Nr. 2 (März 1994): 348–60. http://dx.doi.org/10.1006/jmaa.1994.1090.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
50

Nguyen, Giang T., und Oscar Peralta. „Rate of strong convergence to Markov-modulated Brownian motion“. Journal of Applied Probability 59, Nr. 1 (18.01.2022): 1–16. http://dx.doi.org/10.1017/jpr.2021.30.

Der volle Inhalt der Quelle
Annotation:
AbstractLatouche and Nguyen (2015b) constructed a sequence of stochastic fluid processes and showed that it converges weakly to a Markov-modulated Brownian motion (MMBM). Here, we construct a different sequence of stochastic fluid processes and show that it converges strongly to an MMBM. To the best of our knowledge, this is the first result on strong convergence to a Markov-modulated Brownian motion. Besides implying weak convergence, such a strong approximation constitutes a powerful tool for developing deep results for sophisticated models. Additionally, we prove that the rate of this almost sure convergence is $o(n^{-1/2} \log n)$ . When reduced to the special case of standard Brownian motion, our convergence rate is an improvement over that obtained by a different approximation in Gorostiza and Griego (1980), which is $o(n^{-1/2}(\log n)^{5/2})$ .
APA, Harvard, Vancouver, ISO und andere Zitierweisen
Wir bieten Rabatte auf alle Premium-Pläne für Autoren, deren Werke in thematische Literatursammlungen aufgenommen wurden. Kontaktieren Sie uns, um einen einzigartigen Promo-Code zu erhalten!

Zur Bibliographie