To see the other types of publications on this topic, follow the link: Markov processes.

Journal articles on the topic 'Markov processes'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 journal articles for your research on the topic 'Markov processes.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.

1

Demenkov, N. P., E. A. Mirkin, and I. A. Mochalov. "Markov and Semi-Markov Processes with Fuzzy States. Part 1. Markov Processes." Informacionnye tehnologii 26, no. 6 (June 23, 2020): 323–34. http://dx.doi.org/10.17587/it.26.323-334.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

FRANZ, UWE. "CLASSICAL MARKOV PROCESSES FROM QUANTUM LÉVY PROCESSES." Infinite Dimensional Analysis, Quantum Probability and Related Topics 02, no. 01 (March 1999): 105–29. http://dx.doi.org/10.1142/s0219025799000060.

Full text
Abstract:
We show how classical Markov processes can be obtained from quantum Lévy processes. It is shown that quantum Lévy processes are quantum Markov processes, and sufficient conditions for restrictions to subalgebras to remain quantum Markov processes are given. A classical Markov process (which has the same time-ordered moments as the quantum process in the vacuum state) exists whenever we can restrict to a commutative subalgebra without losing the quantum Markov property.8 Several examples, including the Azéma martingale, with explicit calculations are presented. In particular, the action of the generator of the classical Markov processes on polynomials or their moments are calculated using Hopf algebra duality.
APA, Harvard, Vancouver, ISO, and other styles
3

Demenkov, N. P., E. A. Mirkin, and I. A. Mochalov. "Markov and Semi-Markov Processes with Fuzzy States. Part 2. Semi-Markov Processes." INFORMACIONNYE TEHNOLOGII 26, no. 7 (July 17, 2020): 387–93. http://dx.doi.org/10.17587/it.26.387-393.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Whittle, P., and M. L. Puterman. "Markov Decision Processes." Journal of the Royal Statistical Society. Series A (Statistics in Society) 158, no. 3 (1995): 636. http://dx.doi.org/10.2307/2983459.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Smith, J. Q., and D. J. White. "Markov Decision Processes." Journal of the Royal Statistical Society. Series A (Statistics in Society) 157, no. 1 (1994): 164. http://dx.doi.org/10.2307/2983520.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

King, Aaron A., Qianying Lin, and Edward L. Ionides. "Markov genealogy processes." Theoretical Population Biology 143 (February 2022): 77–91. http://dx.doi.org/10.1016/j.tpb.2021.11.003.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Thomas, L. C., D. J. White, and Martin L. Puterman. "Markov Decision Processes." Journal of the Operational Research Society 46, no. 6 (June 1995): 792. http://dx.doi.org/10.2307/2584317.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Ephraim, Y., and N. Merhav. "Hidden Markov processes." IEEE Transactions on Information Theory 48, no. 6 (June 2002): 1518–69. http://dx.doi.org/10.1109/tit.2002.1003838.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Bäuerle, Nicole, and Ulrich Rieder. "Markov Decision Processes." Jahresbericht der Deutschen Mathematiker-Vereinigung 112, no. 4 (September 8, 2010): 217–43. http://dx.doi.org/10.1365/s13291-010-0007-2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Wal, J., and J. Wessels. "MARKOV DECISION PROCESSES." Statistica Neerlandica 39, no. 2 (June 1985): 219–33. http://dx.doi.org/10.1111/j.1467-9574.1985.tb01140.x.

Full text
APA, Harvard, Vancouver, ISO, and other styles
11

Thomas, L. C. "Markov Decision Processes." Journal of the Operational Research Society 46, no. 6 (June 1995): 792–93. http://dx.doi.org/10.1057/jors.1995.110.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Frank, T. D. "Nonlinear Markov processes." Physics Letters A 372, no. 25 (June 2008): 4553–55. http://dx.doi.org/10.1016/j.physleta.2008.04.027.

Full text
APA, Harvard, Vancouver, ISO, and other styles
13

Kinateder, Kimberly K. J. "Corner Markov processes." Journal of Theoretical Probability 8, no. 3 (July 1995): 539–47. http://dx.doi.org/10.1007/bf02218043.

Full text
APA, Harvard, Vancouver, ISO, and other styles
14

Brooks, Stephen, and D. J. White. "Markov Decision Processes." Statistician 44, no. 2 (1995): 292. http://dx.doi.org/10.2307/2348465.

Full text
APA, Harvard, Vancouver, ISO, and other styles
15

Fagnola, Franco. "Algebraic Markov processes." Proyecciones (Antofagasta) 18, no. 3 (1999): 13–28. http://dx.doi.org/10.22199/s07160917.1999.0003.00003.

Full text
APA, Harvard, Vancouver, ISO, and other styles
16

Craven, B. D. "Perturbed Markov Processes." Stochastic Models 19, no. 2 (January 5, 2003): 269–85. http://dx.doi.org/10.1081/stm-120020390.

Full text
APA, Harvard, Vancouver, ISO, and other styles
17

White, Chelsea C., and Douglas J. White. "Markov decision processes." European Journal of Operational Research 39, no. 1 (March 1989): 1–16. http://dx.doi.org/10.1016/0377-2217(89)90348-2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
18

Zhenting, Hou, Liu Zaiming, and Zou Jiezhong. "Markov skeleton processes." Chinese Science Bulletin 43, no. 11 (June 1998): 881–89. http://dx.doi.org/10.1007/bf02884605.

Full text
APA, Harvard, Vancouver, ISO, and other styles
19

Franz, Uwe, Volkmar Liebscher, and Stefan Zeiser. "Piecewise-Deterministic Markov Processes as Limits of Markov Jump Processes." Advances in Applied Probability 44, no. 3 (September 2012): 729–48. http://dx.doi.org/10.1239/aap/1346955262.

Full text
Abstract:
A classical result about Markov jump processes states that a certain class of dynamical systems given by ordinary differential equations are obtained as the limit of a sequence of scaled Markov jump processes. This approach fails if the scaling cannot be carried out equally across all entities. In the present paper we present a convergence theorem for such an unequal scaling. In contrast to an equal scaling the limit process is not purely deterministic but still possesses randomness. We show that these processes constitute a rich subclass of piecewise-deterministic processes. Such processes apply in molecular biology where entities often occur in different scales of numbers.
APA, Harvard, Vancouver, ISO, and other styles
20

Franz, Uwe, Volkmar Liebscher, and Stefan Zeiser. "Piecewise-Deterministic Markov Processes as Limits of Markov Jump Processes." Advances in Applied Probability 44, no. 03 (September 2012): 729–48. http://dx.doi.org/10.1017/s0001867800005851.

Full text
Abstract:
A classical result about Markov jump processes states that a certain class of dynamical systems given by ordinary differential equations are obtained as the limit of a sequence of scaled Markov jump processes. This approach fails if the scaling cannot be carried out equally across all entities. In the present paper we present a convergence theorem for such an unequal scaling. In contrast to an equal scaling the limit process is not purely deterministic but still possesses randomness. We show that these processes constitute a rich subclass of piecewise-deterministic processes. Such processes apply in molecular biology where entities often occur in different scales of numbers.
APA, Harvard, Vancouver, ISO, and other styles
21

Buchholz, Peter, and Miklós Telek. "Rational Processes Related to Communicating Markov Processes." Journal of Applied Probability 49, no. 1 (March 2012): 40–59. http://dx.doi.org/10.1239/jap/1331216833.

Full text
Abstract:
We define a class of stochastic processes, denoted as marked rational arrival processes (MRAPs), which is an extension of matrix exponential distributions and rational arrival processes. Continuous-time Markov processes with labeled transitions are a subclass of this more general model class. New equivalence relations between processes are defined, and it is shown that these equivalence relations are natural extensions of strong and weak lumpability and the corresponding bisimulation relations that have been defined for Markov processes. If a general rational process is equivalent to a Markov process, it can be used in numerical analysis techniques instead of the Markov process. This observation allows one to apply MRAPs like Markov processes and since the new equivalence relations are more general than lumpability and bisimulation, it is sometimes possible to find smaller representations of given processes. Finally, we show that the equivalence is preserved by the composition of MRAPs and can therefore be exploited in compositional modeling.
APA, Harvard, Vancouver, ISO, and other styles
22

Buchholz, Peter, and Miklós Telek. "Rational Processes Related to Communicating Markov Processes." Journal of Applied Probability 49, no. 01 (March 2012): 40–59. http://dx.doi.org/10.1017/s0021900200008858.

Full text
Abstract:
We define a class of stochastic processes, denoted as marked rational arrival processes (MRAPs), which is an extension of matrix exponential distributions and rational arrival processes. Continuous-time Markov processes with labeled transitions are a subclass of this more general model class. New equivalence relations between processes are defined, and it is shown that these equivalence relations are natural extensions of strong and weak lumpability and the corresponding bisimulation relations that have been defined for Markov processes. If a general rational process is equivalent to a Markov process, it can be used in numerical analysis techniques instead of the Markov process. This observation allows one to apply MRAPs like Markov processes and since the new equivalence relations are more general than lumpability and bisimulation, it is sometimes possible to find smaller representations of given processes. Finally, we show that the equivalence is preserved by the composition of MRAPs and can therefore be exploited in compositional modeling.
APA, Harvard, Vancouver, ISO, and other styles
23

Iwata, Yukiko. "Constrictive Markov operators induced by Markov processes." Positivity 20, no. 2 (September 3, 2015): 355–67. http://dx.doi.org/10.1007/s11117-015-0360-6.

Full text
APA, Harvard, Vancouver, ISO, and other styles
24

Malinovskii, V. K. "Limit theorems for recurrent semi-Markov processes and Markov renewal processes." Journal of Soviet Mathematics 36, no. 4 (February 1987): 493–502. http://dx.doi.org/10.1007/bf01663460.

Full text
APA, Harvard, Vancouver, ISO, and other styles
25

Fredkin, Donald R., and John A. Rice. "On aggregated Markov processes." Journal of Applied Probability 23, no. 1 (March 1986): 208–14. http://dx.doi.org/10.2307/3214130.

Full text
APA, Harvard, Vancouver, ISO, and other styles
26

Kazak, Jolanta. "Piecewise-deterministic Markov processes." Annales Polonici Mathematici 109, no. 3 (2013): 279–96. http://dx.doi.org/10.4064/ap109-3-4.

Full text
APA, Harvard, Vancouver, ISO, and other styles
27

Lee, P. M., and O. Hernandez-Lerma. "Adaptive Markov Control Processes." Mathematical Gazette 74, no. 470 (December 1990): 417. http://dx.doi.org/10.2307/3618186.

Full text
APA, Harvard, Vancouver, ISO, and other styles
28

Kowalski, Zbigniew S. "Multiple Markov Gaussian processes." Applicationes Mathematicae 48, no. 1 (2021): 65–78. http://dx.doi.org/10.4064/am2411-1-2021.

Full text
APA, Harvard, Vancouver, ISO, and other styles
29

SHIEH, Narn-Rueih. "Collisions of Markov Processes." Tokyo Journal of Mathematics 18, no. 1 (June 1995): 111–21. http://dx.doi.org/10.3836/tjm/1270043612.

Full text
APA, Harvard, Vancouver, ISO, and other styles
30

Hawkes, Alan G. "Markov processes in APL." ACM SIGAPL APL Quote Quad 20, no. 4 (May 1990): 173–85. http://dx.doi.org/10.1145/97811.97843.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Pollett, P. K. "Connecting reversible Markov processes." Advances in Applied Probability 18, no. 4 (December 1986): 880–900. http://dx.doi.org/10.2307/1427254.

Full text
Abstract:
We provide a framework for interconnecting a collection of reversible Markov processes in such a way that the resulting process has a product-form invariant measure with respect to which the process is reversible. A number of examples are discussed including Kingman&s reversible migration process, interconnected random walks and stratified clustering processes.
APA, Harvard, Vancouver, ISO, and other styles
32

Gerontidis, Ioannis I. "Markov population replacement processes." Advances in Applied Probability 27, no. 3 (September 1995): 711–40. http://dx.doi.org/10.2307/1428131.

Full text
Abstract:
We consider a migration process whose singleton process is a time-dependent Markov replacement process. For the singleton process, which may be treated as either open or closed, we study the limiting distribution, the distribution of the time to replacement and related quantities. For a replacement process in equilibrium we obtain a version of Little's law and we provide conditions for reversibility. For the resulting linear population process we characterize exponential ergodicity for two types of environmental behaviour, i.e. either convergent or cyclic, and finally for large population sizes a diffusion approximation analysis is provided.
APA, Harvard, Vancouver, ISO, and other styles
33

Alexopoulos, Christos, Akram A. El-Tannir, and Richard F. Serfozo. "Partition-Reversible Markov Processes." Operations Research 47, no. 1 (February 1999): 125–30. http://dx.doi.org/10.1287/opre.47.1.125.

Full text
APA, Harvard, Vancouver, ISO, and other styles
34

Avrachenkov, Konstantin, Alexey Piunovskiy, and Yi Zhang. "Markov Processes with Restart." Journal of Applied Probability 50, no. 4 (December 2013): 960–68. http://dx.doi.org/10.1239/jap/1389370093.

Full text
Abstract:
We consider a general homogeneous continuous-time Markov process with restarts. The process is forced to restart from a given distribution at time moments generated by an independent Poisson process. The motivation to study such processes comes from modeling human and animal mobility patterns, restart processes in communication protocols, and from application of restarting random walks in information retrieval. We provide a connection between the transition probability functions of the original Markov process and the modified process with restarts. We give closed-form expressions for the invariant probability measure of the modified process. When the process evolves on the Euclidean space, there is also a closed-form expression for the moments of the modified process. We show that the modified process is always positive Harris recurrent and exponentially ergodic with the index equal to (or greater than) the rate of restarts. Finally, we illustrate the general results by the standard and geometric Brownian motions.
APA, Harvard, Vancouver, ISO, and other styles
35

Novak, Stephanie, and Lyman J. Fretwell. "Non‐Markov noise processes." Journal of the Acoustical Society of America 80, S1 (December 1986): S64. http://dx.doi.org/10.1121/1.2023904.

Full text
APA, Harvard, Vancouver, ISO, and other styles
36

Rodrigues, Josemar, N. Balakrishnan, and Patrick Borges. "Markov-Correlated Poisson Processes." Communications in Statistics - Theory and Methods 42, no. 20 (October 18, 2013): 3696–703. http://dx.doi.org/10.1080/03610926.2011.636168.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Larralde, H., F. Leyvraz, and D. P. Sanders. "Metastability in Markov processes." Journal of Statistical Mechanics: Theory and Experiment 2006, no. 08 (August 18, 2006): P08013. http://dx.doi.org/10.1088/1742-5468/2006/08/p08013.

Full text
APA, Harvard, Vancouver, ISO, and other styles
38

jun luo, Shou. "Two-parameter markov processes." Stochastics and Stochastic Reports 40, no. 3-4 (September 1992): 181–93. http://dx.doi.org/10.1080/17442509208833788.

Full text
APA, Harvard, Vancouver, ISO, and other styles
39

Chin, Y. C., and A. J. Baddeley. "Markov interacting component processes." Advances in Applied Probability 32, no. 3 (September 2000): 597–619. http://dx.doi.org/10.1239/aap/1013540233.

Full text
Abstract:
A generalization of Markov point processes is introduced in which interactions occur between connected components of the point pattern. A version of the Hammersley-Clifford characterization theorem is proved which states that a point process is a Markov interacting component process if and only if its density function is a product of interaction terms associated with cliques of connected components. Integrability and superpositional properties of the processes are shown and a pairwise interaction example is used for detailed exploration.
APA, Harvard, Vancouver, ISO, and other styles
40

Cai, Haiyan. "Piecewise deterministic Markov processes." Stochastic Analysis and Applications 11, no. 3 (January 1993): 255–74. http://dx.doi.org/10.1080/07362999308809317.

Full text
APA, Harvard, Vancouver, ISO, and other styles
41

ACCARDI, LUIGI, and ANILESH MOHARI. "TIME REFLECTED MARKOV PROCESSES." Infinite Dimensional Analysis, Quantum Probability and Related Topics 02, no. 03 (September 1999): 397–425. http://dx.doi.org/10.1142/s0219025799000230.

Full text
Abstract:
A classical stochastic process which is Markovian for its past filtration is also Markovian for its future filtration. We show with a counterexample based on quantum liftings of a finite state classical Markov chain that this property cannot hold in the category of expected Markov processes. Using a duality theory for von Neumann algebras with weights, developed by Petz on the basis of previous results by Groh and Kümmerer, we show that a quantum version of this symmetry can be established in the category of weak Markov processes in the sense of Bhat and Parthasarathy. Here time reversal is implemented by an anti-unitary operator and a weak Markov process is time reversal invariant if and only if the associated semigroup coincides with its Petz dual. This construction allows one to extend to the quantum case, both for backward and forward processes, the Misra–Prigogine–Courbage internal time operator and to show that the two operators are intertwined by the time reversal anti-automorphism.
APA, Harvard, Vancouver, ISO, and other styles
42

Fredkin, Donald R., and John A. Rice. "On aggregated Markov processes." Journal of Applied Probability 23, no. 01 (March 1986): 208–14. http://dx.doi.org/10.1017/s0021900200106412.

Full text
APA, Harvard, Vancouver, ISO, and other styles
43

Halibard, Moishe, and Ido Kanter. "Markov processes and linguistics." Physica A: Statistical Mechanics and its Applications 249, no. 1-4 (January 1998): 525–35. http://dx.doi.org/10.1016/s0378-4371(97)00512-8.

Full text
APA, Harvard, Vancouver, ISO, and other styles
44

Chin, Y. C., and A. J. Baddeley. "Markov interacting component processes." Advances in Applied Probability 32, no. 03 (September 2000): 597–619. http://dx.doi.org/10.1017/s0001867800010144.

Full text
Abstract:
A generalization of Markov point processes is introduced in which interactions occur between connected components of the point pattern. A version of the Hammersley-Clifford characterization theorem is proved which states that a point process is a Markov interacting component process if and only if its density function is a product of interaction terms associated with cliques of connected components. Integrability and superpositional properties of the processes are shown and a pairwise interaction example is used for detailed exploration.
APA, Harvard, Vancouver, ISO, and other styles
45

Pollett, P. K. "Connecting reversible Markov processes." Advances in Applied Probability 18, no. 04 (December 1986): 880–900. http://dx.doi.org/10.1017/s0001867800016190.

Full text
Abstract:
We provide a framework for interconnecting a collection of reversible Markov processes in such a way that the resulting process has a product-form invariant measure with respect to which the process is reversible. A number of examples are discussed including Kingman&s reversible migration process, interconnected random walks and stratified clustering processes.
APA, Harvard, Vancouver, ISO, and other styles
46

Gerontidis, Ioannis I. "Markov population replacement processes." Advances in Applied Probability 27, no. 03 (September 1995): 711–40. http://dx.doi.org/10.1017/s0001867800027129.

Full text
Abstract:
We consider a migration process whose singleton process is a time-dependent Markov replacement process. For the singleton process, which may be treated as either open or closed, we study the limiting distribution, the distribution of the time to replacement and related quantities. For a replacement process in equilibrium we obtain a version of Little's law and we provide conditions for reversibility. For the resulting linear population process we characterize exponential ergodicity for two types of environmental behaviour, i.e. either convergent or cyclic, and finally for large population sizes a diffusion approximation analysis is provided.
APA, Harvard, Vancouver, ISO, and other styles
47

Avrachenkov, Konstantin, Alexey Piunovskiy, and Yi Zhang. "Markov Processes with Restart." Journal of Applied Probability 50, no. 04 (December 2013): 960–68. http://dx.doi.org/10.1017/s0021900200013735.

Full text
Abstract:
We consider a general homogeneous continuous-time Markov process with restarts. The process is forced to restart from a given distribution at time moments generated by an independent Poisson process. The motivation to study such processes comes from modeling human and animal mobility patterns, restart processes in communication protocols, and from application of restarting random walks in information retrieval. We provide a connection between the transition probability functions of the original Markov process and the modified process with restarts. We give closed-form expressions for the invariant probability measure of the modified process. When the process evolves on the Euclidean space, there is also a closed-form expression for the moments of the modified process. We show that the modified process is always positive Harris recurrent and exponentially ergodic with the index equal to (or greater than) the rate of restarts. Finally, we illustrate the general results by the standard and geometric Brownian motions.
APA, Harvard, Vancouver, ISO, and other styles
48

Baykal-Gürsoy, M., and K. Gürsoy. "SEMI-MARKOV DECISION PROCESSES." Probability in the Engineering and Informational Sciences 21, no. 4 (October 2007): 635–57. http://dx.doi.org/10.1017/s026996480700037x.

Full text
Abstract:
Considered are semi-Markov decision processes (SMDPs) with finite state and action spaces. We study two criteria: the expected average reward per unit time subject to a sample path constraint on the average cost per unit time and the expected time-average variability. Under a certain condition, for communicating SMDPs, we construct (randomized) stationary policies that are ε-optimal for each criterion; the policy is optimal for the first criterion under the unichain assumption and the policy is optimal and pure for a specific variability function in the second criterion. For general multichain SMDPs, by using a state space decomposition approach, similar results are obtained.
APA, Harvard, Vancouver, ISO, and other styles
49

Desharnais, Josée, Vineet Gupta, Radha Jagadeesan, and Prakash Panangaden. "Approximating labelled Markov processes." Information and Computation 184, no. 1 (July 2003): 160–200. http://dx.doi.org/10.1016/s0890-5401(03)00051-8.

Full text
APA, Harvard, Vancouver, ISO, and other styles
50

Sabbadin, Régis. "Possibilistic Markov decision processes." Engineering Applications of Artificial Intelligence 14, no. 3 (June 2001): 287–300. http://dx.doi.org/10.1016/s0952-1976(01)00007-0.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography