Artículos de revistas sobre el tema "Markov processes"

Siga este enlace para ver otros tipos de publicaciones sobre el tema: Markov processes.

Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros

Elija tipo de fuente:

Consulte los 50 mejores artículos de revistas para su investigación sobre el tema "Markov processes".

Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.

También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.

Explore artículos de revistas sobre una amplia variedad de disciplinas y organice su bibliografía correctamente.

1

Demenkov, N. P., E. A. Mirkin y I. A. Mochalov. "Markov and Semi-Markov Processes with Fuzzy States. Part 1. Markov Processes". Informacionnye tehnologii 26, n.º 6 (23 de junio de 2020): 323–34. http://dx.doi.org/10.17587/it.26.323-334.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
2

FRANZ, UWE. "CLASSICAL MARKOV PROCESSES FROM QUANTUM LÉVY PROCESSES". Infinite Dimensional Analysis, Quantum Probability and Related Topics 02, n.º 01 (marzo de 1999): 105–29. http://dx.doi.org/10.1142/s0219025799000060.

Texto completo
Resumen
We show how classical Markov processes can be obtained from quantum Lévy processes. It is shown that quantum Lévy processes are quantum Markov processes, and sufficient conditions for restrictions to subalgebras to remain quantum Markov processes are given. A classical Markov process (which has the same time-ordered moments as the quantum process in the vacuum state) exists whenever we can restrict to a commutative subalgebra without losing the quantum Markov property.8 Several examples, including the Azéma martingale, with explicit calculations are presented. In particular, the action of the generator of the classical Markov processes on polynomials or their moments are calculated using Hopf algebra duality.
Los estilos APA, Harvard, Vancouver, ISO, etc.
3

Demenkov, N. P., E. A. Mirkin y I. A. Mochalov. "Markov and Semi-Markov Processes with Fuzzy States. Part 2. Semi-Markov Processes". INFORMACIONNYE TEHNOLOGII 26, n.º 7 (17 de julio de 2020): 387–93. http://dx.doi.org/10.17587/it.26.387-393.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
4

Whittle, P. y M. L. Puterman. "Markov Decision Processes." Journal of the Royal Statistical Society. Series A (Statistics in Society) 158, n.º 3 (1995): 636. http://dx.doi.org/10.2307/2983459.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
5

Smith, J. Q. y D. J. White. "Markov Decision Processes." Journal of the Royal Statistical Society. Series A (Statistics in Society) 157, n.º 1 (1994): 164. http://dx.doi.org/10.2307/2983520.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
6

King, Aaron A., Qianying Lin y Edward L. Ionides. "Markov genealogy processes". Theoretical Population Biology 143 (febrero de 2022): 77–91. http://dx.doi.org/10.1016/j.tpb.2021.11.003.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
7

Thomas, L. C., D. J. White y Martin L. Puterman. "Markov Decision Processes." Journal of the Operational Research Society 46, n.º 6 (junio de 1995): 792. http://dx.doi.org/10.2307/2584317.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
8

Ephraim, Y. y N. Merhav. "Hidden Markov processes". IEEE Transactions on Information Theory 48, n.º 6 (junio de 2002): 1518–69. http://dx.doi.org/10.1109/tit.2002.1003838.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
9

Bäuerle, Nicole y Ulrich Rieder. "Markov Decision Processes". Jahresbericht der Deutschen Mathematiker-Vereinigung 112, n.º 4 (8 de septiembre de 2010): 217–43. http://dx.doi.org/10.1365/s13291-010-0007-2.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
10

Wal, J. y J. Wessels. "MARKOV DECISION PROCESSES". Statistica Neerlandica 39, n.º 2 (junio de 1985): 219–33. http://dx.doi.org/10.1111/j.1467-9574.1985.tb01140.x.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
11

Thomas, L. C. "Markov Decision Processes". Journal of the Operational Research Society 46, n.º 6 (junio de 1995): 792–93. http://dx.doi.org/10.1057/jors.1995.110.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
12

Frank, T. D. "Nonlinear Markov processes". Physics Letters A 372, n.º 25 (junio de 2008): 4553–55. http://dx.doi.org/10.1016/j.physleta.2008.04.027.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
13

Kinateder, Kimberly K. J. "Corner Markov processes". Journal of Theoretical Probability 8, n.º 3 (julio de 1995): 539–47. http://dx.doi.org/10.1007/bf02218043.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
14

Brooks, Stephen y D. J. White. "Markov Decision Processes." Statistician 44, n.º 2 (1995): 292. http://dx.doi.org/10.2307/2348465.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
15

Fagnola, Franco. "Algebraic Markov processes". Proyecciones (Antofagasta) 18, n.º 3 (1999): 13–28. http://dx.doi.org/10.22199/s07160917.1999.0003.00003.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
16

Craven, B. D. "Perturbed Markov Processes". Stochastic Models 19, n.º 2 (5 de enero de 2003): 269–85. http://dx.doi.org/10.1081/stm-120020390.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
17

White, Chelsea C. y Douglas J. White. "Markov decision processes". European Journal of Operational Research 39, n.º 1 (marzo de 1989): 1–16. http://dx.doi.org/10.1016/0377-2217(89)90348-2.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
18

Zhenting, Hou, Liu Zaiming y Zou Jiezhong. "Markov skeleton processes". Chinese Science Bulletin 43, n.º 11 (junio de 1998): 881–89. http://dx.doi.org/10.1007/bf02884605.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
19

Franz, Uwe, Volkmar Liebscher y Stefan Zeiser. "Piecewise-Deterministic Markov Processes as Limits of Markov Jump Processes". Advances in Applied Probability 44, n.º 3 (septiembre de 2012): 729–48. http://dx.doi.org/10.1239/aap/1346955262.

Texto completo
Resumen
A classical result about Markov jump processes states that a certain class of dynamical systems given by ordinary differential equations are obtained as the limit of a sequence of scaled Markov jump processes. This approach fails if the scaling cannot be carried out equally across all entities. In the present paper we present a convergence theorem for such an unequal scaling. In contrast to an equal scaling the limit process is not purely deterministic but still possesses randomness. We show that these processes constitute a rich subclass of piecewise-deterministic processes. Such processes apply in molecular biology where entities often occur in different scales of numbers.
Los estilos APA, Harvard, Vancouver, ISO, etc.
20

Franz, Uwe, Volkmar Liebscher y Stefan Zeiser. "Piecewise-Deterministic Markov Processes as Limits of Markov Jump Processes". Advances in Applied Probability 44, n.º 03 (septiembre de 2012): 729–48. http://dx.doi.org/10.1017/s0001867800005851.

Texto completo
Resumen
A classical result about Markov jump processes states that a certain class of dynamical systems given by ordinary differential equations are obtained as the limit of a sequence of scaled Markov jump processes. This approach fails if the scaling cannot be carried out equally across all entities. In the present paper we present a convergence theorem for such an unequal scaling. In contrast to an equal scaling the limit process is not purely deterministic but still possesses randomness. We show that these processes constitute a rich subclass of piecewise-deterministic processes. Such processes apply in molecular biology where entities often occur in different scales of numbers.
Los estilos APA, Harvard, Vancouver, ISO, etc.
21

Buchholz, Peter y Miklós Telek. "Rational Processes Related to Communicating Markov Processes". Journal of Applied Probability 49, n.º 1 (marzo de 2012): 40–59. http://dx.doi.org/10.1239/jap/1331216833.

Texto completo
Resumen
We define a class of stochastic processes, denoted as marked rational arrival processes (MRAPs), which is an extension of matrix exponential distributions and rational arrival processes. Continuous-time Markov processes with labeled transitions are a subclass of this more general model class. New equivalence relations between processes are defined, and it is shown that these equivalence relations are natural extensions of strong and weak lumpability and the corresponding bisimulation relations that have been defined for Markov processes. If a general rational process is equivalent to a Markov process, it can be used in numerical analysis techniques instead of the Markov process. This observation allows one to apply MRAPs like Markov processes and since the new equivalence relations are more general than lumpability and bisimulation, it is sometimes possible to find smaller representations of given processes. Finally, we show that the equivalence is preserved by the composition of MRAPs and can therefore be exploited in compositional modeling.
Los estilos APA, Harvard, Vancouver, ISO, etc.
22

Buchholz, Peter y Miklós Telek. "Rational Processes Related to Communicating Markov Processes". Journal of Applied Probability 49, n.º 01 (marzo de 2012): 40–59. http://dx.doi.org/10.1017/s0021900200008858.

Texto completo
Resumen
We define a class of stochastic processes, denoted as marked rational arrival processes (MRAPs), which is an extension of matrix exponential distributions and rational arrival processes. Continuous-time Markov processes with labeled transitions are a subclass of this more general model class. New equivalence relations between processes are defined, and it is shown that these equivalence relations are natural extensions of strong and weak lumpability and the corresponding bisimulation relations that have been defined for Markov processes. If a general rational process is equivalent to a Markov process, it can be used in numerical analysis techniques instead of the Markov process. This observation allows one to apply MRAPs like Markov processes and since the new equivalence relations are more general than lumpability and bisimulation, it is sometimes possible to find smaller representations of given processes. Finally, we show that the equivalence is preserved by the composition of MRAPs and can therefore be exploited in compositional modeling.
Los estilos APA, Harvard, Vancouver, ISO, etc.
23

Iwata, Yukiko. "Constrictive Markov operators induced by Markov processes". Positivity 20, n.º 2 (3 de septiembre de 2015): 355–67. http://dx.doi.org/10.1007/s11117-015-0360-6.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
24

Malinovskii, V. K. "Limit theorems for recurrent semi-Markov processes and Markov renewal processes". Journal of Soviet Mathematics 36, n.º 4 (febrero de 1987): 493–502. http://dx.doi.org/10.1007/bf01663460.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
25

Fredkin, Donald R. y John A. Rice. "On aggregated Markov processes". Journal of Applied Probability 23, n.º 1 (marzo de 1986): 208–14. http://dx.doi.org/10.2307/3214130.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
26

Kazak, Jolanta. "Piecewise-deterministic Markov processes". Annales Polonici Mathematici 109, n.º 3 (2013): 279–96. http://dx.doi.org/10.4064/ap109-3-4.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
27

Lee, P. M. y O. Hernandez-Lerma. "Adaptive Markov Control Processes". Mathematical Gazette 74, n.º 470 (diciembre de 1990): 417. http://dx.doi.org/10.2307/3618186.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
28

Kowalski, Zbigniew S. "Multiple Markov Gaussian processes". Applicationes Mathematicae 48, n.º 1 (2021): 65–78. http://dx.doi.org/10.4064/am2411-1-2021.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
29

SHIEH, Narn-Rueih. "Collisions of Markov Processes". Tokyo Journal of Mathematics 18, n.º 1 (junio de 1995): 111–21. http://dx.doi.org/10.3836/tjm/1270043612.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
30

Hawkes, Alan G. "Markov processes in APL". ACM SIGAPL APL Quote Quad 20, n.º 4 (mayo de 1990): 173–85. http://dx.doi.org/10.1145/97811.97843.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
31

Pollett, P. K. "Connecting reversible Markov processes". Advances in Applied Probability 18, n.º 4 (diciembre de 1986): 880–900. http://dx.doi.org/10.2307/1427254.

Texto completo
Resumen
We provide a framework for interconnecting a collection of reversible Markov processes in such a way that the resulting process has a product-form invariant measure with respect to which the process is reversible. A number of examples are discussed including Kingman&s reversible migration process, interconnected random walks and stratified clustering processes.
Los estilos APA, Harvard, Vancouver, ISO, etc.
32

Gerontidis, Ioannis I. "Markov population replacement processes". Advances in Applied Probability 27, n.º 3 (septiembre de 1995): 711–40. http://dx.doi.org/10.2307/1428131.

Texto completo
Resumen
We consider a migration process whose singleton process is a time-dependent Markov replacement process. For the singleton process, which may be treated as either open or closed, we study the limiting distribution, the distribution of the time to replacement and related quantities. For a replacement process in equilibrium we obtain a version of Little's law and we provide conditions for reversibility. For the resulting linear population process we characterize exponential ergodicity for two types of environmental behaviour, i.e. either convergent or cyclic, and finally for large population sizes a diffusion approximation analysis is provided.
Los estilos APA, Harvard, Vancouver, ISO, etc.
33

Alexopoulos, Christos, Akram A. El-Tannir y Richard F. Serfozo. "Partition-Reversible Markov Processes". Operations Research 47, n.º 1 (febrero de 1999): 125–30. http://dx.doi.org/10.1287/opre.47.1.125.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
34

Avrachenkov, Konstantin, Alexey Piunovskiy y Yi Zhang. "Markov Processes with Restart". Journal of Applied Probability 50, n.º 4 (diciembre de 2013): 960–68. http://dx.doi.org/10.1239/jap/1389370093.

Texto completo
Resumen
We consider a general homogeneous continuous-time Markov process with restarts. The process is forced to restart from a given distribution at time moments generated by an independent Poisson process. The motivation to study such processes comes from modeling human and animal mobility patterns, restart processes in communication protocols, and from application of restarting random walks in information retrieval. We provide a connection between the transition probability functions of the original Markov process and the modified process with restarts. We give closed-form expressions for the invariant probability measure of the modified process. When the process evolves on the Euclidean space, there is also a closed-form expression for the moments of the modified process. We show that the modified process is always positive Harris recurrent and exponentially ergodic with the index equal to (or greater than) the rate of restarts. Finally, we illustrate the general results by the standard and geometric Brownian motions.
Los estilos APA, Harvard, Vancouver, ISO, etc.
35

Novak, Stephanie y Lyman J. Fretwell. "Non‐Markov noise processes". Journal of the Acoustical Society of America 80, S1 (diciembre de 1986): S64. http://dx.doi.org/10.1121/1.2023904.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
36

Rodrigues, Josemar, N. Balakrishnan y Patrick Borges. "Markov-Correlated Poisson Processes". Communications in Statistics - Theory and Methods 42, n.º 20 (18 de octubre de 2013): 3696–703. http://dx.doi.org/10.1080/03610926.2011.636168.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
37

Larralde, H., F. Leyvraz y D. P. Sanders. "Metastability in Markov processes". Journal of Statistical Mechanics: Theory and Experiment 2006, n.º 08 (18 de agosto de 2006): P08013. http://dx.doi.org/10.1088/1742-5468/2006/08/p08013.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
38

jun luo, Shou. "Two-parameter markov processes". Stochastics and Stochastic Reports 40, n.º 3-4 (septiembre de 1992): 181–93. http://dx.doi.org/10.1080/17442509208833788.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
39

Chin, Y. C. y A. J. Baddeley. "Markov interacting component processes". Advances in Applied Probability 32, n.º 3 (septiembre de 2000): 597–619. http://dx.doi.org/10.1239/aap/1013540233.

Texto completo
Resumen
A generalization of Markov point processes is introduced in which interactions occur between connected components of the point pattern. A version of the Hammersley-Clifford characterization theorem is proved which states that a point process is a Markov interacting component process if and only if its density function is a product of interaction terms associated with cliques of connected components. Integrability and superpositional properties of the processes are shown and a pairwise interaction example is used for detailed exploration.
Los estilos APA, Harvard, Vancouver, ISO, etc.
40

Cai, Haiyan. "Piecewise deterministic Markov processes". Stochastic Analysis and Applications 11, n.º 3 (enero de 1993): 255–74. http://dx.doi.org/10.1080/07362999308809317.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
41

ACCARDI, LUIGI y ANILESH MOHARI. "TIME REFLECTED MARKOV PROCESSES". Infinite Dimensional Analysis, Quantum Probability and Related Topics 02, n.º 03 (septiembre de 1999): 397–425. http://dx.doi.org/10.1142/s0219025799000230.

Texto completo
Resumen
A classical stochastic process which is Markovian for its past filtration is also Markovian for its future filtration. We show with a counterexample based on quantum liftings of a finite state classical Markov chain that this property cannot hold in the category of expected Markov processes. Using a duality theory for von Neumann algebras with weights, developed by Petz on the basis of previous results by Groh and Kümmerer, we show that a quantum version of this symmetry can be established in the category of weak Markov processes in the sense of Bhat and Parthasarathy. Here time reversal is implemented by an anti-unitary operator and a weak Markov process is time reversal invariant if and only if the associated semigroup coincides with its Petz dual. This construction allows one to extend to the quantum case, both for backward and forward processes, the Misra–Prigogine–Courbage internal time operator and to show that the two operators are intertwined by the time reversal anti-automorphism.
Los estilos APA, Harvard, Vancouver, ISO, etc.
42

Fredkin, Donald R. y John A. Rice. "On aggregated Markov processes". Journal of Applied Probability 23, n.º 01 (marzo de 1986): 208–14. http://dx.doi.org/10.1017/s0021900200106412.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
43

Halibard, Moishe y Ido Kanter. "Markov processes and linguistics". Physica A: Statistical Mechanics and its Applications 249, n.º 1-4 (enero de 1998): 525–35. http://dx.doi.org/10.1016/s0378-4371(97)00512-8.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
44

Chin, Y. C. y A. J. Baddeley. "Markov interacting component processes". Advances in Applied Probability 32, n.º 03 (septiembre de 2000): 597–619. http://dx.doi.org/10.1017/s0001867800010144.

Texto completo
Resumen
A generalization of Markov point processes is introduced in which interactions occur between connected components of the point pattern. A version of the Hammersley-Clifford characterization theorem is proved which states that a point process is a Markov interacting component process if and only if its density function is a product of interaction terms associated with cliques of connected components. Integrability and superpositional properties of the processes are shown and a pairwise interaction example is used for detailed exploration.
Los estilos APA, Harvard, Vancouver, ISO, etc.
45

Pollett, P. K. "Connecting reversible Markov processes". Advances in Applied Probability 18, n.º 04 (diciembre de 1986): 880–900. http://dx.doi.org/10.1017/s0001867800016190.

Texto completo
Resumen
We provide a framework for interconnecting a collection of reversible Markov processes in such a way that the resulting process has a product-form invariant measure with respect to which the process is reversible. A number of examples are discussed including Kingman&s reversible migration process, interconnected random walks and stratified clustering processes.
Los estilos APA, Harvard, Vancouver, ISO, etc.
46

Gerontidis, Ioannis I. "Markov population replacement processes". Advances in Applied Probability 27, n.º 03 (septiembre de 1995): 711–40. http://dx.doi.org/10.1017/s0001867800027129.

Texto completo
Resumen
We consider a migration process whose singleton process is a time-dependent Markov replacement process. For the singleton process, which may be treated as either open or closed, we study the limiting distribution, the distribution of the time to replacement and related quantities. For a replacement process in equilibrium we obtain a version of Little's law and we provide conditions for reversibility. For the resulting linear population process we characterize exponential ergodicity for two types of environmental behaviour, i.e. either convergent or cyclic, and finally for large population sizes a diffusion approximation analysis is provided.
Los estilos APA, Harvard, Vancouver, ISO, etc.
47

Avrachenkov, Konstantin, Alexey Piunovskiy y Yi Zhang. "Markov Processes with Restart". Journal of Applied Probability 50, n.º 04 (diciembre de 2013): 960–68. http://dx.doi.org/10.1017/s0021900200013735.

Texto completo
Resumen
We consider a general homogeneous continuous-time Markov process with restarts. The process is forced to restart from a given distribution at time moments generated by an independent Poisson process. The motivation to study such processes comes from modeling human and animal mobility patterns, restart processes in communication protocols, and from application of restarting random walks in information retrieval. We provide a connection between the transition probability functions of the original Markov process and the modified process with restarts. We give closed-form expressions for the invariant probability measure of the modified process. When the process evolves on the Euclidean space, there is also a closed-form expression for the moments of the modified process. We show that the modified process is always positive Harris recurrent and exponentially ergodic with the index equal to (or greater than) the rate of restarts. Finally, we illustrate the general results by the standard and geometric Brownian motions.
Los estilos APA, Harvard, Vancouver, ISO, etc.
48

Baykal-Gürsoy, M. y K. Gürsoy. "SEMI-MARKOV DECISION PROCESSES". Probability in the Engineering and Informational Sciences 21, n.º 4 (octubre de 2007): 635–57. http://dx.doi.org/10.1017/s026996480700037x.

Texto completo
Resumen
Considered are semi-Markov decision processes (SMDPs) with finite state and action spaces. We study two criteria: the expected average reward per unit time subject to a sample path constraint on the average cost per unit time and the expected time-average variability. Under a certain condition, for communicating SMDPs, we construct (randomized) stationary policies that are ε-optimal for each criterion; the policy is optimal for the first criterion under the unichain assumption and the policy is optimal and pure for a specific variability function in the second criterion. For general multichain SMDPs, by using a state space decomposition approach, similar results are obtained.
Los estilos APA, Harvard, Vancouver, ISO, etc.
49

Desharnais, Josée, Vineet Gupta, Radha Jagadeesan y Prakash Panangaden. "Approximating labelled Markov processes". Information and Computation 184, n.º 1 (julio de 2003): 160–200. http://dx.doi.org/10.1016/s0890-5401(03)00051-8.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
50

Sabbadin, Régis. "Possibilistic Markov decision processes". Engineering Applications of Artificial Intelligence 14, n.º 3 (junio de 2001): 287–300. http://dx.doi.org/10.1016/s0952-1976(01)00007-0.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
Ofrecemos descuentos en todos los planes premium para autores cuyas obras están incluidas en selecciones literarias temáticas. ¡Contáctenos para obtener un código promocional único!

Pasar a la bibliografía