Segui questo link per vedere altri tipi di pubblicazioni sul tema: Markov processes.

Articoli di riviste sul tema "Markov processes"

Cita una fonte nei formati APA, MLA, Chicago, Harvard e in molti altri stili

Scegli il tipo di fonte:

Vedi i top-50 articoli di riviste per l'attività di ricerca sul tema "Markov processes".

Accanto a ogni fonte nell'elenco di riferimenti c'è un pulsante "Aggiungi alla bibliografia". Premilo e genereremo automaticamente la citazione bibliografica dell'opera scelta nello stile citazionale di cui hai bisogno: APA, MLA, Harvard, Chicago, Vancouver ecc.

Puoi anche scaricare il testo completo della pubblicazione scientifica nel formato .pdf e leggere online l'abstract (il sommario) dell'opera se è presente nei metadati.

Vedi gli articoli di riviste di molte aree scientifiche e compila una bibliografia corretta.

1

Demenkov, N. P., E. A. Mirkin e I. A. Mochalov. "Markov and Semi-Markov Processes with Fuzzy States. Part 1. Markov Processes". Informacionnye tehnologii 26, n. 6 (23 giugno 2020): 323–34. http://dx.doi.org/10.17587/it.26.323-334.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
2

FRANZ, UWE. "CLASSICAL MARKOV PROCESSES FROM QUANTUM LÉVY PROCESSES". Infinite Dimensional Analysis, Quantum Probability and Related Topics 02, n. 01 (marzo 1999): 105–29. http://dx.doi.org/10.1142/s0219025799000060.

Testo completo
Abstract (sommario):
We show how classical Markov processes can be obtained from quantum Lévy processes. It is shown that quantum Lévy processes are quantum Markov processes, and sufficient conditions for restrictions to subalgebras to remain quantum Markov processes are given. A classical Markov process (which has the same time-ordered moments as the quantum process in the vacuum state) exists whenever we can restrict to a commutative subalgebra without losing the quantum Markov property.8 Several examples, including the Azéma martingale, with explicit calculations are presented. In particular, the action of the generator of the classical Markov processes on polynomials or their moments are calculated using Hopf algebra duality.
Gli stili APA, Harvard, Vancouver, ISO e altri
3

Demenkov, N. P., E. A. Mirkin e I. A. Mochalov. "Markov and Semi-Markov Processes with Fuzzy States. Part 2. Semi-Markov Processes". INFORMACIONNYE TEHNOLOGII 26, n. 7 (17 luglio 2020): 387–93. http://dx.doi.org/10.17587/it.26.387-393.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
4

Whittle, P., e M. L. Puterman. "Markov Decision Processes." Journal of the Royal Statistical Society. Series A (Statistics in Society) 158, n. 3 (1995): 636. http://dx.doi.org/10.2307/2983459.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
5

Smith, J. Q., e D. J. White. "Markov Decision Processes." Journal of the Royal Statistical Society. Series A (Statistics in Society) 157, n. 1 (1994): 164. http://dx.doi.org/10.2307/2983520.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
6

King, Aaron A., Qianying Lin e Edward L. Ionides. "Markov genealogy processes". Theoretical Population Biology 143 (febbraio 2022): 77–91. http://dx.doi.org/10.1016/j.tpb.2021.11.003.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
7

Thomas, L. C., D. J. White e Martin L. Puterman. "Markov Decision Processes." Journal of the Operational Research Society 46, n. 6 (giugno 1995): 792. http://dx.doi.org/10.2307/2584317.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
8

Ephraim, Y., e N. Merhav. "Hidden Markov processes". IEEE Transactions on Information Theory 48, n. 6 (giugno 2002): 1518–69. http://dx.doi.org/10.1109/tit.2002.1003838.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
9

Bäuerle, Nicole, e Ulrich Rieder. "Markov Decision Processes". Jahresbericht der Deutschen Mathematiker-Vereinigung 112, n. 4 (8 settembre 2010): 217–43. http://dx.doi.org/10.1365/s13291-010-0007-2.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
10

Wal, J., e J. Wessels. "MARKOV DECISION PROCESSES". Statistica Neerlandica 39, n. 2 (giugno 1985): 219–33. http://dx.doi.org/10.1111/j.1467-9574.1985.tb01140.x.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
11

Thomas, L. C. "Markov Decision Processes". Journal of the Operational Research Society 46, n. 6 (giugno 1995): 792–93. http://dx.doi.org/10.1057/jors.1995.110.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
12

Frank, T. D. "Nonlinear Markov processes". Physics Letters A 372, n. 25 (giugno 2008): 4553–55. http://dx.doi.org/10.1016/j.physleta.2008.04.027.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
13

Kinateder, Kimberly K. J. "Corner Markov processes". Journal of Theoretical Probability 8, n. 3 (luglio 1995): 539–47. http://dx.doi.org/10.1007/bf02218043.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
14

Brooks, Stephen, e D. J. White. "Markov Decision Processes." Statistician 44, n. 2 (1995): 292. http://dx.doi.org/10.2307/2348465.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
15

Fagnola, Franco. "Algebraic Markov processes". Proyecciones (Antofagasta) 18, n. 3 (1999): 13–28. http://dx.doi.org/10.22199/s07160917.1999.0003.00003.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
16

Craven, B. D. "Perturbed Markov Processes". Stochastic Models 19, n. 2 (5 gennaio 2003): 269–85. http://dx.doi.org/10.1081/stm-120020390.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
17

White, Chelsea C., e Douglas J. White. "Markov decision processes". European Journal of Operational Research 39, n. 1 (marzo 1989): 1–16. http://dx.doi.org/10.1016/0377-2217(89)90348-2.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
18

Zhenting, Hou, Liu Zaiming e Zou Jiezhong. "Markov skeleton processes". Chinese Science Bulletin 43, n. 11 (giugno 1998): 881–89. http://dx.doi.org/10.1007/bf02884605.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
19

Franz, Uwe, Volkmar Liebscher e Stefan Zeiser. "Piecewise-Deterministic Markov Processes as Limits of Markov Jump Processes". Advances in Applied Probability 44, n. 3 (settembre 2012): 729–48. http://dx.doi.org/10.1239/aap/1346955262.

Testo completo
Abstract (sommario):
A classical result about Markov jump processes states that a certain class of dynamical systems given by ordinary differential equations are obtained as the limit of a sequence of scaled Markov jump processes. This approach fails if the scaling cannot be carried out equally across all entities. In the present paper we present a convergence theorem for such an unequal scaling. In contrast to an equal scaling the limit process is not purely deterministic but still possesses randomness. We show that these processes constitute a rich subclass of piecewise-deterministic processes. Such processes apply in molecular biology where entities often occur in different scales of numbers.
Gli stili APA, Harvard, Vancouver, ISO e altri
20

Franz, Uwe, Volkmar Liebscher e Stefan Zeiser. "Piecewise-Deterministic Markov Processes as Limits of Markov Jump Processes". Advances in Applied Probability 44, n. 03 (settembre 2012): 729–48. http://dx.doi.org/10.1017/s0001867800005851.

Testo completo
Abstract (sommario):
A classical result about Markov jump processes states that a certain class of dynamical systems given by ordinary differential equations are obtained as the limit of a sequence of scaled Markov jump processes. This approach fails if the scaling cannot be carried out equally across all entities. In the present paper we present a convergence theorem for such an unequal scaling. In contrast to an equal scaling the limit process is not purely deterministic but still possesses randomness. We show that these processes constitute a rich subclass of piecewise-deterministic processes. Such processes apply in molecular biology where entities often occur in different scales of numbers.
Gli stili APA, Harvard, Vancouver, ISO e altri
21

Buchholz, Peter, e Miklós Telek. "Rational Processes Related to Communicating Markov Processes". Journal of Applied Probability 49, n. 1 (marzo 2012): 40–59. http://dx.doi.org/10.1239/jap/1331216833.

Testo completo
Abstract (sommario):
We define a class of stochastic processes, denoted as marked rational arrival processes (MRAPs), which is an extension of matrix exponential distributions and rational arrival processes. Continuous-time Markov processes with labeled transitions are a subclass of this more general model class. New equivalence relations between processes are defined, and it is shown that these equivalence relations are natural extensions of strong and weak lumpability and the corresponding bisimulation relations that have been defined for Markov processes. If a general rational process is equivalent to a Markov process, it can be used in numerical analysis techniques instead of the Markov process. This observation allows one to apply MRAPs like Markov processes and since the new equivalence relations are more general than lumpability and bisimulation, it is sometimes possible to find smaller representations of given processes. Finally, we show that the equivalence is preserved by the composition of MRAPs and can therefore be exploited in compositional modeling.
Gli stili APA, Harvard, Vancouver, ISO e altri
22

Buchholz, Peter, e Miklós Telek. "Rational Processes Related to Communicating Markov Processes". Journal of Applied Probability 49, n. 01 (marzo 2012): 40–59. http://dx.doi.org/10.1017/s0021900200008858.

Testo completo
Abstract (sommario):
We define a class of stochastic processes, denoted as marked rational arrival processes (MRAPs), which is an extension of matrix exponential distributions and rational arrival processes. Continuous-time Markov processes with labeled transitions are a subclass of this more general model class. New equivalence relations between processes are defined, and it is shown that these equivalence relations are natural extensions of strong and weak lumpability and the corresponding bisimulation relations that have been defined for Markov processes. If a general rational process is equivalent to a Markov process, it can be used in numerical analysis techniques instead of the Markov process. This observation allows one to apply MRAPs like Markov processes and since the new equivalence relations are more general than lumpability and bisimulation, it is sometimes possible to find smaller representations of given processes. Finally, we show that the equivalence is preserved by the composition of MRAPs and can therefore be exploited in compositional modeling.
Gli stili APA, Harvard, Vancouver, ISO e altri
23

Iwata, Yukiko. "Constrictive Markov operators induced by Markov processes". Positivity 20, n. 2 (3 settembre 2015): 355–67. http://dx.doi.org/10.1007/s11117-015-0360-6.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
24

Malinovskii, V. K. "Limit theorems for recurrent semi-Markov processes and Markov renewal processes". Journal of Soviet Mathematics 36, n. 4 (febbraio 1987): 493–502. http://dx.doi.org/10.1007/bf01663460.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
25

Fredkin, Donald R., e John A. Rice. "On aggregated Markov processes". Journal of Applied Probability 23, n. 1 (marzo 1986): 208–14. http://dx.doi.org/10.2307/3214130.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
26

Kazak, Jolanta. "Piecewise-deterministic Markov processes". Annales Polonici Mathematici 109, n. 3 (2013): 279–96. http://dx.doi.org/10.4064/ap109-3-4.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
27

Lee, P. M., e O. Hernandez-Lerma. "Adaptive Markov Control Processes". Mathematical Gazette 74, n. 470 (dicembre 1990): 417. http://dx.doi.org/10.2307/3618186.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
28

Kowalski, Zbigniew S. "Multiple Markov Gaussian processes". Applicationes Mathematicae 48, n. 1 (2021): 65–78. http://dx.doi.org/10.4064/am2411-1-2021.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
29

SHIEH, Narn-Rueih. "Collisions of Markov Processes". Tokyo Journal of Mathematics 18, n. 1 (giugno 1995): 111–21. http://dx.doi.org/10.3836/tjm/1270043612.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
30

Hawkes, Alan G. "Markov processes in APL". ACM SIGAPL APL Quote Quad 20, n. 4 (maggio 1990): 173–85. http://dx.doi.org/10.1145/97811.97843.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
31

Pollett, P. K. "Connecting reversible Markov processes". Advances in Applied Probability 18, n. 4 (dicembre 1986): 880–900. http://dx.doi.org/10.2307/1427254.

Testo completo
Abstract (sommario):
We provide a framework for interconnecting a collection of reversible Markov processes in such a way that the resulting process has a product-form invariant measure with respect to which the process is reversible. A number of examples are discussed including Kingman&s reversible migration process, interconnected random walks and stratified clustering processes.
Gli stili APA, Harvard, Vancouver, ISO e altri
32

Gerontidis, Ioannis I. "Markov population replacement processes". Advances in Applied Probability 27, n. 3 (settembre 1995): 711–40. http://dx.doi.org/10.2307/1428131.

Testo completo
Abstract (sommario):
We consider a migration process whose singleton process is a time-dependent Markov replacement process. For the singleton process, which may be treated as either open or closed, we study the limiting distribution, the distribution of the time to replacement and related quantities. For a replacement process in equilibrium we obtain a version of Little's law and we provide conditions for reversibility. For the resulting linear population process we characterize exponential ergodicity for two types of environmental behaviour, i.e. either convergent or cyclic, and finally for large population sizes a diffusion approximation analysis is provided.
Gli stili APA, Harvard, Vancouver, ISO e altri
33

Alexopoulos, Christos, Akram A. El-Tannir e Richard F. Serfozo. "Partition-Reversible Markov Processes". Operations Research 47, n. 1 (febbraio 1999): 125–30. http://dx.doi.org/10.1287/opre.47.1.125.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
34

Avrachenkov, Konstantin, Alexey Piunovskiy e Yi Zhang. "Markov Processes with Restart". Journal of Applied Probability 50, n. 4 (dicembre 2013): 960–68. http://dx.doi.org/10.1239/jap/1389370093.

Testo completo
Abstract (sommario):
We consider a general homogeneous continuous-time Markov process with restarts. The process is forced to restart from a given distribution at time moments generated by an independent Poisson process. The motivation to study such processes comes from modeling human and animal mobility patterns, restart processes in communication protocols, and from application of restarting random walks in information retrieval. We provide a connection between the transition probability functions of the original Markov process and the modified process with restarts. We give closed-form expressions for the invariant probability measure of the modified process. When the process evolves on the Euclidean space, there is also a closed-form expression for the moments of the modified process. We show that the modified process is always positive Harris recurrent and exponentially ergodic with the index equal to (or greater than) the rate of restarts. Finally, we illustrate the general results by the standard and geometric Brownian motions.
Gli stili APA, Harvard, Vancouver, ISO e altri
35

Novak, Stephanie, e Lyman J. Fretwell. "Non‐Markov noise processes". Journal of the Acoustical Society of America 80, S1 (dicembre 1986): S64. http://dx.doi.org/10.1121/1.2023904.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
36

Rodrigues, Josemar, N. Balakrishnan e Patrick Borges. "Markov-Correlated Poisson Processes". Communications in Statistics - Theory and Methods 42, n. 20 (18 ottobre 2013): 3696–703. http://dx.doi.org/10.1080/03610926.2011.636168.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
37

Larralde, H., F. Leyvraz e D. P. Sanders. "Metastability in Markov processes". Journal of Statistical Mechanics: Theory and Experiment 2006, n. 08 (18 agosto 2006): P08013. http://dx.doi.org/10.1088/1742-5468/2006/08/p08013.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
38

jun luo, Shou. "Two-parameter markov processes". Stochastics and Stochastic Reports 40, n. 3-4 (settembre 1992): 181–93. http://dx.doi.org/10.1080/17442509208833788.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
39

Chin, Y. C., e A. J. Baddeley. "Markov interacting component processes". Advances in Applied Probability 32, n. 3 (settembre 2000): 597–619. http://dx.doi.org/10.1239/aap/1013540233.

Testo completo
Abstract (sommario):
A generalization of Markov point processes is introduced in which interactions occur between connected components of the point pattern. A version of the Hammersley-Clifford characterization theorem is proved which states that a point process is a Markov interacting component process if and only if its density function is a product of interaction terms associated with cliques of connected components. Integrability and superpositional properties of the processes are shown and a pairwise interaction example is used for detailed exploration.
Gli stili APA, Harvard, Vancouver, ISO e altri
40

Cai, Haiyan. "Piecewise deterministic Markov processes". Stochastic Analysis and Applications 11, n. 3 (gennaio 1993): 255–74. http://dx.doi.org/10.1080/07362999308809317.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
41

ACCARDI, LUIGI, e ANILESH MOHARI. "TIME REFLECTED MARKOV PROCESSES". Infinite Dimensional Analysis, Quantum Probability and Related Topics 02, n. 03 (settembre 1999): 397–425. http://dx.doi.org/10.1142/s0219025799000230.

Testo completo
Abstract (sommario):
A classical stochastic process which is Markovian for its past filtration is also Markovian for its future filtration. We show with a counterexample based on quantum liftings of a finite state classical Markov chain that this property cannot hold in the category of expected Markov processes. Using a duality theory for von Neumann algebras with weights, developed by Petz on the basis of previous results by Groh and Kümmerer, we show that a quantum version of this symmetry can be established in the category of weak Markov processes in the sense of Bhat and Parthasarathy. Here time reversal is implemented by an anti-unitary operator and a weak Markov process is time reversal invariant if and only if the associated semigroup coincides with its Petz dual. This construction allows one to extend to the quantum case, both for backward and forward processes, the Misra–Prigogine–Courbage internal time operator and to show that the two operators are intertwined by the time reversal anti-automorphism.
Gli stili APA, Harvard, Vancouver, ISO e altri
42

Fredkin, Donald R., e John A. Rice. "On aggregated Markov processes". Journal of Applied Probability 23, n. 01 (marzo 1986): 208–14. http://dx.doi.org/10.1017/s0021900200106412.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
43

Halibard, Moishe, e Ido Kanter. "Markov processes and linguistics". Physica A: Statistical Mechanics and its Applications 249, n. 1-4 (gennaio 1998): 525–35. http://dx.doi.org/10.1016/s0378-4371(97)00512-8.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
44

Chin, Y. C., e A. J. Baddeley. "Markov interacting component processes". Advances in Applied Probability 32, n. 03 (settembre 2000): 597–619. http://dx.doi.org/10.1017/s0001867800010144.

Testo completo
Abstract (sommario):
A generalization of Markov point processes is introduced in which interactions occur between connected components of the point pattern. A version of the Hammersley-Clifford characterization theorem is proved which states that a point process is a Markov interacting component process if and only if its density function is a product of interaction terms associated with cliques of connected components. Integrability and superpositional properties of the processes are shown and a pairwise interaction example is used for detailed exploration.
Gli stili APA, Harvard, Vancouver, ISO e altri
45

Pollett, P. K. "Connecting reversible Markov processes". Advances in Applied Probability 18, n. 04 (dicembre 1986): 880–900. http://dx.doi.org/10.1017/s0001867800016190.

Testo completo
Abstract (sommario):
We provide a framework for interconnecting a collection of reversible Markov processes in such a way that the resulting process has a product-form invariant measure with respect to which the process is reversible. A number of examples are discussed including Kingman&s reversible migration process, interconnected random walks and stratified clustering processes.
Gli stili APA, Harvard, Vancouver, ISO e altri
46

Gerontidis, Ioannis I. "Markov population replacement processes". Advances in Applied Probability 27, n. 03 (settembre 1995): 711–40. http://dx.doi.org/10.1017/s0001867800027129.

Testo completo
Abstract (sommario):
We consider a migration process whose singleton process is a time-dependent Markov replacement process. For the singleton process, which may be treated as either open or closed, we study the limiting distribution, the distribution of the time to replacement and related quantities. For a replacement process in equilibrium we obtain a version of Little's law and we provide conditions for reversibility. For the resulting linear population process we characterize exponential ergodicity for two types of environmental behaviour, i.e. either convergent or cyclic, and finally for large population sizes a diffusion approximation analysis is provided.
Gli stili APA, Harvard, Vancouver, ISO e altri
47

Avrachenkov, Konstantin, Alexey Piunovskiy e Yi Zhang. "Markov Processes with Restart". Journal of Applied Probability 50, n. 04 (dicembre 2013): 960–68. http://dx.doi.org/10.1017/s0021900200013735.

Testo completo
Abstract (sommario):
We consider a general homogeneous continuous-time Markov process with restarts. The process is forced to restart from a given distribution at time moments generated by an independent Poisson process. The motivation to study such processes comes from modeling human and animal mobility patterns, restart processes in communication protocols, and from application of restarting random walks in information retrieval. We provide a connection between the transition probability functions of the original Markov process and the modified process with restarts. We give closed-form expressions for the invariant probability measure of the modified process. When the process evolves on the Euclidean space, there is also a closed-form expression for the moments of the modified process. We show that the modified process is always positive Harris recurrent and exponentially ergodic with the index equal to (or greater than) the rate of restarts. Finally, we illustrate the general results by the standard and geometric Brownian motions.
Gli stili APA, Harvard, Vancouver, ISO e altri
48

Baykal-Gürsoy, M., e K. Gürsoy. "SEMI-MARKOV DECISION PROCESSES". Probability in the Engineering and Informational Sciences 21, n. 4 (ottobre 2007): 635–57. http://dx.doi.org/10.1017/s026996480700037x.

Testo completo
Abstract (sommario):
Considered are semi-Markov decision processes (SMDPs) with finite state and action spaces. We study two criteria: the expected average reward per unit time subject to a sample path constraint on the average cost per unit time and the expected time-average variability. Under a certain condition, for communicating SMDPs, we construct (randomized) stationary policies that are ε-optimal for each criterion; the policy is optimal for the first criterion under the unichain assumption and the policy is optimal and pure for a specific variability function in the second criterion. For general multichain SMDPs, by using a state space decomposition approach, similar results are obtained.
Gli stili APA, Harvard, Vancouver, ISO e altri
49

Desharnais, Josée, Vineet Gupta, Radha Jagadeesan e Prakash Panangaden. "Approximating labelled Markov processes". Information and Computation 184, n. 1 (luglio 2003): 160–200. http://dx.doi.org/10.1016/s0890-5401(03)00051-8.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
50

Sabbadin, Régis. "Possibilistic Markov decision processes". Engineering Applications of Artificial Intelligence 14, n. 3 (giugno 2001): 287–300. http://dx.doi.org/10.1016/s0952-1976(01)00007-0.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
Offriamo sconti su tutti i piani premium per gli autori le cui opere sono incluse in raccolte letterarie tematiche. Contattaci per ottenere un codice promozionale unico!

Vai alla bibliografia