Siga este enlace para ver otros tipos de publicaciones sobre el tema: Markoc chain.

Artículos de revistas sobre el tema "Markoc chain"

Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros

Elija tipo de fuente:

Consulte los 50 mejores artículos de revistas para su investigación sobre el tema "Markoc chain".

Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.

También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.

Explore artículos de revistas sobre una amplia variedad de disciplinas y organice su bibliografía correctamente.

1

Guyo, X., and C. Hardouin†. "Markow chain markov field dynamics:models and statistics." Statistics 35, no. 4 (2001): 593–627. http://dx.doi.org/10.1080/02331880108802756.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
2

Valenzuela, Mississippi. "Markov chains and applications." Selecciones Matemáticas 9, no. 01 (2022): 53–78. http://dx.doi.org/10.17268/sel.mat.2022.01.05.

Texto completo
Resumen
This work has three important purposes: first it is the study of Markov Chains, the second is to show that Markov chains have different applications and finally it is to model a process of this behaves. Throughout this work we will describe what a Markov chain is, what these processes are for and how these chains are classified. We will describe a Markov Chain, that is, analyze what are the primary elements that make up a Markov chain, among others.
Los estilos APA, Harvard, Vancouver, ISO, etc.
3

Barker, Richard J., and Matthew R. Schofield. "Putting Markov Chains Back into Markov Chain Monte Carlo." Journal of Applied Mathematics and Decision Sciences 2007 (October 30, 2007): 1–13. http://dx.doi.org/10.1155/2007/98086.

Texto completo
Resumen
Markov chain theory plays an important role in statistical inference both in the formulation of models for data and in the construction of efficient algorithms for inference. The use of Markov chains in modeling data has a long history, however the use of Markov chain theory in developing algorithms for statistical inference has only become popular recently. Using mark-recapture models as an illustration, we show how Markov chains can be used for developing demographic models and also in developing efficient algorithms for inference. We anticipate that a major area of future research involving
Los estilos APA, Harvard, Vancouver, ISO, etc.
4

Marcus, Brian, and Selim Tuncel. "The weight-per-symbol polytope and scaffolds of invariants associated with Markov chains." Ergodic Theory and Dynamical Systems 11, no. 1 (1991): 129–80. http://dx.doi.org/10.1017/s0143385700006052.

Texto completo
Resumen
AbstractWe study Markov chains via invariants constructed from periodic orbits. Canonical extensions, based on these invariants, are used to establish a constraint on the degree of finite-to-one block homomorphisms from one Markov chain to another. We construct a polytope from the normalized weights of periodic orbits. Using this polytope, we find canonically-defined induced Markov chains inside the original Markov chain. Each of the invariants associated with these Markov chains gives rise to a scaffold of invariants for the original Markov chain. This is used to obtain counterexamples to the
Los estilos APA, Harvard, Vancouver, ISO, etc.
5

Xiang, Xuyan, Xiao Zhang, and Xiaoyun Mo. "Statistical Identification of Markov Chain on Trees." Mathematical Problems in Engineering 2018 (2018): 1–13. http://dx.doi.org/10.1155/2018/2036248.

Texto completo
Resumen
The theoretical study of continuous-time homogeneous Markov chains is usually based on a natural assumption of a known transition rate matrix (TRM). However, the TRM of a Markov chain in realistic systems might be unknown and might even need to be identified by partially observable data. Thus, an issue on how to identify the TRM of the underlying Markov chain by partially observable information is derived from the great significance in applications. That is what we call the statistical identification of Markov chain. The Markov chain inversion approach has been derived for basic Markov chains
Los estilos APA, Harvard, Vancouver, ISO, etc.
6

APOSTOLOV, S. S., Z. A. MAYZELIS, O. V. USATENKO, and V. A. YAMPOL'SKII. "HIGH-ORDER CORRELATION FUNCTIONS OF BINARY MULTI-STEP MARKOV CHAINS." International Journal of Modern Physics B 22, no. 22 (2008): 3841–53. http://dx.doi.org/10.1142/s0217979208048589.

Texto completo
Resumen
Two approaches to studying the correlation functions of the binary Markov sequences are considered. The first of them is based on the study of probability of occurring different "words" in the sequence. The other one uses recurrence relations for correlation functions. These methods are applied for two important particular classes of the Markov chains. These classes include the Markov chains with permutative conditional probability functions and the additive Markov chains with the small memory functions. The exciting property of the self-similarity (discovered in Phys. Rev. Lett.90, 110601 (20
Los estilos APA, Harvard, Vancouver, ISO, etc.
7

Masuyama, Hiroyuki. "Error Bounds for Augmented Truncations of Discrete-Time Block-Monotone Markov Chains under Geometric Drift Conditions." Advances in Applied Probability 47, no. 1 (2015): 83–105. http://dx.doi.org/10.1239/aap/1427814582.

Texto completo
Resumen
In this paper we study the augmented truncation of discrete-time block-monotone Markov chains under geometric drift conditions. We first present a bound for the total variation distance between the stationary distributions of an original Markov chain and its augmented truncation. We also obtain such error bounds for more general cases, where an original Markov chain itself is not necessarily block monotone but is blockwise dominated by a block-monotone Markov chain. Finally, we discuss the application of our results to GI/G/1-type Markov chains.
Los estilos APA, Harvard, Vancouver, ISO, etc.
8

Masuyama, Hiroyuki. "Error Bounds for Augmented Truncations of Discrete-Time Block-Monotone Markov Chains under Geometric Drift Conditions." Advances in Applied Probability 47, no. 01 (2015): 83–105. http://dx.doi.org/10.1017/s0001867800007710.

Texto completo
Resumen
In this paper we study the augmented truncation of discrete-time block-monotone Markov chains under geometric drift conditions. We first present a bound for the total variation distance between the stationary distributions of an original Markov chain and its augmented truncation. We also obtain such error bounds for more general cases, where an original Markov chain itself is not necessarily block monotone but is blockwise dominated by a block-monotone Markov chain. Finally, we discuss the application of our results to GI/G/1-type Markov chains.
Los estilos APA, Harvard, Vancouver, ISO, etc.
9

Ledoux, James. "A geometric invariant in weak lumpability of finite Markov chains." Journal of Applied Probability 34, no. 4 (1997): 847–58. http://dx.doi.org/10.2307/3215001.

Texto completo
Resumen
We consider weak lumpability of finite homogeneous Markov chains, which is when a lumped Markov chain with respect to a partition of the initial state space is also a homogeneous Markov chain. We show that weak lumpability is equivalent to the existence of a direct sum of polyhedral cones that is positively invariant by the transition probability matrix of the original chain. It allows us, in a unified way, to derive new results on lumpability of reducible Markov chains and to obtain spectral properties associated with lumpability.
Los estilos APA, Harvard, Vancouver, ISO, etc.
10

Ledoux, James. "A geometric invariant in weak lumpability of finite Markov chains." Journal of Applied Probability 34, no. 04 (1997): 847–58. http://dx.doi.org/10.1017/s0021900200101561.

Texto completo
Resumen
We consider weak lumpability of finite homogeneous Markov chains, which is when a lumped Markov chain with respect to a partition of the initial state space is also a homogeneous Markov chain. We show that weak lumpability is equivalent to the existence of a direct sum of polyhedral cones that is positively invariant by the transition probability matrix of the original chain. It allows us, in a unified way, to derive new results on lumpability of reducible Markov chains and to obtain spectral properties associated with lumpability.
Los estilos APA, Harvard, Vancouver, ISO, etc.
11

Verbeken, Brecht, and Marie-Anne Guerry. "Attainability for Markov and Semi-Markov Chains." Mathematics 12, no. 8 (2024): 1227. http://dx.doi.org/10.3390/math12081227.

Texto completo
Resumen
When studying Markov chain models and semi-Markov chain models, it is useful to know which state vectors n, where each component ni represents the number of entities in the state Si, can be maintained or attained. This question leads to the definitions of maintainability and attainability for (time-homogeneous) Markov chain models. Recently, the definition of maintainability was extended to the concept of state reunion maintainability (SR-maintainability) for semi-Markov chains. Within the framework of semi-Markov chains, the states are subdivided further into seniority-based states. State reu
Los estilos APA, Harvard, Vancouver, ISO, etc.
12

Mueen, Zeina. "Application of the One-Step Second-Derivative Method for Solving the Transient Distribution in Markov Chain." Engineering, Technology & Applied Science Research 15, no. 2 (2025): 20595–600. https://doi.org/10.48084/etasr.9485.

Texto completo
Resumen
Markov chains are an application of stochastic models in operation research, helping the analysis and optimization of processes with random events and transitions. The method that will be deployed to obtain the transient solution to a Markov chain problem is an important part of this process. The present paper introduces a novel Ordinary Differential Equation (ODE) approach to solve the Markov chain problem. The probability distribution of a continuous-time Markov chain with an infinitesimal generator at a given time is considered, which is a resulting solution of the Chapman-Kolmogorov differ
Los estilos APA, Harvard, Vancouver, ISO, etc.
13

Takemura, Akimichi, and Hisayuki Hara. "Markov chain Monte Carlo test of toric homogeneous Markov chains." Statistical Methodology 9, no. 3 (2012): 392–406. http://dx.doi.org/10.1016/j.stamet.2011.10.004.

Texto completo
Los estilos APA, Harvard, Vancouver, ISO, etc.
14

Zhou, Xinyi. "Applications of Markov Chain in the Field of Computer Science." Theoretical and Natural Science 100, no. 1 (2025): 150–57. https://doi.org/10.54254/2753-8818/2025.21986.

Texto completo
Resumen
In the era of digital transformation, dealing with big data that alters over time is necessary. Markov chain is a fundamental concept in the field of stochastic processes for modeling systems that evolve probabilistically over time, and especially it can be used in computer science for data analysis. This paper focused on analyzing the applications of Markov chain in predicting cloud service trusted state and network traffic. The main problem addressed is how to integrate Markov chains into the complicated computation systems. By employing discrete Markov processes, hidden Markov chains, and f
Los estilos APA, Harvard, Vancouver, ISO, etc.
15

BOUCHER, THOMAS R., and DAREN B. H. CLINE. "PIGGYBACKING THRESHOLD PROCESSES WITH A FINITE STATE MARKOV CHAIN." Stochastics and Dynamics 09, no. 02 (2009): 187–204. http://dx.doi.org/10.1142/s0219493709002622.

Texto completo
Resumen
The state-space representations of certain nonlinear autoregressive time series are general state Markov chains. The transitions of a general state Markov chain among regions in its state-space can be modeled with the transitions among states of a finite state Markov chain. Stability of the time series is then informed by the stationary distributions of the finite state Markov chain. This approach generalizes some previous results.
Los estilos APA, Harvard, Vancouver, ISO, etc.
16

Xia, Sitian. "Research on the Efficacy of Markov Chains in Analyzing and Predicting Chemical Reaction Rates." Theoretical and Natural Science 55, no. 1 (2024): 85–89. http://dx.doi.org/10.54254/2753-8818/55/20240208.

Texto completo
Resumen
Abstract. The Markov chain is a powerful mathematical model used to describe stochastic processes in which the probability of each event depends only on the state attained in the previous event. The chemical reaction rate is a measure of how quickly or slowly a chemical reaction proceeds in multiple stages. This research paper uses the case study of enzymes to investigate the efficacy of the Markov chain in analyzing and predicting chemical reaction rates. By modeling chemical reaction pathways using Markov chains, this study tries to explore how Markov chains help capture and predict a chemic
Los estilos APA, Harvard, Vancouver, ISO, etc.
17

Rydén, Tobias. "On identifiability and order of continuous-time aggregated Markov chains, Markov-modulated Poisson processes, and phase-type distributions." Journal of Applied Probability 33, no. 3 (1996): 640–53. http://dx.doi.org/10.2307/3215346.

Texto completo
Resumen
An aggregated Markov chain is a Markov chain for which some states cannot be distinguished from each other by the observer. In this paper we consider the identifiability problem for such processes in continuous time, i.e. the problem of determining whether two parameters induce identical laws for the observable process or not. We also study the order of a continuous-time aggregated Markov chain, which is the minimum number of states needed to represent it. In particular, we give a lower bound on the order. As a by-product, we obtain results of this kind also for Markov-modulated Poisson proces
Los estilos APA, Harvard, Vancouver, ISO, etc.
18

Rydén, Tobias. "On identifiability and order of continuous-time aggregated Markov chains, Markov-modulated Poisson processes, and phase-type distributions." Journal of Applied Probability 33, no. 03 (1996): 640–53. http://dx.doi.org/10.1017/s0021900200100087.

Texto completo
Resumen
An aggregated Markov chain is a Markov chain for which some states cannot be distinguished from each other by the observer. In this paper we consider the identifiability problem for such processes in continuous time, i.e. the problem of determining whether two parameters induce identical laws for the observable process or not. We also study the order of a continuous-time aggregated Markov chain, which is the minimum number of states needed to represent it. In particular, we give a lower bound on the order. As a by-product, we obtain results of this kind also for Markov-modulated Poisson proces
Los estilos APA, Harvard, Vancouver, ISO, etc.
19

Dixit, Purushottam D. "Introducing User-Prescribed Constraints in Markov Chains for Nonlinear Dimensionality Reduction." Neural Computation 31, no. 5 (2019): 980–97. http://dx.doi.org/10.1162/neco_a_01184.

Texto completo
Resumen
Stochastic kernel-based dimensionality-reduction approaches have become popular in the past decade. The central component of many of these methods is a symmetric kernel that quantifies the vicinity between pairs of data points and a kernel-induced Markov chain on the data. Typically, the Markov chain is fully specified by the kernel through row normalization. However, in many cases, it is desirable to impose user-specified stationary-state and dynamical constraints on the Markov chain. Unfortunately, no systematic framework exists to impose such user-defined constraints. Here, based on our pre
Los estilos APA, Harvard, Vancouver, ISO, etc.
20

Zhong, Pingping, Weiguo Yang, and Peipei Liang. "THE ASYMPTOTIC EQUIPARTITION PROPERTY FOR ASYMPTOTIC CIRCULAR MARKOV CHAINS." Probability in the Engineering and Informational Sciences 24, no. 2 (2010): 279–88. http://dx.doi.org/10.1017/s0269964809990271.

Texto completo
Resumen
In this article, we study the asymptotic equipartition property (AEP) for asymptotic circular Markov chains. First, the definition of an asymptotic circular Markov chain is introduced. Then by applying the limit property for the bivariate functions of nonhomogeneous Markov chains, the strong limit theorem on the frequencies of occurrence of states for asymptotic circular Markov chains is established. Next, the strong law of large numbers on the frequencies of occurrence of states for asymptotic circular Markov chains is obtained. Finally, we prove the AEP for asymptotic circular Markov chains.
Los estilos APA, Harvard, Vancouver, ISO, etc.
21

Qi-feng, Yao, Dong Yun, and Wang Zhong-Zhi. "An Entropy Rate Theorem for a Hidden Inhomogeneous Markov Chain." Open Statistics & Probability Journal 8, no. 1 (2017): 19–26. http://dx.doi.org/10.2174/1876527001708010019.

Texto completo
Resumen
Objective: The main object of our study is to extend some entropy rate theorems to a Hidden Inhomogeneous Markov Chain (HIMC) and establish an entropy rate theorem under some mild conditions. Introduction: A hidden inhomogeneous Markov chain contains two different stochastic processes; one is an inhomogeneous Markov chain whose states are hidden and the other is a stochastic process whose states are observable. Materials and Methods: The proof of theorem requires some ergodic properties of an inhomogeneous Markov chain, and the flexible application of the properties of norm and the bounded con
Los estilos APA, Harvard, Vancouver, ISO, etc.
22

Glynn, Peter W., and Chang-Han Rhee. "Exact estimation for Markov chain equilibrium expectations." Journal of Applied Probability 51, A (2014): 377–89. http://dx.doi.org/10.1239/jap/1417528487.

Texto completo
Resumen
We introduce a new class of Monte Carlo methods, which we call exact estimation algorithms. Such algorithms provide unbiased estimators for equilibrium expectations associated with real-valued functionals defined on a Markov chain. We provide easily implemented algorithms for the class of positive Harris recurrent Markov chains, and for chains that are contracting on average. We further argue that exact estimation in the Markov chain setting provides a significant theoretical relaxation relative to exact simulation methods.
Los estilos APA, Harvard, Vancouver, ISO, etc.
23

Glynn, Peter W., and Chang-Han Rhee. "Exact estimation for Markov chain equilibrium expectations." Journal of Applied Probability 51, A (2014): 377–89. http://dx.doi.org/10.1017/s0021900200021392.

Texto completo
Resumen
We introduce a new class of Monte Carlo methods, which we call exact estimation algorithms. Such algorithms provide unbiased estimators for equilibrium expectations associated with real-valued functionals defined on a Markov chain. We provide easily implemented algorithms for the class of positive Harris recurrent Markov chains, and for chains that are contracting on average. We further argue that exact estimation in the Markov chain setting provides a significant theoretical relaxation relative to exact simulation methods.
Los estilos APA, Harvard, Vancouver, ISO, etc.
24

Lekgari, Mokaedi V. "Maximal Coupling Procedure and Stability of Continuous-Time Markov Chains." Bulletin of Mathematical Sciences and Applications 10 (November 2014): 30–37. http://dx.doi.org/10.18052/www.scipress.com/bmsa.10.30.

Texto completo
Resumen
In this study we first investigate the stability of subsampled discrete Markov chains through the use of the maximal coupling procedure. This is an extension of the available results on Markov chains and is realized through the analysis of the subsampled chain ΦΤn, where {Τn, nєZ+}is an increasing sequence of random stopping times. Then the similar results are realized for the stability of countable-state Continuous-time Markov processes by employing the skeleton-chain method.
Los estilos APA, Harvard, Vancouver, ISO, etc.
25

Nuel, Grégory. "Pattern Markov Chains: Optimal Markov Chain Embedding Through Deterministic Finite Automata." Journal of Applied Probability 45, no. 1 (2008): 226–43. http://dx.doi.org/10.1239/jap/1208358964.

Texto completo
Resumen
In the framework of patterns in random texts, the Markov chain embedding techniques consist of turning the occurrences of a pattern over an order-m Markov sequence into those of a subset of states into an order-1 Markov chain. In this paper we use the theory of language and automata to provide space-optimal Markov chain embedding using the new notion of pattern Markov chains (PMCs), and we give explicit constructive algorithms to build the PMC associated to any given pattern problem. The interest of PMCs is then illustrated through the exact computation of P-values whose complexity is discusse
Los estilos APA, Harvard, Vancouver, ISO, etc.
26

Nuel, Grégory. "Pattern Markov Chains: Optimal Markov Chain Embedding Through Deterministic Finite Automata." Journal of Applied Probability 45, no. 01 (2008): 226–43. http://dx.doi.org/10.1017/s0021900200004083.

Texto completo
Resumen
In the framework of patterns in random texts, the Markov chain embedding techniques consist of turning the occurrences of a pattern over an order-m Markov sequence into those of a subset of states into an order-1 Markov chain. In this paper we use the theory of language and automata to provide space-optimal Markov chain embedding using the new notion of pattern Markov chains (PMCs), and we give explicit constructive algorithms to build the PMC associated to any given pattern problem. The interest of PMCs is then illustrated through the exact computation of P-values whose complexity is discusse
Los estilos APA, Harvard, Vancouver, ISO, etc.
27

Mohamed, Mohamed, Mohamed Bisher Zeina, and Yasin Karmouta. "Classification of States for Literal Neutrosophic and Plithogenic Markov Chains." Journal of Neutrosophic and Fuzzy Systems 08, no. 2 (2024): 49–61. http://dx.doi.org/10.54216/jnfs.080206.

Texto completo
Resumen
In this paper we represent many classifications of neutrosophic and plithogenic Markov Chains states including absorbent states, inessential and essential states, recurrent states and communicated states. We prove that if a state (i) according to a neutrosophic Markov Chain with neutrosophic transition matrix is classified as any of the previous classifications then it is also classified as the same classification in classical scene to two Markov Chains defined with transition matrices respectively. Also, we prove that if a state (i) according to a plithogenic Markov Chain with plithogenic tra
Los estilos APA, Harvard, Vancouver, ISO, etc.
28

Choi, Michael C. H., and Pierre Patie. "Analysis of non-reversible Markov chains via similarity orbits." Combinatorics, Probability and Computing 29, no. 4 (2020): 508–36. http://dx.doi.org/10.1017/s0963548320000024.

Texto completo
Resumen
AbstractIn this paper we develop an in-depth analysis of non-reversible Markov chains on denumerable state space from a similarity orbit perspective. In particular, we study the class of Markov chains whose transition kernel is in the similarity orbit of a normal transition kernel, such as that of birth–death chains or reversible Markov chains. We start by identifying a set of sufficient conditions for a Markov chain to belong to the similarity orbit of a birth–death chain. As by-products, we obtain a spectral representation in terms of non-self-adjoint resolutions of identity in the sense of
Los estilos APA, Harvard, Vancouver, ISO, etc.
29

Papastathopoulos, I., K. Strokorb, J. A. Tawn, and A. Butler. "Extreme events of Markov chains." Advances in Applied Probability 49, no. 1 (2017): 134–61. http://dx.doi.org/10.1017/apr.2016.82.

Texto completo
Resumen
Abstract The extremal behaviour of a Markov chain is typically characterised by its tail chain. For asymptotically dependent Markov chains, existing formulations fail to capture the full evolution of the extreme event when the chain moves out of the extreme tail region, and, for asymptotically independent chains, recent results fail to cover well-known asymptotically independent processes, such as Markov processes with a Gaussian copula between consecutive values. We use more sophisticated limiting mechanisms that cover a broader class of asymptotically independent processes than current metho
Los estilos APA, Harvard, Vancouver, ISO, etc.
30

Li, Wenxi, and Zhongzhi Wang. "A NOTE ON RÉNYI'S ENTROPY RATE FOR TIME-INHOMOGENEOUS MARKOV CHAINS." Probability in the Engineering and Informational Sciences 33, no. 4 (2018): 579–90. http://dx.doi.org/10.1017/s026996481800044x.

Texto completo
Resumen
AbstractIn this note, we use the Perron–Frobenius theorem to obtain the Rényi's entropy rate for a time-inhomogeneous Markov chain whose transition matrices converge to a primitive matrix. As direct corollaries, we also obtain the Rényi's entropy rate for asymptotic circular Markov chain and the Rényi's divergence rate between two time-inhomogeneous Markov chains.
Los estilos APA, Harvard, Vancouver, ISO, etc.
31

Politis, Dimitris N. "Markov Chains in Many Dimensions." Advances in Applied Probability 26, no. 3 (1994): 756–74. http://dx.doi.org/10.2307/1427819.

Texto completo
Resumen
A generalization of the notion of a stationary Markov chain in more than one dimension is proposed, and is found to be a special class of homogeneous Markov random fields. Stationary Markov chains in many dimensions are shown to possess a maximum entropy property, analogous to the corresponding property for Markov chains in one dimension. In addition, a representation of Markov chains in many dimensions is provided, together with a method for their generation that converges to their stationary distribution.
Los estilos APA, Harvard, Vancouver, ISO, etc.
32

Politis, Dimitris N. "Markov Chains in Many Dimensions." Advances in Applied Probability 26, no. 03 (1994): 756–74. http://dx.doi.org/10.1017/s0001867800026537.

Texto completo
Resumen
A generalization of the notion of a stationary Markov chain in more than one dimension is proposed, and is found to be a special class of homogeneous Markov random fields. Stationary Markov chains in many dimensions are shown to possess a maximum entropy property, analogous to the corresponding property for Markov chains in one dimension. In addition, a representation of Markov chains in many dimensions is provided, together with a method for their generation that converges to their stationary distribution.
Los estilos APA, Harvard, Vancouver, ISO, etc.
33

Zhou, Hua, and Kenneth Lange. "Composition Markov chains of multinomial type." Advances in Applied Probability 41, no. 1 (2009): 270–91. http://dx.doi.org/10.1239/aap/1240319585.

Texto completo
Resumen
Suppose that n identical particles evolve according to the same marginal Markov chain. In this setting we study chains such as the Ehrenfest chain that move a prescribed number of randomly chosen particles at each epoch. The product chain constructed by this device inherits its eigenstructure from the marginal chain. There is a further chain derived from the product chain called the composition chain that ignores particle labels and tracks the numbers of particles in the various states. The composition chain in turn inherits its eigenstructure and various properties such as reversibility from
Los estilos APA, Harvard, Vancouver, ISO, etc.
34

Zhou, Hua, and Kenneth Lange. "Composition Markov chains of multinomial type." Advances in Applied Probability 41, no. 01 (2009): 270–91. http://dx.doi.org/10.1017/s0001867800003220.

Texto completo
Resumen
Suppose that n identical particles evolve according to the same marginal Markov chain. In this setting we study chains such as the Ehrenfest chain that move a prescribed number of randomly chosen particles at each epoch. The product chain constructed by this device inherits its eigenstructure from the marginal chain. There is a further chain derived from the product chain called the composition chain that ignores particle labels and tracks the numbers of particles in the various states. The composition chain in turn inherits its eigenstructure and various properties such as reversibility from
Los estilos APA, Harvard, Vancouver, ISO, etc.
35

Marpaung, Jonathan Liviera, Friska Maharani Simanjuntak, and Raissa Maulida. "The Markov Chain Method in Predicting Market Share for Smartphone Brand Usage Among University Students." Journal of Mathematics Technology and Education 3, no. 1 (2024): 1–8. https://doi.org/10.32734/jomte.v3i1.16856.

Texto completo
Resumen
As technology develops, smartphones have become one of the technological devices that influence modern life. Markov chains are a method used to model systems that experience transitions from one state to another with certain probabilities. In this research, the Markov chain aims to predict the market share of each smartphone brand using the Markov Chain Analysis method. The results of Markov chain analysis show that the iPhone is predicted to become a brand that is in great demand by users, especially the younger generation, until 2031.
Los estilos APA, Harvard, Vancouver, ISO, etc.
36

Lorek, Paweł. "Antiduality and Möbius monotonicity: generalized coupon collector problem." ESAIM: Probability and Statistics 23 (2019): 739–69. http://dx.doi.org/10.1051/ps/2019004.

Texto completo
Resumen
For a given absorbing Markov chain X* on a finite state space, a chain X is a sharp antidual of X* if the fastest strong stationary time (FSST) of X is equal, in distribution, to the absorption time of X*. In this paper, we show a systematic way of finding such an antidual based on some partial ordering of the state space. We use a theory of strong stationary duality developed recently for Möbius monotone Markov chains. We give several sharp antidual chains for Markov chain corresponding to a generalized coupon collector problem. As a consequence – utilizing known results on the limiting distr
Los estilos APA, Harvard, Vancouver, ISO, etc.
37

Hsiau, Shoou-Ren. "Compound Poisson limit theorems for Markov chains." Journal of Applied Probability 34, no. 1 (1997): 24–34. http://dx.doi.org/10.2307/3215171.

Texto completo
Resumen
This paper establishes a compound Poisson limit theorem for the sum of a sequence of multi-state Markov chains. Our theorem generalizes an earlier one by Koopman for the two-state Markov chain. Moreover, a similar approach is used to derive a limit theorem for the sum of the k th-order two-state Markov chain.
Los estilos APA, Harvard, Vancouver, ISO, etc.
38

Hsiau, Shoou-Ren. "Compound Poisson limit theorems for Markov chains." Journal of Applied Probability 34, no. 01 (1997): 24–34. http://dx.doi.org/10.1017/s002190020010066x.

Texto completo
Resumen
This paper establishes a compound Poisson limit theorem for the sum of a sequence of multi-state Markov chains. Our theorem generalizes an earlier one by Koopman for the two-state Markov chain. Moreover, a similar approach is used to derive a limit theorem for the sum of the k th-order two-state Markov chain.
Los estilos APA, Harvard, Vancouver, ISO, etc.
39

Boys, R. J., and D. A. Henderson. "On Determining the Order of Markov Dependence of an Observed Process Governed by a Hidden Markov Model." Scientific Programming 10, no. 3 (2002): 241–51. http://dx.doi.org/10.1155/2002/683164.

Texto completo
Resumen
This paper describes a Bayesian approach to determining the order of a finite state Markov chain whose transition probabilities are themselves governed by a homogeneous finite state Markov chain. It extends previous work on homogeneous Markov chains to more general and applicable hidden Markov models. The method we describe uses a Markov chain Monte Carlo algorithm to obtain samples from the (posterior) distribution for both the order of Markov dependence in the observed sequence and the other governing model parameters. These samples allow coherent inferences to be made straightforwardly in c
Los estilos APA, Harvard, Vancouver, ISO, etc.
40

Altman, Eitan, Konstantin E. Avrachenkov, and Rudesindo Núñez-Queija. "Perturbation analysis for denumerable Markov chains with application to queueing models." Advances in Applied Probability 36, no. 3 (2004): 839–53. http://dx.doi.org/10.1239/aap/1093962237.

Texto completo
Resumen
We study the parametric perturbation of Markov chains with denumerable state spaces. We consider both regular and singular perturbations. By the latter we mean that transition probabilities of a Markov chain, with several ergodic classes, are perturbed such that (rare) transitions among the different ergodic classes of the unperturbed chain are allowed. Singularly perturbed Markov chains have been studied in the literature under more restrictive assumptions such as strong recurrence ergodicity or Doeblin conditions. We relax these conditions so that our results can be applied to queueing model
Los estilos APA, Harvard, Vancouver, ISO, etc.
41

Altman, Eitan, Konstantin E. Avrachenkov, and Rudesindo Núñez-Queija. "Perturbation analysis for denumerable Markov chains with application to queueing models." Advances in Applied Probability 36, no. 03 (2004): 839–53. http://dx.doi.org/10.1017/s0001867800013148.

Texto completo
Resumen
We study the parametric perturbation of Markov chains with denumerable state spaces. We consider both regular and singular perturbations. By the latter we mean that transition probabilities of a Markov chain, with several ergodic classes, are perturbed such that (rare) transitions among the different ergodic classes of the unperturbed chain are allowed. Singularly perturbed Markov chains have been studied in the literature under more restrictive assumptions such as strong recurrence ergodicity or Doeblin conditions. We relax these conditions so that our results can be applied to queueing model
Los estilos APA, Harvard, Vancouver, ISO, etc.
42

Zhao, Yiqiang Q., Wei Li, and W. John Braun. "Infinite block-structured transition matrices and their properties." Advances in Applied Probability 30, no. 2 (1998): 365–84. http://dx.doi.org/10.1239/aap/1035228074.

Texto completo
Resumen
In this paper, we study Markov chains with infinite state block-structured transition matrices, whose states are partitioned into levels according to the block structure, and various associated measures. Roughly speaking, these measures involve first passage times or expected numbers of visits to certain levels without hitting other levels. They are very important and often play a key role in the study of a Markov chain. Necessary and/or sufficient conditions are obtained for a Markov chain to be positive recurrent, recurrent, or transient in terms of these measures. Results are obtained for g
Los estilos APA, Harvard, Vancouver, ISO, etc.
43

Zhao, Yiqiang Q., Wei Li, and W. John Braun. "Infinite block-structured transition matrices and their properties." Advances in Applied Probability 30, no. 02 (1998): 365–84. http://dx.doi.org/10.1017/s0001867800047339.

Texto completo
Resumen
In this paper, we study Markov chains with infinite state block-structured transition matrices, whose states are partitioned into levels according to the block structure, and various associated measures. Roughly speaking, these measures involve first passage times or expected numbers of visits to certain levels without hitting other levels. They are very important and often play a key role in the study of a Markov chain. Necessary and/or sufficient conditions are obtained for a Markov chain to be positive recurrent, recurrent, or transient in terms of these measures. Results are obtained for g
Los estilos APA, Harvard, Vancouver, ISO, etc.
44

Janssen, A., and J. Segers. "Markov Tail Chains." Journal of Applied Probability 51, no. 4 (2014): 1133–53. http://dx.doi.org/10.1239/jap/1421763332.

Texto completo
Resumen
The extremes of a univariate Markov chain with regularly varying stationary marginal distribution and asymptotically linear behavior are known to exhibit a multiplicative random walk structure called the tail chain. In this paper we extend this fact to Markov chains with multivariate regularly varying marginal distributions inRd. We analyze both the forward and the backward tail process and show that they mutually determine each other through a kind of adjoint relation. In a broader setting, we will show that even for non-Markovian underlying processes a Markovian forward tail chain always imp
Los estilos APA, Harvard, Vancouver, ISO, etc.
45

Janssen, A., and J. Segers. "Markov Tail Chains." Journal of Applied Probability 51, no. 04 (2014): 1133–53. http://dx.doi.org/10.1017/s0001867800012027.

Texto completo
Resumen
The extremes of a univariate Markov chain with regularly varying stationary marginal distribution and asymptotically linear behavior are known to exhibit a multiplicative random walk structure called the tail chain. In this paper we extend this fact to Markov chains with multivariate regularly varying marginal distributions in R d . We analyze both the forward and the backward tail process and show that they mutually determine each other through a kind of adjoint relation. In a broader setting, we will show that even for non-Markovian underlying processes a Markovian forward tail chain always
Los estilos APA, Harvard, Vancouver, ISO, etc.
46

Janssen, A., and J. Segers. "Markov Tail Chains." Journal of Applied Probability 51, no. 04 (2014): 1133–53. http://dx.doi.org/10.1017/s002190020001202x.

Texto completo
Resumen
The extremes of a univariate Markov chain with regularly varying stationary marginal distribution and asymptotically linear behavior are known to exhibit a multiplicative random walk structure called the tail chain. In this paper we extend this fact to Markov chains with multivariate regularly varying marginal distributions in R d . We analyze both the forward and the backward tail process and show that they mutually determine each other through a kind of adjoint relation. In a broader setting, we will show that even for non-Markovian underlying processes a Markovian forward tail chain always
Los estilos APA, Harvard, Vancouver, ISO, etc.
47

BOWEN, LEWIS. "Non-abelian free group actions: Markov processes, the Abramov–Rohlin formula and Yuzvinskii’s formula." Ergodic Theory and Dynamical Systems 30, no. 6 (2009): 1629–63. http://dx.doi.org/10.1017/s0143385709000844.

Texto completo
Resumen
AbstractThis paper introduces Markov chains and processes over non-abelian free groups and semigroups. We prove a formula for the f-invariant of a Markov chain over a free group in terms of transition matrices that parallels the classical formula for the entropy a Markov chain. Applications include free group analogues of the Abramov–Rohlin formula for skew-product actions and Yuzvinskii’s addition formula for algebraic actions.
Los estilos APA, Harvard, Vancouver, ISO, etc.
48

Golomoziy, Vitaliy, and Olha Moskanova. "Polynomial Recurrence of Time-inhomogeneous Markov Chains." Austrian Journal of Statistics 52, SI (2023): 40–53. http://dx.doi.org/10.17713/ajs.v52isi.1752.

Texto completo
Resumen
This paper is devoted to establishing conditions that guarantee the existence of a p-th moment of the time it takes for a timeinhomogeneous Markov chain to hit some set C. We modified the well-known Drift Condition from the theory of homogeneous Markov chains. We demonstrated that the inhomogeneous Markov chain may be polynomially recurrent while exhibiting different dynamics from its homogeneous counterpart.
Los estilos APA, Harvard, Vancouver, ISO, etc.
49

Huang, Huilin, and Weiguo Yang. "Limit theorems for asymptotic circular mth-order Markov chains indexed by an m-rooted homogeneous tree." Filomat 33, no. 6 (2019): 1817–32. http://dx.doi.org/10.2298/fil1906817h.

Texto completo
Resumen
In this paper, we give the definition of an asymptotic circularmth-order Markov chain indexed by an m rooted homogeneous tree. By applying the limit property for a sequence of multi-variables functions of a nonhomogeneous Markov chain indexed by such tree, we estabish the strong law of large numbers and the asymptotic equipartition property (AEP) for asymptotic circular mth-order finite Markov chains indexed by this homogeneous tree. As a corollary, we can obtain the strong law of large numbers and AEP about the mth-order finite nonhomogeneous Markov chain indexed by the m rooted homogeneous t
Los estilos APA, Harvard, Vancouver, ISO, etc.
50

Agboola, Sunday Olanrewaju. "The Performance Measure Analysis on the Irreducibility in Markov Chain States Classification." Journal of Scientific and Engineering Research 8, no. 12 (2021): 109–17. https://doi.org/10.5281/zenodo.10620054.

Texto completo
Resumen
<strong>Abstract</strong> The transitions in Markov chain are assumed to occur instantaneously and the future evolution of the system depends only on its current state and not on its past history, then the system may be represented by a Markov process. Even when the system does not possess this Markov property explicitly, it is often possible to construct a corresponding implicit representation. In this study, the irreducible Markov chain where all states are positive recurrent, null recurrent and transient are investigated, in order to provide an insight into the performance measures in irred
Los estilos APA, Harvard, Vancouver, ISO, etc.
Ofrecemos descuentos en todos los planes premium para autores cuyas obras están incluidas en selecciones literarias temáticas. ¡Contáctenos para obtener un código promocional único!