Optimal decision procedures for finite Markov chains. Part III: General convex systems

1973 ◽  
Vol 5 (3) ◽  
pp. 541-553 ◽  
Author(s):  
John Bather

This paper is concerned with the general problem of finding an optimal transition matrix for a finite Markov chain, where the probabilities for each transition must be chosen from a given convex family of distributions. The immediate cost is determined by this choice, but it is required to minimise the average expected cost in the long run. The problem is investigated by classifying the states according to the accessibility relations between them. If an optimal policy exists, it can be found by considering the convex subsystems associated with the states at different levels in the classification scheme.

1973 ◽  
Vol 5 (03) ◽  
pp. 541-553 ◽  
Author(s):  
John Bather

This paper is concerned with the general problem of finding an optimal transition matrix for a finite Markov chain, where the probabilities for each transition must be chosen from a given convex family of distributions. The immediate cost is determined by this choice, but it is required to minimise the average expected cost in the long run. The problem is investigated by classifying the states according to the accessibility relations between them. If an optimal policy exists, it can be found by considering the convex subsystems associated with the states at different levels in the classification scheme.


Author(s):  
Julian Keilson ◽  
David M. G. Wishart

Introduction. Two previous papers (1,2) have dealt with additive processes defined on finite Markov chains. Such a process in discrete time may be treated as a bivariate Markov process {R(k), X(k)}. The process R(k) is an irreducible Markov chain on states r = 1, 2, …, R governed by a stochastic transition matrix B0 with components brs. The marginal process X(k) ‘defined’ on the chain R(k) is a sum of random increments ξ(i) dependent on the chain, i.e. if the ith transition takes the chain from state r to state s, ξ,(i) is chosen from a distribution function Drs(x) indexed by r and s. The distribution of the bivariate process may be represented by a vector F(x, k) with componentsThese are generated recursively by the relationwhere the increment matrix distribution B(x) has components brsDrs(x). We denote the nth moment of B(x) by Bn = ∫xndB(x), so that B0 = B(∞).


2019 ◽  
Vol 44 (3) ◽  
pp. 282-308 ◽  
Author(s):  
Brian G. Vegetabile ◽  
Stephanie A. Stout-Oswald ◽  
Elysia Poggi Davis ◽  
Tallie Z. Baram ◽  
Hal S. Stern

Predictability of behavior is an important characteristic in many fields including biology, medicine, marketing, and education. When a sequence of actions performed by an individual can be modeled as a stationary time-homogeneous Markov chain the predictability of the individual’s behavior can be quantified by the entropy rate of the process. This article compares three estimators of the entropy rate of finite Markov processes. The first two methods directly estimate the entropy rate through estimates of the transition matrix and stationary distribution of the process. The third method is related to the sliding-window Lempel–Ziv compression algorithm. The methods are compared via a simulation study and in the context of a study of interactions between mothers and their children.


2019 ◽  
Vol 29 (08) ◽  
pp. 1431-1449
Author(s):  
John Rhodes ◽  
Anne Schilling

We show that the stationary distribution of a finite Markov chain can be expressed as the sum of certain normal distributions. These normal distributions are associated to planar graphs consisting of a straight line with attached loops. The loops touch only at one vertex either of the straight line or of another attached loop. Our analysis is based on our previous work, which derives the stationary distribution of a finite Markov chain using semaphore codes on the Karnofsky–Rhodes and McCammond expansion of the right Cayley graph of the finite semigroup underlying the Markov chain.


1973 ◽  
Vol 5 (3) ◽  
pp. 521-540 ◽  
Author(s):  
John Bather

A Markov process in discrete time with a finite state space is controlled by choosing the transition probabilities from a given convex family of distributions depending on the present state. The immediate cost is prescribed for each choice and it is required to minimise the average expected cost over an infinite future. The paper considers a special case of this general problem and provides the foundation for a general solution. The main result is that an optimal policy exists if each state of the system can be reached with positive probability from any other state by choosing a suitable policy.


2012 ◽  
Vol 263-266 ◽  
pp. 2339-2343
Author(s):  
Ying Ming Jin

This paper analyzes the convergence deviation of surrogate assisted (1+1)EA. A model of surrogate assisted (1+1)EA can be built by the finite markov chain, then we got the transition matrix of this algorithm. The deviation of surrogate model can be expressed by the perturbation of transition matrix. So we can estimate the convergence deviation with the method of matrix perturbation analysis. Analyzing of the convergence changes brought by surrogate model’s deviations can help us to have a better select of the surrogate model.


Author(s):  
J. L. Mott

SynopsisIn this note we study the asymptotic behaviour of a product of matrices where Pj is a matrix of transition probabilities in a non-homogeneous finite Markov chain. We give conditions that (i) the rows of P(n) tend to identity and that (ii) P(n) tends to a limit matrix with identical rows.


Author(s):  
H. D. Miller

SummaryLet {kr} (r = 0, 1, 2, …; 1 ≤ kr ≤ h) be a positively regular, finite Markov chain with transition matrix P = (pjk). For each possible transition j → k let gjk(x)(− ∞ ≤ x ≤ ∞) be a given distribution function. The sequence of random variables {ξr} is defined where ξr has the distribution gjk(x) if the rth transition takes the chain from state j to state k. It is supposed that each distribution gjk(x) admits a two-sided Laplace-Stieltjes transform in a real t-interval surrounding t = 0. Let P(t) denote the matrix {Pjkmjk(t)}. It is shown, using probability arguments, that I − sP(t) admits a Wiener-Hopf type of factorization in two ways for suitable values of s where the plus-factors are non-singular, bounded and have regular elements in a right half of the complex t-plane and the minus-factors have similar properties in an overlapping left half-plane (Theorem 1).


1967 ◽  
Vol 4 (03) ◽  
pp. 496-507 ◽  
Author(s):  
J. N. Darroch ◽  
K. W. Morris

Let T denote a subset of the possible transitions between the states of a finite Markov chain and let Yk denote the time of the kth occurrence of a T-transition. Formulae are derived for the generating functions of Yk , of Yj + k — Yj and of Yj + k — Yj in the limit as j → ∞, for both discrete-time and continuoustime chains. Several particular cases are briefly discussed.


1982 ◽  
Vol 19 (03) ◽  
pp. 685-691 ◽  
Author(s):  
Atef M. Abdel-moneim ◽  
Frederick W. Leysieffer

Criteria are given to determine whether a given finite Markov chain can be lumped weakly with respect to a given partition of its state space. These conditions are given in terms of solution classes of systems of linear equations associated with the transition probability matrix of the Markov chain and the given partition.


Sign in / Sign up

Export Citation Format

Share Document