A simple proof for optimality of (s, S) policies in infinite-horizon inventory systems

1991 ◽  
Vol 28 (4) ◽  
pp. 802-810 ◽  
Author(s):  
Yu-Sheng Zheng

The classical proofs for the existence of a stationary (s, S) inventory policy that minimizes the total discounted or average cost over an infinite horizon are lengthy because they depend heavily on the optimality results for corresponding finite-horizon models. This note presents a simpler alternative. Since optimal stationary (s, S) policies are relatively simple to characterize, it is easy to construct a solution to the optimality equation which is satisfied by an (s, S) policy or an equivalent variant thereof. For the discounted model, the proof characterizes an (s, S) policy that is optimal for all initial inventory positions. This policy can be generated by a simple existing algorithm. For the average-cost model, the optimality proof is completed with some additional arguments, which are simple but novel, to overcome the normal difficulties encountered in models with unbounded one-step expected costs.

1991 ◽  
Vol 28 (04) ◽  
pp. 802-810 ◽  
Author(s):  
Yu-Sheng Zheng

The classical proofs for the existence of a stationary (s, S) inventory policy that minimizes the total discounted or average cost over an infinite horizon are lengthy because they depend heavily on the optimality results for corresponding finite-horizon models. This note presents a simpler alternative. Since optimal stationary (s, S) policies are relatively simple to characterize, it is easy to construct a solution to the optimality equation which is satisfied by an (s, S) policy or an equivalent variant thereof. For the discounted model, the proof characterizes an (s, S) policy that is optimal for all initial inventory positions. This policy can be generated by a simple existing algorithm. For the average-cost model, the optimality proof is completed with some additional arguments, which are simple but novel, to overcome the normal difficulties encountered in models with unbounded one-step expected costs.


2011 ◽  
Vol 2011 ◽  
pp. 1-11
Author(s):  
Epaminondas G. Kyriakidis

We introduce a Markov decision process in continuous time for the optimal control of a simple symmetrical immigration-emigration process by the introduction of total catastrophes. It is proved that a particular control-limit policy is average cost optimal within the class of all stationary policies by verifying that the relative values of this policy are the solution of the corresponding optimality equation.


1988 ◽  
Vol 2 (3) ◽  
pp. 321-328 ◽  
Author(s):  
Laurence A. Baxter ◽  
Eui Yong Lee

The state of a system is modelled by Brownian motion with negative drift and an absorbing barrier at the origin. A repairman arrives according to a Poisson process of rate λ. If the state of the system at arrival of the repairman does not exceed a certain threshold, he/she increases it by a random amount, otherwise no action is taken. Costs are assigned to each visit of the repairman, to each repair, and to the system being in state 0. It is shown that there exists a unique arrival rate λ which minimizes the average cost per unit time over an infinite horizon.


1980 ◽  
Vol 17 (04) ◽  
pp. 996-1003
Author(s):  
D. R. Robinson

It is known that when costs are unbounded satisfaction of the appropriate dynamic programming ‘optimality' equation by a policy is not sufficient to guarantee its average optimality. A ‘lowest-order potential' condition is introduced which, along with the dynamic programming equation, is sufficient to establish the optimality of the policy. Also, it is shown that under fairly general conditions, if the lowest-order potential condition is not satisfied there exists a non-memoryless policy with smaller average cost than the policy satisfying the dynamic programming equation.


Sign in / Sign up

Export Citation Format

Share Document