Numerical methods for solving Markov chains are in general inefficient if the state space of the chain is very large (or infinite) and lacking a simple repeating structure. One alternative to solving such chains is to construct models that are simple to analyze and that provide bounds for a reward function of interest. We present a new bounding method for Markov chains inspired by Markov reward theory; our method constructs bounds by redirecting selected sets of transitions, facilitating an intuitive interpretation of the modifications on the original system. We show that our method is compatible with strong aggregation of Markov chains; thus we can obtain bounds for the initial chain by analyzing a much smaller chain. We illustrate our method on a problem of order fill rates for an inventory system of service tools.
|Place of Publication||Eindhoven|
|Number of pages||34|
|Publication status||Published - 2009|