Cost Optimality

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 8358 Experts worldwide ranked by ideXlab platform

Oscar Vegaamaya - One of the best experts on this subject based on the ideXlab platform.

Anna Jaśkiewicz - One of the best experts on this subject based on the ideXlab platform.

E. Fernandez-gaucherand - One of the best experts on this subject based on the ideXlab platform.

  • Markov decision processes with risk-sensitive criteria: dynamic programming operators and discounted stochastic games
    Proceedings of the 40th IEEE Conference on Decision and Control (Cat. No.01CH37228), 2001
    Co-Authors: Rolando Cavazos-cadena, E. Fernandez-gaucherand
    Abstract:

    We study discrete-time Markov decision processes with denumerable state space and bounded Costs per stage. It is assumed that the decision maker exhibits a constant sensitivity to risk, and that the performance of a control policy is measured by a (long-run) risk-sensitive average Cost criterion. Besides standard continuity-compactness conditions, the basic structural constraint on the decision model is that the transition law satisfies a simultaneous Doeblin condition. Within this framework, the main objective is to study the existence of bounded solutions to the risk-sensitive average Cost Optimality equation. Our main result guarantees a bounded solution to the Optimality equation only if the risk sensitivity coefficient /spl lambda/ is small enough and, via a detailed example, it can be shown that such a conclusion cannot be extended to arbitrary values of /spl lambda/. Our results are in opposition to previous claims in the literature, but agree with recent results obtained via a direct probabilistic analysis. A key analysis tool developed in the paper is the definition of an appropriate operator with contractive properties, analogous to the dynamic programming operator in Bellman's equation, and a family of (value) functions with a discounted stochastic games interpretation.

  • Convex stochastic control problems
    [1992] Proceedings of the 31st IEEE Conference on Decision and Control, 1992
    Co-Authors: E. Fernandez-gaucherand, A. Arapostathis, S.i. Marcus
    Abstract:

    The solution of the infinite horizon stochastic control problem under certain criteria, the functional characterization and computation of optimal values and policies, is related to two dynamic programming-like functional equations: the discounted Cost Optimality equation (DCOE) and the average Cost Optimality equation (ACOE). The authors consider what useful properties, shared by large and important problem classes, can be used to show that an ACOE holds, and how these properties can be exploited to aid in the development of tractable algorithmic solutions. They address this issue by concentrating on structured solutions to stochastic control models. By a structured solution is meant a model for which value functions and/or optimal policies have some special dependence on the (initial) state. The focus is on convexity properties of the value function.

Andrzej S Nowak - One of the best experts on this subject based on the ideXlab platform.

Rolando Cavazoscadena - One of the best experts on this subject based on the ideXlab platform.

  • solution to the risk sensitive average Cost Optimality equation in a class of markov decision processes with finite state space
    Mathematical Methods of Operations Research, 2003
    Co-Authors: Rolando Cavazoscadena
    Abstract:

    This work concerns discrete-time Markov decision processes with finite state space and bounded Costs per stage. The decision maker ranks random Costs via the expectation of the utility function associated to a constant risk sensitivity coefficient, and the performance of a control policy is measured by the corresponding (long-run) risk-sensitive average Cost criterion. The main structural restriction on the system is the following communication assumption: For every pair of states x and y, there exists a policy π, possibly depending on x and y, such that when the system evolves under π starting at x, the probability of reaching y is positive. Within this framework, the paper establishes the existence of solutions to the Optimality equation whenever the constant risk sensitivity coefficient does not exceed certain positive value. Copyright Springer-Verlag Berlin Heidelberg 2003

  • controlled markov chains with risk sensitive criteria average Cost Optimality equations and optimal solutions
    Mathematical Methods of Operations Research, 1999
    Co-Authors: Rolando Cavazoscadena, E Fernandezgaucherand
    Abstract:

    We study controlled Markov chains with denumerable state space and bounded Costs per stage. A (long-run) risk-sensitive average Cost criterion, associated to an exponential utility function with a constant risk sensitivity coefficient, is used as a performance measure. The main assumption on the probabilistic structure of the model is that the transition law satisfies a simultaneous Doeblin condition. Working within this framework, the main results obtained can be summarized as follows: If the constant risk-sensitivity coefficient is small enough, then an associated Optimality equation has a bounded solution with a constant value for the optimal risk-sensitive average Cost; in addition, under further standard continuity-compactness assumptions, optimal stationary policies are obtained. However, it is also shown that the above conclusions fail to hold, in general, for large enough values of the risk-sensitivity coefficient. Our results therefore disprove previous claims on this topic. Also of importance is the fact that our developments are very much self-contained and employ only basic probabilistic and analysis principles. Copyright Springer-Verlag Berlin Heidelberg 1999

  • a counterexample on the Optimality equation in markov decision chains with the average Cost criterion
    Systems & Control Letters, 1991
    Co-Authors: Rolando Cavazoscadena
    Abstract:

    Abstract We consider Markov decision processes with denumerable state space and finite control sets; the performance index of a control policy is a long-run expected average Cost criterion and the Cost function is bounded below. For these models, the existence of average optimal stationary policies was recently established in [11] under very general assumptions. Such a result was obtained via an Optimality inequality. Here, we use a simple example to prove that the conditions in [11] do not imply the existence of a solution to the average Cost Optimality equation.