Approximate Solutions To Constrained Risk-Sensitive Markov Decision Processes

This paper considers the problem of finding near-optimal Markovian randomized (MR) policies for finite-state-action, infinite-horizon, constrained risk-sensitive Markov decision processes (CRSMDPs). Constraints are in the form of standard expected discounted cost functions as well as expected risk-s...

Full description

Saved in:
Bibliographic Details
Main Authors M, Uday Kumar, Bhat, Sanjay P, Kavitha, Veeraruna, Hemachandra, Nandyala
Format Journal Article
LanguageEnglish
Published 29.09.2022
Subjects
Online AccessGet full text

Cover

Loading…
Abstract This paper considers the problem of finding near-optimal Markovian randomized (MR) policies for finite-state-action, infinite-horizon, constrained risk-sensitive Markov decision processes (CRSMDPs). Constraints are in the form of standard expected discounted cost functions as well as expected risk-sensitive discounted cost functions over finite and infinite horizons. The main contribution is to show that the problem possesses a solution if it is feasible, and to provide two methods for finding an approximate solution in the form of an ultimately stationary (US) MR policy. The latter is achieved through two approximating finite-horizon CRSMDPs which are constructed from the original CRSMDP by time-truncating the original objective and constraint cost functions, and suitably perturbing the constraint upper bounds. The first approximation gives a US policy which is $\epsilon$-optimal and feasible for the original problem, while the second approximation gives a near-optimal US policy whose violation of the original constraints is bounded above by a specified $\epsilon$. A key step in the proofs is an appropriate choice of a metric that makes the set of infinite-horizon MR policies and the feasible regions of the three CRSMDPs compact, and the objective and constraint functions continuous. A linear-programming-based formulation for solving the approximating finite-horizon CRSMDPs is also given.
AbstractList This paper considers the problem of finding near-optimal Markovian randomized (MR) policies for finite-state-action, infinite-horizon, constrained risk-sensitive Markov decision processes (CRSMDPs). Constraints are in the form of standard expected discounted cost functions as well as expected risk-sensitive discounted cost functions over finite and infinite horizons. The main contribution is to show that the problem possesses a solution if it is feasible, and to provide two methods for finding an approximate solution in the form of an ultimately stationary (US) MR policy. The latter is achieved through two approximating finite-horizon CRSMDPs which are constructed from the original CRSMDP by time-truncating the original objective and constraint cost functions, and suitably perturbing the constraint upper bounds. The first approximation gives a US policy which is $\epsilon$-optimal and feasible for the original problem, while the second approximation gives a near-optimal US policy whose violation of the original constraints is bounded above by a specified $\epsilon$. A key step in the proofs is an appropriate choice of a metric that makes the set of infinite-horizon MR policies and the feasible regions of the three CRSMDPs compact, and the objective and constraint functions continuous. A linear-programming-based formulation for solving the approximating finite-horizon CRSMDPs is also given.
Author M, Uday Kumar
Kavitha, Veeraruna
Bhat, Sanjay P
Hemachandra, Nandyala
Author_xml – sequence: 1
  givenname: Uday Kumar
  surname: M
  fullname: M, Uday Kumar
– sequence: 2
  givenname: Sanjay P
  surname: Bhat
  fullname: Bhat, Sanjay P
– sequence: 3
  givenname: Veeraruna
  surname: Kavitha
  fullname: Kavitha, Veeraruna
– sequence: 4
  givenname: Nandyala
  surname: Hemachandra
  fullname: Hemachandra, Nandyala
BackLink https://doi.org/10.48550/arXiv.2209.14963$$DView paper in arXiv
BookMark eNotj8tOwzAUBb2ABRQ-gBX-gQS_4sTLKrwqtQLR7KMb50ayWuzIDlH5e0JhNZujo5lrcuGDR0LuOMtVVRTsAeLJzbkQzORcGS2vyG49jjGc3CdMSPfh-DW54BNtAq0XThGcx55-uHTI9uiTm9yMdAfxEGb6iNalZU7fY7CYEqYbcjnAMeHtP1ekeX5q6tds-_ayqdfbDHQpM94b1KICpqtCWMWQYyeKAUDowRiteCE7tFWpeCektRJ626sSFQAzw4Barsj93-25px3joh-_29-u9twlfwCAiEty
ContentType Journal Article
Copyright http://arxiv.org/licenses/nonexclusive-distrib/1.0
Copyright_xml – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0
DBID AKZ
GOX
DOI 10.48550/arxiv.2209.14963
DatabaseName arXiv Mathematics
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2209_14963
GroupedDBID AKZ
GOX
ID FETCH-LOGICAL-a673-1d9e628a06852c40e1eb25faa26f9964153bec8741b23cc3adcd47e4aa09ffe63
IEDL.DBID GOX
IngestDate Mon Jan 08 05:39:58 EST 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a673-1d9e628a06852c40e1eb25faa26f9964153bec8741b23cc3adcd47e4aa09ffe63
OpenAccessLink https://arxiv.org/abs/2209.14963
ParticipantIDs arxiv_primary_2209_14963
PublicationCentury 2000
PublicationDate 2022-09-29
PublicationDateYYYYMMDD 2022-09-29
PublicationDate_xml – month: 09
  year: 2022
  text: 2022-09-29
  day: 29
PublicationDecade 2020
PublicationYear 2022
Score 1.8619912
SecondaryResourceType preprint
Snippet This paper considers the problem of finding near-optimal Markovian randomized (MR) policies for finite-state-action, infinite-horizon, constrained...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Mathematics - Optimization and Control
Title Approximate Solutions To Constrained Risk-Sensitive Markov Decision Processes
URI https://arxiv.org/abs/2209.14963
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1NSwMxEB3anryIolI_ycFrMM1mk-6xqLUIVdAKvZV8TKAIrbS19Oc7ya4fFy972M1h85LdN5O8vAG4DhUGVCi4N3RRwStehRh5aaKTjl7dZ7um8ZMevanHaTltAfs-C2NXu_m29gd26xspRUXfMk2SNrSlTJKth-dpvTmZrbia9r_tKMbMt_6QxPAA9pvojg3q4TiEFi6OYDxIvt27OcWGyH6WodhkyVK1zFyjAQN7ma_f-WuSk6cfEEtnaJZbdteUwGGNoB_XxzAZ3k9uR7ypYsCtNgXvERpa9q3Q_VJ6JbBHuWwZrZU6Uq5B_FkQjH0idicL7wtCJyiDylpRxYi6OIHOYrnALjD0vSBcWWDpFEUxRCyI2hivhaNMszKn0M19n33URhWzBMssw3L2_6Nz2JNJ0p91UxfQ2aw-8ZKIduOuMtpfGH-AQw
link.rule.ids 228,230,786,891
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Approximate+Solutions+To+Constrained+Risk-Sensitive+Markov+Decision+Processes&rft.au=M%2C+Uday+Kumar&rft.au=Bhat%2C+Sanjay+P&rft.au=Kavitha%2C+Veeraruna&rft.au=Hemachandra%2C+Nandyala&rft.date=2022-09-29&rft_id=info:doi/10.48550%2Farxiv.2209.14963&rft.externalDocID=2209_14963