Approximate Solutions To Constrained Risk-Sensitive Markov Decision Processes
This paper considers the problem of finding near-optimal Markovian randomized (MR) policies for finite-state-action, infinite-horizon, constrained risk-sensitive Markov decision processes (CRSMDPs). Constraints are in the form of standard expected discounted cost functions as well as expected risk-s...
Saved in:
Main Authors | , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
29.09.2022
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | This paper considers the problem of finding near-optimal Markovian randomized
(MR) policies for finite-state-action, infinite-horizon, constrained
risk-sensitive Markov decision processes (CRSMDPs). Constraints are in the form
of standard expected discounted cost functions as well as expected
risk-sensitive discounted cost functions over finite and infinite horizons. The
main contribution is to show that the problem possesses a solution if it is
feasible, and to provide two methods for finding an approximate solution in the
form of an ultimately stationary (US) MR policy. The latter is achieved through
two approximating finite-horizon CRSMDPs which are constructed from the
original CRSMDP by time-truncating the original objective and constraint cost
functions, and suitably perturbing the constraint upper bounds. The first
approximation gives a US policy which is $\epsilon$-optimal and feasible for
the original problem, while the second approximation gives a near-optimal US
policy whose violation of the original constraints is bounded above by a
specified $\epsilon$. A key step in the proofs is an appropriate choice of a
metric that makes the set of infinite-horizon MR policies and the feasible
regions of the three CRSMDPs compact, and the objective and constraint
functions continuous. A linear-programming-based formulation for solving the
approximating finite-horizon CRSMDPs is also given. |
---|---|
AbstractList | This paper considers the problem of finding near-optimal Markovian randomized
(MR) policies for finite-state-action, infinite-horizon, constrained
risk-sensitive Markov decision processes (CRSMDPs). Constraints are in the form
of standard expected discounted cost functions as well as expected
risk-sensitive discounted cost functions over finite and infinite horizons. The
main contribution is to show that the problem possesses a solution if it is
feasible, and to provide two methods for finding an approximate solution in the
form of an ultimately stationary (US) MR policy. The latter is achieved through
two approximating finite-horizon CRSMDPs which are constructed from the
original CRSMDP by time-truncating the original objective and constraint cost
functions, and suitably perturbing the constraint upper bounds. The first
approximation gives a US policy which is $\epsilon$-optimal and feasible for
the original problem, while the second approximation gives a near-optimal US
policy whose violation of the original constraints is bounded above by a
specified $\epsilon$. A key step in the proofs is an appropriate choice of a
metric that makes the set of infinite-horizon MR policies and the feasible
regions of the three CRSMDPs compact, and the objective and constraint
functions continuous. A linear-programming-based formulation for solving the
approximating finite-horizon CRSMDPs is also given. |
Author | M, Uday Kumar Kavitha, Veeraruna Bhat, Sanjay P Hemachandra, Nandyala |
Author_xml | – sequence: 1 givenname: Uday Kumar surname: M fullname: M, Uday Kumar – sequence: 2 givenname: Sanjay P surname: Bhat fullname: Bhat, Sanjay P – sequence: 3 givenname: Veeraruna surname: Kavitha fullname: Kavitha, Veeraruna – sequence: 4 givenname: Nandyala surname: Hemachandra fullname: Hemachandra, Nandyala |
BackLink | https://doi.org/10.48550/arXiv.2209.14963$$DView paper in arXiv |
BookMark | eNotj8tOwzAUBb2ABRQ-gBX-gQS_4sTLKrwqtQLR7KMb50ayWuzIDlH5e0JhNZujo5lrcuGDR0LuOMtVVRTsAeLJzbkQzORcGS2vyG49jjGc3CdMSPfh-DW54BNtAq0XThGcx55-uHTI9uiTm9yMdAfxEGb6iNalZU7fY7CYEqYbcjnAMeHtP1ekeX5q6tds-_ayqdfbDHQpM94b1KICpqtCWMWQYyeKAUDowRiteCE7tFWpeCektRJ626sSFQAzw4Barsj93-25px3joh-_29-u9twlfwCAiEty |
ContentType | Journal Article |
Copyright | http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
Copyright_xml | – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
DBID | AKZ GOX |
DOI | 10.48550/arxiv.2209.14963 |
DatabaseName | arXiv Mathematics arXiv.org |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
DeliveryMethod | fulltext_linktorsrc |
ExternalDocumentID | 2209_14963 |
GroupedDBID | AKZ GOX |
ID | FETCH-LOGICAL-a673-1d9e628a06852c40e1eb25faa26f9964153bec8741b23cc3adcd47e4aa09ffe63 |
IEDL.DBID | GOX |
IngestDate | Mon Jan 08 05:39:58 EST 2024 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-a673-1d9e628a06852c40e1eb25faa26f9964153bec8741b23cc3adcd47e4aa09ffe63 |
OpenAccessLink | https://arxiv.org/abs/2209.14963 |
ParticipantIDs | arxiv_primary_2209_14963 |
PublicationCentury | 2000 |
PublicationDate | 2022-09-29 |
PublicationDateYYYYMMDD | 2022-09-29 |
PublicationDate_xml | – month: 09 year: 2022 text: 2022-09-29 day: 29 |
PublicationDecade | 2020 |
PublicationYear | 2022 |
Score | 1.8619912 |
SecondaryResourceType | preprint |
Snippet | This paper considers the problem of finding near-optimal Markovian randomized
(MR) policies for finite-state-action, infinite-horizon, constrained... |
SourceID | arxiv |
SourceType | Open Access Repository |
SubjectTerms | Mathematics - Optimization and Control |
Title | Approximate Solutions To Constrained Risk-Sensitive Markov Decision Processes |
URI | https://arxiv.org/abs/2209.14963 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1NSwMxEB3anryIolI_ycFrMM1mk-6xqLUIVdAKvZV8TKAIrbS19Oc7ya4fFy972M1h85LdN5O8vAG4DhUGVCi4N3RRwStehRh5aaKTjl7dZ7um8ZMevanHaTltAfs-C2NXu_m29gd26xspRUXfMk2SNrSlTJKth-dpvTmZrbia9r_tKMbMt_6QxPAA9pvojg3q4TiEFi6OYDxIvt27OcWGyH6WodhkyVK1zFyjAQN7ma_f-WuSk6cfEEtnaJZbdteUwGGNoB_XxzAZ3k9uR7ypYsCtNgXvERpa9q3Q_VJ6JbBHuWwZrZU6Uq5B_FkQjH0idicL7wtCJyiDylpRxYi6OIHOYrnALjD0vSBcWWDpFEUxRCyI2hivhaNMszKn0M19n33URhWzBMssw3L2_6Nz2JNJ0p91UxfQ2aw-8ZKIduOuMtpfGH-AQw |
link.rule.ids | 228,230,786,891 |
linkProvider | Cornell University |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Approximate+Solutions+To+Constrained+Risk-Sensitive+Markov+Decision+Processes&rft.au=M%2C+Uday+Kumar&rft.au=Bhat%2C+Sanjay+P&rft.au=Kavitha%2C+Veeraruna&rft.au=Hemachandra%2C+Nandyala&rft.date=2022-09-29&rft_id=info:doi/10.48550%2Farxiv.2209.14963&rft.externalDocID=2209_14963 |