Publication:
Markov decision processes: Monotonicity of optimal policy in exponential and quasi-hyperbolic discounting parameters

dc.contributor.coauthorKılıç, Hakan (58787428000)
dc.contributor.coauthorCanbolat, Pelin G. (54949950000)
dc.contributor.coauthorGunes, Evrim Didem (55940271800)
dc.date.accessioned2025-12-31T08:22:26Z
dc.date.available2025-12-31
dc.date.issued2025
dc.description.abstractIntertemporal preferences of decision makers, i.e., the way they discount delayed utilities, impact their decisions. Empirical evidence suggests that individuals commonly have hyperbolic discounting preferences. This can result in time-inconsistent behavior, e.g., procrastination, which may be a barrier to adopting preventive behavior such as machine maintenance and patient adherence to treatment. In this paper, we theoretically compare the actions of individuals based on their discounting characteristics. We consider the Hyperbolic Discounting (HD) model, which is more representative of individual behavior than Exponential Discounting (ED). We formulate a discrete-time finite-horizon Markov decision process with Quasi-Hyperbolic Discounting (QHD), an analytically tractable function representing HD and present sufficient conditions that ensure the monotonicity of the optimal policy in the discounting parameters. We consider submodular maximization or supermodular maximization problems. Our paper is the first to investigate the monotonicity of the optimal policy in QHD parameters for these problems. Moreover, we compare the optimal actions under ED and QHD. We apply our results to the settings of machine maintenance, individual health behavior and inventory control. We provide numerical examples that show there might not be monotonicity if our sufficient conditions are not met. Also, we explore the discrepancy between the expected total exponentially-discounted rewards of the actions obtained from QHD and of the actions that are optimal under ED, and observe that this discrepancy is affected mainly by the present bias. © 2025 Elsevier B.V., All rights reserved.
dc.description.fulltextYes
dc.description.harvestedfromManual
dc.description.indexedbyScopus
dc.description.publisherscopeInternational
dc.description.readpublishN/A
dc.description.sponsoredbyTubitakEuN/A
dc.identifier.doi10.1016/j.ejor.2025.09.013
dc.identifier.embargoNo
dc.identifier.issn0377-2217
dc.identifier.quartileN/A
dc.identifier.scopus2-s2.0-105016809149
dc.identifier.urihttps://doi.org/10.1016/j.ejor.2025.09.013
dc.identifier.urihttps://hdl.handle.net/20.500.14288/31659
dc.keywordsExponential Discounting
dc.keywordsMarkov Decision Processes
dc.keywordsOptimal Policy Monotonicity
dc.keywordsQuasi-hyperbolic Discounting
dc.keywordsBehavioral Research
dc.keywordsDecision Making
dc.keywordsMarkov Processes
dc.keywordsPatient Treatment
dc.keywordsPreventive Maintenance
dc.keywordsCondition
dc.keywordsExponential Discounting
dc.keywordsExponentials
dc.keywordsHyperbolic Discounting
dc.keywordsMachine Maintenance
dc.keywordsMarkov Decision Processes
dc.keywordsMonotonicity
dc.keywordsOptimal Policies
dc.keywordsOptimal Policy Monotonicity
dc.keywordsQuasi-hyperbolic Discounting
dc.keywordsHyperbolic Functions
dc.language.isoeng
dc.publisherElsevier B.V.
dc.relation.affiliationKoç University
dc.relation.collectionKoç University Institutional Repository
dc.relation.ispartofEuropean Journal of Operational Research
dc.relation.openaccessYes
dc.rightsCC BY-NC-ND (Attribution-NonCommercial-NoDerivs)
dc.rights.urihttps://creativecommons.org/licenses/by-nc-nd/4.0/
dc.titleMarkov decision processes: Monotonicity of optimal policy in exponential and quasi-hyperbolic discounting parameters
dc.typeJournal Article
dspace.entity.typePublication

Files