EconPapers    
Economics at your fingertips  
 

Time-Inhomogeneous Volatility Aversion for Financial Applications of Reinforcement Learning

Federico Cacciamani, Roberto Daluiso, Marco Pinciroli, Michele Trapletti and Edoardo Vittori

Papers from arXiv.org

Abstract: In finance, sequential decision problems are often faced, for which reinforcement learning (RL) emerges as a promising tool for optimisation without the need of analytical tractability. However, the objective of classical RL is the expected cumulated reward, while financial applications typically require a trade-off between return and risk. In this work, we focus on settings where one cares about the time split of the total return, ruling out most risk-aware generalisations of RL which optimise a risk measure defined on the latter. We notice that a preference for homogeneous splits, which we found satisfactory for hedging, can be unfit for other problems, and therefore propose a new risk metric which still penalises uncertainty of the single rewards, but allows for an arbitrary planning of their target levels. We study the properties of the resulting objective and the generalisation of learning algorithms to optimise it. Finally, we show numerical results on toy examples.

Date: 2026-02
New Economics Papers: this item is included in nep-upt
References: Add references at CitEc
Citations:

Downloads: (external link)
http://arxiv.org/pdf/2602.12030 Latest version (application/pdf)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:arx:papers:2602.12030

Access Statistics for this paper

More papers in Papers from arXiv.org
Bibliographic data for series maintained by arXiv administrators ().

 
Page updated 2026-02-24
Handle: RePEc:arx:papers:2602.12030