导航
学术报告|
当前位置:首页  科学研究  学术报告
报告题目: Reinforcement Learning Algorithms for Entropy-Regularized HJB Equations with Model Uncertainty
报 告 人: Gaozhan Wang
报告人所在单位: University of Southern California
报告日期: 2025-05-26
报告时间: 10:00-11:00
报告地点: 光华东主楼2001
   
报告摘要:

In this paper we investigate the convergence of the Policy Iteration Algorithm (PIA) for a class of general continuous-time entropy-regularized stochastic control problems. In particular, instead of employing sophisticated PDE estimates for the iterative PDEs involved in the algorithm (see, e.g., Huang-Wang-Zhou (2023)), we shall provide a simple proof from scratch for the convergence of the PIA. Our approach builds on probabilistic representation formulae for solutions of PDEs and their derivatives. Moreover, in the finite horizon model and in the infinite horizon model with large discount factor, the similar arguments lead to a super-exponential rate of convergence without tear. Finally, with some extra efforts we show that our approach can be extended to the diffusion control case in the one dimensional setting, also with a super-exponential rate of convergence. In addition, building upon this approach we address model uncertainty and propose a new algorithm that does not require the explicit functional form of the coefficients, relying instead on observable values. This significantly enhances the applicability of our results in practical scenarios.

Gaozhan Wang20250526.pdf

   
本年度学院报告总序号: 1055

Copyright © |2012 复旦大学数学科学学院版权所有 沪ICP备042465