Speaker
Hans Harder
(University of Paderborn)
Description
The value function plays a crucial role as a measure for the cumulative future reward an agent receives in both reinforcement learning and optimal control. It is therefore of interest to study how similar the values of neighboring states are, i.e. to investigate the continuity of the value function. We do so by providing and verifying upper bounds on the value function's modulus of continuity. Additionally, we show that the value function is always Hölder continuous under relatively weak assumptions on the underlying system.
Authors
Hans Harder
(University of Paderborn)
Dr
Sebastian Peitz
(University of Paderborn)