Addressing Function Approximation Error in Actor-Critic Methods
Date:
In value-based reinforcement learning methods, function approximation errors are known to lead to overestimated value estimates and sub-optimal policies.
And the problem persists in an actor-critic setting.
This paper proposes novel mechanisms to minimize its effects on both the actor and the critic.
Leave a Comment