Quasi-Newton Compatible Actor-Critic for Deterministic Policies
By: Arash Bahari Kordabad , Dean Brandner , Sebastien Gros and more
Potential Business Impact:
Teaches computers to learn faster by watching mistakes.
In this paper, we propose a second-order deterministic actor-critic framework in reinforcement learning that extends the classical deterministic policy gradient method to exploit curvature information of the performance function. Building on the concept of compatible function approximation for the critic, we introduce a quadratic critic that simultaneously preserves the true policy gradient and an approximation of the performance Hessian. A least-squares temporal difference learning scheme is then developed to estimate the quadratic critic parameters efficiently. This construction enables a quasi-Newton actor update using information learned by the critic, yielding faster convergence compared to first-order methods. The proposed approach is general and applicable to any differentiable policy class. Numerical examples demonstrate that the method achieves improved convergence and performance over standard deterministic actor-critic baselines.
Similar Papers
First-order Sobolev Reinforcement Learning
Machine Learning (CS)
Teaches computers to learn faster and more reliably.
Nonlinear discretizations and Newton's method: characterizing stationary points of regression objectives
Machine Learning (CS)
Makes AI learn faster by using better math.
Mitigating Estimation Bias with Representation Learning in TD Error-Driven Regularization
Machine Learning (CS)
Teaches robots to learn better by balancing risks.