Actor-Critic Methods


Actor-Critic Methods

  • Actor-critic agent is an agent that uses function approximation to learn a policy and value function
  • Employ two NN
    • One for Actor
    • One for Policy
  • Advantages:
    • more stable than value-based agents
    • Need fewer samples than policy-based agents

Generalized Advantage Estimation (GAE)

DDPG

  • Deep Deterministic Policy Gradient
  • DDPG is a DQN method for continuous action spaces
  • DDPG uses a replay buffer and soft updates to the target networks

Read