This paper introduces the Actor-Critic using Kronecker-factored Trust Region (ACKTR), a scalable optimization method for deep reinforcement learning that employs a Kronecker-factored approximation to natural policy gradients. ACKTR enhances both actor and critic updates, achieving significant improvements in sample efficiency and final performance in various environments, notably outperforming traditional methods such as A2C and TRPO. The authors claim that ACKTR is the first scalable implementation of trust region optimization for actor-critic methods, making it suitable for complex tasks directly from raw pixel inputs.