Skip to content

Latest commit

 

History

History
executable file
·
48 lines (39 loc) · 2.92 KB

README.md

File metadata and controls

executable file
·
48 lines (39 loc) · 2.92 KB

Visualizing and Understanding Atari Agents

Link to paper. Blog post.

Sam Greydanus. October 2017. MIT License. Explainable AI Project. Supported by DARPA.

Strong agents

breakout-tunneling.gif pong-killshot.gif spaceinv-aiming.gif

The Breakout agent executes a tunneling strategy. The Pong agent makes a kill shot. The SpaceInvaders agent displays an aiming strategy. (All agents are LSTM-CNN trained via A3C; Blue=actor, Red=critic)

Overfit agents

Learning

breakout-learning The Breakout agent learns a tunneling stategy (each frame is separated by ~1e7 frames of training).

About

A quick comparison of our method to Jacobian-based saliency: Jupyter notebook

Abstract. Deep reinforcement learning (deep RL) agents have achieved remarkable success in a broad range of game-playing and continuous control tasks. While these agents are effective at maximizing rewards, it is often unclear what strategies they use to do so. In this paper, we take a step toward explaining deep RL agents through a case study in three Atari 2600 environments. In particular, we focus on understanding agents in terms of their visual attentional patterns during decision making. To this end, we introduce a method for generating rich saliency maps and use it to explain 1) what strong agents attend to 2) whether agents are making decisions for the right or wrong reasons, and 3) how agents evolve during the learning phase. We also test our method on non-expert human subjects and find that it improves their ability to reason about these agents. Our techniques are general and, though we focus on Atari, our long-term objective is to produce tools that explain any deep RL policy.

Pretrained models

We provide pretrained models for obtaining results like those in the "Strong agents" and "Overfit agents" sections of the paper. These models were obtained using this repo (default hyperparameters).

  1. Download from this link
  2. Unzip the file in this directory

Dependencies

All code is written in Python 3.6. You will need:

  • NumPy
  • SciPy
  • Matplotlib
  • PyTorch 0.2: easier to write and debug than TensorFlow :)
  • Jupyter