Skip to content

Latest commit

 

History

History
28 lines (20 loc) · 1.24 KB

README.md

File metadata and controls

28 lines (20 loc) · 1.24 KB

Implicit Q-Learning

This repository contains a PyTorch re-implementation of Offline Reinforcement Learning with Implicit Q-Learning by Ilya Kostrikov, Ashvin Nair, and Sergey Levine.

For the official repository, please use: https://github.com/ikostrikov/implicit_q_learning

This code can be used for offline RL, or for offline RL followed by online finetuning. Negative epochs are offline RL, positive epochs are online (the agent is actively collecting data and adding it to the replay buffer).

If you use this code for your research, please consider citing the paper:

@article{kostrikov2021iql,
    title={Offline Reinforcement Learning with Implicit Q-Learning},
    author={Ilya Kostrikov and Ashvin Nair and Sergey Levine},
    year={2021},
    archivePrefix={arXiv},
    primaryClass={cs.LG}
}

Tests

To run quick versions of these experiments to test if the code matches exactly as the results below, you can run the tests in tests/regression/iql

Mujoco results

Mujoco results

Antmaze results

Ant-maze results