A collection of GPU-accelerated parallel game simulators for reinforcement learning (RL)
Note
If you find this project helpful, we would be grateful for your support through a GitHub star to help us grow the community and motivate further development!
v1
"tic_tac_toe"
v0
Each environment is versioned, and the version is incremented when there are changes that affect the performance of agents or when there are changes that are not backward compatible with the API. If you want to pursue complete reproducibility, we recommend that you check the version of Pgx and each environment as follows:
>>> pgx.__version__
'1.0.0'
>>> env.version
'v0'
Pgx is intended to complement these JAX-native environments with (classic) board game suits:
Combining Pgx with these JAX-native algorithms/implementations might be an interesting direction:
Currently, some environments, including Go and chess, do not perform well on TPUs. Please use GPUs instead.
If you use Pgx in your work, please cite our paper:
@inproceedings{koyamada2023pgx,
title={Pgx: Hardware-Accelerated Parallel Game Simulators for Reinforcement Learning},
author={Koyamada, Sotetsu and Okano, Shinri and Nishimori, Soichiro and Murata, Yu and Habara, Keigo and Kita, Haruka and Ishii, Shin},
booktitle={Advances in Neural Information Processing Systems},
pages={45716--45743},
volume={36},
year={2023}
}
Apache-2.0