A simple two-player environment for openai/gym
Go to file
Nicolai Czempin e8dcc0fb4f Merge pull request #2 from tsu-nera/master
Add diagonals in game over conditions
2017-06-13 15:20:31 +02:00
examples Provide different policy for player 1 and -1 2017-04-04 18:16:57 +02:00
gym_tic_tac_toe checking the diagonals outside the i loop 2017-06-13 19:57:32 +09:00
.gitignore Initial commit 2017-03-31 00:36:55 +02:00
LICENSE Initial commit 2017-03-31 00:36:55 +02:00
README.md Update README.md 2017-04-03 21:55:40 +02:00
setup.py Add basic TTT 2017-04-03 23:51:35 +02:00

gym-tic-tac-toe

An example of a custom environment for https://github.com/openai/gym.

I want to try out self-play in a Reinforcement Learning context. Rather than the board game environments on openai/gym right now, which are "single-player" by providing a built-in opponent, I want to create an agent that learns a strategy by playing against itself, so it will try to maximize the reward for "player 1" and minimize it for "player 2".

The canonical example of a simple two player game is Tic Tac Toe, also known as Noughts and Crosses.