-
Notifications
You must be signed in to change notification settings - Fork 20
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
60e673f
commit 8c15608
Showing
18 changed files
with
1,349 additions
and
696 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,22 +1,33 @@ | ||
""" | ||
`PantheonRL <https://github.com/Stanford-ILIAD/PantheonRL>`_ is a package for training and testing multi-agent reinforcement learning environments. The goal of PantheonRL is to provide a modular and extensible framework for training agent policies, fine-tuning agent policies, ad-hoc pairing of agents, and more. | ||
`PantheonRL <https://github.com/Stanford-ILIAD/PantheonRL>`_ is a | ||
package for training and testing multi-agent reinforcement learning | ||
environments. The goal of PantheonRL is to provide a modular and | ||
extensible framework for training agent policies, fine-tuning agent | ||
policies, ad-hoc pairing of agents, and more. | ||
PantheonRL is built to support Stable-Baselines3 (SB3), allowing direct access to many of SB3's standard RL training algorithms such as PPO. PantheonRL currently follows a decentralized training paradigm -- each agent is equipped with its own replay buffer and update algorithm. The agents objects are designed to be easily manipulable. They can be saved, loaded and plugged into different training procedures such as self-play, ad-hoc / cross-play, round-robin training, or finetuning. | ||
PantheonRL is built to support Stable-Baselines3 (SB3), allowing | ||
direct access to many of SB3's standard RL training algorithms such as | ||
PPO. PantheonRL currently follows a decentralized training paradigm -- | ||
each agent is equipped with its own replay buffer and update | ||
algorithm. The agents objects are designed to be easily | ||
manipulable. They can be saved, loaded and plugged into different | ||
training procedures such as self-play, ad-hoc / cross-play, | ||
round-robin training, or finetuning. | ||
""" | ||
import pantheonrl.envs | ||
|
||
from pantheonrl.common.agents import ( | ||
Agent, | ||
StaticPolicyAgent, | ||
OnPolicyAgent, | ||
OffPolicyAgent | ||
OffPolicyAgent, | ||
) | ||
|
||
from pantheonrl.common.multiagentenv import ( | ||
DummyEnv, | ||
MultiAgentEnv, | ||
TurnBasedEnv, | ||
SimultaneousEnv | ||
SimultaneousEnv, | ||
) | ||
|
||
from pantheonrl.common.observation import Observation |
Oops, something went wrong.