Introduction
Basic Description of the Agents Module
Three variants of PPO are currently implemented:
Method
Python Module
Notes
Action Space
Observation Space
Compatible Environments
PPO Distributed Centralized Critic
ppo_distributed
Centralized Training with Decentralized Execution
Box / Discrete
Tuple
1.1 - 1.7, 2.1 - 2.6
PPO OneStep
ppo_one_step
Optimized Implementation for Single Step RL
Box / Discrete
Box
3.1
PPO Alternating Optimization
ppo_altopt
PPO with Bi-Level Optimization
Tuple[Box / Discrete, Box / Discrete]
Tuple[Box, Tuple]
3.2 - 3.5
To view the training configurations run:
$ python -m sdriving.agents.<module>.train --help
Last updated
Was this helpful?