Available Agents
The SLC package includes a collection of robust RL algorithms accompanied by their less stable baselines. These algorithms are designed with non-recurrent MLP actor-critic models, making them well-suited for fully observable RL environments that do not rely on image data, such as the gymnasium Mujoco and stable-gym environments. The implementation follows a modular approach, allowing for seamless adaptation to different types of environments and neural network architectures.
Stable Agents
Important
As explained in the installation section of the documentation,
although the opt_type
algorithm variable can be used to train on standard
gymnasium environments, the stable RL agents require a positive definite
cost function to guarantee stability (and robustness). Several custom environments with
positive definite cost functions can be found in the stable-gym and
ros-gazebo-gym packages. When using the latter, make sure to set
positive_reward
to True
.
The SLC package currently contains the following theoretically stable RL algorithms:
Unstable Agents
The SLC package currently contains the following (unstable) baseline RL algorithms: