About

We're a one-term Winter 2024 comps team of six people, advised by Anna Rafferty.

Our Project

We were tasked with learning about how safety in Reinforcement Learning (RL) works.

Deliverables

If you head over to the Downloads tab, you can see our final paper and our GitHub.

Main Focus

We wanted to investigate how altering hyperparameters of typical safe reinforcement learning algorithms would impact returns and costs for different agents and tasks. We used Omnisafe, an infrastructural framework for Safe Reinforcement Learning to run all of our experiments. They provide pre-built algorithm implementations, safety adapters, environments from Safety Gym, and documentation on how to use their software. OmniSafe also integreates with Weights and Biases, a logging tool for ML research. We chose to run our tests on SafetyHumanoidVelocityv1 and SafetyCarGoal2v0 as the have different safety measures and different levels of complexity. This allows us to test for differet definitions of safety and compare algorithm performance.

Links to both resources:

OmniSafe
Omnisafe Logo
Weights and Biases
Weights and Biases Logo