HASARD (Harnessing Safe Reinforcement Learning with Doom) is a benchmark for Safe Reinforcement Learning within complex, egocentric perception 3D environments derived from the classic DOOM video game. It features 6 diverse scenarios each spanning across 3 levels of difficulty.
| Scenario | Level 1 | Level 2 | Level 3 |
|---|---|---|---|
| Armament Burden | ![]() |
![]() |
![]() |
| Detonator’s Dilemma | ![]() |
![]() |
![]() |
| Volcanic Venture | ![]() |
![]() |
![]() |
| Precipice Plunge | ![]() |
![]() |
![]() |
| Collateral Damage | ![]() |
![]() |
![]() |
| Remedy Rush | ![]() |
![]() |
![]() |
- Egocentric Perception: Agents learn solely from first-person pixel observations under partial observability.
- Beyond Simple Navigation: Whereas prior benchmarks merely require the agent to reach goal locations on flat surfaces while avoiding obstacles, HASARD necessitates comprehending complex environment dynamics, anticipating the movement of entities, and grasping spatial relationships.
- Dynamic Environments: HASARD features random spawns, unpredictably moving units, and terrain that is constantly moving or periodically changing.
- Difficulty Levels: Higher levels go beyond parameter adjustments, introducing entirely new elements and mechanics.
- Reward-Cost Trade-offs: Rewards and costs are closely intertwined, with tightening cost budget necessitating a sacrifice of rewards.
- Safety Constraints: Each scenario features a hard constraint setting, where any error results in immediate in-game penalties.
- Focus on Safety: Achieving high rewards is straightforward, but doing so while staying within the safety budget demands learning complex and nuanced behaviors.
HASARD enables overlaying a heatmap of the agent's most frequently visited locations providing further insights into its policy and behavior within the environment. These examples show how an agent navigates Volcanic Venture, Remedy Rush, and Armament Burden during the course of training:
HASARD supports augmented observation modes for further visual analysis. By utilizing privileged game state information, it can generate simplified observation representations, such as segmenting objects in the scene or rendering the environment displaying only depth from surroundings.
HASARD supports modular installation to install only the dependencies you need:
# Core dependencies only (environments and basic functionality)
pip install HASARD
# With sample-factory support for training RL agents
pip install HASARD[sample-factory]
# With results analysis and plotting tools
pip install HASARD[results]
# Full installation with all optional dependencies
pip install HASARD[sample-factory,results]To install from source:
git clone https://github.com/TTomilin/HASARD
cd HASARD
pip install . # or pip install .[sample-factory,results] for extrasTo get started with HASARD, here's a minimal example of running a task environment.
This script can also be found in run_env.py:
import hasard
env = hasard.make('RemedyRushLevel1-v0')
env.reset()
terminated = truncated = False
steps = total_cost = total_reward = 0
while not (terminated or truncated):
action = env.action_space.sample()
state, reward, terminated, truncated, info = env.step(action)
env.render()
steps += 1
total_cost += info['cost']
total_reward += reward
print(f"Episode finished in {steps} steps. Reward: {total_reward:.2f}. Cost: {total_cost:.2f}")
env.close()For highly parallelized training of Safe RL agents on HASARD environments, and to reproduce the results from the paper,
refer to sample_factory for detailed usage instructions and examples.
HASARD environments are built on top of the ViZDoom platform.
Our Safe RL baseline methods are implemented in Sample-Factory.
Our experiments were managed using WandB.
If you use our work in your research, please cite it as follows:
@inproceedings{tomilin2025hasard,
title={HASARD: A Benchmark for Vision-Based Safe Reinforcement Learning in Embodied Agents},
author={T. Tomilin, M. Fang, and M. Pechenizkiy},
booktitle={The Thirteenth International Conference on Learning Representations},
year={2025}
}
























