██████╗ ██╗ ██╗██████╗ ██████╗ ██████╗ ██╗ ██╔══██╗╚██╗ ██╔╝██╔══██╗██╔═══██╗██╔══██╗██║ ██████╔╝ ╚████╔╝ ██████╔╝██║ ██║██████╔╝██║ ██╔═══╝ ╚██╔╝ ██╔══██╗██║ ██║██╔══██╗██║ ██║ ██║ ██║ ██║╚██████╔╝██║ ██║███████╗ ╚═╝ ╚═╝ ╚═╝ ╚═╝ ╚═════╝ ╚═╝ ╚═╝╚══════╝
PyroRL is a new reinforcement learning environment built for the simulation of wildfire evacuation. Check out the docs and the demo.
First, install our package:
pip install pyrorl
To use our wildfire evacuation environment, define the dimensions of your grid, where the populated areas are, the paths, and which populated areas can use which path. See an example below.
# Create environment
kwargs = {
'num_rows': num_rows,
'num_cols': num_cols,
'populated_areas': populated_areas,
'paths': paths,
'paths_to_pops': paths_to_pops
}
env = gymnasium.make('pyrorl/PyroRL-v0', **kwargs)
# Run a simple loop of the environment
env.reset()
for _ in range(10):
# Take action and observation
action = env.action_space.sample()
observation, reward, terminated, truncated, info = env.step(action)
# Render environment and print reward
env.render()
print("Reward: " + str(reward))
A compiled visualization of numerous iterations is seen below. For more examples, check out the examples/
folder.
For information on how to contribute, check out our contribution guide.