Closed draabe closed 4 months ago
Hi!
Thank you for your valuable feedback. Here are my responses to the points you've raised:
Major Points:
Outdated "How to Run" Section: Apologies for the confusion regarding the outdated "How to run" section. The algorithms repository had some modifications to its structure to align with the documentation, but these changes had not been merged at the time of your review. Now, everything is updated and in sync. The scripts prefixed with "train" and "test" provide real demos on how to combine our project with the well-established Reinforcement Learning library, RLlib.
Clarifying Functionality Between Repositories: Alongside developing the DSSE environment, we conducted research that resulted in the algorithms repository. While these algorithms are specifically designed to use our environment, the project submitted for review focuses solely on the environment itself. The DSSE-algorithms repository serves as a practical demonstration of the scientific and research applications for the software developed.
Documentation and API Reference:
There are no additional hidden functionalities beyond the env
objects. The package is built on top of the PettingZoo interface, extending reinforcement learning applications to the domain of autonomous drone SAR missions.
Minor Points:
GDAL Dependency: Thank you for highlighting the GDAL dependency issue. It indeed requires Microsoft Visual C++ 14.0 or greater for building. We will add this information to the documentation to ensure users are aware of this requirement.
Random Policy Function: The "random_policy" function is designed to abstract the concept of a model or function that chooses actions within the environment's action space. After training a model with RLlib, you can use it to select an action with the following code:
from ray.rllib.algorithms.ppo import PPO
model = PPO.from_checkpoint(checkpoint_path)
action = model.compute_single_action(observation)
Here, checkpoint_path
refers to the directory where RLlib saves the state of the model during training. This is essential for preserving progress in case of long training processes that could take many hours.
I will ensure to include additional guidance on how to customize the "random_policy" function to meet specific needs.
Thank you again for your insightful feedback. If there are any further questions or clarifications needed, please let me know. I look forward to your continued review and any additional suggestions you might have.
Great, thanks!
Hi @renatex333,
I've checked the repository for its functionality, and here are my points w.r.t to the revision. In general, installation and core functionality works fine on my machine, and I like the overall structure of the package. The documentation also covers the most important points. I'm having some difficulties running the algorithms with the package being split into two repositories. Here's a few things I noticed:
Major
env
object for high- or low-level usage? This is difficult to see as there seems to be no dedicated API reference in your documentation. If everything is limited to theenv
object, I feel that the documentation in theSearch Environment
andCoverage Environment
are sufficiently explanatory for usage, but maybe I'm missing some functions?Minor
pip install DSSE
Quick Start
code and then building the explanation of functionality around it. The only thing I did not find sufficient information on (when I want to modify it) is the interface of therandom_policy
function - what does this one need to do in order to use it beyond the sample you provide?Could you kindly help adressing these points? Let me know if anything is unclear.