Unofficial name: S.T.A.R. this if you have G.U.T.S.
Example of a Target being Detected | Desert mountainous Search Region |
---|---|
Overhead View of the Scene on top | The fullsize robots pictured in the massive region |
A known terrain map informs stealthy behaviour such as moving through regions of greater concealment (for eg. between/close around mountains) as well as efficient search behaviours such as searching in better hiding places.
Published in Conference on Robot Learning (CoRL) 2023
@inproceedings{
bakshi2023stealthy,
title={Stealthy Terrain-Aware Multi-Agent Active Search},
author={Nikhil Angad Bakshi and Jeff Schneider},
booktitle={7th Annual Conference on Robot Learning},
year={2023},
url={https://openreview.net/forum?id=eE3fsO5Mi2}
}
In order to run the repo the dependencies are ROS melodic and python2.7 with numpy, matplotlib and PIL.
To recreate the experiments or just run the algorithm you should use docker:
- Install docker following this link: https://docs.docker.com/engine/install/
- Recommended OS is Ubuntu 18.04
- Then run the following command to download the necessary image and start the container
bash start_docker.sh
- Inside the docker container:
cd /home/user/src
bash init.sh
- Ignore any pip install errors that occur if they do, then:
cd zone_recon
source devel/setup.bash
USER=user ROS_LOG_DIR=/home/user/src/zone_recon/logs bash test_mysim.sh 100 star 5 4
- The syntax for the above command is
test_mysim.sh <runtime budget in seconds> <algorithm: star, rsi, guts, coverage, random> <number of targets> <number of agents: 1 2 4 or 8>
More detailed instructions to follow to recreate experiments. - Feel free to try any of combinations. The runs can be seen in the logging folder such that every robot has logs (replicating the real life system).
- The results for the paper were generated using the
experiments{_multi,_simplesim}.sh
scripts and they take several days for realistic simulations and several hours for the simplified simulator. - The main code for STAR, GUTS, RSI, COVERAGE and RANDOM policies tested for adversarial multi agent active search in this paper can be found written in src/zone_recon/src/waypoint_planner/src/waypoint_planner/active_search.py.
N. A. Bakshi and J. Schneider are with the Robotics Institute, School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213
(c) Nikhil Angad Bakshi 2023 ([email protected])
This material is based upon work supported by the U.S. Army Research Office and the U.S. Army Futures Command under Contract No. W911NF-20-D-0002. Authors would like to acknowledge the contributions of Conor Igoe, Tejus Gupta and Arundhati Banerjee to the development of the STAR algorithm. Additionally the work on the physical platforms and true-to-life simulations were enabled thanks to Herman Herman, Jesse Holdaway, Prasanna Kannappan, Luis Ernesto Navarro-Serment, Maxfield Kassel and Trenton Tabor.