Conference Paper/Proceeding/Abstract 231 views
Predator-Prey Q-Learning Based Collaborative Coverage Path Planning for Swarm Robotics
Michael Watson,
Hans Ren,
Farshad Arvin,
Junyan Hu
2024 Annual Conference Towards Autonomous Robotic Systems (TAROS)
Swansea University Author: Hans Ren
Full text not available from this repository: check for access using links below.
Abstract
Coverage Path Planning (CPP) is an effective approach to let intelligent robots cover an area by finding feasible paths through the environment. In this paper, we focus on using reinforcement learning to learn about a given environment and find the most efficient path that explores all target points...
Published in: | 2024 Annual Conference Towards Autonomous Robotic Systems (TAROS) |
---|---|
ISSN: | 0302-9743 |
Published: |
Springer
2024
|
Online Access: |
Check full text
|
URI: | https://cronfa.swan.ac.uk/Record/cronfa66908 |
Abstract: |
Coverage Path Planning (CPP) is an effective approach to let intelligent robots cover an area by finding feasible paths through the environment. In this paper, we focus on using reinforcement learning to learn about a given environment and find the most efficient path that explores all target points. To overcome the limitations caused by stan- dard Q-learning based CPP that often fall into a local optimum and may be in-efficient in large-scale environments, two methods of improve- ment are considered, i.e., the use of a robot swarm working towards the same goal and the augmenting of the Q-learning algorithm to include a predator-prey based reward system. Existing predator-prey based reward systems provide rewards the further away an agent is from its predator, the paper adapts this concept to work within a robot swarm by simulat- ing each agent of the swarm as both predator and prey. Simulation case studies and comparisons with the standard Q-learning show that the proposed method has a superior coverage performance in complicated environments. |
---|---|
College: |
Faculty of Science and Engineering |