Back to Search
Start Over
Low-Cost Multi-Agent Navigation via Reinforcement Learning With Multi-Fidelity Simulator
- Source :
- IEEE Access, Vol 9, Pp 84773-84782 (2021)
- Publication Year :
- 2021
- Publisher :
- IEEE, 2021.
-
Abstract
- In recent years, reinforcement learning (RL) has been widely used to solve multi-agent navigation tasks, and a high-fidelity level for the simulator is critical to narrow the gap between simulation and real-world tasks. However, high-fidelity simulators have high sampling costs and bottleneck the training model-free RL algorithms. Hence, we propose a Multi-Fidelity Simulator framework to train Multi-Agent Reinforcement Learning (MFS-MARL), reducing the total data cost with samples generated by a low-fidelity simulator. We apply the depth-first search to obtain local feasible policies on the low-fidelity simulator as expert policies to help the original reinforcement learning algorithm explore. We built a multi-vehicle simulator with variable fidelity levels to test the proposed method and compared it with the vanilla Soft Actor-Critic (SAC) and expert actor methods. The results show that our method can effectively obtain local feasible policies and can achieve a 23% cost reduction in multi-agent navigation tasks.
- Subjects :
- 0209 industrial biotechnology
General Computer Science
multi-robot systems
Computer science
media_common.quotation_subject
Fidelity
02 engineering and technology
010501 environmental sciences
01 natural sciences
Bottleneck
020901 industrial engineering & automation
Reinforcement learning
multi-fidelity simulators
General Materials Science
Simulation
0105 earth and related environmental sciences
media_common
Deep reinforcement learning
Robot kinematics
General Engineering
intelligent robots
TK1-9971
Cost reduction
Variable (computer science)
Task analysis
Robot
Electrical engineering. Electronics. Nuclear engineering
Subjects
Details
- Language :
- English
- ISSN :
- 21693536
- Volume :
- 9
- Database :
- OpenAIRE
- Journal :
- IEEE Access
- Accession number :
- edsair.doi.dedup.....9261cf19e7029bc13d7f53287ade91ea