Back to Search Start Over

Olfactory-Based Navigation via Model-Based Reinforcement Learning and Fuzzy Inference Methods

Authors :
Lingxiao Wang
Shuo Pang
Jin-long Li
Source :
IEEE Transactions on Fuzzy Systems. 29:3014-3027
Publication Year :
2021
Publisher :
Institute of Electrical and Electronics Engineers (IEEE), 2021.

Abstract

This article presents an olfactory-based navigation algorithm for using a mobile robot to locate an odor source in a turbulent flow environment. We analogize the odor source localization as a reinforcement learning problem. During the odor plume tracing process, the belief state in a partially observable Markov decision process model is adapted to generate a source probability map that estimates possible odor source locations, and a hidden Markov model is employed to produce a plume distribution map that premises plume propagation areas. Both source and plume estimations are fed to the robot, and a decision-making approach based on fuzzy inference is designed to dynamically fuse information from two maps and to balance the exploitation and exploration of the search. After assigning the fused information to reward functions, a value iteration based path planning algorithm is presented to solve for the optimal action policy. Comparing to other commonly used olfactory-based navigation algorithms, such as moth-inspired and Bayesian inference methods, simulation results show that the proposed method is more intelligent and efficient.

Details

ISSN :
19410034 and 10636706
Volume :
29
Database :
OpenAIRE
Journal :
IEEE Transactions on Fuzzy Systems
Accession number :
edsair.doi...........eafd19b7f5c7d33f25d977ca1c2f896b
Full Text :
https://doi.org/10.1109/tfuzz.2020.3011741