Reinforcement learning-based mobile robot navigation


ALTUNTAŞ N., Imal E., Emanet N., Ozturk C. N.

TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, cilt.24, sa.3, ss.1747-1767, 2016 (SCI-Expanded) identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 24 Sayı: 3
  • Basım Tarihi: 2016
  • Doi Numarası: 10.3906/elk-1311-129
  • Dergi Adı: TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Scopus, TR DİZİN (ULAKBİM)
  • Sayfa Sayıları: ss.1747-1767
  • Anahtar Kelimeler: Reinforcement learning, temporal difference, eligibility traces, Sarsa, Q-learning, mobile robot navigation, obstacle avoidance
  • Bursa Uludağ Üniversitesi Adresli: Hayır

Özet

In recent decades, reinforcement learning (RL) has been widely used in different research fields ranging from psychology to computer science. The unfeasibility of sampling all possibilities for continuous-state problems and the absence of an explicit teacher make RL algorithms preferable for supervised learning in the machine learning area, as the optimal control problem has become a popular subject of research. In this study, a system is proposed to solve mobile robot navigation by opting for the most popular two RL algorithms, Sarsa(lambda) and Q(lambda). The proposed system, developed in MATLAB, uses state and action sets, defined in a novel way, to increase performance. The system can guide the mobile robot to a desired goal by avoiding obstacles with a high success rate in both simulated and real environments. Additionally, it is possible to observe the effects of the initial parameters used by the RL methods, e.g., lambda, on learning, and also to make comparisons between the performances of Sarsa(lambda) and Q(lambda) algorithms.