arXiv Analytics

Sign in

arXiv:2206.08881 [cs.AI]AbstractReferencesReviewsResources

Logic-based Reward Shaping for Multi-Agent Reinforcement Learning

Ingy ElSayed-Aly, Lu Feng

Published 2022-06-17Version 1

Reinforcement learning (RL) relies heavily on exploration to learn from its environment and maximize observed rewards. Therefore, it is essential to design a reward function that guarantees optimal learning from the received experience. Previous work has combined automata and logic based reward shaping with environment assumptions to provide an automatic mechanism to synthesize the reward function based on the task. However, there is limited work on how to expand logic-based reward shaping to Multi-Agent Reinforcement Learning (MARL). The environment will need to consider the joint state in order to keep track of other agents if the task requires cooperation, thus suffering from the curse of dimensionality with respect to the number of agents. This project explores how logic-based reward shaping for MARL can be designed for different scenarios and tasks. We present a novel method for semi-centralized logic-based MARL reward shaping that is scalable in the number of agents and evaluate it in multiple scenarios.

Related articles: Most relevant | Search more
arXiv:2211.01527 [cs.AI] (Published 2022-11-03)
Sensor Control for Information Gain in Dynamic, Sparse and Partially Observed Environments
arXiv:2405.18733 [cs.AI] (Published 2024-05-29)
Efficient Learning in Chinese Checkers: Comparing Parameter Sharing in Multi-Agent Reinforcement Learning
arXiv:2312.09009 [cs.AI] (Published 2023-12-14)
Adaptive parameter sharing for multi-agent reinforcement learning