Authors:
Toshihiro Matsui
and
Hiroshi Matsuo
Affiliation:
Nagoya Institute of Technology, Japan
Keyword(s):
Multiagent System, Reinforcement Learning, Distributed Constraint Optimization, Unfairness, Leximin.
Related
Ontology
Subjects/Areas/Topics:
Agents
;
Artificial Intelligence
;
Artificial Intelligence and Decision Support Systems
;
Cooperation and Coordination
;
Distributed and Mobile Software Systems
;
Distributed Problem Solving
;
Enterprise Information Systems
;
Knowledge Engineering and Ontology Development
;
Knowledge-Based Systems
;
Multi-Agent Systems
;
Software Engineering
;
Symbolic Systems
Abstract:
Reinforcement learning has been studied for cooperative learning and optimization methods in multiagent systems. In several frameworks of multiagent reinforcement learning, the system’s whole problem is decomposed into local problems for agents. To choose an appropriate cooperative action, the agents perform an optimization method that can be performed in a distributed manner. While the conventional goal of the learning is the maximization of the total rewards among agents, in practical resource allocation problems, unfairness among agents is critical. In several recent studies of decentralized optimization methods, unfairness was considered a criterion. We address an action selection method based on leximin criteria, which reduces the unfairness among agents, in decentralized reinforcement learning. We experimentally evaluated the effects and influences of the proposed approach on classes of sensor network problems.