Project/Area Number |
07680402
|
Research Category |
Grant-in-Aid for Scientific Research (C)
|
Allocation Type | Single-year Grants |
Section | 一般 |
Research Field |
Intelligent informatics
|
Research Institution | THE UNIVERSITY OF TOKUSHIMA |
Principal Investigator |
ONO Norihiko THE UNIVERSITY OF TOKUSHIMA,FACULTY OF ENGINEERING,ASSOCIATE PROFESSOR, 工学部, 助教授 (60194594)
|
Project Period (FY) |
1995 – 1996
|
Project Status |
Completed (Fiscal Year 1996)
|
Budget Amount *help |
¥2,300,000 (Direct Cost: ¥2,300,000)
Fiscal Year 1996: ¥800,000 (Direct Cost: ¥800,000)
Fiscal Year 1995: ¥1,500,000 (Direct Cost: ¥1,500,000)
|
Keywords | MULTI-AGENT SYSTEMS / ARTIFICIAL LIFE / REINFORCEMENT-LEARNING / MACHINE LEARNING / DISTRIBUTED ARTIFICIAL INTELLIGENCE / GENETIC ALGORITHMS / 遺伝的アルゴリズム |
Research Abstract |
In attempting to let artificial organisms or simple reactive robots synthesize some coordinated behavior, several researchers in the fields of artificial life and robotics have applied monolithic reinforcement-learning algorithms to multi-agent learning problems. In most of these applications, only a small number of learning agents are engaged in their joint tasks and accordingly the state space for each agent is relatively small. This is the reason why monolithic reinforcement-learning algorithms have been successfully applied to these multi-agent learning problems. However, these straightforward applications of reinforcement-learning algorithms do not successfully scale up to more complex multi-agent learning problems, where not a few learning agents are engaged in some coordinated tasks. In such a multi-agent problem domain, agents should appropriately behave according to not only sensory information produced by the physical environment itself but also that produced by other agents, and hence the state space for each reinforcement-learning agent grows exponentially in the number of agents operating in the same environment. Even simple multi-agent learning problems are computationally intractable by the monolithic reinfocrement-learning approaches. We consider a modified version of the Pursuit Problem as such a multi-agent learning problem, and show how successfully modular Q-learning prey-pursuing agents synthesize coordinated decision policies needed to capture a randomly-moving prey agent. Multi-agent learning is an extremely difficult problem in general, and the results we obtained strongly-rely on specific attributes of the problem. But the results are quite encouraging and suggest that our modular reinforcement-learning approach is promising in studying adaptive behavior of multiple autonomous agents.
|