World model based multi-agent proximal policy optimization framework for multi-agent pathfinding

dc.contributor.authorChung, Jaehoon
dc.contributor.supervisorNajjaran, Homayoun
dc.date.accessioned2024-09-16T20:42:05Z
dc.date.available2024-09-16T20:42:05Z
dc.date.issued2024
dc.degree.departmentDepartment of Mechanical Engineering
dc.degree.levelMaster of Applied Science MASc
dc.description.abstractMulti-agent pathfinding plays a crucial role in various robot applications. Recently, deep reinforcement learning methods have been adopted to solve large-scale planning problems in a decentralized manner. Nonetheless, such approaches pose challenges such as non-stationarity and partial observability. This thesis addresses these challenges by introducing a centralized communication block into a multi-agent proximal policy optimization framework. The evaluation is conducted in a simulation based environment, featuring continuous state and action spaces. The simulator consists of a vectorized 2D physics engine where agents are bound by the laws of physics. Within the framework, a World model is utilized to extract and abstract representation features from the global map, leveraging the global context to enhance the training process. This approach involves decoupling the feature extractor from the agent training process, enabling a more accurate representation of the global state that remains unbiased by the actions of the agents. Furthermore, the modularized approach offers the flexibility to replace the representation model with another model or modify tasks within the global map without the retraining of the agents. The empirical study demonstrates the effectiveness of the proposed approach by comparing three proximal policy optimization-based multi-agent pathfinding frameworks. The results indicate that utilizing an autoencoder-based state representation model as the centralized communication model sufficiently provides the global context. Additionally, introducing centralized communication block improves performance and the generalization capability of agent policies.
dc.description.scholarlevelGraduate
dc.identifier.bibliographicCitation@article{chung2024learning, title={Learning team-based navigation: a review of deep reinforcement learning techniques for multi-agent pathfinding}, author={Chung, Jaehoon and Fayyad, Jamil and Younes, Younes Al and Najjaran, Homayoun}, journal={Artificial Intelligence Review}, volume={57}, number={2}, pages={41}, year={2024}, publisher={Springer} }
dc.identifier.urihttps://hdl.handle.net/1828/20436
dc.languageEnglisheng
dc.language.isoen
dc.rightsAvailable to the World Wide Web
dc.subjectDeep Reinforcement Learning
dc.subjectMulti-agent Pathfinding
dc.subjectAutoencoder
dc.subjectMobile Robot
dc.subjectMulti-agent Reinforcement Learning
dc.titleWorld model based multi-agent proximal policy optimization framework for multi-agent pathfinding
dc.typeThesis

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Chung_Jaehoon_MAsc_2024.pdf
Size:
10.64 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.62 KB
Format:
Item-specific license agreed upon to submission
Description: