single-rb.php

JRM Vol.37 No.6 pp. 1452-1460
doi: 10.20965/jrm.2025.p1452
(2025)

Paper:

Communication-Free Adaptive Swarm Robotic System: LLM-Based Decision Making and MARL-Based Multi-Policy Control

Takahiro Yoshida and Yuichiro Sueoka ORCID Icon

Department of Mechanical Engineering, Graduate School of Engineering, The University of Osaka
2-1 Yamadaoka, Suita, Osaka 565-0871, Japan

Received:
April 7, 2025
Accepted:
August 12, 2025
Published:
December 20, 2025
Keywords:
communication-free adaptive swarm robotic system, large language models, deep multi-agent reinforcement learning
Abstract

Swarm robotic systems consist of a large number of distributed autonomous robots that coordinate their actions to accomplish diverse tasks beyond the capabilities of a single robot. These systems have recently been considered for deployment in disaster scenarios, where communication is often unstable, making it necessary to achieve adaptive cooperative behavior without relying on explicit communication between robots. In the context of multi-robot systems—including swarm robotic systems—some studies have explored approaches utilizing large language models (LLMs) or other learning-based methods, but few have proposed systems that enable communication-free coordination. In this paper, we propose a system incorporating a novel method that combines high-level decision-making via LLM-based policy selection—guided by questionnaire-style prompts—with low-level control using multiple MARL-trained policies. We consider a complex task scenario in which robots search for a target object and transport it to a designated destination. To evaluate the method, we define implicit consensus as a condition in which a robot selects the same policy as its nearby robots without any explicit communication. The effectiveness of the proposed method is demonstrated through simulated task execution, with particular emphasis on implicit consensus as a key evaluation metric.

LM-driven adaptive swarm robotic system

LM-driven adaptive swarm robotic system

Cite this article as:
T. Yoshida and Y. Sueoka, “Communication-Free Adaptive Swarm Robotic System: LLM-Based Decision Making and MARL-Based Multi-Policy Control,” J. Robot. Mechatron., Vol.37 No.6, pp. 1452-1460, 2025.
Data files:
References
  1. [1] E. Şahin, “Swarm robotics: From sources of inspiration to domains of application,” Int. Workshop on Swarm Robotics, pp. 10-20, 2004. https://doi.org/10.1007/978-3-540-30552-1_2
  2. [2] K. Nagatani, M. Abe, K. Osuka, P. jo Chun, T. Okatani, M. Nishio, S. Chikushi, T. Matsubara, Y. Ikemoto, and H. Asama, “Innovative technologies for infrastructure construction and maintenance through collaborative robots based on an open design approach,” Advanced Robotics, Vol.35, No.11, pp. 715-722, 2021. https://doi.org/10.1080/01691864.2021.1929471
  3. [3] P. Li, Z. An, S. Abrar, and L. Zhou, “Large language models for multi-robot systems: A survey,” arXiv preprint, arXiv:2502.03814, 2025. https://doi.org/10.48550/arXiv.2502.03814
  4. [4] J. Chen, C. Yu, X. Zhou, T. Xu, Y. Mu, M. Hu, W. Shao, Y. Wang, G. Li, and L. Shao, “Emos: Embodiment-aware heterogeneous multi-robot operating system with llm agents,” arXiv preprint, arXiv:2410.22662, 2024. https://doi.org/10.48550/arXiv.2410.22662
  5. [5] K. Liu, Z. Tang, D. Wang, Z. Wang, X. Li, and B. Zhao, “Coherent: Collaboration of heterogeneous multi-robot system with large language models,” arXiv preprint, arXiv:2409.15146, 2024. https://doi.org/10.48550/arXiv.2409.15146
  6. [6] V. Strobel, M. Dorigo, and M. Fritz, “Llm2swarm: Robot swarms that responsively reason, plan, and collaborate through llms,” arXiv preprint, arXiv:2410.11387, 2024. https://doi.org/10.48550/arXiv.2410.11387
  7. [7] B. Yu, H. Kasaei, and M. Cao, “Co-navgpt: Multi-robot cooperative visual semantic navigation using large language models,” arXiv preprint, arXiv:2310.07937, 2023. https://doi.org/10.48550/arXiv.2310.07937
  8. [8] Y. Wang, R. Xiao, J. Y. L. Kasahara, R. Yajima, K. Nagatani, A. Yamashita, and H. Asama, “Dart-llm: Dependency-aware multi-robot task decomposition and execution using large language models,” arXiv preprint, arXiv:2411.09022, 2024. https://doi.org/10.48550/arXiv.2411.09022
  9. [9] S. S. Kannan, V. L. N. Venkatesh, and B.-C. Min, “Smart-llm: Smart multi-agent robot task planning using large language models,” arXiv preprint, arXiv:2309.10062, 2023. https://doi.org/10.48550/arXiv.2309.10062
  10. [10] Y. Lakhnati, M. Pascher, and J. Gerken, “Exploring a gpt-based large language model for variable autonomy in a VR-based human-robot teaming simulation,” Frontiers in Robotics and AI, Vol.11, Article No.1347538, 2024. https://doi.org/10.3389/frobt.2024.1347538
  11. [11] M. Ahn, M. G. Arenas, M. Bennice, N. Brown, C. Chan, B. David, A. Francis, G. Gonzalez, R. Hessmer, T. Jackson, N. J. Joshi, D. Lam, T.-W. E. Lee, A. Luong, S. Maddineni, H. Patel, J. Peralta, J. Quiambao, D. Reyes, R. M. J. Ruano, D. Sadigh, P. Sanketi, L. Takayama, P. Vodenski, and F. Xia, “Vader: Visual affordance detection and error recovery for multi robot human collaboration,” arXiv preprint, arXiv:2405.16021, 2024. https://doi.org/10.48550/arXiv.2405.16021
  12. [12] K. Garg, S. Zhang, J. Arkin, and C. Fan, “Foundation models to the rescue: Deadlock resolution in connected multi-robot systems,” arXiv preprint, arXiv:2404.06413, 2024. https://doi.org/10.48550/arXiv.2404.06413
  13. [13] S. Morad, A. Shankar, J. Blumenkamp, and A. Prorok, “Language-conditioned offline RL for multi-robot navigation,” arXiv preprint, arXiv:2407.20164, 2024. https://doi.org/10.48550/arXiv.2407.20164
  14. [14] T. Godfrey, W. Hunt, and M. D. Soorati, “Marlin: Multi-agent reinforcement learning guided by language-based inter-robot negotiation,” arXiv preprint, arXiv:2410.14383, 2025. https://doi.org/10.48550/arXiv.2410.14383
  15. [15] J. Orr and A. Dutta, “Multi-agent deep reinforcement learning for multi-robot applications: A survey,” Sensors, Vol.23, No.7, Article No.3625, 2023. https://doi.org/10.3390/s23073625
  16. [16] T. Niwa, K. Shibata, and T. Jimbo, “Multi-agent reinforcement learning and individuality analysis for cooperative transportation with obstacle removal,” F. Matsuno, S. Azuma, and M. Yamamoto (Eds.), “Distributed Autonomous Robotic Systems,” pp. 202-213, Springer, 2022. https://doi.org/10.1007/978-3-030-92790-5_16
  17. [17] G. Eoh and T.-H. Park, “Cooperative object transportation using curriculum-based deep reinforcement learning,” Sensors, Vol.21, No.14, Article No.4780, 2021. https://doi.org/10.3390/s21144780
  18. [18] L. Zhang, Y. Sun, A. Barth, and O. Ma, “Decentralized control of multi-robot system in cooperative object transportation using deep reinforcement learning,” IEEE Access, Vol.8, pp. 184109-184119, 2020. https://doi.org/10.1109/ACCESS.2020.3025287
  19. [19] Y. Sueoka, T. Yoshida, and K. Osuka, “Reinforcement learning of scalable, flexible and robust cooperative transport behavior using the transformer encoder,” A. Nilles, K. H. Petersen, T. L. Lam, A. Prorok, M. Rubenstein, and M. Otte (Eds.), “Distributed Autonomous Robotic Systems,” Springer, 2025. https://doi.org/10.1007/978-3-032-04584-3_20
  20. [20] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” 31st Conf. on Neural Information Processing Systems (NIPS 2017), 2017.
  21. [21] G. Serapio-García, M. Safdari, C. Crepy, L. Sun, S. Fitz, P. Romero, M. Abdulhai, A. Faust, and M. Matarić, “Personality traits in large language models,” arXiv preprint, arXiv:2307.00184, 2023. https://doi.org/10.48550/arXiv.2307.00184
  22. [22] C. Yu, A. Velu, E. Vinitsky, J. Gao, Y. Wang, A. Bayen, and Y. Wu, “The surprising effectiveness of PPO in cooperative, multi-agent games,” arXiv preprint, arXiv:2103.01955, 2021. https://doi.org/10.48550/arXiv.2103.01955
  23. [23] J. Schulman, P. Moritz, S. Levine, M. Jordan, and P. Abbeel, “High-dimensional continuous control using generalized advantage estimation,” arXiv preprint, arXiv:1506.02438, 2015. https://doi.org/10.48550/arXiv.1506.02438

*This site is desgined based on HTML5 and CSS3 for modern browsers, e.g. Chrome, Firefox, Safari, Edge, Opera.

Last updated on Dec. 19, 2025