Improving Cooperation in Collaborative Embodied AI
By: Hima Jacob Leven Suprabha , Laxmi Nag Laxminarayan Nagesh , Ajith Nair and more
Potential Business Impact:
AI agents work together better using smart instructions.
The integration of Large Language Models (LLMs) into multiagent systems has opened new possibilities for collaborative reasoning and cooperation with AI agents. This paper explores different prompting methods and evaluates their effectiveness in enhancing agent collaborative behaviour and decision-making. We enhance CoELA, a framework designed for building Collaborative Embodied Agents that leverage LLMs for multi-agent communication, reasoning, and task coordination in shared virtual spaces. Through systematic experimentation, we examine different LLMs and prompt engineering strategies to identify optimised combinations that maximise collaboration performance. Furthermore, we extend our research by integrating speech capabilities, enabling seamless collaborative voice-based interactions. Our findings highlight the effectiveness of prompt optimisation in enhancing collaborative agent performance; for example, our best combination improved the efficiency of the system running with Gemma3 by 22% compared to the original CoELA system. In addition, the speech integration provides a more engaging user interface for iterative system development and demonstrations.
Similar Papers
Cross-Lingual Prompt Steerability: Towards Accurate and Robust LLM Behavior across Languages
Computation and Language
Makes AI understand and work in many languages.
CoLa: Learning to Interactively Collaborate with Large Language Models
Computation and Language
AI learns to guide other AI to solve problems.
The Future of MLLM Prompting is Adaptive: A Comprehensive Experimental Evaluation of Prompt Engineering Methods for Robust Multimodal Performance
Artificial Intelligence
Teaches AI to understand pictures and words better.