Motivation. The scaling of model parameters has unlocked the groundbreaking capabilities of foundation models. Likewise, in human society, scaling and collaboration across individuals, organizations, companies, and nations amplify collective intelligence to unprecedented levels, enabling remarkable achievements that would be impossible for individuals alone, such as space exploration and autonomy. Could this principle of scaling also apply to the growth in the number of agents? Multi-agent systems may offer a promising path forward. By progressively integrating more agents, multi-agent systems can activate diverse functionalities within these foundation model-powered generalist agents and coordinate a broader range of complementary functionalities. This synergy fosters improved problem-solving, adaptability, and decision-making capabilities. As the multi-agent system scales, it has a huge potential to achieve enhanced capabilities and tackle increasingly complex tasks, offering a promising solution toward the ultimate goal of achieving artificial general intelligence (AGI).

Background. With the advent of large foundation models, including large language models (LLMs) and visual language models (VLMs), research in multi-agent systems has progressed from relying on specialized models to harnessing the versatile capabilities of more generalized LLM/VLM-powered agents. As a community, we are embracing this paradigm shift to develop general-purpose multi-agent systems. These systems aim to leverage collaboration to enhance individual agent capabilities, ultimately advancing toward sophisticated AI agents that can serve as versatile human assistants. This field has already seen numerous promising developments. For instance, multi-agent systems for social simulation are used to generate diverse synthetic data, further enhancing the capabilities of foundation models and also investigating social behaviors. Moreover, multi-agent systems emulate human standards of processing through agentic workflows, significantly advancing a wide range of complex and urgent real-world applications, including mathematics, software development, web and mobile operations, embodied manipulation, and navigation, among others. These developments underscore the vast potential of multi-agent systems powered by foundation models to tackle increasingly complex challenges across diverse domains.

Our Workshop. This workshop seeks to bring together researchers from diverse disciplines to explore emerging topics in multi-agent systems powered by foundation models (LLMs, VLMs, and MMLMs). By complementing existing workshops focused on individual intelligence, it introduces a new perspective centered on advancing agent capabilities through multi-agent collaboration and their integration into human life. The primary objective of this workshop is to foster meaningful discussions and collaborations that address critical scientific questions surrounding the development and application of such systems.


Call for papers

This workshop aims to deepen understanding and offer innovative perspectives on the growth and collaboration of LLM/VLM-powered agents in multi-agent contexts. To foster an inclusive environment for discussion and debate, we welcome speakers and panelists from diverse backgrounds and expertise. Our lineup features distinguished researchers alongside emerging investigators who have made significant contributions to the field. Spotlight and poster sessions will highlight new ideas, key challenges, and retrospective insights related to the workshop’s themes. We strive for our participant selection to reflect the dynamic and diverse landscape of machine learning and AI. We will explore a range of topics in this workshop, including, but not limited to, the following areas:

Multi-Agent Simulation:

Simulations in society, games, psychology, economics, and politics.

Multi-Agent Datasets and Benchmarks:

The creation of datasets and benchmarks to inspire innovation and provide more effective evaluation of the enhanced capabilities of multi-agent systems.

Multi-Agent Orchestration and Efficiency:

Multi-agent workflows, collaboration graphs, and communication protocols.

Human-Agent Collaboration:

Exploring interactions and synergies between humans and agents to foster the development of more trustworthy and socially friendly agents.

Multi-Agent Applications:

Math, software development, open question answering, web agents, os agents, mobile agents, embodied agents for navigation, exploration, manipulation, autonomous systems, robotics, medical agents.

Reinforcement Learning Methods for Multi-Agent Systems:

Advance the multi-agent system from the interactive environment feedback.

Symbolic Learning Methods for Multi-Agent Systems:

Theoretical framework for multi-agent systems.

Submission Guide

Submission Platform: Submit your papers here: OpenReview Submission Site

Submission Requirements:
Use the official LaTeX template of ICML 2025 (Style Files)
Papers must be prepared and submitted as a single PDF: 8 pages for the main paper, with unlimited pages for references and appendices (reviewers are not obliged to read the appendices)
All submissions must be anonymized, which will be reviewed in a double-blind manner

Non-Archival Policy: Submissions will not be indexed or have archival proceedings. We welcome NeurIPS 2025 submissions.

Best Paper Award: Best paper award will be announced at the workshop.

Key Dates
📄 Paper Submission Open: Apr 20, 2025 🐰Happy Easter!
📄 Paper Submission Deadline: May 26, 2025
📄 Acceptance Notification: June 9, 2025
📄 Camera-Ready Deadline: June 30, 2025
📄 Workshop Date: July 18, 2025

Workshop Schedule

Time Session Duration Details
9:00AM - 9:10AM Opening Remarks 10 min Welcome and Introduction to the Workshop
9:10AM - 9:30AM Invited Talk 1 20 min Talk1 - Yejin Choi
9:40AM - 10:00AM Invited Talk 2 20 min Talk2 - Natasha Jaques
10:10AM - 10:40AM Oral Presentations 1 30 min 15 min * 2
10:45AM - 11:45AM Poster Session & Coffee Socials 1 60 min Poster Session & Coffee Socials1
11:45AM - 1:00PM Lunch Break 75 min Time for lunch and informal discussions
1:00PM - 1:20PM Invited Talk 3 20 min Talk3 - Yilun Du
1:30PM - 1:40PM Invited Talk 4 10 min Talk4 - Diyi Yang
1:50PM - 2:20PM Oral Presentations 2 30 min Oral Presentations2
2:25PM - 3:25PM Poster Session & Coffee Socials 2 60 min Networking and refreshments
3:30PM - 3:50PM Invited Talk 5 20 min Talk5 - Mengdi Wang
4:00PM - 4:20PM Invited Talk 6 20 min Talk6 - Yulia Tsvetkov
4:30PM - 5:00PM Panel Discussion 30 min Interactive session with experts - Joon Sung Park, Mingchen Zhuge, Chen Qian
5:00PM - 5:15PM Awards and Conclusive Remarks 15 min Concluding the workshop and award announcements

Challenge

Multi-Agent Embodied Intelligence Challenge @ ICML 2025

Advancing Embodied AI beyond single-agent settings, the Multi-Agent Embodied Intelligence Challenge, hosted at ICML 2025, invites the community to explore intelligent coordination in physically grounded environments. This challenge focuses on multi-agent manipulation via imitation learning, underpinned by compositional reasoning and scalable learning architectures.

Hosted by the MARS-EAI initiative, which focuses on multi-agent embodied AI, this competition is built upon RoboFactory -- a novel simulation platform and benchmark tailored for collaborative robotic manipulation. RoboFactory introduces a constraint-driven framework where agent behavior is guided by logical, spatial, and temporal constraints. These compositional constraints are critical for enabling safe, efficient, and interpretable collaboration among embodied agents in shared spaces.

Participants will train policies in a suite of challenging scenarios, ranging from dual-arm stacking to long-horizon cooperative tasks involving up to four agents. The core challenge lies in learning coordination policies that can handle uncertainty, leverage egocentric and global observations, and scale with the number of interacting agents.

Key highlights:

  • Compositional Imitation Learning: Participants must generate structured, constraint-aware behavior from demonstrations.
  • Benchmark Integration: Built atop ManiSkill3 and aligned with real-world application, RoboFactory enables rich, diverse task settings.
  • Evaluation Metrics: Performance will be assessed on task success, data efficiency, constraint compliance, and generalization under scene randomization.

We invite researchers from robotics, imitation learning, multi-agent RL, and foundation model communities to participate. Top teams will be recognized at ICML 2025 and contribute to a shared vision of scalable, collaborative embodied intelligence.

Speakers and panelists

Yejin Choi

Senior Director at NVIDIA, Professor at Stanford

Natasha Jaques

Assistant Professor, University of Washington, Senior Research Scientist at Google DeepMind

Yilun Du

Senior Research Scientist at Google Deepmind, Assistant Professor at Harvard

Diyi Yang

Assistant Professor at Stanford

Mengdi Wang

Associate Professor at Princeton University

Yulia Tsvetkov

Associate Professor at University of Washington

Joon Sung Park

PhD Student at Stanford

Mingchen Zhuge

PhD Candidate at KAUST

Chen Qian

Associate Professor at Shanghai Jiao Tong University

Workshop Organizers

Zhenfei Yin

PhD at USYD, Visiting Researcher at Oxford

Yue Hu

Postdoctoral Fellow at University of Michigan

Siheng Chen

Associate Professor at Shanghai Jiao Tong University

Bernadette Bucher

Assistant Professor at University of Michigan

Rui Ye

PhD Candidate at Shanghai Jiao Tong University

Chahyon Ku

PhD Student at University of Michigan

Katrina Ashton

PhD Candidate at University of Pennsylvania

Juan Carlos Niebles

Research Director at Salesforce AI Research, Adjunct Professor at Stanford

Lei Bai

Research Scientist at Shanghai AI Laboratory

Dawn Song

Professor at University of California, Berkeley

Philip Torr

Professor at University of Oxford

Contact us

Email us at masworks2025@gmail.com | zhenfei.yin@sydney.edu.au | huyu@umich.edu
This is also an event hosted by [MASWorks Community]