Learning-based Multi-Agent Planning

We are interested in learning-based approaches for coordinating teams of agents in complex environments. Our focus includes combining various machine learning techniques with classical planning methods to achieve scalable and robust multi-agent coordination.

Research Directions

Bayesian Learning for Multi-Agent Systems

We explore Bayesian learning approaches to model uncertainty in multi-agent environments. This includes learning probabilistic models of agent behaviors, environmental dynamics, and interaction patterns to improve planning and coordination.

Reinforcement Learning for Multi-Agent Coordination

Reinforcement learning offers a promising approach for learning coordination policies directly from experience. We are interested in:

  • Multi-agent reinforcement learning (MARL): Learning policies for teams of agents
  • Combining RL with Large Neighborhood Search: Using learned policies to guide search algorithms
  • Imitation learning: Learning from expert demonstrations for lifelong multi-agent path finding
  • Policy learning for heterogeneous agent teams: Handling agents with different capabilities

Large Language Models for Multi-Agent Planning

Large Language Models (LLMs) have shown remarkable capabilities in reasoning and planning. We investigate:

  • LLM-guided task planning: Using LLMs to decompose complex multi-agent tasks
  • Natural language interfaces: Enabling human-AI collaboration in multi-agent scenarios
  • Emergent communication protocols: Learning communication strategies through LLM-based agents

Diffusion Models and Flow Matching

Generative models like diffusion models and flow matching offer new perspectives on multi-agent planning:

  • Generative path planning: Using diffusion models to generate diverse, high-quality paths
  • Flow matching for coordination: Modeling agent flows in continuous spaces
  • Combining with classical methods: Integrating generative models with search-based planning

Key Challenges

  • Scalability: Applying learning methods to large-scale multi-agent systems
  • Generalization: Ensuring learned policies work across different scenarios
  • Safety: Guaranteeing safe coordination when using learned policies
  • Interpretability: Understanding and explaining learned coordination strategies

Back to Research page