Research
All intelligence is collective intelligence.
Our aim is to extend single robot capabilities through collaborative strategies. We develop data-driven approaches to hard coordination problems — from teaching agents how to communicate and cooperate, to co-designing the environments they work in — and transfer these policies from simulation to real robots for fast, online decision-making.
Themes
Agile, dense aerial swarms
Dense, agile swarm flight with downwash control, cooperative docking, scalable kinodynamic planning, and outdoor deployment.
Co-design of agents and environments
Making robots smarter by also redesigning the spaces they work in — discovering that adding obstacles and embedding sensors can dramatically improve coordination.
Compositionality in multi-robot AI
Moving beyond monolithic models toward modular, specialized agents that combine at runtime for flexible and scalable multi-robot intelligence.
Decentralized coordination
Scaling multi-agent coordination from two robots to hundreds — without central planners, and transferring those policies from simulation onto physical teams.
Learned communication
Differentiable inter-agent communication — letting robots learn what to share with their teammates directly from the downstream task.
MARL and GNN foundations
Building the algorithmic foundations for multi-agent learning — graph neural network architectures, scalable MARL algorithms, and principled tools for controlling team diversity.
Platforms and tools
Open-source simulators, benchmarking libraries, and custom hardware platforms — VMAS, BenchMARL, RoboMaster, Raven, and Sanity.
Robot-task co-optimization
Making robots smarter by also redesigning the spaces they work in — discovering that adding obstacles and embedding sensors into the environment can dramatically improve coordination.
Lab videos
Demos, paper presentations, and experiment recordings from our research.
Papers
2026
- Mar 16PAPER npj RoboticsConcrete multi-agent path planning enabling kinodynamically aggressive maneuvers
Keisuke Okumura, Guang Yang, Zhan Gao, Heedo Woo, Amanda Prorok
We present concrete planning, a hybrid approach that captures real-world continuous dynamics while maintaining scalable guaranteed planning via discrete search. The framework integrates advances in robot dynamics learning, optimal control, and anytime complete planning into a modular system deployed with 40 robots — 20 aerial, 8 ground, and 12 obstacle robots — operating in a compact laboratory space.
- Jan 20PAPER ICLR 2026When Is Diversity Rewarded in Cooperative Multi-Agent Learning?
Michael Amir, Matteo Bettini, Amanda Prorok
An investigation of when and why behavioral diversity benefits cooperative multi-agent learning, establishing conditions under which heterogeneous policies outperform homogeneous ones.
- Jan 20PAPER ICLR 2026Pairwise is Not Enough: Hypergraph Neural Networks for Multi-Agent Pathfinding
Rishabh Jain, Keisuke Okumura, Michael Amir, Pietro Liò, Amanda Prorok
Pairwise interactions captured by standard GNNs miss higher-order dependencies in multi-agent pathfinding. This work introduces hypergraph neural networks that model group-level constraints, improving solution quality for dense navigation scenarios.
- Jan 20PAPER ICLR 2026Remotely Detectable Robot Policy Watermarking
Michael Amir, Manon Flageat, Amanda Prorok
A method for embedding remotely detectable watermarks into robot policies, enabling verification of policy provenance without requiring access to the model parameters.
- Nov 2025PAPER AAAI 2026Graph Attention-Guided Search for Dense Multi-Agent Pathfinding
Rishabh Jain, Keisuke Okumura, Michael Amir, Amanda Prorok
Historically, learning-based approaches to multi-agent pathfinding have struggled to outperform classical search-based methods. In this work, we introduce a hybrid approach that combines GNNs with a classical search algorithm, to achieve superior performance to both purely learning-based and classical methods.
2025
- Dec 2025PAPER NeurIPS 2025No-Regret Thompson Sampling for Finite-Horizon Markov Decision Processes with Gaussian Processes
Jasmine Bayrooti, Sattar Vakili, Amanda Prorok, Carl Henrik Ek
A no-regret Thompson sampling algorithm for finite-horizon Markov decision processes using Gaussian process models, providing efficient exploration in model-based reinforcement learning settings.
- Sep 2025PAPER Science RoboticsExtending robot minds through collective learning
Amanda Prorok
The current trend toward generalist robot behaviors powered by massive, monolithic AI models is unsustainable. This viewpoint article argues for a paradigm shift toward collective robotic intelligence — a "mixture-of-robots" approach where diverse, specialized agents learn and work together, inspired by natural systems where cooperation among specialized components leads to greater intelligence and resilience.
- Aug 2025PAPER CoRL 2025ReCoDe: Reinforcement Learning-based Dynamic Constraint Design for Multi-Agent Coordination
Michael Amir, Guang Yang, Zhan Gao, Keisuke Okumura, Heedo Woo, Amanda Prorok
A hybrid, decentralized framework combining optimization-based control with adaptation provided by MARL. Rather than discarding expert controllers, ReCoDe improves them by learning additional dynamic constraints that capture subtler behaviors — for example, constraining agent movements to prevent congestion in cluttered scenarios.
- Jul 2025PAPER IROS 2025D4orm: Multi-Robot Trajectories with Dynamics-aware Diffusion Denoised Deformations
Yixiao Zhang, Keisuke Okumura, Heedo Woo, Ajay Shankar, Amanda Prorok
An optimization method for generating kinodynamically feasible and collision-free multi-robot trajectories that exploits an incremental denoising scheme from diffusion models. Evaluated for differential-drive and holonomic teams with up to 16 robots in 2D and 3D worlds.
- May 2025PAPER ICRA 2025DVM-SLAM: Decentralized Visual Monocular Simultaneous Localization and Mapping for Multi-Agent Systems
Joshua Bird, Jan Blumenkamp, Amanda Prorok
A decentralized visual monocular SLAM system for multi-agent systems, enabling collaborative mapping and localization without centralized infrastructure.
- May 2025PAPER ICRA 2025Language-Conditioned Offline RL for Multi-Robot Navigation
Steven Morad, Ajay Shankar, Jan Blumenkamp, Amanda Prorok
Offline reinforcement learning approach conditioned on natural language commands for decentralized multi-robot navigation, bridging high-level human instructions and low-level robot coordination policies.
- Apr 2025PAPER ICLR 2025Efficient Model-Based Reinforcement Learning Through Optimistic Thompson Sampling
Jasmine Bayrooti, Carl Henrik Ek, Amanda Prorok
An efficient model-based reinforcement learning method using optimistic Thompson sampling to balance exploration and exploitation in complex sequential decision-making problems.
- Mar 2025PAPER IEEE Transactions on RoboticsCo-Optimizing Reconfigurable Environments and Policies for Decentralized Multi-Agent Navigation
Zhan Gao, Guang Yang, Amanda Prorok
Treats the environment as a co-decision variable alongside agent policies, jointly optimizing both the physical layout and the decentralized navigation policies for multi-agent systems.
2024
- Dec 2024PAPER JMLRBenchMARL: Benchmarking Multi-Agent Reinforcement Learning
Matteo Bettini, Amanda Prorok, Vincent Moens
A standardized library for benchmarking multi-agent reinforcement learning, enabling seamless mixing and matching of MARL algorithms, tasks, and models while maintaining rigorous reproducibility and standardization.
- Nov 2024PAPER CoRL 2024Provably Safe Online Multi-Agent Navigation in Unknown Environments
Zhan Gao, Guang Yang, Jasmine Bayrooti, Amanda Prorok
A provably safe online method for multi-agent navigation in unknown environments, combining control barrier functions with decentralized planning to guarantee collision avoidance while maintaining liveness.
- Oct 2024PAPER CoRL 2024CoViS-Net: A Cooperative Visual Spatial Foundation Model for Multi-Robot Applications
Jan Blumenkamp, Steven Morad, Jennifer Gielis, Amanda Prorok
A decentralized visual spatial foundation model enabling real-time, platform-agnostic pose estimation and spatial comprehension for autonomous robots. Provides accurate pose estimates and local bird's-eye-view without needing camera overlap, deployed on wheeled platforms and a quadruped.
- Jul 2024PAPER ICML 2024Controlling Behavioral Diversity in Multi-Agent Reinforcement Learning
Matteo Bettini, Ryan Kortvelesy, Amanda Prorok
DiCo (Diversity Control) controls behavioral diversity to an exact value of a given metric by representing policies as the sum of a parameter-shared component and dynamically scaled per-agent components. Applied directly to the policy architecture, leaving the learning objective unchanged.
- May 2024PAPER DARS 2024The Cambridge RoboMaster: An Agile Multi-Robot Research Platform
Jan Blumenkamp, Steven Morad, Jennifer Gielis, Amanda Prorok
An agile multi-robot research platform built on DJI RoboMaster hardware, designed for real-world experiments in decentralized multi-agent coordination, supporting GNN-based policy deployment and sim-to-real transfer.
2023
- Dec 2023PAPER NeurIPS 2023Generalised f-Mean Aggregation for Graph Neural Networks
Ryan Kortvelesy, Steven Morad, Amanda Prorok
A generalized f-mean aggregation framework for graph neural networks that subsumes common aggregation functions (sum, mean, max) as special cases while enabling learnable, task-adaptive aggregation strategies.
- Dec 2023PAPER NeurIPS 2023Reinforcement Learning with Fast and Forgetful Memory
Steven Morad, Ryan Kortvelesy, Stephan Liwicki, Amanda Prorok
A new memory model for RL that serves as a super-efficient plug-in replacement for RNNs and transformers. Runs up to two orders of magnitude faster with linear space complexity, and sets new records on the POPGym benchmark for partially observable RL.
- May 2023PAPER AAMAS 2023Heterogeneous multi-robot reinforcement learning
Matteo Bettini, Ajay Shankar, Amanda Prorok
We study cooperative multi-robot tasks where the team is composed of agents with structurally different observations, action spaces, and reward functions. We introduce HetGPPO, a parameter-sharing paradigm enabling heterogeneous behaviors in multi-agent reinforcement learning with graph neural networks, achieving superior performance over role-blind baselines.
- May 2023PAPER ICLR 2023POPGym: Benchmarking Partially Observable Reinforcement Learning
Steven Morad, Ryan Kortvelesy, Matteo Bettini, Stephan Liwicki, Amanda Prorok
A comprehensive benchmark suite for partially observable reinforcement learning, providing a unified evaluation framework across diverse POMDP environments to compare memory-based RL architectures.
This page highlights recent work. For the complete record going back to 2008: