HydraMARL
A foundational study of parameter sharing in cooperative multi-agent reinforcement learning. Existing methods entangle sharing with auxiliary machinery (clustering, adapters, diversity bonuses), confounding what's actually driving performance. HydraMARL disentangles them on a minimal shared-backbone + per-agent-head architecture, and gives the first principled characterization of the speed-vs-specialization spectrum in cooperative MARL.
When many AI agents learn together, parts of their neural networks can be shared. We isolate how much sharing actually helps, and expose the real trade-off between learning fast and learning to specialize.