Scale-Plan: Scalable Language-Enabled Task Planning for Heterogeneous Multi-Robot Teams

Scale-Plan is a scalable framework that leverages large language models to filter irrelevant perceptual information and construct compact, task-relevant representations from natural language instructions, thereby enabling efficient and reliable long-horizon planning for heterogeneous multi-robot teams while outperforming existing baselines on the new MAT2-THOR benchmark.

Piyush Gupta, Sangjae Bae, Jiachen Li, David IseleWed, 11 Ma🤖 cs.AI

NarrativeLoom: Enhancing Creative Storytelling through Multi-Persona Collaborative Improvisation

The paper introduces NarrativeLoom, a multi-persona collaborative storytelling system grounded in Campbell's theory of blind variation and selective retention, which a controlled study of 50 participants shows significantly enhances the novelty, diversity, and overall creativity of co-authored stories compared to existing tools, particularly benefiting novice writers through structured scaffolding.

Yuxi Ma, Yongqian Peng, Fengyuan Yang, Siyu Zha, Chi Zhang, Zixia Jia, Zilong Zheng, Yixin ZhuTue, 10 Ma💻 cs

Randomise Alone, Reach as a Team

This paper investigates concurrent graph games with distributed randomization where team players lack a shared random source, establishing that memoryless strategies suffice for the threshold problem (placing it in R\exists\mathbb{R} and proving NP-hardness) and that almost-sure reachability is NP-complete, while introducing the IRATL logic and a corresponding solver.

Léonard Brice, Thomas A. Henzinger, Alipasha Montaseri, Ali Shafiee, K. S. ThejaswiniTue, 10 Ma💻 cs

MAS-Orchestra: Understanding and Improving Multi-Agent Reasoning Through Holistic Orchestration and Controlled Benchmarks

This paper introduces MAS-Orchestra, a training-time framework that optimizes multi-agent system orchestration via function-calling reinforcement learning, alongside the MASBENCH benchmark, to demonstrate that multi-agent benefits are task-dependent and to achieve significant performance gains with over 10x efficiency on complex reasoning tasks.

Zixuan Ke, Yifei Ming, Austin Xu, Ryan Chin, Xuan-Phi Nguyen, Prathyusha Jwalapuram, Jiayu Wang, Semih Yavuz, Caiming Xiong, Shafiq JotyTue, 10 Ma💬 cs.CL

FOR-Prompting: From Objection to Revision via an Asymmetric Prompting Protocol

The paper introduces FOR-Prompting, a model-agnostic, asymmetric prompting protocol that enhances reasoning and iterative refinement across diverse tasks by structuring interactions between a Defender, a Questioner, and an optional Host, enabling even small models to achieve performance comparable to or better than standard baselines without requiring training or access to model internals.

He Zhang, Anzhou Zhang, Jian DaiTue, 10 Ma💬 cs.CL

Characterizing MARL for Energy Control: A Multi-KPI Benchmark on the CityLearn Environment

This paper establishes a comprehensive multi-KPI benchmark for Multi-Agent Reinforcement Learning in urban energy management using the CityLearn environment, demonstrating that Decentralized Training with Decentralized Execution (DTDE) consistently outperforms Centralized Training with Decentralized Execution (CTDE) in both average and worst-case performance while offering greater resilience and sustainability.

Aymen Khouja, Imen Jendoubi, Oumayma Mahjoub, Oussama Mahfoudhi, Ruan De Kock, Siddarth Singh, Claude FormanekTue, 10 Ma🤖 cs.LG

LatentMem: Customizing Latent Memory for Multi-Agent Systems

This paper introduces LatentMem, a learnable multi-agent memory framework that addresses memory homogenization and information overload by using an experience bank and a memory composer to generate customized, token-efficient latent memories, further optimized via Latent Memory Policy Optimization (LMPO) to significantly enhance multi-agent system performance.

Muxin Fu, Xiangyuan Xue, Yafu Li, Zefeng He, Siyuan Huang, Xiaoye Qu, Yu Cheng, Yang YangTue, 10 Ma🤖 cs.LG

Behavioral Inference at Scale: The Fundamental Asymmetry Between Motivations and Belief Systems

Through large-scale experiments with over 1.5 million LLM-generated behavioral sequences, this paper reveals a fundamental asymmetry in behavioral inference where agent motivations are nearly perfectly recoverable while belief systems remain largely opaque due to inherent information-theoretic limits and architectural constraints, particularly within a "neutral zone" of behavioral ambiguity.

Jason Starace, Terence SouleTue, 10 Ma🤖 cs.LG