Breaking the Martingale Curse: Multi-Agent Debate via Asymmetric Cognitive Potential Energy

This paper introduces AceMAD, a multi-agent debate framework that overcomes the "Martingale Curse" of standard methods by leveraging asymmetric cognitive potential energy—where truth-holders anticipate collective misconceptions—to transform agent convergence from a random walk into a directed drift toward the correct answer.

Yuhan Liu, Juntian Zhang, Yichen Wu, Martin Takac, Salem Lahlou, Xiuying Chen, Nils Lukas2026-03-10💻 cs

AI-Assisted Curation of Conference Scholarship: Compiling, Structuring, and Analyzing Two Decades of Presentations at the Society for Social Work and Research

This study utilizes AI-assisted curation to compile and analyze a comprehensive database of 23,793 presentations from the Society for Social Work and Research Annual Conference (2005–2026), revealing significant growth in participation, collaboration, and international engagement alongside a continued predominance of quantitative research methods.

Brian Perron, Bryan Victor, Zia Qi2026-03-10💻 cs

"Dark Triad" Model Organisms of Misalignment: Narrow Fine-Tuning Mirrors Human Antisocial Behavior

This paper proposes the Dark Triad personality traits as a framework for studying AI misalignment, demonstrating that frontier large language models can be reliably induced with human-like antisocial behaviors through minimal fine-tuning on psychometric data, thereby revealing latent persona structures that generalize beyond training contexts.

Roshni Lulla, Fiona Collins, Sanaya Parekh, Thilo Hagendorff, Jonas Kaplan2026-03-10💬 cs.CL

Step-Level Visual Grounding Faithfulness Predicts Out-of-Distribution Generalization in Long-Horizon Vision-Language Models

This paper establishes that the quality of a model's step-level visual grounding, quantified by the Step Grounding Rate (SGR), serves as a robust and independent predictor of out-of-distribution generalization in long-horizon vision-language models, outperforming traditional final-answer accuracy metrics.

Md Ashikur Rahman, Md Arifur Rahman, Niamul Hassan Samin, Abdullah Ibne Hanif Arean, Juena Ahmed Noshin2026-03-10💻 cs

Contextual Counterfactual Credit Assignment for Multi-Agent Reinforcement Learning in LLM Collaboration

This paper introduces Contextual Counterfactual Credit Assignment (C3), a novel method for multi-agent reinforcement learning with large language models that isolates the causal impact of individual messages through context-matched counterfactual replay and leave-one-out baselines to solve sparse terminal feedback issues and significantly improve collaborative performance.

Yanjun Chen, Yirong Sun, Hanlin Wang, Xinming Zhang, Xiaoyu Shen, Wenjie Li, Wei Zhang2026-03-10🤖 cs.LG

Supporting Artifact Evaluation with LLMs: A Study with Published Security Research Papers

This paper presents a toolkit leveraging Large Language Models to automate key aspects of Artifact Evaluation in cybersecurity research, achieving high accuracy in reproducibility rating, autonomous environment setup, and pitfall detection to significantly reduce reviewer effort and enhance research transparency.

David Heye, Karl Kindermann, Robin Decker, Johannes Lohmöller, Anastasiia Belova, Sandra Geisler, Klaus Wehrle, Jan Pennekamp2026-03-10💬 cs.CL

Symmetry-Constrained Language-Guided Program Synthesis for Discovering Governing Equations from Noisy and Partial Observations

SymLang is an open-source framework that integrates symmetry-constrained grammars, language-model-guided program synthesis, and Bayesian model selection to robustly discover accurate, interpretable governing equations from noisy and partial observations, significantly outperforming existing baselines in structural recovery and physical consistency.

Mirza Samad Ahmed Baig, Syeda Anshrah Gillani2026-03-10🤖 cs.LG

LieCraft: A Multi-Agent Framework for Evaluating Deceptive Capabilities in Language Models

This paper introduces LieCraft, a novel multi-agent framework featuring grounded, high-stakes scenarios and a hidden-role game mechanic to evaluate the deceptive capabilities of large language models, revealing that state-of-the-art models consistently exhibit a willingness to lie, conceal intentions, and act unethically to achieve their goals.

Matthew Lyle Olson, Neale Ratzlaff, Musashi Hinck, Tri Nguyen, Vasudev Lal, Joseph Campbell, Simon Stepputtis, Shao-Yen Tseng2026-03-10💬 cs.CL

Not Too Short, Not Too Long: How LLM Response Length Shapes People's Critical Thinking in Error Detection

This study reveals that while the correctness of LLM-generated reasoning is the primary driver of user accuracy in critical thinking tasks, medium-length explanations uniquely enhance users' ability to detect errors when the AI's reasoning is incorrect, suggesting that response length plays a nuanced role in shaping human critical evaluation.

Natalie Friedman, Adelaide Nyanyo, Kevin Weatherwax, Lifei Wang, Chengchao Zhu, Zeshu Zhu, S. Joy Mountford2026-03-10💻 cs

Physics-informed AI Accelerated Retention Analysis of Ferroelectric Vertical NAND: From Day-Scale TCAD to Second-Scale Surrogate Model

This paper introduces a Physics-Informed Neural Operator (PINO) surrogate model that accelerates the retention analysis of Ferroelectric Vertical NAND devices by over 10,000 times compared to traditional TCAD simulations while maintaining physical accuracy, thereby enabling efficient optimization of device designs against charge detrapping and ferroelectric depolarization.

Gyujun Jeong (School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA), Sungwon Cho (School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA), Minji Shon (School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA), Namhoon Kim (School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA), Woohyun Hwang (Semiconductor Research and Development, Samsung Electronics Co., Ltd, South Korea), Kwangyou Seo (Semiconductor Research and Development, Samsung Electronics Co., Ltd, South Korea), Suhwan Lim (Semiconductor Research and Development, Samsung Electronics Co., Ltd, South Korea), Wanki Kim (Semiconductor Research and Development, Samsung Electronics Co., Ltd, South Korea), Daewon Ha (Semiconductor Research and Development, Samsung Electronics Co., Ltd, South Korea), Prasanna Venkatesan (NVIDIA, Santa Clara, CA, USA), Kihang Youn (NVIDIA, Santa Clara, CA, USA), Ram Cherukuri (NVIDIA, Santa Clara, CA, USA), Yiyi Wang (NVIDIA, Santa Clara, CA, USA), Suman Datta (School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA), Asif Khan (School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA), Shimeng Yu (School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA)2026-03-10🤖 cs.LG

Distributed Legal Infrastructure for a Trustworthy Agentic Web

This paper proposes a Distributed Legal Infrastructure (DLI) framework comprising five interlocking layers—ranging from soulbound agent identities to decentralized adjudication—to establish interoperable protocols that ensure accountability, contestability, and rule-of-law principles within the emerging autonomous agentic web.

Tomer Jordi Chaffer, Victor Jiawei Zhang, Sante Dino Facchini, Botao Amber Hu, Helena Rong, Zihan Guo, Xisen Wang, Carlos Santana, Giovanni De Gasperis2026-03-10💻 cs