Safe RLHF Beyond Expectation: Stochastic Dominance for Universal Spectral Risk Control

This paper proposes Risk-sensitive Alignment via Dominance (RAD), a novel Safe RLHF framework that replaces traditional expected cost constraints with First-Order Stochastic Dominance constraints within an Optimal Transport framework to universally control spectral risk measures, thereby achieving superior robustness against tail risks and out-of-distribution failures while maintaining helpfulness.

Yaswanth Chittepu, Ativ Joshi, Rajarshi Bhattacharjee, Scott Niekum2026-03-12🤖 cs.LG

Contact Coverage-Guided Exploration for General-Purpose Dexterous Manipulation

This paper proposes Contact Coverage-Guided Exploration (CCGE), a general-purpose exploration method that leverages contact state counters and energy-based rewards to guide dexterous hands in discovering diverse contact patterns, thereby significantly improving training efficiency and real-world transferability across complex manipulation tasks.

Zixuan Liu, Ruoyi Qiao, Chenrui Tie, Xuanwei Liu, Yunfan Lou, Chongkai Gao, Zhixuan Xu, Lin Shao2026-03-12🤖 cs.AI

GroundCount: Grounding Vision-Language Models with Object Detection for Mitigating Counting Hallucinations

GroundCount proposes a framework that augments Vision-Language Models with explicit spatial grounding from object detection models to significantly mitigate counting hallucinations, demonstrating that structured prompt-based integration outperforms feature-level fusion and yields consistent accuracy improvements across most architectures.

Boyuan Chen, Minghao Shao, Siddharth Garg, Ramesh Karri, Muhammad Shafique2026-03-12🤖 cs.AI

RCTs & Human Uplift Studies: Methodological Challenges and Practical Solutions for Frontier AI Evaluation

This paper synthesizes findings from interviews with 16 experts to identify methodological challenges in applying randomized controlled trials to evaluate frontier AI's impact on human performance and proposes practical solutions to address validity issues in high-stakes decision-making.

Patricia Paskov, Kevin Wei, Shen Zhou Hong, Dan Bateyko, Xavier Roberts-Gaal, Carson Ezell, Gailius Praninskas, Valerie Chen, Umang Bhatt, Ella Guest2026-03-12🤖 cs.AI

Does AI See like Art Historians? Interpreting How Vision Language Models Recognize Artistic Style

Through an interdisciplinary collaboration between computer scientists and art historians, this paper employs latent-space decomposition and quantitative analysis to reveal that Vision Language Models predict artistic styles using concepts that are largely coherent and relevant to human experts, often aligning with art historical reasoning even when utilizing formally interpreted features.

Marvin Limpijankit, Milad Alshomary, Yassin Oulad Daoud, Amith Ananthram, Tim Trombley, Elias Stengel-Eskin, Mohit Bansal, Noam M. Elcott, Kathleen McKeown2026-03-12🤖 cs.AI

Neural Field Thermal Tomography: A Differentiable Physics Framework for Non-Destructive Evaluation

The paper introduces Neural Field Thermal Tomography (NeFTY), a differentiable physics framework that parameterizes 3D material diffusivity as a continuous neural field optimized via a rigorous numerical solver to achieve high-resolution, quantitative reconstruction of subsurface defects from transient surface temperature measurements, overcoming the limitations of traditional 1D approximations and soft-constrained PINNs.

Tao Zhong, Yixun Hu, Dongzhe Zheng, Aditya Sood, Christine Allen-Blanchette2026-03-12🔬 cond-mat.mtrl-sci

Enhancing Heterogeneous Multi-Agent Cooperation in Decentralized MARL via GNN-driven Intrinsic Rewards

This paper proposes CoHet, a novel algorithm that leverages Graph Neural Network-driven intrinsic rewards to enable effective decentralized learning and cooperation among heterogeneous multi-agent systems despite challenges like partial observability and reward sparsity, demonstrating superior performance over state-of-the-art methods in standard benchmarks.

Jahir Sadik Monon, Deeparghya Dutta Barua, Md. Mosaddek Khan2026-03-11🤖 cs.AI

Sparse Variational Student-t Processes for Heavy-tailed Modeling

This paper introduces Sparse Variational Student-t Processes (SVTP), a scalable framework that extends sparse inducing point methods to Student-t processes via novel inference algorithms and natural gradient optimization, achieving superior robustness to outliers and heavy-tailed data with significantly faster convergence and lower prediction error compared to sparse Gaussian processes on large datasets.

Jian Xu, Delu Zeng, John Paisley2026-03-11🤖 cs.AI

Robust Training of Neural Networks at Arbitrary Precision and Sparsity

This paper introduces a unified framework that models quantization and sparsification as additive noise to derive a principled, noise-corrective gradient path, enabling the stable training of neural networks at arbitrary low precisions and sparsity levels without relying on heuristic estimators like the Straight-Through Estimator.

Chengxi Ye, Grace Chu, Yanfeng Liu, Yichi Zhang, Lukasz Lew, Li Zhang, Mark Sandler, Andrew Howard2026-03-11🤖 cs.AI