Dual-Robust Cross-Domain Offline Reinforcement Learning Against Dynamics Shifts

This paper introduces DROCO, a novel dual-robust cross-domain offline reinforcement learning algorithm that addresses both train-time and test-time dynamics shifts by employing a robust cross-domain Bellman operator alongside dynamic value penalty and Huber loss to enhance policy robustness and prevent value estimation errors.

Zhongjian Qiao, Rui Yang, Jiafei Lyu, Xiu Li, Zhongxiang Dai, Zhuoran Yang, Siyang Gao, Shuang Qiu2026-03-10🤖 cs.LG

Evolving Diffusion and Flow Matching Policies for Online Reinforcement Learning

The paper introduces GoRL, an algorithm-agnostic framework that stabilizes online reinforcement learning with expressive generative policies by decoupling optimization in a tractable latent space from action synthesis via a conditional generative decoder, achieving superior performance on challenging continuous-control tasks.

Chubin Zhang, Zhenglin Wan, Feng Chen, Fuchao Yang, Lang Feng, Yaxin Zhou, Xingrui Yu, Yang You, Ivor Tsang, Bo An2026-03-10🤖 cs.LG

Two-Step Data Augmentation for Masked Face Detection and Recognition: Turning Fake Masks to Real

This paper presents a two-step generative data augmentation framework combining rule-based mask warping and unpaired image-to-image translation to address the scarcity of masked face datasets, achieving performance improvements with minimal training data while explicitly noting its origins as a resource-constrained coursework project that lacked downstream quantitative evaluation.

Yan Yang, George Bebis, Mircea Nicolescu2026-03-10🤖 cs.LG

Latent Sculpting for Zero-Shot Generalization: A Manifold Learning Approach to Out-of-Distribution Anomaly Detection

The paper proposes "Latent Sculpting," a hierarchical two-stage architecture that combines a Transformer-based encoder with a Binary Latent Sculpting loss and a Masked Autoregressive Flow to enforce explicit geometric boundaries on benign data, achieving robust zero-shot generalization and high detection rates for out-of-distribution cyberattacks on the CIC-IDS-2017 benchmark.

Rajeeb Thapa Chhetri, Saurab Thapa, Avinash Kumar, Zhixiong Chen2026-03-10🤖 cs.LG

Certifying the Right to Be Forgotten: Primal-Dual Optimization for Sample and Label Unlearning in Vertical Federated Learning

This paper proposes FedORA, a primal-dual optimization framework that enables efficient and theoretically certified sample and label unlearning in Vertical Federated Learning by introducing a novel uncertainty-promoting loss function and adaptive strategies to minimize computational overhead while preserving model utility.

Yu Jiang, Xindi Tong, Ziyao Liu, Xiaoxi Zhang, Kwok-Yan Lam, Chee Wei Tan2026-03-10🤖 cs.LG

Sparse Offline Reinforcement Learning with Corruption Robustness

This paper proposes actor-critic methods with sparse robust estimator oracles to achieve the first non-vacuous guarantees for learning near-optimal policies in high-dimensional sparse offline reinforcement learning under strong data corruption and single-policy concentrability, overcoming the limitations of traditional Least Square Value Iteration approaches in such regimes.

Nam Phuong Tran, Andi Nika, Goran Radanovic, Long Tran-Thanh, Debmalya Mandal2026-03-10🤖 cs.LG

Reliable Grid Forecasting: State Space Models for Safety-Critical Energy Systems

This paper introduces an operator-legible evaluation framework centered on under-prediction risk to demonstrate that standard accuracy metrics fail to capture safety-critical grid forecasting needs, revealing that while explicit weather integration improves reliability, unconstrained probabilistic models often induce "fake safety" through excessive inflation, a problem solved by new Bias/OPR-constrained objectives.

Sunki Hong, Jisoo Lee2026-03-10⚡ eess

From Mice to Trains: Amortized Bayesian Inference on Graph Data

This paper proposes an amortized Bayesian inference framework for graph-structured data that combines permutation-invariant summary networks with neural posterior estimators to enable fast, likelihood-free inference on node, edge, and graph-level parameters, demonstrating its effectiveness through evaluations on synthetic benchmarks and real-world applications in biology and logistics.

Svenja Jedhoff, Elizaveta Semenova, Aura Raulo, Anne Meyer, Paul-Christian Bürkner2026-03-10🤖 cs.LG

DevBench: A Realistic, Developer-Informed Benchmark for Code Generation Models

DevBench is a realistic, telemetry-driven benchmark comprising 1,800 instances across six languages that evaluates LLMs on code completion tasks with a focus on ecological validity, contamination-free assessment, and detailed diagnostic insights to guide practical model selection and development.

Pareesa Ameneh Golnari, Adarsh Kumarappan, Wen Wen, Xiaoyu Liu, Gabriel Ryan, Yuting Sun, Shengyu Fu, Elsie Nallipogu2026-03-10🤖 cs.LG