Latent Sculpting for Zero-Shot Generalization: A Manifold Learning Approach to Out-of-Distribution Anomaly Detection

The paper proposes "Latent Sculpting," a hierarchical two-stage architecture that combines a Transformer-based encoder with a Binary Latent Sculpting loss and a Masked Autoregressive Flow to enforce explicit geometric boundaries on benign data, achieving robust zero-shot generalization and high detection rates for out-of-distribution cyberattacks on the CIC-IDS-2017 benchmark.

Rajeeb Thapa Chhetri, Saurab Thapa, Avinash Kumar, Zhixiong Chen2026-03-10🤖 cs.LG

Certifying the Right to Be Forgotten: Primal-Dual Optimization for Sample and Label Unlearning in Vertical Federated Learning

This paper proposes FedORA, a primal-dual optimization framework that enables efficient and theoretically certified sample and label unlearning in Vertical Federated Learning by introducing a novel uncertainty-promoting loss function and adaptive strategies to minimize computational overhead while preserving model utility.

Yu Jiang, Xindi Tong, Ziyao Liu, Xiaoxi Zhang, Kwok-Yan Lam, Chee Wei Tan2026-03-10🤖 cs.LG

Sparse Offline Reinforcement Learning with Corruption Robustness

This paper proposes actor-critic methods with sparse robust estimator oracles to achieve the first non-vacuous guarantees for learning near-optimal policies in high-dimensional sparse offline reinforcement learning under strong data corruption and single-policy concentrability, overcoming the limitations of traditional Least Square Value Iteration approaches in such regimes.

Nam Phuong Tran, Andi Nika, Goran Radanovic, Long Tran-Thanh, Debmalya Mandal2026-03-10🤖 cs.LG

Reliable Grid Forecasting: State Space Models for Safety-Critical Energy Systems

This paper introduces an operator-legible evaluation framework centered on under-prediction risk to demonstrate that standard accuracy metrics fail to capture safety-critical grid forecasting needs, revealing that while explicit weather integration improves reliability, unconstrained probabilistic models often induce "fake safety" through excessive inflation, a problem solved by new Bias/OPR-constrained objectives.

Sunki Hong, Jisoo Lee2026-03-10⚡ eess

From Mice to Trains: Amortized Bayesian Inference on Graph Data

This paper proposes an amortized Bayesian inference framework for graph-structured data that combines permutation-invariant summary networks with neural posterior estimators to enable fast, likelihood-free inference on node, edge, and graph-level parameters, demonstrating its effectiveness through evaluations on synthetic benchmarks and real-world applications in biology and logistics.

Svenja Jedhoff, Elizaveta Semenova, Aura Raulo, Anne Meyer, Paul-Christian Bürkner2026-03-10🤖 cs.LG

DevBench: A Realistic, Developer-Informed Benchmark for Code Generation Models

DevBench is a realistic, telemetry-driven benchmark comprising 1,800 instances across six languages that evaluates LLMs on code completion tasks with a focus on ecological validity, contamination-free assessment, and detailed diagnostic insights to guide practical model selection and development.

Pareesa Ameneh Golnari, Adarsh Kumarappan, Wen Wen, Xiaoyu Liu, Gabriel Ryan, Yuting Sun, Shengyu Fu, Elsie Nallipogu2026-03-10🤖 cs.LG

ELSA: Efficient LLM-Centric Split Aggregation for Privacy-Aware Hierarchical Federated Learning over the Network Edge

ELSA is a novel framework that integrates split learning and hierarchical federated learning with client clustering, dynamic model splitting, and privacy-preserving communication sketches to enable efficient, robust, and privacy-aware fine-tuning of large language models on resource-constrained edge networks.

Xiaohong Yang, Tong Xie, Minghui Liwang, Chikai Shang, Yang Lu, Zhenzhen Jiao, Liqun Fu, Seyyedali Hosseinalipour2026-03-10🤖 cs.LG

Continuous-Flow Data-Rate-Aware CNN Inference on FPGA

This paper proposes a novel data-rate-aware continuous-flow architecture for CNN inference on FPGAs that mitigates hardware underutilization caused by data reduction in pooling and strided convolution layers by interleaving signals and sharing resources, thereby enabling the high-throughput implementation of complex models like MobileNet on a single device.

Tobias Habermann, Michael Mecik, Zhenyu Wang, César David Vera, Martin Kumm, Mario Garrido2026-03-10🤖 cs.LG

MeanCache: From Instantaneous to Average Velocity for Accelerating Flow Matching Inference

MeanCache is a training-free framework that accelerates Flow Matching inference by replacing instantaneous velocity caching with an average-velocity approach using cached Jacobian-vector products and a trajectory-stability scheduling strategy, achieving significant speedups (up to 4.56X) while maintaining high generation quality across models like FLUX.1 and HunyuanVideo.

Huanlin Gao, Ping Chen, Fuyuan Shi, Ruijia Wu, Li YanTao, Qiang Hui, Yuren You, Ting Lu, Chao Tan, Shaoan Zhao, Zhaoxiang Liu, Fang Zhao, Kai Wang, Shiguo Lian2026-03-10🤖 cs.LG

Transferable Graph Condensation from the Causal Perspective

This paper proposes TGCC, a novel causal-invariance-based graph dataset condensation method that extracts domain-invariant features and injects them via spectral contrastive learning to significantly improve performance in cross-task and cross-domain scenarios while maintaining state-of-the-art results in single-task settings.

Huaming Du, Yijie Huang, Su Yao, Yiying Wang, Yueyang Zhou, Jingwen Yang, Jinshi Zhang, Han Ji, Yu Zhao, Guisong Liu, Hegui Zhang, Carl Yang, Gang Kou2026-03-10🤖 cs.LG