Deep Randomized Distributed Function Computation (DeepRDFC): Neural Distributed Channel Simulation

This paper proposes a deep learning-based autoencoder architecture for the Randomized Distributed Function Computation (RDFC) framework that minimizes the total variation distance to an unknown target distribution using only data samples, demonstrating superior communication efficiency compared to traditional data compression methods, particularly under limited common randomness.

Didrik Bergström, Onur Günlü2026-03-12🔢 math

Prioritizing Gradient Sign Over Modulus: An Importance-Aware Framework for Wireless Federated Learning

This paper proposes Sign-Prioritized FL (SP-FL), a novel wireless federated learning framework that enhances model training reliability under resource constraints by prioritizing the transmission of gradient signs through a hierarchical resource allocation scheme, achieving up to 9.96% higher accuracy than existing methods on the CIFAR-10 dataset.

Yiyang Yue, Jiacheng Yao, Wei Xu, Zhaohui Yang, George K. Karagiannidis, Dusit Niyato2026-03-12⚡ eess

Evaluating randomized smoothing as a defense against adversarial attacks in trajectory prediction

This paper proposes and evaluates randomized smoothing as an effective, simple, and computationally efficient defense mechanism that significantly enhances the robustness of trajectory prediction models against adversarial attacks without compromising their accuracy in standard settings.

Julian F. Schumann, Eduardo Figueiredo, Frederik Baymler Mathiesen, Luca Laurenti, Jens Kober, Arkady Zgonnikov2026-03-12🤖 cs.LG

Towards Cold-Start Drafting and Continual Refining: A Value-Driven Memory Approach with Application to NPU Kernel Synthesis

The paper introduces EvoKernel, a self-evolving agentic framework that leverages value-driven memory and reinforcement learning to overcome data scarcity in NPU kernel synthesis, significantly improving model correctness and achieving substantial speedups through automated drafting and iterative refinement.

Yujie Zheng, Zhuo Li, Shengtao Zhang, Hanjing Wang, Junjie Sheng, Jiaqian Wang, Junchi Yan, Weinan Zhang, Ying Wen, Bo Tang, Muning Wen2026-03-12🤖 cs.LG

V0.5V_{0.5}: Generalist Value Model as a Prior for Sparse RL Rollouts

The paper proposes V0.5V_{0.5}, a novel method that dynamically fuses a Generalist Value Model's prior with sparse RL rollouts via real-time statistical testing to minimize baseline estimation error, thereby achieving faster convergence and over 10% performance gains on mathematical reasoning benchmarks compared to GRPO and DAPO.

Yi-Kai Zhang, Yueqing Sun, Hongyan Hao, Qi Gu, Xunliang Cai, De-Chuan Zhan, Han-Jia Ye2026-03-12🤖 cs.LG

6ABOS: An Open-Source Atmospheric Correction Framework for the EnMAP Hyperspectral Mission Based on 6S

This paper introduces 6ABOS, an open-source Python framework that leverages the 6S radiative transfer model and Google Earth Engine to automate the atmospheric correction of EnMAP hyperspectral imagery, successfully validating its accuracy in retrieving water-leaving reflectance over diverse Mediterranean reservoirs.

Gabriel Caballero Cañas, Bárbara Alvado Arranz, Xavier Sòria-Perpinyà, Antonio Ruiz-Verdú, Jesús Delegido, José Moreno2026-03-12🤖 cs.LG

SNPgen: Phenotype-Supervised Genotype Representation and Synthetic Data Generation via Latent Diffusion

SNPgen is a two-stage conditional latent diffusion framework that generates privacy-preserving, phenotype-aligned synthetic genotype data, enabling machine learning models trained on synthetic samples to achieve predictive performance comparable to those trained on real data while maintaining strict privacy guarantees and preserving key genetic structures.

Andrea Lampis, Michela Carlotta Massi, Nicola Pirastu, Francesca Ieva, Matteo Matteucci, Emanuele Di Angelantonio2026-03-12🧬 q-bio

LAtte: Hyperbolic Lorentz Attention for Cross-Subject EEG Classification

The paper proposes LAtte, a novel cross-subject EEG classification framework that combines an InceptionTime-based encoder with a Lorentz Attention Module to learn shared baseline patterns and subject-specific embeddings, thereby achieving robust generalization and superior performance over state-of-the-art methods on multiple datasets.

Johannes Burchert, Ahmad Bdeir, Tom Hanika, Lars Schmidt-Thieme, Niels Landwehr2026-03-12🤖 cs.LG

Dynamics-Predictive Sampling for Active RL Finetuning of Large Reasoning Models

This paper introduces Dynamics-Predictive Sampling (DPS), a method that models prompt solving progress as a dynamical system to predict and select informative training samples via online Bayesian inference, thereby significantly reducing the computational overhead of extensive rollouts while accelerating and improving the reinforcement learning finetuning of large reasoning models.

Yixiu Mao, Yun Qu, Qi Wang, Heming Zou, Xiangyang Ji2026-03-12🤖 cs.LG

Ergodicity in reinforcement learning

This paper argues that the standard expected value objective in reinforcement learning is inadequate for non-ergodic environments where individual agent performance matters, and it explores the relationship between ergodic reward processes and Markov chains while presenting solutions to optimize long-term performance for single trajectories.

Dominik Baumann, Erfaun Noorani, Arsenii Mustafin, Xinyi Sheng, Bert Verbruggen, Arne Vanhoyweghen, Vincent Ginis, Thomas B. Schön2026-03-12🤖 cs.LG

LookaheadKV: Fast and Accurate KV Cache Eviction by Glimpsing into the Future without Generation

LookaheadKV is a lightweight KV cache eviction framework that achieves fast and accurate long-context inference by using parameter-efficient modules to predict future token importance without the computational overhead of explicit draft generation, thereby outperforming existing methods in both accuracy and speed.

Jinwoo Ahn, Ingyu Seong, Akhil Kedia, Junhan Kim, Hyemi Jang, Kangwook Lee, Yongkweon Jeon2026-03-12🤖 cs.LG