Evaluating randomized smoothing as a defense against adversarial attacks in trajectory prediction

This paper proposes and evaluates randomized smoothing as an effective, simple, and computationally efficient defense mechanism that significantly enhances the robustness of trajectory prediction models against adversarial attacks without compromising their accuracy in standard settings.

Julian F. Schumann, Eduardo Figueiredo, Frederik Baymler Mathiesen, Luca Laurenti, Jens Kober, Arkady Zgonnikov2026-03-12🤖 cs.LG

Towards Cold-Start Drafting and Continual Refining: A Value-Driven Memory Approach with Application to NPU Kernel Synthesis

The paper introduces EvoKernel, a self-evolving agentic framework that leverages value-driven memory and reinforcement learning to overcome data scarcity in NPU kernel synthesis, significantly improving model correctness and achieving substantial speedups through automated drafting and iterative refinement.

Yujie Zheng, Zhuo Li, Shengtao Zhang, Hanjing Wang, Junjie Sheng, Jiaqian Wang, Junchi Yan, Weinan Zhang, Ying Wen, Bo Tang, Muning Wen2026-03-12🤖 cs.LG

V0.5V_{0.5}: Generalist Value Model as a Prior for Sparse RL Rollouts

The paper proposes V0.5V_{0.5}, a novel method that dynamically fuses a Generalist Value Model's prior with sparse RL rollouts via real-time statistical testing to minimize baseline estimation error, thereby achieving faster convergence and over 10% performance gains on mathematical reasoning benchmarks compared to GRPO and DAPO.

Yi-Kai Zhang, Yueqing Sun, Hongyan Hao, Qi Gu, Xunliang Cai, De-Chuan Zhan, Han-Jia Ye2026-03-12🤖 cs.LG

6ABOS: An Open-Source Atmospheric Correction Framework for the EnMAP Hyperspectral Mission Based on 6S

This paper introduces 6ABOS, an open-source Python framework that leverages the 6S radiative transfer model and Google Earth Engine to automate the atmospheric correction of EnMAP hyperspectral imagery, successfully validating its accuracy in retrieving water-leaving reflectance over diverse Mediterranean reservoirs.

Gabriel Caballero Cañas, Bárbara Alvado Arranz, Xavier Sòria-Perpinyà, Antonio Ruiz-Verdú, Jesús Delegido, José Moreno2026-03-12🤖 cs.LG

SNPgen: Phenotype-Supervised Genotype Representation and Synthetic Data Generation via Latent Diffusion

SNPgen is a two-stage conditional latent diffusion framework that generates privacy-preserving, phenotype-aligned synthetic genotype data, enabling machine learning models trained on synthetic samples to achieve predictive performance comparable to those trained on real data while maintaining strict privacy guarantees and preserving key genetic structures.

Andrea Lampis, Michela Carlotta Massi, Nicola Pirastu, Francesca Ieva, Matteo Matteucci, Emanuele Di Angelantonio2026-03-12🧬 q-bio

LAtte: Hyperbolic Lorentz Attention for Cross-Subject EEG Classification

The paper proposes LAtte, a novel cross-subject EEG classification framework that combines an InceptionTime-based encoder with a Lorentz Attention Module to learn shared baseline patterns and subject-specific embeddings, thereby achieving robust generalization and superior performance over state-of-the-art methods on multiple datasets.

Johannes Burchert, Ahmad Bdeir, Tom Hanika, Lars Schmidt-Thieme, Niels Landwehr2026-03-12🤖 cs.LG

Dynamics-Predictive Sampling for Active RL Finetuning of Large Reasoning Models

This paper introduces Dynamics-Predictive Sampling (DPS), a method that models prompt solving progress as a dynamical system to predict and select informative training samples via online Bayesian inference, thereby significantly reducing the computational overhead of extensive rollouts while accelerating and improving the reinforcement learning finetuning of large reasoning models.

Yixiu Mao, Yun Qu, Qi Wang, Heming Zou, Xiangyang Ji2026-03-12🤖 cs.LG

Ergodicity in reinforcement learning

This paper argues that the standard expected value objective in reinforcement learning is inadequate for non-ergodic environments where individual agent performance matters, and it explores the relationship between ergodic reward processes and Markov chains while presenting solutions to optimize long-term performance for single trajectories.

Dominik Baumann, Erfaun Noorani, Arsenii Mustafin, Xinyi Sheng, Bert Verbruggen, Arne Vanhoyweghen, Vincent Ginis, Thomas B. Schön2026-03-12🤖 cs.LG

LookaheadKV: Fast and Accurate KV Cache Eviction by Glimpsing into the Future without Generation

LookaheadKV is a lightweight KV cache eviction framework that achieves fast and accurate long-context inference by using parameter-efficient modules to predict future token importance without the computational overhead of explicit draft generation, thereby outperforming existing methods in both accuracy and speed.

Jinwoo Ahn, Ingyu Seong, Akhil Kedia, Junhan Kim, Hyemi Jang, Kangwook Lee, Yongkweon Jeon2026-03-12🤖 cs.LG

NCAA Bracket Prediction Using Machine Learning and Combinatorial Fusion Analysis

This paper demonstrates that applying Combinatorial Fusion Analysis (CFA) to generate team rankings for the 2024 NCAA tournament yields a 74.60% prediction accuracy, outperforming the best of ten popular public ranking systems and validating CFA as an effective paradigm for enhancing sports prediction precision.

Yuanhong Wu, Isaiah Smith, Tushar Marwah, Michael Schroeter, Mohamed Rahouti, D. Frank Hsu2026-03-12🤖 cs.LG

Historical Consensus: Preventing Posterior Collapse via Iterative Selection of Gaussian Mixture Priors

This paper introduces Historical Consensus Training, an iterative method that eliminates posterior collapse in Variational Autoencoders by progressively refining Gaussian Mixture Model priors to create a stable parameter barrier that prevents the degeneration of latent variables, achieving robust representations without relying on specific architectural constraints or hyperparameter tuning.

Zegu Zhang, Jian Zhang2026-03-12🤖 cs.LG

Safe RLHF Beyond Expectation: Stochastic Dominance for Universal Spectral Risk Control

This paper proposes Risk-sensitive Alignment via Dominance (RAD), a novel Safe RLHF framework that replaces traditional expected cost constraints with First-Order Stochastic Dominance constraints within an Optimal Transport framework to universally control spectral risk measures, thereby achieving superior robustness against tail risks and out-of-distribution failures while maintaining helpfulness.

Yaswanth Chittepu, Ativ Joshi, Rajarshi Bhattacharjee, Scott Niekum2026-03-12🤖 cs.LG

When should we trust the annotation? Selective prediction for molecular structure retrieval from mass spectra

This paper introduces a selective prediction framework for molecular structure retrieval from mass spectra that leverages retrieval-level uncertainty and distribution-free risk control to allow models to abstain from low-confidence predictions, thereby ensuring annotations meet specified error rate constraints in high-stakes applications.

Mira Jürgens, Gaetan De Waele, Morteza Rakhshaninejad, Willem Waegeman2026-03-12📊 stat