SNPgen: Phenotype-Supervised Genotype Representation and Synthetic Data Generation via Latent Diffusion

SNPgen is a two-stage conditional latent diffusion framework that generates privacy-preserving, phenotype-aligned synthetic genotype data, enabling machine learning models trained on synthetic samples to achieve predictive performance comparable to those trained on real data while maintaining strict privacy guarantees and preserving key genetic structures.

Andrea Lampis, Michela Carlotta Massi, Nicola Pirastu, Francesca Ieva, Matteo Matteucci, Emanuele Di Angelantonio2026-03-12🧬 q-bio

LAtte: Hyperbolic Lorentz Attention for Cross-Subject EEG Classification

The paper proposes LAtte, a novel cross-subject EEG classification framework that combines an InceptionTime-based encoder with a Lorentz Attention Module to learn shared baseline patterns and subject-specific embeddings, thereby achieving robust generalization and superior performance over state-of-the-art methods on multiple datasets.

Johannes Burchert, Ahmad Bdeir, Tom Hanika, Lars Schmidt-Thieme, Niels Landwehr2026-03-12🤖 cs.LG

Dynamics-Predictive Sampling for Active RL Finetuning of Large Reasoning Models

This paper introduces Dynamics-Predictive Sampling (DPS), a method that models prompt solving progress as a dynamical system to predict and select informative training samples via online Bayesian inference, thereby significantly reducing the computational overhead of extensive rollouts while accelerating and improving the reinforcement learning finetuning of large reasoning models.

Yixiu Mao, Yun Qu, Qi Wang, Heming Zou, Xiangyang Ji2026-03-12🤖 cs.LG

Ergodicity in reinforcement learning

This paper argues that the standard expected value objective in reinforcement learning is inadequate for non-ergodic environments where individual agent performance matters, and it explores the relationship between ergodic reward processes and Markov chains while presenting solutions to optimize long-term performance for single trajectories.

Dominik Baumann, Erfaun Noorani, Arsenii Mustafin, Xinyi Sheng, Bert Verbruggen, Arne Vanhoyweghen, Vincent Ginis, Thomas B. Schön2026-03-12🤖 cs.LG

LookaheadKV: Fast and Accurate KV Cache Eviction by Glimpsing into the Future without Generation

LookaheadKV is a lightweight KV cache eviction framework that achieves fast and accurate long-context inference by using parameter-efficient modules to predict future token importance without the computational overhead of explicit draft generation, thereby outperforming existing methods in both accuracy and speed.

Jinwoo Ahn, Ingyu Seong, Akhil Kedia, Junhan Kim, Hyemi Jang, Kangwook Lee, Yongkweon Jeon2026-03-12🤖 cs.LG

NCAA Bracket Prediction Using Machine Learning and Combinatorial Fusion Analysis

This paper demonstrates that applying Combinatorial Fusion Analysis (CFA) to generate team rankings for the 2024 NCAA tournament yields a 74.60% prediction accuracy, outperforming the best of ten popular public ranking systems and validating CFA as an effective paradigm for enhancing sports prediction precision.

Yuanhong Wu, Isaiah Smith, Tushar Marwah, Michael Schroeter, Mohamed Rahouti, D. Frank Hsu2026-03-12🤖 cs.LG

Historical Consensus: Preventing Posterior Collapse via Iterative Selection of Gaussian Mixture Priors

This paper introduces Historical Consensus Training, an iterative method that eliminates posterior collapse in Variational Autoencoders by progressively refining Gaussian Mixture Model priors to create a stable parameter barrier that prevents the degeneration of latent variables, achieving robust representations without relying on specific architectural constraints or hyperparameter tuning.

Zegu Zhang, Jian Zhang2026-03-12🤖 cs.LG

Safe RLHF Beyond Expectation: Stochastic Dominance for Universal Spectral Risk Control

This paper proposes Risk-sensitive Alignment via Dominance (RAD), a novel Safe RLHF framework that replaces traditional expected cost constraints with First-Order Stochastic Dominance constraints within an Optimal Transport framework to universally control spectral risk measures, thereby achieving superior robustness against tail risks and out-of-distribution failures while maintaining helpfulness.

Yaswanth Chittepu, Ativ Joshi, Rajarshi Bhattacharjee, Scott Niekum2026-03-12🤖 cs.LG

When should we trust the annotation? Selective prediction for molecular structure retrieval from mass spectra

This paper introduces a selective prediction framework for molecular structure retrieval from mass spectra that leverages retrieval-level uncertainty and distribution-free risk control to allow models to abstain from low-confidence predictions, thereby ensuring annotations meet specified error rate constraints in high-stakes applications.

Mira Jürgens, Gaetan De Waele, Morteza Rakhshaninejad, Willem Waegeman2026-03-12📊 stat

Bio-Inspired Self-Supervised Learning for Wrist-worn IMU Signals

This paper introduces a bio-inspired self-supervised learning framework for wrist-worn IMU signals that leverages motor control submovement theory to tokenize motion into segments, enabling a Transformer encoder to outperform existing baselines in human activity recognition through masked segment reconstruction on a large-scale NHANES dataset.

Prithviraj Tarale, Kiet Chu, Abhishek Varghese, Kai-Chun Liu, Maxwell A Xu, Mohit Iyyer, Sunghoon I. Lee2026-03-12🤖 cs.LG

FRIEND: Federated Learning for Joint Optimization of multi-RIS Configuration and Eavesdropper Intelligent Detection in B5G Networks

This paper proposes a privacy-preserving Federated Learning framework that jointly optimizes multi-RIS configuration and eavesdropper detection in B5G cell-free mmWave networks, achieving a 30% improvement in secrecy rate while maintaining high detection accuracy through collaborative DCNN training on local Channel State Information.

Maria Lamprini A. Bartsioka, Ioannis A. Bartsiokas, Anastasios K. Papazafeiropoulos, Maria A. Seimeni, Dimitra I. Kaklamani, Iakovos S. Venieris2026-03-12🤖 cs.LG

Federated Learning-driven Beam Management in LEO 6G Non-Terrestrial Networks

This paper proposes a Federated Learning framework for LEO 6G Non-Terrestrial Networks that leverages High-Altitude Platform Stations to distribute beam selection tasks, demonstrating that a Graph Neural Network model outperforms a Multi-Layer Perceptron in prediction accuracy and stability, especially at low elevation angles.

Maria Lamprini Bartsioka, Ioannis A. Bartsiokas, Athanasios D. Panagopoulos, Dimitra I. Kaklamani, Iakovos S. Venieris2026-03-12🔬 physics

The Discrete Charm of the MLP: Binary Routing of Continuous Signals in Transformer Feed-Forward Layers

This paper demonstrates that MLP layers in transformer models function as binary routing switches that direct continuous signals through distinct computational paths based on consensus and exception-handling neuron architectures, a mechanism that explains the limitations of smooth polynomial approximations and is validated by significant causal performance differences.

Peter Balogh2026-03-12🤖 cs.LG