Bio-Inspired Self-Supervised Learning for Wrist-worn IMU Signals

This paper introduces a bio-inspired self-supervised learning framework for wrist-worn IMU signals that leverages motor control submovement theory to tokenize motion into segments, enabling a Transformer encoder to outperform existing baselines in human activity recognition through masked segment reconstruction on a large-scale NHANES dataset.

Prithviraj Tarale, Kiet Chu, Abhishek Varghese, Kai-Chun Liu, Maxwell A Xu, Mohit Iyyer, Sunghoon I. Lee2026-03-12🤖 cs.LG

FRIEND: Federated Learning for Joint Optimization of multi-RIS Configuration and Eavesdropper Intelligent Detection in B5G Networks

This paper proposes a privacy-preserving Federated Learning framework that jointly optimizes multi-RIS configuration and eavesdropper detection in B5G cell-free mmWave networks, achieving a 30% improvement in secrecy rate while maintaining high detection accuracy through collaborative DCNN training on local Channel State Information.

Maria Lamprini A. Bartsioka, Ioannis A. Bartsiokas, Anastasios K. Papazafeiropoulos, Maria A. Seimeni, Dimitra I. Kaklamani, Iakovos S. Venieris2026-03-12🤖 cs.LG

Federated Learning-driven Beam Management in LEO 6G Non-Terrestrial Networks

This paper proposes a Federated Learning framework for LEO 6G Non-Terrestrial Networks that leverages High-Altitude Platform Stations to distribute beam selection tasks, demonstrating that a Graph Neural Network model outperforms a Multi-Layer Perceptron in prediction accuracy and stability, especially at low elevation angles.

Maria Lamprini Bartsioka, Ioannis A. Bartsiokas, Athanasios D. Panagopoulos, Dimitra I. Kaklamani, Iakovos S. Venieris2026-03-12🔬 physics

The Discrete Charm of the MLP: Binary Routing of Continuous Signals in Transformer Feed-Forward Layers

This paper demonstrates that MLP layers in transformer models function as binary routing switches that direct continuous signals through distinct computational paths based on consensus and exception-handling neuron architectures, a mechanism that explains the limitations of smooth polynomial approximations and is validated by significant causal performance differences.

Peter Balogh2026-03-12🤖 cs.LG

MCMC Informed Neural Emulators for Uncertainty Quantification in Dynamical Systems

This paper introduces an MCMC-informed neural emulator framework that decouples uncertainty quantification from network architecture by incorporating model-parameter distributions as training inputs, thereby enabling computationally efficient and accurate surrogate modeling for dynamical systems while avoiding exhaustive sampling and unphysical parameter evaluations.

Heikki Haario, Zhi-Song Liu, Martin Simon, Hendrik Weichel2026-03-12🤖 cs.LG

Bayesian Optimization with Gaussian Processes to Accelerate Stationary Point Searches

This paper presents a unified Bayesian optimization framework using Gaussian processes with derivative observations and advanced extensions like Optimal Transport and random Fourier features to efficiently accelerate the search for minima and saddle points on potential energy surfaces, bridging theoretical formulation with practical implementation through accompanying Rust code.

Rohit Goswami (Institute IMX and Lab-COSMO, École polytechnique fédérale de Lausanne)2026-03-12📊 stat

Factorized Neural Implicit DMD for Parametric Dynamics

This paper proposes Factorized Neural Implicit DMD, a data-driven method that parameterizes the Koopman operator's spectral decomposition via a physics-coded neural field to decouple spatial modes and temporal evolution, thereby enabling stable long-term rollouts, parameter generalization, and spectral analysis for high-dimensional nonlinear dynamical systems.

Siyuan Chen, Zhecheng Wang, Yixin Chen, Yue Chang, Peter Yichen Chen, Eitan Grinspun, Jonathan Panuelos2026-03-12🤖 cs.LG

Neural Field Thermal Tomography: A Differentiable Physics Framework for Non-Destructive Evaluation

The paper introduces Neural Field Thermal Tomography (NeFTY), a differentiable physics framework that parameterizes 3D material diffusivity as a continuous neural field optimized via a rigorous numerical solver to achieve high-resolution, quantitative reconstruction of subsurface defects from transient surface temperature measurements, overcoming the limitations of traditional 1D approximations and soft-constrained PINNs.

Tao Zhong, Yixun Hu, Dongzhe Zheng, Aditya Sood, Christine Allen-Blanchette2026-03-12🔬 cond-mat.mtrl-sci

XConv: Low-memory stochastic backpropagation for convolutional layers

XConv is a drop-in replacement for standard convolutional layers that significantly reduces memory usage during training by storing compressed activations and approximating weight gradients via randomized trace estimation, while maintaining performance comparable to exact gradient methods without imposing architectural constraints or requiring codebase modifications.

Anirudh Thatipelli, Jeffrey Sam, Mathias Louboutin, Ali Siahkoohi, Rongrong Wang, Felix J. Herrmann2026-03-11🤖 cs.LG

A Survey on Decentralized Federated Learning

This survey systematically reviews decentralized federated learning methods from 2018 to early 2026, categorizing them into traditional distributed and blockchain-based architectures, proposing a unified challenge-driven taxonomy, and outlining future research directions to address security, privacy, and system-level trade-offs in coordinator-free settings.

Edoardo Gabrielli, Anthony Di Pietro, Dario Fenoglio, Giovanni Pica, Gabriele Tolomei2026-03-11🤖 cs.LG

Polynomially Over-Parameterized Convolutional Neural Networks Contain Structured Strong Winning Lottery Tickets

This paper proves that randomly initialized, polynomially over-parameterized convolutional neural networks contain structured subnetworks capable of approximating smaller networks without training, by developing new mathematical tools to overcome previous limitations in analyzing the Strong Lottery Ticket Hypothesis for structured pruning.

Arthur da Cunha, Francesco d'Amore, Emanuele Natale2026-03-11🤖 cs.LG

Enhancing Computational Efficiency in Multiscale Systems Using Deep Learning of Coordinates and Flow Maps

This paper proposes a deep learning framework that jointly discovers optimal coordinates and flow maps to enable precise, computationally efficient time-stepping for multiscale systems, achieving state-of-the-art predictive accuracy with reduced costs on complex models like the Fitzhugh-Nagumo neuron and Kuramoto-Sivashinsky equations.

Asif Hamid, Danish Rafiq, Shahkar Ahmad Nahvi, Mohammad Abid Bazaz2026-03-11🤖 cs.LG