MCMC Informed Neural Emulators for Uncertainty Quantification in Dynamical Systems

This paper introduces an MCMC-informed neural emulator framework that decouples uncertainty quantification from network architecture by incorporating model-parameter distributions as training inputs, thereby enabling computationally efficient and accurate surrogate modeling for dynamical systems while avoiding exhaustive sampling and unphysical parameter evaluations.

Heikki Haario, Zhi-Song Liu, Martin Simon, Hendrik Weichel2026-03-12🤖 cs.LG

Bayesian Optimization with Gaussian Processes to Accelerate Stationary Point Searches

This paper presents a unified Bayesian optimization framework using Gaussian processes with derivative observations and advanced extensions like Optimal Transport and random Fourier features to efficiently accelerate the search for minima and saddle points on potential energy surfaces, bridging theoretical formulation with practical implementation through accompanying Rust code.

Rohit Goswami (Institute IMX and Lab-COSMO, École polytechnique fédérale de Lausanne)2026-03-12📊 stat

Factorized Neural Implicit DMD for Parametric Dynamics

This paper proposes Factorized Neural Implicit DMD, a data-driven method that parameterizes the Koopman operator's spectral decomposition via a physics-coded neural field to decouple spatial modes and temporal evolution, thereby enabling stable long-term rollouts, parameter generalization, and spectral analysis for high-dimensional nonlinear dynamical systems.

Siyuan Chen, Zhecheng Wang, Yixin Chen, Yue Chang, Peter Yichen Chen, Eitan Grinspun, Jonathan Panuelos2026-03-12🤖 cs.LG

Neural Field Thermal Tomography: A Differentiable Physics Framework for Non-Destructive Evaluation

The paper introduces Neural Field Thermal Tomography (NeFTY), a differentiable physics framework that parameterizes 3D material diffusivity as a continuous neural field optimized via a rigorous numerical solver to achieve high-resolution, quantitative reconstruction of subsurface defects from transient surface temperature measurements, overcoming the limitations of traditional 1D approximations and soft-constrained PINNs.

Tao Zhong, Yixun Hu, Dongzhe Zheng, Aditya Sood, Christine Allen-Blanchette2026-03-12🔬 cond-mat.mtrl-sci

XConv: Low-memory stochastic backpropagation for convolutional layers

XConv is a drop-in replacement for standard convolutional layers that significantly reduces memory usage during training by storing compressed activations and approximating weight gradients via randomized trace estimation, while maintaining performance comparable to exact gradient methods without imposing architectural constraints or requiring codebase modifications.

Anirudh Thatipelli, Jeffrey Sam, Mathias Louboutin, Ali Siahkoohi, Rongrong Wang, Felix J. Herrmann2026-03-11🤖 cs.LG

A Survey on Decentralized Federated Learning

This survey systematically reviews decentralized federated learning methods from 2018 to early 2026, categorizing them into traditional distributed and blockchain-based architectures, proposing a unified challenge-driven taxonomy, and outlining future research directions to address security, privacy, and system-level trade-offs in coordinator-free settings.

Edoardo Gabrielli, Anthony Di Pietro, Dario Fenoglio, Giovanni Pica, Gabriele Tolomei2026-03-11🤖 cs.LG

Polynomially Over-Parameterized Convolutional Neural Networks Contain Structured Strong Winning Lottery Tickets

This paper proves that randomly initialized, polynomially over-parameterized convolutional neural networks contain structured subnetworks capable of approximating smaller networks without training, by developing new mathematical tools to overcome previous limitations in analyzing the Strong Lottery Ticket Hypothesis for structured pruning.

Arthur da Cunha, Francesco d'Amore, Emanuele Natale2026-03-11🤖 cs.LG

Enhancing Computational Efficiency in Multiscale Systems Using Deep Learning of Coordinates and Flow Maps

This paper proposes a deep learning framework that jointly discovers optimal coordinates and flow maps to enable precise, computationally efficient time-stepping for multiscale systems, achieving state-of-the-art predictive accuracy with reduced costs on complex models like the Fitzhugh-Nagumo neuron and Kuramoto-Sivashinsky equations.

Asif Hamid, Danish Rafiq, Shahkar Ahmad Nahvi, Mohammad Abid Bazaz2026-03-11🤖 cs.LG

Sparse Variational Student-t Processes for Heavy-tailed Modeling

This paper introduces Sparse Variational Student-t Processes (SVTP), a scalable framework that extends sparse inducing point methods to Student-t processes via novel inference algorithms and natural gradient optimization, achieving superior robustness to outliers and heavy-tailed data with significantly faster convergence and lower prediction error compared to sparse Gaussian processes on large datasets.

Jian Xu, Delu Zeng, John Paisley2026-03-11🤖 cs.AI

Robust Training of Neural Networks at Arbitrary Precision and Sparsity

This paper introduces a unified framework that models quantization and sparsification as additive noise to derive a principled, noise-corrective gradient path, enabling the stable training of neural networks at arbitrary low precisions and sparsity levels without relying on heuristic estimators like the Straight-Through Estimator.

Chengxi Ye, Grace Chu, Yanfeng Liu, Yichi Zhang, Lukasz Lew, Li Zhang, Mark Sandler, Andrew Howard2026-03-11🤖 cs.AI

ARLBench: Flexible and Efficient Benchmarking for Hyperparameter Optimization in Reinforcement Learning

The paper introduces ARLBench, a flexible and efficient benchmark for hyperparameter optimization in reinforcement learning that utilizes a representative subset of tasks to enable cost-effective comparisons of diverse AutoRL methods and lower the barrier to entry for researchers with limited compute resources.

Jannis Becktepe, Julian Dierkes, Carolin Benjamins, Aditya Mohan, David Salinas, Raghu Rajan, Frank Hutter, Holger Hoos, Marius Lindauer, Theresa Eimer2026-03-11🤖 cs.LG

DRUPI: Dataset Reduction Using Privileged Information

The paper introduces DRUPI (Dataset Condensation using Privileged Information), a framework that enhances dataset condensation by synthesizing auxiliary privileged information, such as feature or attention labels, alongside reduced data to significantly improve model training performance across various benchmarks.

Shaobo Wang, Youxin Jiang, Tianle Niu, Yantai Yang, Ruiji Zhang, Shuhao Hu, Shuaiyu Zhang, Chenghao Sun, Weiya Li, Conghui He, Xuming Hu, Linfeng Zhang2026-03-11🤖 cs.AI