CbLDM: A Diffusion Model for recovering nanostructure from atomic pair distribution function

This paper proposes CbLDM, a Condition-based Latent Diffusion Model that utilizes conditional priors and Laplacian matrices to effectively and stably recover the nanostructures of monometallic nanoparticles from their atomic pair distribution functions, addressing the highly ill-posed nature of the inverse problem.

Jiarui Cao, Zhiyang Zhang, Heming Wang, Jun Xu, Ling Lan, Simon J. L. Billinge, Ran Gu2026-03-10🔬 cond-mat.mtrl-sci

Entropy-Driven Curriculum for Multi-Task Training in Human Mobility Prediction

This paper proposes a unified training framework that combines entropy-driven curriculum learning, which sequences training from simple to complex trajectories based on Lempel-Ziv compression, with multi-task learning to simultaneously optimize location, distance, and direction predictions, thereby achieving state-of-the-art performance and significantly faster convergence in human mobility prediction.

Tianye Fang, Xuanshu Luo, Martin Werner2026-03-10🤖 cs.LG

Synthetic data for ratemaking: imputation-based methods vs adversarial networks and autoencoders

This paper benchmarks Multivariate Imputation by Chained Equations (MICE) against deep generative models like Variational Autoencoders and Conditional Tabular GANs for synthetic ratemaking data, finding that MICE offers a simpler yet high-fidelity alternative that effectively preserves statistical distributions and supports robust Generalized Linear Model training.

Yevhen Havrylenko, Meelis Käärik, Artur Tuttar2026-03-10🤖 cs.LG

Faster Gradient Methods for Highly-Smooth Stochastic Bilevel Optimization

This paper proposes the F²SA-pp method, which utilizes pp-th order finite differences to achieve a nearly optimal O~(pϵ4p/2)\tilde{\mathcal{O}}(p \epsilon^{-4-p/2}) complexity for finding ϵ\epsilon-stationary points in stochastic bilevel optimization with highly smooth objectives, thereby improving upon previous first-order bounds and matching the fundamental lower limit.

Lesi Chen, Junru Li, El Mahdi Chayti, Jingzhao Zhang2026-03-10🤖 cs.LG

Behavioral Inference at Scale: The Fundamental Asymmetry Between Motivations and Belief Systems

Through large-scale experiments with over 1.5 million LLM-generated behavioral sequences, this paper reveals a fundamental asymmetry in behavioral inference where agent motivations are nearly perfectly recoverable while belief systems remain largely opaque due to inherent information-theoretic limits and architectural constraints, particularly within a "neutral zone" of behavioral ambiguity.

Jason Starace, Terence Soule2026-03-10🤖 cs.LG

Physics-Aware Neural Operators for Direct Inversion in 3D Photoacoustic Tomography

The paper introduces PANO, a physics-aware neural operator that performs direct, single-pass inversion of raw sensor data into high-quality 3D photoacoustic images, outperforming traditional algorithms and enabling real-time reconstruction across diverse sparse acquisition settings to facilitate the clinical translation of 3D PACT.

Jiayun Wang, Yousuf Aborahama, Arya Khokhar, Yang Zhang, Chuwei Wang, Karteekeya Sastry, Julius Berner, Yilin Luo, Boris Bonev, Zongyi Li, Kamyar Azizzadenesheli, Lihong V. Wang, Anima Anandkumar2026-03-10🤖 cs.LG

ORIC: Benchmarking Object Recognition under Contextual Incongruity in Large Vision-Language Models

This paper introduces the ORIC framework and benchmark to evaluate and improve Large Vision-Language Models' object recognition capabilities under contextual incongruity, demonstrating that such scenarios significantly degrade performance and that targeted Visual Reinforcement Fine-Tuning can effectively mitigate these failures.

Zhaoyang Li, Zhan Ling, Yuchen Zhou, Litian Gong, Erdem Bıyık, Hao Su2026-03-10🤖 cs.LG

ORN-CBF: Learning Observation-conditioned Residual Neural Control Barrier Functions via Hypernetworks

This paper proposes ORN-CBF, a hypernetwork-based learning framework that utilizes Hamilton-Jacobi reachability analysis to generate observation-conditioned neural control barrier functions, ensuring rigorous safety guarantees and improved generalization in partially observable environments through simulation and hardware experiments.

Bojan Derajic, Sebastian Bernhard, Wolfgang Hönig2026-03-10🤖 cs.LG

AEGIS: Authentic Edge Growth In Sparsity for Link Prediction in Edge-Sparse Bipartite Knowledge Graphs

The paper introduces AEGIS, an edge-only augmentation framework that resamples existing training edges to enhance link prediction in edge-sparse bipartite knowledge graphs, demonstrating that authenticity-constrained resampling preserves data integrity while semantic KNN augmentation further boosts performance when node descriptions are available.

Hugh Xuechen Liu, Kıvanç Tatar2026-03-10🤖 cs.LG

CLAD-Net: Continual Activity Recognition in Multi-Sensor Wearable Systems

CLAD-Net is a continual learning framework for wearable human activity recognition that combines a self-supervised transformer for long-term memory and a supervised CNN with knowledge distillation to effectively mitigate catastrophic forgetting and handle label scarcity across diverse subjects.

Reza Rahimi Azghan, Gautham Krishna Gudur, Mohit Malu, Edison Thomaz, Giulia Pedrielli, Pavan Turaga, Hassan Ghasemzadeh2026-03-10🤖 cs.LG

Generative Evolutionary Meta-Solver (GEMS): Scalable Surrogate-Free Multi-Agent Reinforcement Learning

The paper introduces Generative Evolutionary Meta-Solver (GEMS), a scalable, surrogate-free multi-agent reinforcement learning framework that replaces explicit policy populations with a compact generator and latent anchors to achieve significantly faster training, lower memory usage, and higher rewards than traditional methods like PSRO while maintaining game-theoretic guarantees.

Alakh Sharma, Gaurish Trivedi, Kartikey Singh Bhandari, Yash Sinha, Dhruv Kumar, Pratik Narang, Jagat Sesh Challa2026-03-10🤖 cs.LG

Overlap-Adaptive Regularization for Conditional Average Treatment Effect Estimation

This paper introduces Overlap-Adaptive Regularization (OAR), a novel method that enhances the performance of existing CATE meta-learners in low-overlap regions by proportionally increasing regularization based on overlap weights, while offering flexible, debiased variants that preserve Neyman-orthogonality for robust inference.

Valentyn Melnychuk, Dennis Frauen, Jonas Schweisthal, Stefan Feuerriegel2026-03-10🤖 cs.LG