Morphological-Symmetry-Equivariant Heterogeneous Graph Neural Network for Robotic Dynamics Learning

This paper introduces MS-HGNN, a morphological-symmetry-equivariant heterogeneous graph neural network that integrates robotic kinematic structures and symmetries as architectural constraints to achieve high generalizability and efficiency in learning dynamics for various multi-body systems, with its effectiveness validated through formal proofs and experiments on quadruped robots.

Fengze Xie, Sizhe Wei, Yue Song, Yisong Yue, Lu GanWed, 11 Ma🤖 cs.LG

Prognostics for Autonomous Deep-Space Habitat Health Management under Multiple Unknown Failure Modes

This paper proposes an unsupervised prognostics framework that utilizes unlabeled run-to-failure data to simultaneously identify latent failure modes and select informative sensors, thereby enabling accurate remaining useful life prediction for autonomous deep-space habitats under multiple unknown failure conditions.

Benjamin Peters, Ayush Mohanty, Xiaolei Fang, Stephen K. Robinson, Nagi GebraeelWed, 11 Ma🤖 cs.LG

Adaptive and Stratified Subsampling for High-Dimensional Robust Estimation

This paper introduces Adaptive Importance Sampling and Stratified Subsampling estimators that achieve minimax-optimal rates for robust high-dimensional sparse regression under heavy-tailed noise, contamination, and temporal dependence, while also providing fully specified de-biasing procedures for valid confidence intervals and demonstrating superior empirical performance over uniform subsampling.

Prateek Mittal, Joohi ChauhanWed, 11 Ma🤖 cs.LG

Enhancing Computational Efficiency in Multiscale Systems Using Deep Learning of Coordinates and Flow Maps

This paper proposes a deep learning framework that jointly discovers optimal coordinates and flow maps to enable precise, computationally efficient time-stepping for multiscale systems, achieving state-of-the-art predictive accuracy with reduced costs on complex models like the Fitzhugh-Nagumo neuron and Kuramoto-Sivashinsky equations.

Asif Hamid, Danish Rafiq, Shahkar Ahmad Nahvi, Mohammad Abid BazazWed, 11 Ma🤖 cs.LG

SA2^{2}GFM: Enhancing Robust Graph Foundation Models with Structure-Aware Semantic Augmentation

This paper introduces SA2^{2}GFM, a robust Graph Foundation Model framework that enhances domain-adaptive representations and generalization by integrating structure-aware semantic augmentation, an information bottleneck mechanism, and expert adaptive routing to effectively mitigate domain noise, structural perturbations, and adversarial attacks.

Junhua Shi, Qingyun Sun, Haonan Yuan, Xingcheng FuWed, 11 Ma🤖 cs.LG

Bradley-Terry Policy Optimization for Generative Preference Modeling

This paper introduces Bradley-Terry Policy Optimization (BTPO), a novel framework that derives a consistent Monte Carlo gradient estimator to effectively train large language models with chain-of-thought reasoning on non-verifiable pairwise preference tasks, overcoming the limitations of existing heuristic RL approaches.

Shengyu Feng, Yun He, Shuang Ma, Beibin Li, Yuanhao Xiong, Songlin Li, Karishma Mandyam, Julian Katz-Samuels, Shengjie Bi, Licheng Yu, Hejia Zhang, Karthik Abinav Sankararaman, Han Fang, Yiming Yang, Manaal FaruquiWed, 11 Ma🤖 cs.LG

Improved Robustness of Deep Reinforcement Learning for Control of Time-Varying Systems by Bounded Extremum Seeking

This paper proposes a hybrid control framework that combines Deep Reinforcement Learning (DRL) with robust model-independent bounded extremum seeking to enhance the stability and adaptability of controlling nonlinear time-varying systems, demonstrating its effectiveness through numerical simulations and the automatic tuning of a particle accelerator.

Shaifalee Saxena, Alan Williams, Rafael Fierro, Alexander ScheinkerWed, 11 Ma🤖 cs.LG

ZeroSiam: An Efficient Asymmetry for Test-Time Entropy Optimization without Collapse

This paper introduces ZeroSiam, an efficient asymmetric Siamese architecture that prevents model collapse during test-time entropy minimization by employing asymmetric divergence alignment, thereby enhancing adaptation and reasoning performance across diverse vision and language tasks with negligible overhead.

Guohao Chen, Shuaicheng Niu, Deyu Chen, Jiahao Yang, Zitian Zhang, Mingkui Tan, Pengcheng Wu, Zhiqi ShenWed, 11 Ma🤖 cs.LG

A Surrogate model for High Temperature Superconducting Magnets to Predict Current Distribution with Neural Network

This paper presents a fully connected residual neural network (FCRN) surrogate model trained on finite element method data to rapidly and accurately predict current density distributions and optimize the design of large-scale high-temperature superconducting magnets, overcoming the computational limitations of traditional simulations.

Mianjun Xiao, Peng Song, Yulong Liu, Cedric Korte, Ziyang Xu, Jiale Gao, Jiaqi Lu, Haoyang Nie, Qiantong Deng, Timing QuWed, 11 Ma🤖 cs.LG

Iterative In-Context Learning to Enhance LLMs Abstract Reasoning: The Case-Study of Algebraic Tasks

This paper proposes an iterative in-context learning methodology that optimizes few-shot example selection to significantly enhance large language models' systematic generalization and reasoning capabilities on algebraic tasks with non-standard rules, revealing that simpler examples can sometimes outperform complex ones.

Stefano Fioravanti, Matteo Zavatteri, Roberto Confalonieri, Kamyar Zeinalipour, Paolo Frazzetto, Alessandro Sperduti, Nicolò NavarinWed, 11 Ma🤖 cs.LG