Scalable physics-informed deep generative model for solving forward and inverse stochastic differential equations

This paper proposes a scalable physics-informed deep generative model (sPI-GeM) that overcomes the limitations of existing methods by effectively solving forward and inverse stochastic differential equations in high-dimensional stochastic and spatial spaces through a combination of physics-informed basis networks and a deep generative model.

Shaoqian Zhou, Wen You, Ling Guo + 1 more2026-03-05🔬 physics

Convergence, Sticking and Escape: Stochastic Dynamics Near Critical Points in SGD

This paper analyzes the convergence and escape dynamics of Stochastic Gradient Descent in one-dimensional landscapes, establishing that while SGD reliably converges to local minima, it may linger near local maxima depending on noise variance and geometry, with specific results provided for the probability of escaping sharp maxima to neighboring minima.

Dmitry Dudukalov, Artem Logachov, Vladimir Lotov + 3 more2026-03-05🤖 cs.LG

A Copula Based Supervised Filter for Feature Selection in Diabetes Risk Prediction Using Machine Learning

This paper proposes a computationally efficient supervised filter based on a Gumbel-copula implied upper-tail concordance score to identify features that are simultaneously extreme with the positive class, demonstrating its effectiveness in ranking clinically relevant predictors for diabetes risk across large-scale and clinical datasets while outperforming standard filters and matching strong baselines.

Agnideep Aich, Md Monzur Murshed, Sameera Hewage + 1 more2026-03-05🤖 cs.LG

Boosting In-Context Learning in LLMs Through the Lens of Classical Supervised Learning

This paper introduces Supervised Calibration (SC), a loss-minimization framework that enhances In-Context Learning in Large Language Models by learning optimal per-class affine transformations to correct systematic biases and alter decision boundary orientations, thereby achieving state-of-the-art performance across multiple models and datasets.

Korel Gundem, Juncheng Dong, Dennis Zhang + 2 more2026-03-05🤖 cs.AI

Learning in an Echo Chamber: Online Learning with Replay Adversary

This paper introduces the Online Learning in the Replay Setting to model systems training on self-annotated data, establishing the Extended Threshold dimension as the exact measure of learnability and proving that while proper learners may fail catastrophically, specific improper algorithms can achieve optimal mistake bounds against replay adversaries.

Daniil Dmitriev, Harald Eskelund Franck, Carolin Heinzler + 1 more2026-03-05🤖 cs.LG

Implicit Bias of Per-sample Adam on Separable Data: Departure from the Full-batch Regime

This paper demonstrates that the implicit bias of per-sample Adam on separable data can deviate from the full-batch \ell_\infty-max-margin behavior, potentially converging to the 2\ell_2-max-margin classifier or a data-adaptive Mahalanobis-norm margin depending on the dataset, whereas Signum consistently converges to the \ell_\infty-max-margin regardless of batch size.

Beomhan Baek, Minhak Song, Chulhee Yun2026-03-05🤖 cs.AI

Synthetic Augmentation in Imbalanced Learning: When It Helps, When It Hurts, and How Much to Add

This paper establishes a unified statistical framework demonstrating that synthetic augmentation in imbalanced learning is not universally beneficial, revealing that its efficacy and optimal quantity depend on local data symmetry and generator alignment, and proposing a Validation-Tuned Synthetic Size (VTSS) strategy to empirically determine the best augmentation level.

Zhengchi Ma, Anru R. Zhang2026-03-05🤖 cs.LG