Understanding the Dynamics of Demonstration Conflict in In-Context Learning

This paper investigates how large language models process conflicting demonstrations in in-context learning, revealing a two-phase computational structure where early layers encode both correct and incorrect rules while late layers commit to predictions, and identifies specific attention heads responsible for this vulnerability that can be mitigated through targeted ablation to significantly improve performance.

Difan Jiao, Di Wang, Lijie Hu2026-03-06💻 cs

Towards Explainable Deep Learning for Ship Trajectory Prediction in Inland Waterways

This study proposes an interpretable LSTM-based model for predicting ship trajectories in inland waterways that incorporates trained ship domain parameters to analyze attention mechanisms, revealing that while the model achieves competitive accuracy, its attention weights do not fully align with expected causal relationships between interacting vessels.

Tom Legel, Dirk Söffker, Roland Schätzle + 1 more2026-03-06💻 cs

Dictionary Based Pattern Entropy for Causal Direction Discovery

This paper introduces Dictionary Based Pattern Entropy (DPE), a novel framework that combines Algorithmic and Shannon Information Theories to infer causal directions and identify driving subpatterns in symbolic sequences by quantifying how compact, rule-based patterns in a cause systematically reduce uncertainty in an effect, demonstrating robust performance across diverse synthetic and real-world datasets.

Harikrishnan N B, Shubham Bhilare, Aditi Kathpalia + 1 more2026-03-06🔢 math

Standing on the Shoulders of Giants: Rethinking EEG Foundation Model Pretraining via Multi-Teacher Distillation

This paper proposes the Multi-Teacher Distillation Pretraining (MTDP) framework, which leverages representations from established vision and time-series foundation models to efficiently bootstrap EEG foundation models, achieving superior performance across diverse downstream tasks with only 25% of the data required by traditional self-supervised methods.

Chenqi Li, Yu Liu, Shuo Zhang + 2 more2026-03-06💻 cs

Projected Hessian Learning: Fast Curvature Supervision for Accurate Machine-Learning Interatomic Potentials

The paper introduces Projected Hessian Learning (PHL), a scalable framework that enables efficient, curvature-informed training of machine-learning interatomic potentials by utilizing stochastic Hessian-vector products instead of explicit Hessian matrices, thereby achieving full-second-order accuracy with significantly reduced computational cost and memory requirements.

Austin Rodriguez, Justin S. Smith, Sakib Matin + 3 more2026-03-06🔬 physics

Weather-Related Crash Risk Forecasting: A Deep Learning Approach for Heterogenous Spatiotemporal Data

This study proposes a deep learning framework that utilizes an ensemble of ConvLSTM models trained on overlapping spatial grids to effectively forecast weather-related traffic crash risk by capturing complex spatiotemporal dependencies and heterogeneity, demonstrating superior performance over baseline models in North Carolina's diverse high-risk zones.

Abimbola Ogungbire, Srinivas Pulugurtha2026-03-06💻 cs