Do Models See in Line with Human Vision? Probing the Correspondence Between LVLM Representations and EEG Signals

This paper demonstrates that Large Vision Language Models (LVLMs) develop human-aligned visual representations by quantifying their correspondence with EEG signals, revealing that intermediate layers, multimodal architecture, and downstream visual performance are key drivers of this neural alignment.

Xin Xiao, Yang Lei, Haoyang Zeng, Xiao Sun, Xinyi Jiang, Yu Tian, Hao Wu, Kaiwen Wei, Jiang Zhong2026-03-10💻 cs

Retrieval-Augmented Anatomical Guidance for Text-to-CT Generation

This paper proposes a retrieval-augmented framework for text-to-CT generation that leverages a 3D vision-language encoder to retrieve semantically related clinical cases and their anatomical annotations as structural proxies, thereby enhancing image fidelity and spatial controllability in a realistic inference setting without requiring ground-truth annotations.

Daniele Molino, Camillo Maria Caruso, Paolo Soda, Valerio Guarrasi2026-03-10💻 cs

Adaptive Tracking Control of Euler-Lagrange Systems with Time-Varying State and Input Constraints

This paper proposes an adaptive control framework for Euler-Lagrange systems that guarantees user-defined time-varying state and input constraints under uncertainties and disturbances by integrating a time-varying barrier Lyapunov function with a saturated control law, supported by an offline feasibility condition and validated through real-time helicopter experiments.

Poulomee Ghosh, Shubhendu Bhasin2026-03-10💻 cs

Human-AI Divergence in Ego-centric Action Recognition under Spatial and Spatiotemporal Manipulations

This paper presents a large-scale comparative study using the Epic ReduAct dataset and over 3,000 human participants to demonstrate that while humans rely on sparse, semantically critical cues for egocentric action recognition, state-of-the-art AI models degrade more gradually by depending on contextual and low-level features, revealing fundamental divergences in how humans and machines process spatial and spatiotemporal information.

Sadegh Rahmaniboldaji, Filip Rybansky, Quoc C. Vuong, Anya C. Hurlbert, Frank Guerin, Andrew Gilbert2026-03-10💻 cs

CORE-Acu: Structured Reasoning Traces and Knowledge Graph Safety Verification for Acupuncture Clinical Decision Support

CORE-Acu is a neuro-symbolic framework for acupuncture clinical decision support that integrates structured reasoning traces, a knowledge graph-based safety verification system, and a specialized loss function to ensure interpretable, hallucination-free, and strictly safe treatment recommendations, outperforming standard LLMs with zero observed safety violations.

Liuyi Xu, Yun Guo, Ming Chen, Zihan Dun, Yining Qian, An-Yang Lu, Shuang Li, Lijun Liu2026-03-10💻 cs

Turn Complexity of Context-free Languages, Pushdown Automata and One-Counter Automata

This paper investigates the computational complexity of context-free, pushdown, and one-counter automata based on the number of "turns" (switches between increasing and decreasing stack height) in accepting computations, proving that determining whether this number is bounded is undecidable, establishing non-recursive trade-offs between automata types, and demonstrating an infinite hierarchy of complexity classes defined by sublinear turn bounds.

Giovanni Pighizzini2026-03-10💻 cs

The coordination between TSO and DSO in the context of energy transition - A review

This paper reviews and analyzes various coordination schemes between Transmission and Distribution System Operators (TSOs and DSOs) to effectively integrate Distributed Energy Resources, aiming to maintain system balance and prevent network congestion while overcoming technical and market challenges associated with the ongoing energy transition.

Hang Nguyen, Koen Kok, Trung Thai Tran, Phuong H. Nguyen2026-03-10💻 cs

Hierarchical Multi-Modal Planning for Fixed-Altitude Sparse Target Search and Sampling

This paper introduces HIMoS, a hierarchical multi-modal planning framework that enables Autonomous Underwater Vehicles to efficiently search for and sample sparse benthic targets like coral colonies at a fixed altitude by integrating a global topological route optimizer with a local differentiable belief propagation planner, thereby outperforming traditional exhaustive and adaptive sampling strategies in high-fidelity simulations.

Lingpeng Chen, Yuchen Zheng, Apple Pui-Yi Chui, Junfeng Wu, Ziyang Hong2026-03-10💻 cs

The Complexity of Extending Storylines with Minimum Local Crossing Number

This paper investigates the computational complexity of extending fixed storyline layouts by inserting missing characters to minimize the local crossing number, proving the problem is W[1]-hard when parameterized by the number of inserted characters and active characters, but fixed-parameter tractable when parameterized by the sum of active characters and the local crossing number.

Alexander Dobler, Siddharth Gupta, Philipp Kindermann, Fabrizio Montecchiani, Martin Nöllenburg2026-03-10💻 cs

PhaForce: Phase-Scheduled Visual-Force Policy Learning with Slow Planning and Fast Correction for Contact-Rich Manipulation

PhaForce is a phase-scheduled visuomotor policy that enhances contact-rich manipulation by coordinating a slow, vision-dominant diffusion planner with a fast, force-driven corrector to enable high-frequency, phase-aware residual corrections, achieving an 86% success rate and superior adaptability compared to existing baselines.

Mingxin Wang, Zhirun Yue, Renhao Lu, Yizhe Li, Zihan Wang, Guoping Pan, Kangkang Dong, Jun Cheng, Yi Cheng, Houde Liu2026-03-10💻 cs

Local-Global Prompt Learning via Sparse Optimal Transport

The paper proposes SOT-GLP, a novel few-shot adaptation method for vision-language models that employs shared sparse optimal transport to partition visual regions among class-specific local prompts while maintaining global alignment, thereby achieving state-of-the-art performance in both classification accuracy and out-of-distribution detection by preserving the native feature geometry.

Deniz Kizaro\u{g}lu, Ülku Tuncer Küçüktas, Emre Çakmakyurdu, Alptekin Temizel2026-03-10💻 cs

Δ\DeltaVLA: Prior-Guided Vision-Language-Action Models via World Knowledge Variation

This paper introduces Δ\DeltaVLA, a prior-guided framework that enhances robotic manipulation by modeling discrete world-knowledge variations relative to an explicit current state prior, rather than predicting absolute future states, thereby achieving state-of-the-art performance and efficiency through its novel components: the Prior-Guided World Knowledge Extractor, Latent World Variation Quantization, and Conditional Variation Attention.

Yijie Zhu, Jie He, Rui Shao, Kaishen Yuan, Tao Tan, Xiaochen Yuan, Zitong Yu2026-03-10💻 cs