Open-World Motion Forecasting

This paper introduces "Open-World Motion Forecasting," an end-to-end class-incremental framework that predicts future trajectories directly from camera images while mitigating catastrophic forgetting through pseudo-labeling with vision-language models and a novel query feature variance-based replay strategy, enabling continual adaptation to evolving object taxonomies in real-world autonomous driving.

Nicolas Schischka, Nikhil Gosala, B Ravi Kiran, Senthil Yogamani, Abhinav Valada2026-03-11🤖 cs.AI

AI Act Evaluation Benchmark: An Open, Transparent, and Reproducible Evaluation Dataset for NLP and RAG Systems

This paper introduces an open, transparent, and reproducible dataset and methodology for evaluating NLP and RAG systems on EU AI Act compliance, featuring tasks like risk classification and obligation generation that leverage large language models to address regulatory ambiguities and achieve high performance scores.

Athanasios Davvetas, Michael Papademas, Xenia Ziouvelou, Vangelis Karkaletsis2026-03-11🤖 cs.AI

A Guideline-Aware AI Agent for Zero-Shot Target Volume Auto-Delineation

This paper introduces OncoAgent, a novel guideline-aware AI agent that achieves zero-shot, training-free auto-delineation of clinical target volumes by converting textual clinical guidelines into 3D contours, demonstrating superior adaptability and physician preference over traditional supervised deep learning models.

Yoon Jo Kim, Wonyoung Cho, Jongmin Lee, Han Joo Chae, Hyunki Park, Sang Hoon Seo, Noh Jae Myung, Kyungmi Yang, Dongryul Oh, Jin Sung Kim2026-03-11🤖 cs.AI

An Empirical Study and Theoretical Explanation on Task-Level Model-Merging Collapse

This paper identifies and theoretically explains "task-level merging collapse," a phenomenon where incompatible task representations cause catastrophic performance degradation in merged LLMs, demonstrating that representational incompatibility—not parameter-space conflicts—is the primary driver of failure and establishing fundamental limits on task mergeability via rate-distortion theory.

Yuan Cao, Dezhi Ran, Yuzhe Guo, Mengzhou Wu, Simin Chen, Linyi Li, Wei Yang, Tao Xie2026-03-11🤖 cs.AI

EvoDriveVLA: Evolving Autonomous Driving Vision-Language-Action Model via Collaborative Perception-Planning Distillation

EvoDriveVLA is a novel Vision-Language-Action model for autonomous driving that overcomes perception degradation and planning instability through a collaborative distillation framework combining self-anchored visual constraints and oracle-guided trajectory optimization to achieve state-of-the-art performance.

Jiajun Cao, Xiaoan Zhang, Xiaobao Wei, Liyuqiu Huang, Wang Zijian, Hanzhen Zhang, Zhengyu Jia, Wei Mao, Hao Wang, Xianming Liu, Shuchang Zhou Liu, Yang Wang, Shanghang Zhang2026-03-11🤖 cs.AI

Efficiently Aligning Draft Models via Parameter- and Data-Efficient Adaptation

This paper introduces Efficient Draft Adaptation (EDA), a parameter- and data-efficient framework that restores speculative decoding performance on fine-tuned target models through a decoupled architecture, data regeneration strategy, and sample selection mechanism, achieving superior acceptance lengths with significantly reduced training costs compared to full retraining.

Luxi Lin, Zhihang Lin, Zhanpeng Zeng, Yuhao Chen, Qingyu Zhang, Jixiang Luo, Xuelong Li, Rongrong Ji2026-03-11🤖 cs.AI

Enhancing Debunking Effectiveness through LLM-based Personality Adaptation

This study proposes and evaluates a novel methodology for enhancing fake news debunking by using Large Language Models to generate personalized messages tailored to Big Five personality traits, demonstrating that such targeted approaches generally increase persuasiveness while highlighting both the potential and ethical implications of automated, personality-driven disinformation correction.

Pietro Dell'Oglio, Alessandro Bondielli, Francesco Marcelloni, Lucia C. Passaro2026-03-11🤖 cs.AI

Routing without Forgetting

The paper introduces Routing without Forgetting (RwF), a transformer architecture that addresses Online Continual Learning by replacing iterative gradient-based specialization with dynamic, single-step associative retrieval of input-conditioned prompts via energy-based layers, thereby achieving superior performance on class-incremental benchmarks without explicit task identifiers.

Alessio Masano, Giovanni Bellitto, Dipam Goswani, Joost Van de Weijer, Concetto Spampinato2026-03-11🤖 cs.AI

A Variational Latent Equilibrium for Learning in Cortex

This paper proposes a biologically plausible, local learning framework for time-continuous neuronal networks that approximates backpropagation through time by deriving real-time error dynamics from a prospective energy function, thereby unifying and extending the Generalized Latent Equilibrium model to enable spatiotemporal credit assignment consistent with brain circuitry.

Simon Brandt, Paul Haider, Walter Senn, Federico Benitez, Mihai A. Petrovici2026-03-11🤖 cs.AI