Prism-Δ\Delta: Differential Subspace Steering for Prompt Highlighting in Large Language Models

PRISM-Δ\Delta is a novel prompt highlighting method that steers large language models by decomposing cross-covariance matrices to isolate discriminative signals and eliminate shared patterns, achieving superior performance across multiple benchmarks and long-context tasks while maintaining low computational overhead.

Yuyao Ge, Shenghua Liu, Yiwei Wang, Tianyu Liu, Baolong Bi, Lingrui Mei, Jiayu Yao, Jiafeng Guo, Xueqi ChengThu, 12 Ma💬 cs.CL

HeartAgent: An Autonomous Agent System for Explainable Differential Diagnosis in Cardiology

HeartAgent is an autonomous, cardiology-specific agent system that leverages specialized sub-agents and curated data to deliver explainable, high-accuracy differential diagnoses, significantly outperforming existing AI methods and enhancing clinical decision-making when assisting human experts.

Shuang Zhou, Kai Yu, Song Wang, Wenya Xie, Zaifu Zhan, Meng-Han Tsai, Yuen-Hei Chung, Shutong Hou, Huixue Zhou, Min Zeng, Bhavadharini Ramu, Lin Yee Chen, Feng Xie, Rui ZhangThu, 12 Ma💬 cs.CL

Interpretable Chinese Metaphor Identification via LLM-Assisted MIPVU Rule Script Generation: A Comparative Protocol Study

This paper introduces an interpretable, LLM-assisted pipeline that operationalizes four distinct metaphor identification protocols as executable rule scripts for Chinese, demonstrating through a comparative study that the choice of protocol is the primary source of variation in identification results while achieving competitive performance with full transparency and reproducibility.

Weihang Huang, Mengna LiuThu, 12 Ma💬 cs.CL

Towards Cold-Start Drafting and Continual Refining: A Value-Driven Memory Approach with Application to NPU Kernel Synthesis

The paper introduces EvoKernel, a self-evolving agentic framework that leverages value-driven memory and reinforcement learning to overcome data scarcity in NPU kernel synthesis, significantly improving model correctness and achieving substantial speedups through automated drafting and iterative refinement.

Yujie Zheng, Zhuo Li, Shengtao Zhang, Hanjing Wang, Junjie Sheng, Jiaqian Wang, Junchi Yan, Weinan Zhang, Ying Wen, Bo Tang, Muning WenThu, 12 Ma🤖 cs.LG

V0.5V_{0.5}: Generalist Value Model as a Prior for Sparse RL Rollouts

The paper proposes V0.5V_{0.5}, a novel method that dynamically fuses a Generalist Value Model's prior with sparse RL rollouts via real-time statistical testing to minimize baseline estimation error, thereby achieving faster convergence and over 10% performance gains on mathematical reasoning benchmarks compared to GRPO and DAPO.

Yi-Kai Zhang, Yueqing Sun, Hongyan Hao, Qi Gu, Xunliang Cai, De-Chuan Zhan, Han-Jia YeThu, 12 Ma🤖 cs.LG

An Extreme Multi-label Text Classification (XMTC) Library Dataset: What if we took "Use of Practical AI in Digital Libraries" seriously?

This paper introduces a large bilingual (English/German) corpus of catalog records annotated with the Integrated Authority File (GND) and a machine-actionable GND taxonomy to enable ontology-aware multi-label classification and agent-assisted cataloging, aiming to develop transparent, authority-anchored AI tools that enhance the efficiency and scalability of subject indexing in digital libraries.

Jennifer D'Souza, Sameer Sadruddin, Maximilian Kähler, Andrea Salfinger, Luca Zaccagna, Francesca Incitti, Lauro Snidaro, Osma SuominenThu, 12 Ma💬 cs.CL

From Images to Words: Efficient Cross-Modal Knowledge Distillation to Language Models from Black-box Teachers

The paper introduces ARMADA, an efficient cross-modal knowledge distillation framework that transfers knowledge from large, potentially black-box vision-language models to language-only models without requiring teacher pre-training or internal access, thereby significantly improving performance across diverse natural language tasks.

Ayan Sengupta, Shantanu Dixit, Md Shad Akhtar, Tanmoy ChakrabortyThu, 12 Ma💬 cs.CL

GLM-OCR Technical Report

GLM-OCR is a compact 0.9B-parameter multimodal model that leverages a Multi-Token Prediction mechanism and a two-stage pipeline to achieve state-of-the-art efficiency and performance in real-world document understanding tasks, making it suitable for both edge and large-scale deployments.

Shuaiqi Duan, Yadong Xue, Weihan Wang, Zhe Su, Huan Liu, Sheng Yang, Guobing Gan, Guo Wang, Zihan Wang, Shengdong Yan, Dexin Jin, Yuxuan Zhang, Guohong Wen, Yanfeng Wang, Yutao Zhang, Xiaohan Zhang, Wenyi Hong, Yukuo Cen, Da Yin, Bin Chen, Wenmeng Yu, Xiaotao Gu, Jie TangThu, 12 Ma💬 cs.CL

LLM2Vec-Gen: Generative Embeddings from Large Language Models

LLM2Vec-Gen introduces a novel self-supervised framework that generates high-quality, interpretable text embeddings by training special tokens to represent an LLM's potential responses, thereby achieving state-of-the-art performance on MTEB while transferring safety and reasoning capabilities without requiring labeled data or a frozen backbone.

Parishad BehnamGhader, Vaibhav Adlakha, Fabian David Schmidt, Nicolas Chapados, Marius Mosbach, Siva ReddyThu, 12 Ma💬 cs.CL