Template-assisted Contrastive Learning of Task-oriented Dialogue Sentence Embeddings

This paper introduces Template-aware Dialogue Sentence Embedding (TaDSE), a novel self-supervised contrastive learning method that leverages easily obtainable token-level template information to generate high-quality sentence embeddings for task-oriented dialogues, achieving significant performance improvements over state-of-the-art methods on five benchmark datasets.

Minsik Oh, Jiwei Li, Guoyin Wang2026-04-14💬 cs.CL

SCITUNE: Aligning Large Language Models with Human-Curated Scientific Multimodal Instructions

The paper introduces SciTune, a framework that aligns large language models with human-curated scientific multimodal instructions, resulting in a model (LLaMA-SciTune) that significantly outperforms state-of-the-art systems on scientific visual and language benchmarks, even surpassing human performance in certain categories.

Sameera Horawalavithana, Sai Munikoti, Ian Stewart, Henry Kvinge, Karl Pazdernik2026-04-14💬 cs.CL

RiTeK: A Dataset for Large Language Models Complex Reasoning over Textual Knowledge Graphs in Medicine

This paper introduces RiTeK, a comprehensive benchmark dataset designed to evaluate and improve Large Language Models' complex reasoning capabilities over medical Textual Knowledge Graphs by addressing current limitations in data scarcity, topological expressiveness, and retrieval performance.

Jiatan Huang, Mingchen Li, Zonghai Yao, Dawei Li, Yuxin Zhang, Zhichao Yang, Yongkang Xiao, Feiyun Ouyang, Xiaohan Li, Shuo Han, Hong Yu2026-04-14💬 cs.CL

LIFT: A Novel Framework for Enhancing Long-Context Understanding of LLMs via Long Input Fine-Tuning

This paper introduces LIFT, a novel framework that enhances the long-context understanding of short-context large language models by dynamically fine-tuning model parameters to absorb long inputs via synthetic tasks, thereby enabling inference without the input context while avoiding quadratic complexity and maintaining low latency.

Yansheng Mao, Yufei Xu, Jiaqi Li, Fanxu Meng, Haotong Yang, Zilong Zheng, Xiyuan Wang, Muhan Zhang2026-04-14💬 cs.CL

Challenging the Boundaries of Reasoning: An Olympiad-Level Math Benchmark for Large Language Models

The paper introduces OlymMATH, a rigorously curated Olympiad-level math benchmark featuring 350 manually sourced problems in dual English and Chinese versions that support both natural language and formal Lean 4 verification to address the saturation of existing evaluations and reveal persistent reasoning gaps in large language models.

Haoxiang Sun, Yingqian Min, Zhipeng Chen, Wayne Xin Zhao, Ji-Rong Wen2026-04-14💬 cs.CL

If an LLM Were a Character, Would It Know Its Own Story? Evaluating Lifelong Learning in LLMs

This paper introduces LIFESTATE-BENCH, a novel benchmark utilizing narrative datasets like Hamlet to evaluate lifelong learning in large language models, revealing that while non-parametric methods outperform parametric ones in managing stateful interactions, all models still struggle with catastrophic forgetting over extended engagements.

Siqi Fan, Xiusheng Huang, Yiqun Yao, Xuezhi Fang, Kang Liu, Peng Han, Shuo Shang, Aixin Sun, Yequan Wang2026-04-14💬 cs.CL

Seeing Through Deception: Uncovering Misleading Creator Intent in Multimodal News with Vision-Language Models

This paper introduces DeceptionDecoded, a large-scale benchmark and intent-guided simulation framework designed to evaluate and improve vision-language models' ability to detect misleading creator intent in multimodal news, addressing their current reliance on superficial cues and enhancing their robustness in misinformation governance.

Jiaying Wu, Fanxiao Li, Zihang Fu, Min-Yen Kan, Bryan Hooi2026-04-14💬 cs.CL

GoT-R1: Unleashing Reasoning Capability of MLLM for Visual Generation with Reinforcement Learning

GoT-R1 is a novel framework that leverages reinforcement learning with a dual-stage multi-dimensional reward system to enhance the semantic-spatial reasoning capabilities of multimodal large language models, significantly improving their ability to generate images from complex prompts involving precise object relationships and attributes.

Chengqi Duan, Rongyao Fang, Yuqing Wang, Kun Wang, Linjiang Huang, Xingyu Zeng, Hongsheng Li, Xihui Liu2026-04-14💬 cs.CL