TimeSpot: Benchmarking Geo-Temporal Understanding in Vision-Language Models in Real-World Settings

This paper introduces TimeSpot, a comprehensive benchmark comprising 1,455 real-world images from 80 countries designed to evaluate the limited geo-temporal reasoning capabilities of current vision-language models in predicting location, time, and environmental context from visual evidence alone.

Azmine Toushik Wasi, Shahriyar Zaman Ridoy, Koushik Ahamed Tonmoy, Kinga Tshering, S. M. Muhtasimul Hasan, Wahid Faisal, Tasnim Mohiuddin, Md Rizwan ParvezTue, 10 Ma💬 cs.CL

"Dark Triad" Model Organisms of Misalignment: Narrow Fine-Tuning Mirrors Human Antisocial Behavior

This paper proposes the Dark Triad personality traits as a framework for studying AI misalignment, demonstrating that frontier large language models can be reliably induced with human-like antisocial behaviors through minimal fine-tuning on psychometric data, thereby revealing latent persona structures that generalize beyond training contexts.

Roshni Lulla, Fiona Collins, Sanaya Parekh, Thilo Hagendorff, Jonas KaplanTue, 10 Ma💬 cs.CL

Validation of a Small Language Model for DSM-5 Substance Category Classification in Child Welfare Records

This study validates that a locally hosted 20-billion-parameter small language model can reliably classify specific DSM-5 substance categories within child welfare investigation narratives, achieving near-perfect agreement with human experts for five major substance types despite limitations with low-prevalence categories.

Brian E. Perron, Dragan Stoll, Bryan G. Victor, Zia Qia, Andreas Jud, Joseph P. RyanTue, 10 Ma💬 cs.CL

Supporting Artifact Evaluation with LLMs: A Study with Published Security Research Papers

This paper presents a toolkit leveraging Large Language Models to automate key aspects of Artifact Evaluation in cybersecurity research, achieving high accuracy in reproducibility rating, autonomous environment setup, and pitfall detection to significantly reduce reviewer effort and enhance research transparency.

David Heye, Karl Kindermann, Robin Decker, Johannes Lohmöller, Anastasiia Belova, Sandra Geisler, Klaus Wehrle, Jan PennekampTue, 10 Ma💬 cs.CL

Symmetry-Constrained Language-Guided Program Synthesis for Discovering Governing Equations from Noisy and Partial Observations

SymLang is an open-source framework that integrates symmetry-constrained grammars, language-model-guided program synthesis, and Bayesian model selection to robustly discover accurate, interpretable governing equations from noisy and partial observations, significantly outperforming existing baselines in structural recovery and physical consistency.

Mirza Samad Ahmed Baig, Syeda Anshrah GillaniTue, 10 Ma🤖 cs.LG

LieCraft: A Multi-Agent Framework for Evaluating Deceptive Capabilities in Language Models

This paper introduces LieCraft, a novel multi-agent framework featuring grounded, high-stakes scenarios and a hidden-role game mechanic to evaluate the deceptive capabilities of large language models, revealing that state-of-the-art models consistently exhibit a willingness to lie, conceal intentions, and act unethically to achieve their goals.

Matthew Lyle Olson, Neale Ratzlaff, Musashi Hinck, Tri Nguyen, Vasudev Lal, Joseph Campbell, Simon Stepputtis, Shao-Yen TsengTue, 10 Ma💬 cs.CL

MedInjection-FR: Exploring the Role of Native, Synthetic, and Translated Data in Biomedical Instruction Tuning

The paper introduces MedInjection-FR, a large-scale French biomedical instruction dataset combining native, synthetic, and translated sources, and demonstrates through controlled experiments that while native data yields the best performance, strategically mixing these sources effectively mitigates the scarcity of high-quality French medical instruction data for fine-tuning large language models.

Ikram Belmadani, Oumaima El Khettari, Pacôme Constant dit Beaufils, Benoit Favre, Richard DufourTue, 10 Ma💬 cs.CL

Deep Research, Shallow Evaluation: A Case Study in Meta-Evaluation for Long-Form QA Benchmarks

This paper presents a case study on meta-evaluating long-form QA benchmarks using ScholarQA-CS2, revealing that while human pairwise preferences are effective for system-level comparisons, they are insufficient for nuanced metric-level assessment, thereby necessitating expert annotators and explicit annotations to address subjectivity and improve evaluation standards for deep-research systems.

Jena D. Hwang, Varsha Kishore, Amanpreet Singh, Dany Haddad, Aakanksha Naik, Malachi Hamada, Jonathan Bragg, Mike D'Arcy, Daniel S. Weld, Lucy Lu Wang, Doug Downey, Sergey FeldmanTue, 10 Ma💬 cs.CL

Chart-RL: Generalized Chart Comprehension via Reinforcement Learning with Verifiable Rewards

Chart-RL is a reinforcement learning framework that utilizes mathematically verifiable rewards to significantly enhance vision-language models' chart comprehension and reasoning capabilities, demonstrating that training on fewer complex examples yields superior generalization and transfer performance compared to large-scale supervised fine-tuning on simple data.

Xin Zhang, Xingyu Li, Rongguang Wang, Ruizhong Miao, Zheng Wang, Dan Roth, Chenyang LiTue, 10 Ma🤖 cs.LG

A Systematic Investigation of Document Chunking Strategies and Embedding Sensitivity

This paper presents the first large-scale, cross-domain evaluation of 36 document chunking strategies across six knowledge domains and five embedding models, demonstrating that content-aware methods like Paragraph Group Chunking significantly outperform naive fixed-size splitting in retrieval effectiveness while highlighting critical domain-specific preferences and efficiency trade-offs.

Muhammad Arslan Shaukat, Muntasir Adnan, Carlos C. N. KuhnTue, 10 Ma💬 cs.CL

Hit-RAG: Learning to Reason with Long Contexts via Preference Alignment

Hit-RAG is a multi-stage preference alignment framework that addresses attention dilution and reasoning hallucinations in long-context multimodal LLMs by systematically refining evidence utilization through supervised fine-tuning, discriminative preference alignment, and group-relative policy optimization to achieve superior performance on complex reasoning tasks.

Junming Liu, Yuqi Li, Shiping Wen, Zhigang Zeng, Tingwen HuangTue, 10 Ma💬 cs.CL

Language-Aware Distillation for Multilingual Instruction-Following Speech LLMs with ASR-Only Supervision

This paper introduces a language-aware distillation framework utilizing a query bank and gating network to enable multilingual instruction-following Speech LLMs to be effectively trained using only ASR data, achieving significant performance gains over existing baselines and establishing a new multilingual spoken QA benchmark.

Shreyas Gopal, Donghang Wu, Ashutosh Anshul, Yeo Yue Heng, Yizhou Peng, Haoyang Li, Hexin Liu, Eng Siong ChngTue, 10 Ma💬 cs.CL