Addressing the Ecological Fallacy in Larger LMs with Human Context

This paper demonstrates that addressing the ecological fallacy by modeling an author's language context through a specific task called HuLM, particularly during fine-tuning (HuFT) or continued pre-training, significantly improves the performance of an 8B Llama model across multiple downstream tasks compared to standard training methods.

Nikita Soni, Dhruv Vijay Kunjadiya, Pratham Piyush Shah, Dikshya Mohanty, H. Andrew Schwartz, Niranjan Balasubramanian2026-03-09🤖 cs.AI

Who We Are, Where We Are: Mental Health at the Intersection of Person, Situation, and Large Language Models

This paper proposes an interpretable modeling approach that integrates person-level psychological traits with situational context features derived from social media data to predict dynamic mental well-being, demonstrating that theory-driven methods offer competitive performance and greater human-understandable insights compared to standard language model embeddings.

Nikita Soni, August Håkan Nilsson, Syeda Mahwish, Vasudha Varadarajan, H. Andrew Schwartz, Ryan L. Boyd2026-03-09🤖 cs.AI

MASFactory: A Graph-centric Framework for Orchestrating LLM-Based Multi-Agent Systems with Vibe Graphing

This paper introduces MASFactory, a graph-centric framework that utilizes a human-in-the-loop "Vibe Graphing" approach to automatically compile natural language intents into executable multi-agent system workflows, thereby addressing challenges in manual implementation, component reuse, and context integration while demonstrating effectiveness across seven public benchmarks.

Yang Liu, Jinxuan Cai, Yishen Li, Qi Meng, Zedi Liu, Xin Li, Chen Qian, Chuan Shi, Cheng Yang2026-03-09🤖 cs.AI

Evaluating Austrian A-Level German Essays with Large Language Models for Automated Essay Scoring

This paper evaluates the performance of four state-of-the-art open-weight Large Language Models on Austrian A-level German essay grading using rubric-based prompts, finding that while they can apply standardized criteria, their low agreement rates with human experts (maximum 40.6% on sub-dimensions and 32.8% on final grades) render them currently unsuitable for real-world automated scoring.

Jonas Kubesch, Lena Huber, Clemens Havas2026-03-09🤖 cs.AI

Experiences Build Characters: The Linguistic Origins and Functional Impact of LLM Personality

This study demonstrates that exposing Large Language Models to domain-specific texts via continued pre-training shapes distinct machine personalities that influence problem-solving, revealing a "Suppression Advantage" where reduced social traits enhance complex reasoning while identifying a bimodal competence peak between "Expressive Generalists" and "Suppressed Specialists."

Xi Wang, Mengdie Zhuang, Jiqun Liu2026-03-09🤖 cs.AI

Contrastive-to-Self-Supervised: A Two-Stage Framework for Script Similarity Learning

This paper proposes a two-stage framework that first trains a contrastive encoder on labeled invented alphabets and then uses teacher-student distillation to learn unsupervised, deformation-invariant embeddings for historically attested scripts, effectively bridging supervised discriminative learning with unsupervised discovery of latent cross-script similarities without requiring ground-truth evolutionary relationships.

Claire Roman, Philippe Meyer2026-03-09🤖 cs.AI

CRIMSON: A Clinically-Grounded LLM-Based Metric for Generative Radiology Report Evaluation

This paper introduces CRIMSON, a clinically grounded evaluation framework for chest X-ray report generation that leverages patient context, guideline-based severity weighting, and a comprehensive error taxonomy to achieve superior alignment with radiologist judgments compared to existing metrics.

Mohammed Baharoon, Thibault Heintz, Siavash Raissi, Mahmoud Alabbad, Mona Alhammad, Hassan AlOmaish, Sung Eun Kim, Oishi Banerjee, Pranav Rajpurkar2026-03-09🤖 cs.AI

LIT-RAGBench: Benchmarking Generator Capabilities of Large Language Models in Retrieval-Augmented Generation

This paper introduces LIT-RAGBench, a comprehensive benchmark dataset and evaluation framework designed to systematically assess Large Language Models' capabilities in Retrieval-Augmented Generation across five critical categories—Integration, Reasoning, Logic, Table, and Abstention—using a mix of human-constructed Japanese questions and curated English translations to guide model selection and development for practical RAG deployments.

Koki Itai, Shunichi Hasegawa, Yuta Yamamoto, Gouki Minegishi, Masaki Otsuki2026-03-09💬 cs.CL

FlashPrefill: Instantaneous Pattern Discovery and Thresholding for Ultra-Fast Long-Context Prefilling

FlashPrefill is a novel framework that achieves ultra-fast long-context prefilling by combining instantaneous block-searching for dynamic sparse patterns with a thresholding mechanism to eliminate long-tail attention scores, delivering up to a 27.78x speedup on 256K sequences while maintaining efficiency on shorter contexts.

Qihang Fan, Huaibo Huang, Zhiying Wu, Juqiu Wang, Bingning Wang, Ran He2026-03-09🤖 cs.AI