A Novel Multi-Agent Architecture to Reduce Hallucinations of Large Language Models in Multi-Step Structural Modeling

This paper proposes a novel multi-agent architecture that automates structural modeling and analysis using OpenSeesPy by decomposing complex tasks into specialized agents to effectively reduce hallucinations and error accumulation, achieving high accuracy and scalability across benchmark frame problems.

Ziheng Geng, Jiachen Liu, Ran Cao, Lu Cheng, Dan M. Frangopol, Minghui Cheng2026-03-10💻 cs

Large Language Model for Discrete Optimization Problems: Evaluation and Step-by-step Reasoning

This paper evaluates the capabilities of various large language models, including Llama-3 and ChatGPT, in solving diverse discrete optimization problems using natural language datasets, revealing that while stronger models generally perform better, Chain-of-Thought reasoning is not universally effective and data augmentation can improve performance on simpler tasks despite introducing instability.

Tianhao Qian, Guilin Qi, Z. Y. Wu, Ran Gu, Xuanyi Liu, Canchen Lyu2026-03-10💬 cs.CL

Hide and Find: A Distributed Adversarial Attack on Federated Graph Learning

The paper proposes FedShift, a novel two-stage "Hide and Find" distributed adversarial attack for Federated Graph Learning that injects hidden shifters to stealthily guide poisoned data toward a target boundary and efficiently generates perturbations, achieving superior effectiveness, robustness against defenses, and a 90% reduction in convergence time compared to existing methods.

Jinshan Liu, Ken Li, Jiazhe Wei, Bin Shi, Bo Dong2026-03-10🤖 cs.LG

DECADE: A Temporally-Consistent Unsupervised Diffusion Model for Enhanced Rb-82 Dynamic Cardiac PET Image Denoising

The paper proposes DECADE, an unsupervised diffusion model that achieves temporally consistent denoising of Rb-82 dynamic cardiac PET images without paired training data, effectively reducing noise while preserving quantitative accuracy for myocardial blood flow and flow reserve metrics.

Yinchi Zhou, Liang Guo, Huidong Xie, Yuexi Du, Ashley Wang, Menghua Xia, Tian Yu, Ramesh Fazzone-Chettiar, Christopher Weyman, Bruce Spottiswoode, Vladimir Panin, Kuangyu Shi, Edward J. Miller, Attila Feher, Albert J. Sinusas, Nicha C. Dvornek, Chi Liu2026-03-10💻 cs

QuadAI at SemEval-2026 Task 3: Ensemble Learning of Hybrid RoBERTa and LLMs for Dimensional Aspect-Based Sentiment Analysis

The QuadAI system for SemEval-2026 Task 3 achieves superior performance in dimensional aspect-based sentiment regression by employing an ensemble learning framework that combines a hybrid RoBERTa encoder with large language models, leveraging the complementary strengths of both architectures to significantly reduce RMSE and improve correlation scores.

A. J. W. de Vink, Filippos Karolos Ventirozos, Natalia Amat-Lefort, Lifeng Han2026-03-10💬 cs.CL

Dual-Metric Evaluation of Social Bias in Large Language Models: Evidence from an Underrepresented Nepali Cultural Context

This study evaluates seven state-of-the-art large language models in the underrepresented Nepali cultural context using a Dual-Metric Bias Assessment framework, revealing that while explicit agreement with biased statements is measurable, implicit generative bias is distinct, follows a non-linear relationship with temperature, and is poorly predicted by agreement metrics, thereby highlighting the critical need for culturally grounded datasets and evaluation strategies.

Ashish Pandey, Tek Raj Chhetri2026-03-10💬 cs.CL

Gradient Iterated Temporal-Difference Learning

This paper introduces Gradient Iterated Temporal-Difference (GTD) learning, a novel algorithm that modifies iterated TD by computing gradients over moving targets to achieve the stability of gradient methods while matching the competitive learning speed of semi-gradient methods across diverse benchmarks like Atari games.

Théo Vincent, Kevin Gerhardt, Yogesh Tripathi, Habib Maraqten, Adam White, Martha White, Jan Peters, Carlo D'Eramo2026-03-10🤖 cs.LG

AI Steerability 360: A Toolkit for Steering Large Language Models

The paper introduces AI Steerability 360, an open-source, Hugging Face-native Python toolkit that provides a unified interface for composing, evaluating, and comparing diverse large language model steering methods across input, structural, state, and output control surfaces.

Erik Miehling, Karthikeyan Natesan Ramamurthy, Praveen Venkateswaran, Irene Ko, Pierre Dognin, Moninder Singh, Tejaswini Pedapati, Avinash Balakrishnan, Matthew Riemer, Dennis Wei, Inge Vejsbjerg, Elizabeth M. Daly, Kush R. Varshney2026-03-10💬 cs.CL

SynPlanResearch-R1: Encouraging Tool Exploration for Deep Research with Synthetic Plans

The paper introduces SynPlanResearch-R1, a framework that synthesizes tool-use trajectories to encourage deeper exploration during supervised fine-tuning, thereby overcoming the limitations of reinforcement learning with verifiable rewards and significantly improving research agent performance across multiple benchmarks.

Hansi Zeng, Zoey Li, Yifan Gao, Chenwei Zhang, Xiaoman Pan, Tao Yang, Fengran Mo, Jiacheng Lin, Xian Li, Jingbo Shang2026-03-10💬 cs.CL

Hospitality-VQA: Decision-Oriented Informativeness Evaluation for Vision-Language Models

This paper introduces a formal framework for "informativeness" and a corresponding hospitality-specific VQA dataset to evaluate Vision-Language Models, revealing that while current models struggle with decision-oriented reasoning, their performance significantly improves with modest domain-specific finetuning.

Jeongwoo Lee, Baek Duhyeong, Eungyeol Han, Soyeon Shin, Gukin han, Seungduk Kim, Jaehyun Jeon, Taewoo Jeong2026-03-10🤖 cs.LG