Automating Skill Acquisition through Large-Scale Mining of Open-Source Agentic Repositories: A Framework for Multi-Agent Procedural Knowledge Extraction

This paper presents a framework for automating the acquisition of specialized procedural agent skills by systematically mining open-source repositories to extract, standardize, and evaluate capabilities like mathematical visualization, demonstrating that such methods can significantly enhance LLM performance in autonomous workflows without requiring model retraining.

Shuzhen Bi, Mengsong Wu, Hao Hao, Keqian Li, Wentao Liu, Siyu Song, Hongbo Zhao, Aimin Zhou2026-03-13🤖 cs.AI

RADAR: Closed-Loop Robotic Data Generation via Semantic Planning and Autonomous Causal Environment Reset

RADAR is a fully autonomous, closed-loop robotic data generation framework that leverages a four-module pipeline—combining vision-language semantic planning, graph neural network policies, automated success evaluation, and a causal state-machine reset mechanism—to overcome human-in-the-loop bottlenecks and achieve high success rates in both simulation and real-world complex manipulation tasks.

Yongzhong Wang, Keyu Zhu, Yong Zhong, Liqiong Wang, Jinyu Yang, Feng Zheng2026-03-13🤖 cs.AI

VisiFold: Long-Term Traffic Forecasting via Temporal Folding Graph and Node Visibility

The paper proposes VisiFold, a novel framework that addresses the computational and dependency challenges of long-term traffic forecasting by introducing a temporal folding graph to consolidate temporal snapshots and a node visibility mechanism to efficiently handle large-scale spatial data, thereby significantly reducing resource consumption while outperforming existing baselines.

Zhiwei Zhang, Xinyi Du, Weihao Wang, Xuanchi Guo, Wenjuan Han2026-03-13🤖 cs.AI

Automated Detection of Malignant Lesions in the Ovary Using Deep Learning Models and XAI

This research utilizes various Convolutional Neural Network architectures and Explainable AI techniques on a histopathology dataset to develop and evaluate an InceptionV3 model that achieves 94% accuracy in the automated detection of malignant ovarian lesions, aiming to improve non-invasive diagnostic procedures.

Md. Hasin Sarwar Ifty, Nisharga Nirjan, Labib Islam, M. A. Diganta, Reeyad Ahmed Ornate, Anika Tasnim, Md. Saiful Islam2026-03-13🤖 cs.AI

You Told Me to Do It: Measuring Instructional Text-induced Private Data Leakage in LLM Agents

This paper identifies and quantifies a critical "Trusted Executor Dilemma" in high-privilege LLM agents, demonstrating through the ReadSecBench benchmark that agents systematically fail to distinguish malicious instructions embedded in documentation from legitimate guidance, leading to high rates of data exfiltration that current defenses cannot reliably detect.

Ching-Yu Kao, Xinfeng Li, Shenyu Dai, Tianze Qiu, Pengcheng Zhou, Eric Hanchen Jiang, Philip Sperl2026-03-13🤖 cs.AI

CreativeBench: Benchmarking and Enhancing Machine Creativity via Self-Evolving Challenges

This paper introduces CreativeBench, a novel benchmark for objectively evaluating machine creativity in code generation through a unified quality-novelty metric, and proposes EvoRePE, an inference-time strategy that leverages self-evolving patterns to enhance creative performance while revealing key insights into how model scaling affects different creativity types.

Zi-Han Wang, Lam Nguyen, Zhengyang Zhao, Mengyue Yang, Chengwei Qin, Yujiu Yang, Linyi Yang2026-03-13🤖 cs.AI

Social, Legal, Ethical, Empathetic and Cultural Norm Operationalisation for AI Agents

This paper proposes a systematic framework for operationalizing social, legal, ethical, empathetic, and cultural (SLEEC) norms into concrete, verifiable requirements for AI agents, while surveying current methods and outlining a research agenda to bridge the gap between abstract normative principles and practical implementation in high-stakes domains.

Radu Calinescu, Ana Cavalcanti, Marsha Chechik, Lina Marsso, Beverley Townsend2026-03-13🤖 cs.AI

AdaFuse: Accelerating Dynamic Adapter Inference via Token-Level Pre-Gating and Fused Kernel Optimization

AdaFuse is a framework that accelerates dynamic adapter inference in Large Language Models by employing a token-level pre-gating strategy to enable a single global routing decision, which is then executed via a custom fused CUDA kernel to reduce decoding latency by over 2.4x while maintaining accuracy.

Qiyang Li, Rui Kong, Yuchen Li, Hengyi Cai, Shuaiqiang Wang, Linghe Kong, Guihai Chen, Dawei Yin2026-03-13🤖 cs.AI

Bielik-Minitron-7B: Compressing Large Language Models via Structured Pruning and Knowledge Distillation for the Polish Language

This paper introduces Bielik-Minitron-7B, a compressed 7.35B-parameter Polish language model created by applying structured pruning and knowledge distillation to the Bielik-11B-v3.0 model, which achieves a 33.4% parameter reduction and up to 50% inference speedup while retaining approximately 90% of the original model's performance.

Remigiusz Kinas, Paweł Kiszczak, Sergio P. Perez, Krzysztof Ociepa, Łukasz Flis, Krzysztof Wróbel, Adrian Gwozdziej2026-03-13💬 cs.CL

Think While Watching: Online Streaming Segment-Level Memory for Multi-Turn Video Reasoning in Multimodal Large Language Models

The paper proposes "Think While Watching," a memory-anchored streaming framework that enables efficient multi-turn video reasoning in multimodal large language models by preserving segment-level memory and overlapping perception with generation, thereby significantly improving accuracy on streaming benchmarks while reducing output tokens.

Lu Wang (The Key Laboratory of Cognition and Decision Intelligence for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China), Zhuoran Jin (The Key Laboratory of Cognition and Decision Intelligence for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China), Yupu Hao (The Key Laboratory of Cognition and Decision Intelligence for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China), Yubo Chen (The Key Laboratory of Cognition and Decision Intelligence for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China), Kang Liu (The Key Laboratory of Cognition and Decision Intelligence for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China), Yulong Ao (Beijing Academy of Artificial Intelligence), Jun Zhao (The Key Laboratory of Cognition and Decision Intelligence for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China)2026-03-13💬 cs.CL

MobileKernelBench: Can LLMs Write Efficient Kernels for Mobile Devices?

This paper introduces MobileKernelBench, a framework revealing that current LLMs struggle to generate efficient mobile kernels due to compilation failures and hallucinations, and proposes MoKA, a multi-agent system that significantly improves compilation success and kernel performance.

Xingze Zou, Jing Wang, Yuhua Zheng, Xueyi Chen, Haolei Bai, Lingcheng Kong, Syed A. R. Abu-Bakar, Zhaode Wang, Chengfei Lv, Haoji Hu, Huan Wang2026-03-13🤖 cs.LG

Fair Learning for Bias Mitigation and Quality Optimization in Paper Recommendation

This paper introduces Fair-PaperRec, an MLP-based model that effectively mitigates demographic biases in paper recommendations by penalizing disparities through intersectional criteria and a customized fairness loss, achieving a significant increase in underrepresented group participation while simultaneously improving overall utility without compromising academic rigor.

Uttamasha Anjally Oyshi, Susan Gauch2026-03-13🤖 cs.AI

Prototype-Based Knowledge Guidance for Fine-Grained Structured Radiology Reporting

The paper proposes ProtoSR, a prototype-based framework that leverages an instruction-tuned LLM to extract visual prototypes from free-text radiology reports, thereby injecting unstructured knowledge to significantly improve fine-grained structured report generation and achieve state-of-the-art performance on the Rad-ReStruct benchmark.

Chantal Pellegrini, Adrian Delchev, Ege Özsoy, Nassir Navab, Matthias Keicher2026-03-13🤖 cs.AI

Effective Resistance Rewiring: A Simple Topological Correction for Over-Squashing

This paper introduces Effective Resistance Rewiring (ERR), a parameter-free topological correction method that iteratively optimizes graph connectivity by adding and removing edges based on global effective resistance to alleviate over-squashing in Graph Neural Networks, while demonstrating that combining this approach with normalization techniques effectively balances the trade-off between improved long-range signal propagation and oversmoothing.

Bertran Miquel-Oliver, Manel Gil-Sorribes, Victor Guallar, Alexis Molina2026-03-13🤖 cs.LG