GATech at AbjadGenEval Shared Task: Multilingual Embeddings for Arabic Machine-Generated Text Classification

The GATech team's approach to the AbjadGenEval shared task utilized a fine-tuned multilingual E5-large encoder with simple mean pooling to achieve an F1 score of 0.75 for detecting AI-generated Arabic text, finding that this stable baseline outperformed complex pooling strategies likely due to data limitations and a distinct length difference between human-written and machine-generated texts.

Ahmed Khaled Khamis2026-03-12💬 cs.CL

Personalized Group Relative Policy Optimization for Heterogenous Preference Alignment

This paper introduces Personalized Group Relative Policy Optimization (P-GRPO), a novel framework that improves alignment with diverse individual preferences by decoupling advantage estimation from batch statistics and normalizing rewards against preference-group-specific histories, thereby overcoming the limitations of standard GRPO in handling heterogeneous user signals.

Jialu Wang, Heinrich Peters, Asad A. Butt, Navid Hashemi, Alireza Hashemi, Pouya M. Ghari, Joseph Hoover, James Rae, Morteza Dehghani2026-03-12🤖 cs.LG

Defining AI Models and AI Systems: A Framework to Resolve the Boundary Problem

This paper addresses the regulatory ambiguity surrounding "AI models" and "AI systems" by proposing clear conceptual and operational definitions that distinguish trained parameters from broader system components, thereby facilitating the precise allocation of obligations across the AI value chain.

Yuanyuan Sun, Timothy Parker, Lara Gierschmann, Sana Shams, Teo Canmetin, Mathieu Duteil, Rokas Gipiškis, Ze Shen Chin2026-03-12🤖 cs.AI

LWM-Temporal: Sparse Spatio-Temporal Attention for Wireless Channel Representation Learning

LWM-Temporal is a task-agnostic foundation model for wireless channels that leverages a novel Sparse Spatio-Temporal Attention mechanism and physics-informed pretraining to learn universal, geometry-consistent embeddings, achieving superior performance in channel prediction across diverse mobility regimes with limited fine-tuning data.

Sadjad Alikhani, Akshay Malhotra, Shahab Hamidi-Rad, Ahmed Alkhateeb2026-03-12🤖 cs.LG

HTM-EAR: Importance-Preserving Tiered Memory with Hybrid Routing under Saturation

HTM-EAR is a hierarchical tiered memory system that combines HNSW-based working memory with archival storage, importance-aware eviction, and hybrid routing to effectively preserve essential information and maintain high retrieval precision under sustained saturation, significantly outperforming traditional LRU approaches while approaching the performance of unbounded oracle memory.

Shubham Kumar Singh2026-03-12🤖 cs.AI

Evaluating Generalization Mechanisms in Autonomous Cyber Attack Agents

This paper evaluates how autonomous cyber attack agents generalize to unseen IP reassignments in enterprise networks, finding that while prompt-driven LLM agents achieve the highest success rates compared to traditional RL and adaptation methods, they incur significant costs in computational efficiency, transparency, and reliability due to issues like repetition loops.

Ondřej Lukáš, Jihoon Shin, Emilia Rivas, Diego Forni, Maria Rigaki, Carlos Catania, Aritran Piplai, Christopher Kiekintveld, Sebastian Garcia2026-03-12💻 cs

Gated Adaptation for Continual Learning in Human Activity Recognition

This paper proposes a parameter-efficient continual learning framework for Human Activity Recognition that mitigates catastrophic forgetting in domain-incremental scenarios by employing channel-wise gated modulation to adapt frozen pretrained representations through bounded diagonal scaling, thereby achieving superior stability and plasticity with minimal parameter updates.

Reza Rahimi Azghan, Gautham Krishna Gudur, Mohit Malu, Edison Thomaz, Giulia Pedrielli, Pavan Turaga, Hassan Ghasemzadeh2026-03-12🤖 cs.LG

Where Do Flow Semantics Reside? A Protocol-Native Tabular Pretraining Paradigm for Encrypted Traffic Classification

This paper addresses the failure of byte-sequence-based masked modeling in encrypted traffic classification by identifying a mismatch in inductive bias and proposing FlowSem-MAE, a protocol-native tabular masked autoencoder that leverages field-specific semantics and temporal patterns to significantly outperform existing methods with substantially less labeled data.

Sizhe Huang, Shujie Yang2026-03-12🤖 cs.AI

OmniGuide: Universal Guidance Fields for Enhancing Generalist Robot Policies

OmniGuide is a universal framework that enhances the performance of generalist Vision-Language-Action (VLA) policies on complex tasks by integrating diverse guidance sources—such as 3D foundation models and human pose estimators—into a unified differentiable energy field that guides action sampling through task-specific attractors and repellers.

Yunzhou Song, Long Le, Yong-Hyun Park, Jie Wang, Junyao Shi, Lingjie Liu, Jiatao Gu, Eric Eaton, Dinesh Jayaraman, Kostas Daniilidis2026-03-12💻 cs

Cluster-Aware Attention-Based Deep Reinforcement Learning for Pickup and Delivery Problems

This paper proposes CAADRL, a cluster-aware deep reinforcement learning framework that leverages hierarchical encoding and a dynamic dual-decoder to efficiently solve Pickup and Delivery Problems by explicitly modeling multi-scale cluster structures, achieving state-of-the-art performance with significantly lower inference latency than collaborative-search baselines.

Wentao Wang, Lifeng Han, Guangyu Zou2026-03-12🤖 cs.LG

Quantization of Ricci Curvature in Information Geometry

This paper resolves a 20-year-old conjecture by proving that the volume-averaged Ricci scalar of binary Bayesian networks is universally quantized to positive half-integers for tree and complete-graph structures via a Beta function cancellation mechanism, while demonstrating that this quantization fails in general due to loop counterexamples and contrasting the positive curvature of discrete networks with the negative curvature of Gaussian DAGs.

Carlos C. Rodriguez2026-03-12🔢 math

Improving Search Agent with One Line of Code

This paper introduces Search Agent Policy Optimization (SAPO), a method that resolves catastrophic training instability in Tool-based Agentic Reinforcement Learning by applying a conditional token-level KL constraint to prevent Importance Sampling Distribution Drift, achieving significant performance gains with only a single line of code modification to standard GRPO.

Jian Li, Dongsheng Chen, Zhenhua Xu, Yizhang Jin, Jiafu Wu, Chengjie Wang, Xiaotong Yuan, Yabiao Wang2026-03-12🤖 cs.LG