Repurposing Backdoors for Good: Ephemeral Intrinsic Proofs for Verifiable Aggregation in Cross-silo Federated Learning

This paper proposes a lightweight, backdoor-based "Intrinsic Proofs" framework that ensures verifiable aggregation in cross-silo federated learning by embedding ephemeral verification signals into model parameters, thereby achieving high detection rates against malicious servers with over 1000x speedup compared to traditional cryptographic methods while preserving client anonymity and final model utility.

Xian Qin, Xue Yang, Xiaohu Tang2026-03-12🤖 cs.AI

EvoSchema: Towards Text-to-SQL Robustness Against Schema Evolution

This paper introduces EvoSchema, a comprehensive benchmark featuring a novel taxonomy of ten schema perturbation types to evaluate and enhance the robustness of text-to-SQL models against real-world database schema evolution, revealing that table-level changes significantly impact performance and demonstrating that training on diverse schema designs improves model resilience.

Tianshu Zhang, Kun Qian, Siddhartha Sahai, Yuan Tian, Shaddy Garg, Huan Sun, Yunyao Li2026-03-12💬 cs.CL

UAV traffic scene understanding: A cross-spectral guided approach and a unified benchmark

This paper proposes CTCNet, a novel cross-spectral guided network featuring a Prototype-Guided Knowledge Embedding module and a Quality-Aware Spectral Compensation module to enhance UAV traffic scene understanding under adverse conditions, accompanied by the introduction of Traffic-VQA, the first large-scale optical-thermal benchmark for cognitive traffic analysis.

Yu Zhang, Zhicheng Zhao, Ze Luo, Chenglong Li, Jin Tang2026-03-12🤖 cs.AI

Deep Randomized Distributed Function Computation (DeepRDFC): Neural Distributed Channel Simulation

This paper proposes a deep learning-based autoencoder architecture for the Randomized Distributed Function Computation (RDFC) framework that minimizes the total variation distance to an unknown target distribution using only data samples, demonstrating superior communication efficiency compared to traditional data compression methods, particularly under limited common randomness.

Didrik Bergström, Onur Günlü2026-03-12🔢 math

Risk-Adjusted Harm Scoring for Automated Red Teaming for LLMs in Financial Services

This paper introduces a risk-aware evaluation framework for Large Language Models in financial services, featuring a domain-specific taxonomy, an automated multi-round red-teaming pipeline, and a Risk-Adjusted Harm Score (RAHS) metric to better capture and quantify severe, operationally actionable security failures that traditional domain-agnostic benchmarks miss.

Fabrizio Dimino, Bhaskarjit Sarmah, Stefano Pasquali2026-03-12💰 q-fin