A Survey on Decentralized Federated Learning

This survey systematically reviews decentralized federated learning methods from 2018 to early 2026, categorizing them into traditional distributed and blockchain-based architectures, proposing a unified challenge-driven taxonomy, and outlining future research directions to address security, privacy, and system-level trade-offs in coordinator-free settings.

Edoardo Gabrielli, Anthony Di Pietro, Dario Fenoglio, Giovanni Pica, Gabriele TolomeiWed, 11 Ma🤖 cs.LG

XConv: Low-memory stochastic backpropagation for convolutional layers

XConv is a drop-in replacement for standard convolutional layers that significantly reduces memory usage during training by storing compressed activations and approximating weight gradients via randomized trace estimation, while maintaining performance comparable to exact gradient methods without imposing architectural constraints or requiring codebase modifications.

Anirudh Thatipelli, Jeffrey Sam, Mathias Louboutin, Ali Siahkoohi, Rongrong Wang, Felix J. HerrmannWed, 11 Ma🤖 cs.LG

From Semantics to Pixels: Coarse-to-Fine Masked Autoencoders for Hierarchical Visual Understanding

The paper proposes C2FMAE, a coarse-to-fine masked autoencoder that resolves the tension between global semantics and local details in self-supervised learning by employing a cascaded decoder and progressive masking curriculum on a newly constructed multi-granular dataset to achieve hierarchical visual understanding and superior performance across various vision tasks.

Wenzhao Xiang, Yue Wu, Hongyang Yu, Feng Gao, Fan Yang, Xilin ChenWed, 11 Ma🤖 cs.LG

What is Missing? Explaining Neurons Activated by Absent Concepts

This paper identifies that deep neural networks frequently encode the absence of concepts to drive neuron activation—a phenomenon largely overlooked by standard explainable AI methods—and proposes simple extensions to attribution and feature visualization techniques to effectively reveal and leverage these "missing" concepts for better model interpretation and debiasing.

Robin Hesse, Simone Schaub-Meyer, Janina Hesse, Bernt Schiele, Stefan RothWed, 11 Ma🤖 cs.LG

SCDP: Learning Humanoid Locomotion from Partial Observations via Mixed-Observation Distillation

The paper introduces Sensor-Conditioned Diffusion Policies (SCDP), a novel framework that enables robust humanoid locomotion using only onboard sensors by distilling privileged full-body knowledge through mixed-observation training and specialized denoising techniques, successfully achieving near-perfect simulation performance and real-world deployment on a G1 robot without explicit state estimation.

Milo Carroll, Tianhu Peng, Lingfan Bao, Chengxu Zhou, Zhibin LiWed, 11 Ma🤖 cs.LG

What Do We Care About in Bandits with Noncompliance? BRACE: Bandits with Recommendations, Abstention, and Certified Effects

This paper introduces BRACE, a parameter-free algorithm for multi-armed bandits with noncompliance that simultaneously optimizes recommendation welfare and treatment learning by performing certified instrumental variable inversion only when identification is strong, otherwise providing honest structural intervals to navigate the trade-offs between mediated and direct-control regimes.

Nicolás Della PennaWed, 11 Ma🤖 cs.LG

You Didn't Have to Say It like That: Subliminal Learning from Faithful Paraphrases

This paper demonstrates that language models can covertly acquire behavioral traits from a teacher model through "subliminal learning" on faithful paraphrases, where the student adopts the teacher's preferences even when the paraphrased content is semantically unrelated or explicitly contradicts those preferences, rendering content-based inspection ineffective.

Isaia Gisler (ETH Zürich), Zhonghao He (University of Cambridge), Tianyi Qiu (Peking University)Wed, 11 Ma🤖 cs.LG

TrainDeeploy: Hardware-Accelerated Parameter-Efficient Fine-Tuning of Small Transformer Models at the Extreme Edge

TrainDeeploy is a novel framework that enables efficient, parameter-efficient on-device fine-tuning of both CNN and Transformer models on ultra-low-power, memory-constrained RISC-V SoCs, achieving significant reductions in memory usage and computational overhead while supporting end-to-end training at the extreme edge.

Run Wang, Victor J. B. Jung, Philip Wiese, Francesco Conti, Alessio Burrello, Luca BeniniWed, 11 Ma🤖 cs.LG

Flow Field Reconstruction via Voronoi-Enhanced Physics-Informed Neural Networks with End-to-End Sensor Placement Optimization

This paper proposes VSOPINN, a novel framework that integrates differentiable Voronoi tessellation with Physics-Informed Neural Networks to enable end-to-end optimization of sensor placement, thereby significantly enhancing the accuracy and robustness of high-fidelity flow field reconstruction under sparse measurements and sensor failures.

Renjie Xiao, Bingteng Sun, Yiling Chen, Lin Lu, Qiang Du, Junqiang ZhuWed, 11 Ma🤖 cs.LG

MM-Zero: Self-Evolving Multi-Model Vision Language Models From Zero Data

MM-Zero is the first RL-based framework to enable Vision Language Models to self-evolve from zero data by employing a multi-role system (Proposer, Coder, and Solver) trained with Group Relative Policy Optimization to generate visual concepts, render them via code, and solve multimodal reasoning tasks without any seed images.

Zongxia Li, Hongyang Du, Chengsong Huang, Xiyang Wu, Lantao Yu, Yicheng He, Jing Xie, Xiaomin Wu, Zhichao Liu, Jiarui Zhang, Fuxiao LiuWed, 11 Ma🤖 cs.LG

Verifying Good Regulator Conditions for Hypergraph Observers: Natural Gradient Learning from Causal Invariance via Established Theorems

This paper verifies that persistent observers in causally invariant hypergraph substrates satisfy the Conant-Ashby Good Regulator Theorem, thereby necessitating internal models that lead to natural gradient descent as the unique learning rule and yielding a model-dependent closed-form formula for Vanchurin's regime parameter α\alpha with a quantum-classical threshold at κ(F)=2\kappa(F)=2.

Max ZhuravlevWed, 11 Ma🤖 cs.LG

Adaptive Active Learning for Online Reliability Prediction of Satellite Electronics

This paper proposes a novel integrated online reliability prediction framework for satellite electronics that combines a Wiener process-based degradation model with a two-stage adaptive active learning strategy to significantly improve prediction accuracy while reducing data requirements under limited and variable operational conditions.

Shixiang Li, Yubin Tian, Dianpeng Wang, Piao Chen, Mengying RenWed, 11 Ma🤖 cs.LG

Quality over Quantity: Demonstration Curation via Influence Functions for Data-Centric Robot Learning

This paper introduces Quality over Quantity (QoQ), a systematic framework that leverages influence functions to automatically curate high-quality robot learning demonstrations by quantifying each sample's contribution to reducing validation loss, thereby significantly improving policy performance over manual or heuristic data selection methods.

Haeone Lee, Taywon Min, Junsu Kim, Sinjae Kang, Fangchen Liu, Lerrel Pinto, Kimin LeeWed, 11 Ma🤖 cs.LG

FlexServe: A Fast and Secure LLM Serving System for Mobile Devices with Flexible Resource Isolation

This paper presents FlexServe, a high-performance and secure LLM serving system for mobile devices that leverages a novel Flexible Resource Isolation mechanism to overcome the significant overhead of ARM TrustZone, achieving up to 10.05× faster time-to-first-token and 24.30× faster multi-model workflow execution compared to baseline designs.

Yinpeng Wu, Yitong Chen, Lixiang Wang, Jinyu Gu, Zhichao Hua, Yubin XiaWed, 11 Ma🤖 cs.LG