CLAD-Net: Continual Activity Recognition in Multi-Sensor Wearable Systems

CLAD-Net is a continual learning framework for wearable human activity recognition that combines a self-supervised transformer for long-term memory and a supervised CNN with knowledge distillation to effectively mitigate catastrophic forgetting and handle label scarcity across diverse subjects.

Reza Rahimi Azghan, Gautham Krishna Gudur, Mohit Malu, Edison Thomaz, Giulia Pedrielli, Pavan Turaga, Hassan Ghasemzadeh2026-03-10🤖 cs.LG

Generative Evolutionary Meta-Solver (GEMS): Scalable Surrogate-Free Multi-Agent Reinforcement Learning

The paper introduces Generative Evolutionary Meta-Solver (GEMS), a scalable, surrogate-free multi-agent reinforcement learning framework that replaces explicit policy populations with a compact generator and latent anchors to achieve significantly faster training, lower memory usage, and higher rewards than traditional methods like PSRO while maintaining game-theoretic guarantees.

Alakh Sharma, Gaurish Trivedi, Kartikey Singh Bhandari, Yash Sinha, Dhruv Kumar, Pratik Narang, Jagat Sesh Challa2026-03-10🤖 cs.LG

Overlap-Adaptive Regularization for Conditional Average Treatment Effect Estimation

This paper introduces Overlap-Adaptive Regularization (OAR), a novel method that enhances the performance of existing CATE meta-learners in low-overlap regions by proportionally increasing regularization based on overlap weights, while offering flexible, debiased variants that preserve Neyman-orthogonality for robust inference.

Valentyn Melnychuk, Dennis Frauen, Jonas Schweisthal, Stefan Feuerriegel2026-03-10🤖 cs.LG

Your Agent May Misevolve: Emergent Risks in Self-evolving LLM Agents

This paper introduces and empirically validates the concept of "misevolution," demonstrating that self-evolving LLM agents face widespread, emergent risks across model, memory, tool, and workflow pathways that can lead to safety degradation and unintended vulnerabilities, thereby highlighting an urgent need for new safety paradigms.

Shuai Shao, Qihan Ren, Chen Qian, Boyi Wei, Dadi Guo, Jingyi Yang, Xinhao Song, Linfeng Zhang, Weinan Zhang, Dongrui Liu, Jing Shao2026-03-10🤖 cs.LG

Pretraining in Actor-Critic Reinforcement Learning for Robot Locomotion

This paper proposes a pretraining-finetuning paradigm for robot locomotion that leverages a task-agnostic exploration strategy to train a Proprioceptive Inverse Dynamics Model (PIDM), which is then used to warm-start actor-critic algorithms like PPO, resulting in significant improvements in sample efficiency and task performance across diverse robot embodiments.

Jiale Fan, Andrei Cramariuc, Tifanny Portela, Marco Hutter2026-03-10🤖 cs.LG