When Sensors Fail: Temporal Sequence Models for Robust PPO under Sensor Drift
This paper proposes augmenting Proximal Policy Optimization with temporal sequence models, particularly Transformers, to enable robust reinforcement learning under sensor drift and partial observability by inferring missing information from history, a claim supported by theoretical bounds on reward degradation and empirical success on MuJoCo benchmarks.