Adversarial Latent-State Training for Robust Policies in Partially Observable Domains
This paper introduces an adversarial latent-initial-state POMDP framework that theoretically establishes a minimax principle and finite-sample guarantees, while empirically demonstrating that targeted adversarial training significantly reduces robustness gaps in partially observable reinforcement learning.