VQ-Style: Disentangling Style and Content in Motion with Residual Quantized Representations
This paper proposes VQ-Style, a novel framework that leverages Residual Vector Quantized Variational Autoencoders combined with contrastive learning and an information leakage loss to effectively disentangle human motion into coarse content and fine style representations, enabling zero-shot style transfer and other applications through a simple Quantized Code Swapping technique.