LatentLens: Revealing Highly Interpretable Visual Tokens in LLMs
This paper introduces LatentLens, a novel interpretability method that reveals the semantic meaning of visual tokens across all layers of Vision-Language Models by matching them to contextualized text representations, demonstrating that visual tokens are far more interpretable than previously believed.