Into the Rabbit Hull: From Task-Relevant Concepts in DINO to Minkowski Geometry

By applying Sparse Autoencoders to DINOv2, this study reveals that task-specific concepts exhibit functional specialization and a non-sparse, locally connected geometry, leading to the proposal of the Minkowski Representation Hypothesis, which posits that vision transformer tokens are formed by convex mixtures of archetypes within conceptual spaces rather than strict linear sparsity.

Thomas Fel, Binxu Wang, Michael A. Lepori + 8 more2026-03-02🤖 cs.AI

Attentive Feature Aggregation or: How Policies Learn to Stop Worrying about Robustness and Attend to Task-Relevant Visual Cues

This paper introduces Attentive Feature Aggregation (AFA), a lightweight pooling mechanism that enhances the robustness of visuomotor policies trained with pre-trained visual representations by learning to selectively attend to task-relevant cues while ignoring visual distractors, thereby outperforming standard approaches in perturbed environments without requiring expensive data augmentation or model fine-tuning.

Nikolaos Tsagkas, Andreas Sochopoulos, Duolikun Danier + 4 more2026-03-02💻 cs

General vs Domain-Specific CNNs: Understanding Pretraining Effects on Brain MRI Tumor Classification

This study demonstrates that for brain tumor classification on limited MRI data, modern general-purpose CNNs pre-trained on diverse large-scale datasets (specifically ConvNeXt-Tiny) outperform specialized medical-domain pre-trained models (RadImageNet DenseNet121), challenging the assumption that domain-specific pre-training always yields superior results in data-scarce medical imaging scenarios.

Helia Abedini, Saba Rahimi, Reza Vaziri2026-03-02🤖 cs.AI