VINO: Video-driven Invariance for Non-contextual Objects via Structural Prior Guided De-contextualization

VINO is a self-supervised learning framework that overcomes the "co-occurrence trap" in dense video by using a teacher-student distillation approach with structural priors to force representations to focus on foreground objects rather than background context, achieving state-of-the-art unsupervised object discovery performance.

Seul-Ki Yeom, Marcel Simon, Eunbin Lee, Tae-Ho KimTue, 10 Ma💻 cs

LEPA: Learning Geometric Equivariance in Satellite Remote Sensing Data with a Predictive Architecture

This paper introduces LEPA, a learned architecture that conditions on geometric augmentations to accurately predict transformed satellite image embeddings, effectively overcoming the limitations of standard interpolation in non-convex geospatial foundation model manifolds and significantly improving geometric adjustment performance.

Erik Scheurer, Rocco Sedona, Stefan Kesselheim, Gabriele CavallaroTue, 10 Ma💻 cs

Variational Flow Maps: Make Some Noise for One-Step Conditional Generation

This paper introduces Variational Flow Maps (VFMs), a framework that enables high-quality, single-step conditional generation and inverse problem solving by learning a noise adapter to align the initial noise distribution with observations, thereby bypassing the need for iterative sampling trajectories required by traditional diffusion models.

Abbas Mammadov, So Takao, Bohan Chen, Ricardo Baptista, Morteza Mardani, Yee Whye Teh, Julius BernerTue, 10 Ma🤖 cs.LG

MAviS: A Multimodal Conversational Assistant For Avian Species

This paper introduces MAviS, a domain-adaptive multimodal conversational assistant for avian species that leverages the newly created MAviS-Dataset and is evaluated on the MAviS-Bench to achieve state-of-the-art performance in fine-grained bird species understanding and multimodal question answering.

Yevheniia Kryklyvets, Mohammed Irfan Kurpath, Sahal Shaji Mullappilly, Jinxing Zhou, Fahad Shabzan Khan, Rao Anwer, Salman Khan, Hisham CholakkalTue, 10 Ma💻 cs

StructSAM: Structure- and Spectrum-Preserving Token Merging for Segment Anything Models

This paper introduces StructSAM, a novel token merging framework that preserves structural boundaries and spectral properties in Segment Anything Models (SAM) by using gradient-based energy scores and grid-based screening to achieve significant computational savings with minimal accuracy loss across natural and medical imaging benchmarks.

Duy M. H. Nguyen, Tuan A. Tran, Duong Nguyen, Siwei Xie, Trung Q. Nguyen, Mai T. N. Truong, Daniel Palenicek, An T. Le, Michael Barz, TrungTin Nguyen, Tuan Dam, Ngan Le, Minh Vu, Khoa Doan, Vien Ngo, Pengtao Xie, James Zou, Daniel Sonntag, Jan Peters, Mathias NiepertTue, 10 Ma🤖 cs.LG

Faster-HEAL: An Efficient and Privacy-Preserving Collaborative Perception Framework for Heterogeneous Autonomous Vehicles

Faster-HEAL is a lightweight, privacy-preserving collaborative perception framework that addresses the challenges of heterogeneous autonomous vehicles by using low-rank visual prompt fine-tuning and pyramid fusion to align diverse features into a unified space, achieving superior detection performance with significantly reduced computational overhead compared to state-of-the-art methods.

Armin Maleki, Hayder RadhaTue, 10 Ma💻 cs

AgrI Challenge: A Data-Centric AI Competition for Cross-Team Validation in Agricultural Vision

The AgrI Challenge introduces a data-centric competition framework featuring Cross-Team Validation to demonstrate that while single-source training suffers from significant generalization gaps in agricultural vision, collaborative multi-source training on independently collected, heterogeneous datasets dramatically improves model robustness and real-world performance.

Mohammed Brahimi, Karim Laabassi, Mohamed Seghir Hadj Ameur, Aicha Boutorh, Badia Siab-Farsi, Amin Khouani, Omar Farouk Zouak, Seif Eddine Bouziane, Kheira Lakhdari, Abdelkader Nabil BenghanemTue, 10 Ma🤖 cs.LG

AQuA: Toward Strategic Response Generation for Ambiguous Visual Questions

This paper introduces AQuA, a fine-grained dataset that categorizes ambiguous visual questions into four levels with corresponding optimal response strategies, demonstrating that fine-tuning Vision-Language Models on this dataset enables them to effectively recognize ambiguity and adaptively generate context-appropriate responses such as seeking clarification or listing alternatives, thereby outperforming existing baselines.

Jihyoung Jang, Hyounghun KimTue, 10 Ma💬 cs.CL