IGLU: The Integrated Gaussian Linear Unit Activation Function

This paper introduces IGLU, a novel parametric activation function derived from a scale mixture of GELU gates that utilizes a Cauchy CDF to provide heavy-tailed gradient properties and robustness against vanishing gradients, alongside a computationally efficient rational approximation (IGLU-Approx) that achieves competitive or superior performance across vision and language tasks compared to standard baselines like ReLU and GELU.

Mingi Kang, Zai Yang, Jeova Farias Sales Rocha Neto2026-03-10🤖 cs.LG

Small Target Detection Based on Mask-Enhanced Attention Fusion of Visible and Infrared Remote Sensing Images

This paper introduces ESM-YOLO+, a lightweight visible-infrared fusion network that employs a Mask-Enhanced Attention Fusion module and training-time Structural Representation enhancement to achieve high-precision small-target detection in complex remote sensing scenes while significantly reducing model complexity compared to baselines.

Qianqian Zhang, Xiaolong Jia, Ahmed M. Abdelmoniem, Li Zhou, Junshe An2026-03-10💻 cs

HIERAMP: Coarse-to-Fine Autoregressive Amplification for Generative Dataset Distillation

This paper proposes HIERAMP, a method that leverages the coarse-to-fine generation capability of Vision Autoregressive (VAR) models to amplify hierarchical semantics through dynamic class token injection, thereby improving dataset distillation performance by better capturing object structures and details without explicitly optimizing global proximity.

Lin Zhao, Xinru Jiang, Xi Xiao, Qihui Fan, Lei Lu, Yanzhi Wang, Xue Lin, Octavia Camps, Pu Zhao, Jianyang Gu2026-03-10💻 cs

Extracting and analyzing 3D histomorphometric features related to perineural and lymphovascular invasion in prostate cancer

This study presents a 3D histomorphometric analysis pipeline using nnU-Net segmentation on optically cleared prostatectomy specimens to extract features related to perineural and lymphovascular invasion, demonstrating that 3D perineural invasion features significantly outperform their 2D counterparts in predicting 5-year biochemical recurrence in prostate cancer.

Sarah S. L. Chow, Rui Wang, Robert B. Serafin, Yujie Zhao, Elena Baraznenok, Xavier Farré, Jennifer Salguero-Lopez, Gan Gao, Huai-Ching Hsieh, Lawrence D. True, Priti Lal, Anant Madabhushi, Jonathan T. C. Liu2026-03-10💻 cs

Virtual Intraoperative CT (viCT): Sequential Anatomic Updates for Modeling Tissue Resection Throughout Endoscopic Sinus Surgery

This paper introduces Virtual Intraoperative CT (viCT), a method that sequentially updates preoperative CT scans during endoscopic sinus surgery by integrating monocular endoscopic video-derived 3D reconstructions to visualize evolving tissue resection boundaries with submillimeter accuracy, thereby addressing the limitations of static image guidance.

Nicole M. Gunderson, Graham J. Harris, Jeremy S. Ruthberg, Pengcheng Chen, Di Mao, Randall A. Bly, Waleed M. Abuzeid, Eric J. Seibel2026-03-10💻 cs

Conditional Unbalanced Optimal Transport Maps: An Outlier-Robust Framework for Conditional Generative Modeling

This paper introduces Conditional Unbalanced Optimal Transport Maps (CUOTM), a robust conditional generative framework that mitigates the outlier sensitivity of classical Conditional Optimal Transport by relaxing distribution-matching constraints via Csiszár divergence penalties while preserving conditioning marginals through a theoretically justified triangular cc-transform parameterization.

Jiwoo Yoon, Kyumin Choi, Jaewoong Choi2026-03-10🤖 cs.LG

Optimizing Multi-Modal Models for Image-Based Shape Retrieval: The Role of Pre-Alignment and Hard Contrastive Learning

This paper proposes a novel approach to image-based shape retrieval that leverages pre-aligned multi-modal encoders and a hard contrastive learning loss to achieve state-of-the-art performance in both zero-shot and supervised settings, eliminating the need for explicit view-based supervision or view synthesis.

Paul Julius Kühn, Cedric Spengler, Michael Weinmann, Arjan Kuijper, Saptarshi Neil Sinha2026-03-10💻 cs

Perception-Aware Multimodal Spatial Reasoning from Monocular Images

This paper proposes a perception-aware multimodal reasoning framework that enhances Vision-Language Models' spatial understanding in monocular driving scenarios by representing objects with Visual Reference Tokens and utilizing a Multimodal Chain-of-Thought dataset, achieving significant performance gains on the SURDS benchmark through standard supervised fine-tuning.

Yanchun Cheng, Rundong Wang, Xulei Yang, Alok Prakash, Daniela Rus, Marcelo H Ang Jr, ShiJie Li2026-03-10💻 cs

ADAS-TO: A Large-Scale Multimodal Naturalistic Dataset and Empirical Characterization of Human Takeovers during ADAS Engagement

This paper introduces ADAS-TO, the first large-scale naturalistic multimodal dataset of 15,659 ADAS-to-manual takeover events from 327 drivers, which combines kinematic and vision-language analysis to characterize safety-critical scenarios and demonstrate that actionable visual cues often precede takeovers by over three seconds.

Yuhang Wang, Yiyao Xu, Jingran Sun, Hao Zhou2026-03-10💻 cs

AdaGen: Learning Adaptive Policy for Image Synthesis

AdaGen introduces a general, learnable framework that employs reinforcement learning with an adversarial reward to dynamically adapt step-specific parameters during iterative image synthesis, thereby overcoming the limitations of static, manually-designed schedules and achieving superior performance across diverse generative models with reduced inference costs.

Zanlin Ni, Yulin Wang, Yeguo Hua, Renping Zhou, Jiayi Guo, Jun Song, Bo Zheng, Gao Huang2026-03-10💻 cs

TrajPred: Trajectory-Conditioned Joint Embedding Prediction for Surgical Instrument-Tissue Interaction Recognition in Vision-Language Models

TrajPred is a novel framework that enhances surgical instrument-tissue interaction recognition in vision-language models by encoding instrument trajectories to capture temporal motion cues and generating fine-grained visual semantic embeddings, thereby significantly improving performance and vision-text alignment on the CholecT50 benchmark.

Jiajun Cheng, Xiaofan Yu, Subarna, Sainan Liu, Shan Lin2026-03-10💻 cs