HydroGEM: A Self Supervised Zero Shot Hybrid TCN Transformer Foundation Model for Continental Scale Streamflow Quality Control

HydroGEM is a self-supervised, zero-shot hybrid TCN-Transformer foundation model that effectively performs continental-scale streamflow quality control by detecting and reconstructing sensor anomalies with high accuracy and cross-national generalization, thereby addressing the scalability limitations of manual hydrological data validation.

Ijaz Ul Haq, Byung Suk Lee, Julia N. Perdrial + 1 more2026-03-06💻 cs

FluenceFormer: Transformer-Driven Multi-Beam Fluence Map Regression for Radiotherapy Planning

This paper introduces FluenceFormer, a transformer-driven, two-stage framework that leverages a physics-informed Fluence-Aware Regression loss to achieve superior, geometry-aware fluence map prediction for radiotherapy planning, significantly outperforming existing CNN and single-stage methods in energy conservation and structural fidelity.

Ujunwa Mgboh, Rafi Ibn Sultan, Joshua Kim + 2 more2026-03-06💻 cs

When Do Tools and Planning Help Large Language Models Think? A Cost- and Latency-Aware Benchmark

This paper presents a cost- and latency-aware benchmark demonstrating that while tool-augmented planning significantly improves accuracy for complex knowledge-intensive tasks like Event-QA, it often incurs prohibitive latency costs and offers no benefit—or even degrades performance—for tasks like persuasive response generation where simple one-shot prompting is more efficient.

Subha Ghoshal, Ali Al-Bustami2026-03-06💻 cs

Where is the multimodal goal post? On the Ability of Foundation Models to Recognize Contextually Important Moments

This paper introduces a new dataset derived from football highlight reels to evaluate foundation models' ability to identify contextually important video moments, revealing that current state-of-the-art models perform near chance levels due to their reliance on single dominant modalities and failure to effectively synthesize cross-modal information.

Aditya K Surikuchi, Raquel Fernández, Sandro Pezzelle2026-03-06💻 cs

Mobility-Embedded POIs: Learning What A Place Is and How It Is Used from Human Movement

This paper introduces Mobility-Embedded POIs (ME-POIs), a framework that enhances general-purpose point-of-interest representations by integrating large-scale human mobility data with language model embeddings to capture both place identity and real-world usage functions, thereby outperforming existing text-only and mobility-only baselines across diverse map enrichment tasks.

Maria Despoina Siampou, Shushman Choudhury, Shang-Ling Hsu + 2 more2026-03-06💻 cs