Crowdsourcing the Frontier: Advancing Hybrid Physics-ML Climate Simulation via a $50,000 Kaggle Competition

This paper demonstrates that a $50,000 Kaggle competition successfully crowdsourced diverse machine learning architectures for subgrid parameterization, which, when coupled with a full-physics climate model, achieved reproducible online stability and state-of-the-art performance, marking a significant milestone in advancing hybrid physics-ML climate simulations.

Jerry Lin, Zeyuan Hu, Tom Beucler, Katherine Frields, Hannah Christensen, Walter Hannah, Helge Heuer, Peter Ukkonnen, Laura A. Mansfield, Tian Zheng, Liran Peng, Ritwik Gupta, Pierre Gentine, Yusef Al-Naher, Mingjiang Duan, Kyo Hattori, Weiliang Ji, Chunhan Li, Kippei Matsuda, Naoki Murakami, Shlomo Ron, Marec Serlin, Hongjian Song, Yuma Tanabe, Daisuke Yamamoto, Jianyao Zhou, Mike PritchardTue, 10 Ma🤖 cs.LG

ForamDeepSlice: A High-Accuracy Deep Learning Framework for Foraminifera Species Classification from 2D Micro-CT Slices

This study introduces ForamDeepSlice, a high-accuracy deep learning framework that combines an ensemble of ConvNeXt-Large and EfficientNetV2-Small models with a rigorous specimen-level split dataset to achieve 95.64% accuracy in classifying foraminifera species from 2D micro-CT slices, while also providing an interactive dashboard for real-time identification and 3D matching.

Abdelghafour Halimi, Ali Alibrahim, Didier Barradas-Bautista, Ronell Sicat, Abdulkader M. AfifiTue, 10 Ma🤖 cs.LG

Dual-Robust Cross-Domain Offline Reinforcement Learning Against Dynamics Shifts

This paper introduces DROCO, a novel dual-robust cross-domain offline reinforcement learning algorithm that addresses both train-time and test-time dynamics shifts by employing a robust cross-domain Bellman operator alongside dynamic value penalty and Huber loss to enhance policy robustness and prevent value estimation errors.

Zhongjian Qiao, Rui Yang, Jiafei Lyu, Xiu Li, Zhongxiang Dai, Zhuoran Yang, Siyang Gao, Shuang QiuTue, 10 Ma🤖 cs.LG

Evolving Diffusion and Flow Matching Policies for Online Reinforcement Learning

The paper introduces GoRL, an algorithm-agnostic framework that stabilizes online reinforcement learning with expressive generative policies by decoupling optimization in a tractable latent space from action synthesis via a conditional generative decoder, achieving superior performance on challenging continuous-control tasks.

Chubin Zhang, Zhenglin Wan, Feng Chen, Fuchao Yang, Lang Feng, Yaxin Zhou, Xingrui Yu, Yang You, Ivor Tsang, Bo AnTue, 10 Ma🤖 cs.LG

Two-Step Data Augmentation for Masked Face Detection and Recognition: Turning Fake Masks to Real

This paper presents a two-step generative data augmentation framework combining rule-based mask warping and unpaired image-to-image translation to address the scarcity of masked face datasets, achieving performance improvements with minimal training data while explicitly noting its origins as a resource-constrained coursework project that lacked downstream quantitative evaluation.

Yan Yang, George Bebis, Mircea NicolescuTue, 10 Ma🤖 cs.LG

Latent Sculpting for Zero-Shot Generalization: A Manifold Learning Approach to Out-of-Distribution Anomaly Detection

The paper proposes "Latent Sculpting," a hierarchical two-stage architecture that combines a Transformer-based encoder with a Binary Latent Sculpting loss and a Masked Autoregressive Flow to enforce explicit geometric boundaries on benign data, achieving robust zero-shot generalization and high detection rates for out-of-distribution cyberattacks on the CIC-IDS-2017 benchmark.

Rajeeb Thapa Chhetri, Saurab Thapa, Avinash Kumar, Zhixiong ChenTue, 10 Ma🤖 cs.LG

Certifying the Right to Be Forgotten: Primal-Dual Optimization for Sample and Label Unlearning in Vertical Federated Learning

This paper proposes FedORA, a primal-dual optimization framework that enables efficient and theoretically certified sample and label unlearning in Vertical Federated Learning by introducing a novel uncertainty-promoting loss function and adaptive strategies to minimize computational overhead while preserving model utility.

Yu Jiang, Xindi Tong, Ziyao Liu, Xiaoxi Zhang, Kwok-Yan Lam, Chee Wei TanTue, 10 Ma🤖 cs.LG