Breaking the Factorization Barrier in Diffusion Language Models

The paper introduces Coupled Discrete Diffusion (CoDD), a hybrid framework that overcomes the "factorization barrier" in diffusion language models by replacing fully factorized outputs with a lightweight probabilistic inference layer, thereby enabling efficient parallel generation of coherent, high-quality text without the prohibitive costs of full joint modeling or reinforcement learning.

Ian Li, Zilei Shao, Benjie Wang, Rose Yu, Guy Van den Broeck, Anji Liu2026-03-11🤖 cs.AI

Reasoning as Gradient: Scaling MLE Agents Beyond Tree Search

This paper introduces \textsc{Gome}, a gradient-based MLE agent that outperforms traditional tree search methods on MLE-Bench by mapping diagnostic reasoning to gradient computation, demonstrating that as LLM reasoning capabilities improve, gradient-based optimization becomes increasingly superior to exhaustive enumeration.

Yifei Zhang, Xu Yang, Xiao Yang, Bowen Xian, Qizheng Li, Shikai Fang, Jingyuan Li, Jian Wang, Mingrui Xu, Weiqing Liu, Jiang Bian2026-03-11🤖 cs.AI

FinTexTS: Financial Text-Paired Time-Series Dataset via Semantic-Based and Multi-Level Pairing

The paper introduces FinTexTS, a large-scale financial text-paired time-series dataset constructed via a novel semantic-based and multi-level pairing framework that overcomes the limitations of simple keyword matching by leveraging LLMs to align news articles with stock prices across macro, sector, related company, and target-company levels, thereby significantly improving stock price forecasting performance.

Jaehoon Lee, Suhwan Park, Tae Yoon Lim, Seunghan Lee, Jun Seo, Dongwan Kang, Hwanil Choi, Minjae Kim, Sungdong Yoo, SoonYoung Lee, Yongjae Lee, Wonbin Ahn2026-03-11🤖 cs.AI

Unveiling the Potential of Quantization with MXFP4: Strategies for Quantization Error Reduction

This paper introduces two software-only techniques, Overflow-Aware Scaling (OAS) and Macro Block Scaling (MBS), that significantly reduce the accuracy gap between the hardware-efficient MXFP4 format and NVIDIA's NVFP4 standard in Large Language Models, achieving near-parity performance with minimal computational overhead.

Jatin Chhugani, Geonhwa Jeong, Bor-Yiing Su, Yunjie Pan, Hanmei Yang, Aayush Ankit, Jiecao Yu, Summer Deng, Yunqing Chen, Nadathur Satish, Changkyu Kim2026-03-11🤖 cs.AI

KernelCraft: Benchmarking for Agentic Close-to-Metal Kernel Generation on Emerging Hardware

KernelCraft introduces the first benchmark evaluating agentic LLM systems that use feedback-driven workflows to automatically generate and optimize low-level kernels for emerging hardware with novel ISAs, demonstrating their ability to produce valid, high-performance code that rivals or exceeds traditional compiler baselines.

Jiayi Nie, Haoran Wu, Yao Lai, Zeyu Cao, Cheng Zhang, Binglei Lou, Erwei Wang, Jianyi Cheng, Timothy M. Jones, Robert Mullins, Rika Antonova, Yiren Zhao2026-03-11🤖 cs.LG

Performance Analysis of Edge and In-Sensor AI Processors: A Comparative Review

This paper reviews the landscape of ultra-low-power edge and in-sensor AI processors and empirically benchmarks a segmentation model on GAP9, STM32N6, and Sony IMX500 platforms to demonstrate that while in-sensor processing offers superior energy-delay performance, different architectures provide distinct trade-offs between latency, energy efficiency, and power budgets.

Luigi Capogrosso, Pietro Bonazzi, Michele Magno2026-03-11🤖 cs.LG

Memory-Augmented Spiking Networks: Synergistic Integration of Complementary Mechanisms for Neuromorphic Vision

This paper demonstrates that synergistically integrating Supervised Contrastive Learning, Hopfield networks, and Hierarchical Gated Recurrent Networks into Spiking Neural Networks achieves optimal neuromorphic vision performance on N-MNIST by balancing accuracy, energy efficiency, and structured neuronal clustering, rather than relying on isolated architectural optimizations.

Effiong Blessing, Chiung-Yi Tseng, Isaac Nkrumah, Junaid Rehman2026-03-11🤖 cs.LG

Sensitivity-Guided Framework for Pruned and Quantized Reservoir Computing Accelerators

This paper presents a sensitivity-guided framework for compressing Reservoir Computing accelerators that systematically balances quantization and pruning to significantly improve hardware efficiency and reduce power consumption on FPGAs while maintaining high model accuracy across various time-series tasks.

Atousa Jafari, Mahdi Taheri, Hassan Ghasemzadeh Mohammadi, Christian Herglotz, Marco Platzner2026-03-11🤖 cs.AI

Robust Parameter and State Estimation in Multiscale Neuronal Systems Using Physics-Informed Neural Networks

This paper presents a physics-informed neural network (PINN) framework that robustly reconstructs hidden state variables and estimates biophysical parameters in multiscale neuronal models using only partial, noisy voltage observations, effectively overcoming the convergence failures and sensitivity issues common in traditional numerical methods.

Changliang Wei, Yangyang Wang, Xueyu Zhu2026-03-11🤖 cs.LG

Permutation-Equivariant 2D State Space Models: Theory and Canonical Architecture for Multivariate Time Series

This paper introduces the Variable-Invariant Two-Dimensional State Space Model (VI 2D SSM) and its unified VI 2D Mamba architecture, which theoretically establish and implement a permutation-equivariant framework for multivariate time series that eliminates artificial variable ordering to achieve state-of-the-art performance with improved structural scalability.

Seungwoo Jeong, Heung-Il Suk2026-03-11🤖 cs.AI