Deep Hierarchical Learning with Nested Subspace Networks for Large Language Models

This paper introduces Nested Subspace Networks (NSNs), a novel architectural paradigm that re-parameterizes linear layers to enable a single large language model to be dynamically adjusted across a continuous spectrum of compute budgets at inference time, achieving a smooth and predictable trade-off between efficiency and performance without requiring retraining or multiple specialist models.

Paulius Rauba, Mihaela van der Schaar2026-03-05🤖 cs.LG

Bridging Computational Social Science and Deep Learning: Cultural Dissemination-Inspired Graph Neural Networks

This paper introduces AxelGNN, a novel Graph Neural Network architecture inspired by Axelrod's cultural dissemination model that utilizes similarity-gated interactions, segment-wise feature copying, and global polarization to effectively address oversmoothing and heterophily challenges while achieving competitive performance across diverse graph types.

Asela Hevapathige2026-03-05🤖 cs.AI