CrossLLM-Mamba: Multimodal State Space Fusion of LLMs for RNA Interaction Prediction

CrossLLM-Mamba is a novel, scalable framework that leverages bidirectional Mamba encoders to model RNA interaction prediction as a dynamic state-space alignment problem, achieving state-of-the-art performance across RNA-protein, RNA-small molecule, and RNA-RNA tasks by capturing context-dependent molecular binding more effectively than static fusion methods.

Rabeya Tus Sadia, Qiang Ye, Qiang Cheng2026-02-27🧬 q-bio

Enhancing Renal Tumor Malignancy Prediction: Deep Learning with Automatic 3D CT Organ Focused Attention

This study introduces a deep learning framework utilizing an Organ Focused Attention (OFA) loss function to accurately predict renal tumor malignancy from 3D CT images without requiring labor-intensive manual segmentation, achieving performance that surpasses conventional segmentation-based models on both private and public datasets.

Zhengkang Fan, Chengkun Sun, Russell Terry + 2 more2026-02-27🤖 cs.AI

MolFM-Lite: Multi-Modal Molecular Property Prediction with Conformer Ensemble Attention and Cross-Modal Fusion

MolFM-Lite is a multi-modal machine learning model that improves molecular property prediction by jointly encoding 1D sequences, 2D graphs, and 3D conformer ensembles through cross-attention fusion and FiLM conditioning, achieving significant performance gains over single-modality baselines on MoleculeNet benchmarks.

Syed Omer Shah, Mohammed Maqsood Ahmed, Danish Mohiuddin Mohammed + 2 more2026-02-27🤖 cs.LG

Pix2Key: Controllable Open-Vocabulary Retrieval with Semantic Decomposition and Self-Supervised Visual Dictionary Learning

Pix2Key is a novel composed image retrieval framework that utilizes semantic decomposition and self-supervised visual dictionary learning to represent queries and candidates as open-vocabulary dictionaries, thereby achieving superior intent-aware matching and diversity-aware reranking without relying on supervised triplets.

Guoyizhe Wei, Yang Jiao, Nan Xi + 4 more2026-02-27💻 cs

HARU-Net: Hybrid Attention Residual U-Net for Edge-Preserving Denoising in Cone-Beam Computed Tomography

This paper introduces HARU-Net, a novel Hybrid Attention Residual U-Net architecture that integrates hybrid attention transformers and residual learning to effectively denoise low-dose Cone-Beam Computed Tomography (CBCT) images while preserving critical anatomical edges, outperforming state-of-the-art methods in both image quality metrics and computational efficiency.

Khuram Naveed, Ruben Pauwels2026-02-27⚡ eess

DisQ-HNet: A Disentangled Quantized Half-UNet for Interpretable Multimodal Image Synthesis Applications to Tau-PET Synthesis from T1 and FLAIR MRI

DisQ-HNet is a novel, interpretable framework that synthesizes tau-PET images from T1 and FLAIR MRI by employing a Partial Information Decomposition-guided vector-quantized encoder and a Half-UNet decoder to disentangle modality contributions while preserving anatomical details and disease-relevant signals for Alzheimer's disease analysis.

Agamdeep S. Chopra, Caitlin Neher, Tianyi Ren + 2 more2026-02-27🤖 cs.AI