Training with Pseudo-Code for Instruction Following
This paper proposes a training-time approach that fine-tunes Large Language Models using instruction-tuning data augmented with pseudo-code representations of natural language instructions, resulting in significant improvements in instruction-following reliability and overall reasoning performance across multiple benchmarks.