Draft-Conditioned Constrained Decoding for Structured Generation in LLMs
The paper proposes Draft-Conditioned Constrained Decoding (DCCD), a training-free two-step inference method that decouples semantic planning from structural enforcement to significantly improve the accuracy and parameter efficiency of structured generation in large language models by mitigating the distortions caused by hard constraints.