arXiv

Distribution-Aligned Sequence Distillation for Superior Long-CoT Reasoning

Shaotian Yan, Kaiyuan Liu, Chen Shen
Jan 15, 2026·07:09·
00:00
07:09
DASD-4B-ThinkingSequence-Level DistillationDistribution AlignmentLong-CoT ReasoningEfficient TrainingKnowledge Distillation

About This Paper

Turn any paper into a podcast

ResearchPod turns research papers into podcasts you can actually follow.

Download on the App Store