arXiv
Distribution-Aligned Sequence Distillation for Superior Long-CoT Reasoning
Shaotian Yan, Kaiyuan Liu, Chen Shen
Jan 15, 2026·07:09·
00:0007:09
DASD-4B-ThinkingSequence-Level DistillationDistribution AlignmentLong-CoT ReasoningEfficient TrainingKnowledge Distillation
About This Paper
Turn any paper into a podcast
ResearchPod turns research papers into podcasts you can actually follow.
Download on the App Store