An implicit CoT framework that simultaneously trains natural language CoT (teacher) within the same model while aligning the hidden state of a specific token (e.g., "The answer is:") before answer generation using L1 (=self-distillation) to compress and replace CoT with a small number of continuous latent thoughts
CODI
Creator
Creator
Seonglae ChoCreated
Created
2025 Dec 27 23:57Editor
Editor
Seonglae ChoEdited
Edited
2025 Dec 27 23:58Refs
Refs
