Ultra-Innerthought is a bilingual (Chinese and English) open-domain R1/o1 style SFT dataset in think-than-response format, containing 2,085,326 dialogues. Unlike current reasoning datasets that mainly focus on mathematics and coding domains, Ultra-Innerthought covers a broader range of fields and includes both Chinese and English languages. We used Deepseek V3 as the model for data synthesis.
{
"id": "dialogue_id",
"conversations": [
{
"user": "user_input",
"inner_thought": "model's inner thought",
"assistant": "model_output"
},
...
],
"data_source": "data_source"
}
Ultra-Innerthought uses the following SFT datasets as raw input and employs Deepseek V3 for data synthesis. We preserved the user input from each round of the original datasets, using Deepseek V3 to first generate a model's Inner thought, and then generate the final response based on that Inner thought. When generating the model's Inner thought, we prompted the model to perform intent clarification, problem decomposition, self-reflection, exploration, and other behaviors. The dataset has approximately a 1:1 ratio of Chinese to English content.
User inputs are sampled from OpenHerms2.5 and its Chinese version translated by Deepseek V3, QwQ-LONGCOT-500K and its Chinese version translated by Deepseek V3, tulu-3-sft-mixture, sharegpt-zh, COIG-CQIA, Wildchat, WizardLM, Moss-inhouse-data, lmsys.