{"id":18944,"date":"2025-06-14T04:08:27","date_gmt":"2025-06-14T04:08:27","guid":{"rendered":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/"},"modified":"2025-06-14T04:08:27","modified_gmt":"2025-06-14T04:08:27","slug":"openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models","status":"publish","type":"post","link":"https:\/\/youzum.net\/ja\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/","title":{"rendered":"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models"},"content":{"rendered":"<h3 class=\"wp-block-heading\"><strong>The Growing Complexity of Reasoning Data Curation<\/strong><\/h3>\n<p>Recent reasoning models, such as DeepSeek-R1 and o3, have shown outstanding performance in mathematical, coding, and scientific areas, utilizing post-training techniques like supervised fine-tuning (SFT) and reinforcement learning (RL). However, the complete methodologies behind these frontier reasoning models are not public, which makes research for building reasoning models difficult. While SFT data curation has become a powerful approach for developing strong reasoning capabilities, most existing efforts explore only limited design choices, such as relying solely on human-written questions or single teacher models. Moreover, exploring the extensive design space of various techniques for generating question-answer pairs requires high costs for teacher inference and model training.<\/p>\n<p>Reasoning traces provided by models such as Gemini, QwQ, and DeepSeek-R1 have enabled knowledge distillation techniques to train smaller reasoning models. Projects like OpenR1, OpenMathReasoning, and OpenCodeReasoning collect questions from public forums and competition sites, while Natural Reasoning utilizes pre-training corpora as seed data. Some efforts, such as S1 and LIMO, focus on manually curating small, high-quality datasets of challenging prompts. Other methods, such as DeepMath-103K and Nvidia Nemotron, introduce innovations across data sourcing, filtering, and scaling stages. RL methods, including AceReason and Skywork-OR1, have enhanced reasoning capabilities beyond traditional SFT methods.<\/p>\n<h3 class=\"wp-block-heading\"><strong>OpenThoughts: A Scalable Framework for SFT Dataset Development<\/strong><\/h3>\n<p>Researchers from Stanford University, the University of Washington, BespokeLabs.ai, Toyota Research Institute, UC Berkeley, and 12 additional organizations have proposed OpenThoughts, a new SOTA open reasoning data recipe. OpenThoughts uses a progressive approach across three iterations: OpenThoughts-114K scales the Sky-T1 pipeline with automated verification, OpenThoughts2-1M enhances data scale through augmented question diversity and synthetic generation strategies, and OpenThoughts3-1.2M incorporates findings from over 1,000 ablation experiments to develop a simple, scalable, and high-performing data curation pipeline. Moreover, the model OpenThinker3-7B achieves state-of-the-art performance among open-data models at the 7B scale.<\/p>\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw?key=X4wUL90redYDlGnyWlpSoQ\" alt=\"\"\/><\/figure>\n<\/div>\n<p>The OpenThoughts3-1.2M is built by ablating each pipeline component independently while maintaining constant conditions across other stages, generating 31,600 data points per strategy and fine-tuning Qwen2.5-7B-Instruct on each resulting dataset. The goal during training is to create the best dataset of question-response pairs for SFT reasoning. Evaluation occurs across eight reasoning benchmarks across mathematics (AIME24, AMC23, MATH500), coding (CodeElo, CodeForces, LiveCodeBench), and science (GPQA Diamond, JEEBench). The experimental design includes a rigorous decontamination process to remove high-similarity samples and maintains a held-out benchmark set for generalization testing. Evalchemy serves as the primary evaluation tool, ensuring consistent evaluation protocols.<\/p>\n<h3 class=\"wp-block-heading\"><strong>Evaluation Insights and Benchmark Performance<\/strong><\/h3>\n<p>The OpenThoughts pipeline evaluation reveals key insights across question sourcing, mixing, filtering, answer filtering, and the teacher model. Question sourcing experiments show that CodeGolf and competitive coding questions achieve the highest performance for code tasks (25.3-27.5 average scores), while LLM-generated and human-written questions excel in mathematics (58.8-58.5 scores), and physics StackExchange questions with chemistry textbook extractions perform best in science (43.2-45.3 scores). Mixing question shows that combining multiple question sources degrades performance, with optimal results of 5% accuracy improvements over diverse mixing strategies. In the teacher model, QwQ-32B outperforms DeepSeek-R1 in knowledge distillation, achieving an accuracy improvement of 1.9-2.6%.<\/p>\n<p>In conclusion, researchers present the OpenThoughts project, showing that systematic experimentation can significantly advance SFT data curation for reasoning models. Researchers developed OpenThoughts3-1.2M, a state-of-the-art open-data reasoning dataset across science, mathematics, and coding domains. The resulting OpenThinker3-7B model achieves superior performance among open-data reasoning models at its scale. However, several limitations remain unexplored, including RL approaches, staged fine-tuning, and curriculum learning strategies. Future research directions include investigating cross-domain transfer effects when optimizing individual domains versus overall performance, and understanding the scaling dynamics as student models approach teacher capabilities.<\/p>\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n<p>Check out the<strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/2506.04178v1\" target=\"_blank\" rel=\"noreferrer noopener\">Paper<\/a><\/strong>, <strong><a href=\"https:\/\/www.openthoughts.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Project Page<\/a><\/strong> and <strong><a href=\"https:\/\/github.com\/open-thoughts\/open-thoughts\" target=\"_blank\" rel=\"noreferrer noopener\">GitHub Page<\/a><em>.<\/em><\/strong>\u00a0All credit for this research goes to the researchers of this project. Also,\u00a0feel free to follow us on\u00a0<strong><a href=\"https:\/\/x.com\/intent\/follow?screen_name=marktechpost\" target=\"_blank\" rel=\"noreferrer noopener\"><mark>Twitter<\/mark><\/a><\/strong>\u00a0and don\u2019t forget to join our\u00a0<strong><a href=\"https:\/\/www.reddit.com\/r\/machinelearningnews\/\" target=\"_blank\" rel=\"noreferrer noopener\">99k+ ML SubReddit<\/a><\/strong>\u00a0and Subscribe to\u00a0<strong><a href=\"https:\/\/www.airesearchinsights.com\/subscribe\" target=\"_blank\" rel=\"noreferrer noopener\">our Newsletter<\/a><\/strong>.<\/p>\n<p>The post <a href=\"https:\/\/www.marktechpost.com\/2025\/06\/13\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/\">OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models<\/a> appeared first on <a href=\"https:\/\/www.marktechpost.com\/\">MarkTechPost<\/a>.<\/p>","protected":false},"excerpt":{"rendered":"<p>The Growing Complexity of Reasoning Data Curation Recent reasoning models, such as DeepSeek-R1 and o3, have shown outstanding performance in mathematical, coding, and scientific areas, utilizing post-training techniques like supervised fine-tuning (SFT) and reinforcement learning (RL). However, the complete methodologies behind these frontier reasoning models are not public, which makes research for building reasoning models difficult. While SFT data curation has become a powerful approach for developing strong reasoning capabilities, most existing efforts explore only limited design choices, such as relying solely on human-written questions or single teacher models. Moreover, exploring the extensive design space of various techniques for generating question-answer pairs requires high costs for teacher inference and model training. Reasoning traces provided by models such as Gemini, QwQ, and DeepSeek-R1 have enabled knowledge distillation techniques to train smaller reasoning models. Projects like OpenR1, OpenMathReasoning, and OpenCodeReasoning collect questions from public forums and competition sites, while Natural Reasoning utilizes pre-training corpora as seed data. Some efforts, such as S1 and LIMO, focus on manually curating small, high-quality datasets of challenging prompts. Other methods, such as DeepMath-103K and Nvidia Nemotron, introduce innovations across data sourcing, filtering, and scaling stages. RL methods, including AceReason and Skywork-OR1, have enhanced reasoning capabilities beyond traditional SFT methods. OpenThoughts: A Scalable Framework for SFT Dataset Development Researchers from Stanford University, the University of Washington, BespokeLabs.ai, Toyota Research Institute, UC Berkeley, and 12 additional organizations have proposed OpenThoughts, a new SOTA open reasoning data recipe. OpenThoughts uses a progressive approach across three iterations: OpenThoughts-114K scales the Sky-T1 pipeline with automated verification, OpenThoughts2-1M enhances data scale through augmented question diversity and synthetic generation strategies, and OpenThoughts3-1.2M incorporates findings from over 1,000 ablation experiments to develop a simple, scalable, and high-performing data curation pipeline. Moreover, the model OpenThinker3-7B achieves state-of-the-art performance among open-data models at the 7B scale. The OpenThoughts3-1.2M is built by ablating each pipeline component independently while maintaining constant conditions across other stages, generating 31,600 data points per strategy and fine-tuning Qwen2.5-7B-Instruct on each resulting dataset. The goal during training is to create the best dataset of question-response pairs for SFT reasoning. Evaluation occurs across eight reasoning benchmarks across mathematics (AIME24, AMC23, MATH500), coding (CodeElo, CodeForces, LiveCodeBench), and science (GPQA Diamond, JEEBench). The experimental design includes a rigorous decontamination process to remove high-similarity samples and maintains a held-out benchmark set for generalization testing. Evalchemy serves as the primary evaluation tool, ensuring consistent evaluation protocols. Evaluation Insights and Benchmark Performance The OpenThoughts pipeline evaluation reveals key insights across question sourcing, mixing, filtering, answer filtering, and the teacher model. Question sourcing experiments show that CodeGolf and competitive coding questions achieve the highest performance for code tasks (25.3-27.5 average scores), while LLM-generated and human-written questions excel in mathematics (58.8-58.5 scores), and physics StackExchange questions with chemistry textbook extractions perform best in science (43.2-45.3 scores). Mixing question shows that combining multiple question sources degrades performance, with optimal results of 5% accuracy improvements over diverse mixing strategies. In the teacher model, QwQ-32B outperforms DeepSeek-R1 in knowledge distillation, achieving an accuracy improvement of 1.9-2.6%. In conclusion, researchers present the OpenThoughts project, showing that systematic experimentation can significantly advance SFT data curation for reasoning models. Researchers developed OpenThoughts3-1.2M, a state-of-the-art open-data reasoning dataset across science, mathematics, and coding domains. The resulting OpenThinker3-7B model achieves superior performance among open-data reasoning models at its scale. However, several limitations remain unexplored, including RL approaches, staged fine-tuning, and curriculum learning strategies. Future research directions include investigating cross-domain transfer effects when optimizing individual domains versus overall performance, and understanding the scaling dynamics as student models approach teacher capabilities. Check out the\u00a0Paper, Project Page and GitHub Page.\u00a0All credit for this research goes to the researchers of this project. Also,\u00a0feel free to follow us on\u00a0Twitter\u00a0and don\u2019t forget to join our\u00a099k+ ML SubReddit\u00a0and Subscribe to\u00a0our Newsletter. The post OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models appeared first on MarkTechPost.<\/p>","protected":false},"author":2,"featured_media":18945,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"pmpro_default_level":"","site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"default","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"var(--ast-global-color-4)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"_pvb_checkbox_block_on_post":false,"footnotes":""},"categories":[52,5,7,1],"tags":[],"class_list":["post-18944","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-club","category-committee","category-news","category-uncategorized","pmpro-has-access"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v25.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models - YouZum<\/title>\n<meta name=\"description\" content=\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/youzum.net\/ja\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/\" \/>\n<meta property=\"og:locale\" content=\"ja_JP\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models - YouZum\" \/>\n<meta property=\"og:description\" content=\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\" \/>\n<meta property=\"og:url\" content=\"https:\/\/youzum.net\/ja\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/\" \/>\n<meta property=\"og:site_name\" content=\"YouZum\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/DroneAssociationTH\/\" \/>\n<meta property=\"article:published_time\" content=\"2025-06-14T04:08:27+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1170\" \/>\n\t<meta property=\"og:image:height\" content=\"676\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"admin NU\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"\u57f7\u7b46\u8005\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin NU\" \/>\n\t<meta name=\"twitter:label2\" content=\"\u63a8\u5b9a\u8aad\u307f\u53d6\u308a\u6642\u9593\" \/>\n\t<meta name=\"twitter:data2\" content=\"3\u5206\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/\"},\"author\":{\"name\":\"admin NU\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c\"},\"headline\":\"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models\",\"datePublished\":\"2025-06-14T04:08:27+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/\"},\"wordCount\":678,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\"},\"image\":{\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png\",\"articleSection\":[\"AI\",\"Committee\",\"News\",\"Uncategorized\"],\"inLanguage\":\"ja\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/\",\"url\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/\",\"name\":\"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models - YouZum\",\"isPartOf\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png\",\"datePublished\":\"2025-06-14T04:08:27+00:00\",\"description\":\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\",\"breadcrumb\":{\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#breadcrumb\"},\"inLanguage\":\"ja\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"ja\",\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#primaryimage\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png\",\"width\":1170,\"height\":676},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/youzum.net\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/yousum.gpucore.co\/#website\",\"url\":\"https:\/\/yousum.gpucore.co\/\",\"name\":\"YouSum\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/yousum.gpucore.co\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"ja\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\",\"name\":\"Drone Association Thailand\",\"url\":\"https:\/\/yousum.gpucore.co\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"ja\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png\",\"width\":300,\"height\":300,\"caption\":\"Drone Association Thailand\"},\"image\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/DroneAssociationTH\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c\",\"name\":\"admin NU\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"ja\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png\",\"caption\":\"admin NU\"},\"url\":\"https:\/\/youzum.net\/ja\/members\/adminnu\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models - YouZum","description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/youzum.net\/ja\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/","og_locale":"ja_JP","og_type":"article","og_title":"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models - YouZum","og_description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","og_url":"https:\/\/youzum.net\/ja\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/","og_site_name":"YouZum","article_publisher":"https:\/\/www.facebook.com\/DroneAssociationTH\/","article_published_time":"2025-06-14T04:08:27+00:00","og_image":[{"width":1170,"height":676,"url":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png","type":"image\/png"}],"author":"admin NU","twitter_card":"summary_large_image","twitter_misc":{"\u57f7\u7b46\u8005":"admin NU","\u63a8\u5b9a\u8aad\u307f\u53d6\u308a\u6642\u9593":"3\u5206"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#article","isPartOf":{"@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/"},"author":{"name":"admin NU","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c"},"headline":"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models","datePublished":"2025-06-14T04:08:27+00:00","mainEntityOfPage":{"@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/"},"wordCount":678,"commentCount":0,"publisher":{"@id":"https:\/\/yousum.gpucore.co\/#organization"},"image":{"@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#primaryimage"},"thumbnailUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png","articleSection":["AI","Committee","News","Uncategorized"],"inLanguage":"ja","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/","url":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/","name":"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models - YouZum","isPartOf":{"@id":"https:\/\/yousum.gpucore.co\/#website"},"primaryImageOfPage":{"@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#primaryimage"},"image":{"@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#primaryimage"},"thumbnailUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png","datePublished":"2025-06-14T04:08:27+00:00","description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","breadcrumb":{"@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#breadcrumb"},"inLanguage":"ja","potentialAction":[{"@type":"ReadAction","target":["https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/"]}]},{"@type":"ImageObject","inLanguage":"ja","@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#primaryimage","url":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png","width":1170,"height":676},{"@type":"BreadcrumbList","@id":"https:\/\/youzum.net\/openthoughts-a-scalable-supervised-fine-tuning-sft-data-curation-pipeline-for-reasoning-models\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/youzum.net\/"},{"@type":"ListItem","position":2,"name":"OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models"}]},{"@type":"WebSite","@id":"https:\/\/yousum.gpucore.co\/#website","url":"https:\/\/yousum.gpucore.co\/","name":"YouSum","description":"","publisher":{"@id":"https:\/\/yousum.gpucore.co\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/yousum.gpucore.co\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"ja"},{"@type":"Organization","@id":"https:\/\/yousum.gpucore.co\/#organization","name":"Drone Association Thailand","url":"https:\/\/yousum.gpucore.co\/","logo":{"@type":"ImageObject","inLanguage":"ja","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/","url":"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png","width":300,"height":300,"caption":"Drone Association Thailand"},"image":{"@id":"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/DroneAssociationTH\/"]},{"@type":"Person","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c","name":"admin NU","image":{"@type":"ImageObject","inLanguage":"ja","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/image\/","url":"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png","caption":"admin NU"},"url":"https:\/\/youzum.net\/ja\/members\/adminnu\/"}]}},"rttpg_featured_image_url":{"full":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png",1170,676,false],"landscape":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png",1170,676,false],"portraits":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png",1170,676,false],"thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2-150x150.png",150,150,true],"medium":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2-300x173.png",300,173,true],"large":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2-1024x592.png",1024,592,true],"1536x1536":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png",1170,676,false],"2048x2048":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2.png",1170,676,false],"trp-custom-language-flag":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2-18x10.png",18,10,true],"woocommerce_thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2-300x300.png",300,300,true],"woocommerce_single":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2-600x347.png",600,347,true],"woocommerce_gallery_thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXcSxGFMcv8U5oqhZl7RzFnzSeiQo-gJUq5rPHSD_qvj7G1zFVsczx10h5io7AqCi7DM7pp77oF5THHPQIdbNtzEU1Jkxq6xmJhcVVTyGdHfbRIbyEkIqb9IqDWtrmITE1huScCAEw-xpzcT2-100x100.png",100,100,true]},"rttpg_author":{"display_name":"admin NU","author_link":"https:\/\/youzum.net\/ja\/members\/adminnu\/"},"rttpg_comment":0,"rttpg_category":"<a href=\"https:\/\/youzum.net\/ja\/category\/ai-club\/\" rel=\"category tag\">AI<\/a> <a href=\"https:\/\/youzum.net\/ja\/category\/committee\/\" rel=\"category tag\">Committee<\/a> <a href=\"https:\/\/youzum.net\/ja\/category\/news\/\" rel=\"category tag\">News<\/a> <a href=\"https:\/\/youzum.net\/ja\/category\/uncategorized\/\" rel=\"category tag\">Uncategorized<\/a>","rttpg_excerpt":"The Growing Complexity of Reasoning Data Curation Recent reasoning models, such as DeepSeek-R1 and o3, have shown outstanding performance in mathematical, coding, and scientific areas, utilizing post-training techniques like supervised fine-tuning (SFT) and reinforcement learning (RL). However, the complete methodologies behind these frontier reasoning models are not public, which makes research for building reasoning models&hellip;","_links":{"self":[{"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/posts\/18944","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/comments?post=18944"}],"version-history":[{"count":0,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/posts\/18944\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/media\/18945"}],"wp:attachment":[{"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/media?parent=18944"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/categories?post=18944"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/tags?post=18944"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}