{"id":16905,"date":"2025-06-06T04:24:48","date_gmt":"2025-06-06T04:24:48","guid":{"rendered":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/"},"modified":"2025-06-06T04:24:48","modified_gmt":"2025-06-06T04:24:48","slug":"teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning","status":"publish","type":"post","link":"https:\/\/youzum.net\/zh\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/","title":{"rendered":"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning"},"content":{"rendered":"<p>Reinforcement finetuning uses reward signals to guide the <a href=\"https:\/\/www.marktechpost.com\/2025\/01\/11\/what-are-large-language-model-llms\/\" target=\"_blank\">large language model<\/a> toward desirable behavior. This method sharpens the model\u2019s ability to produce logical and structured outputs by reinforcing correct responses. Yet, the challenge persists in ensuring that these models also know when not to respond\u2014particularly when faced with incomplete or misleading questions that don\u2019t have a definite answer.<\/p>\n<p>The problem arises when language models, after reinforcement finetuning, begin to lose their ability to refuse to answer unclear or ambiguous queries. Instead of signaling uncertainty, the models tend to produce confidently stated but incorrect responses. This phenomenon, identified in the paper as the \u201challucination tax,\u201d highlights a growing risk. As models are trained to perform better, they may also become more likely to hallucinate answers in situations where silence would be more appropriate. This is especially hazardous in domains that require high trust and precision.<\/p>\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg?key=emUTDo0-GigzW1TFjkQzTg\" alt=\"\"\/><\/figure>\n<\/div>\n<p>Tools currently used in training large language models often overlook the importance of refusal behavior. Reinforcement finetuning frameworks tend to reward only correct answers while penalizing incorrect ones, ignoring cases where a valid response should be no answer at all. The reward systems in use do not sufficiently reinforce refusal, resulting in overconfident models. For instance, the paper shows that refusal rates dropped to near zero across multiple models after standard RFT, demonstrating that current training fails to address hallucination properly.<\/p>\n<p>Researchers from the University of Southern California developed the Synthetic Unanswerable Math (SUM) dataset. SUM introduces implicitly unanswerable math problems by modifying existing questions through criteria such as missing key information or creating logical inconsistencies. The researchers used DeepScaleR as the base dataset and employed the o3-mini model to generate high-quality unanswerable questions. This synthetic dataset aims to teach models to recognize when a problem lacks sufficient information and respond accordingly.<\/p>\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXePH2CCDBEeXu2RvjE7ovm1t0AMej0mEM2dFkNzWjvHWcBboEUnb163W5q1vVu1TUyqj2d-JHAl86yX-88FtORt9quAH4hBaX1dE2SvVdhADM6xTIUykOwuPrPLZVl8WJVenT-_Ww?key=emUTDo0-GigzW1TFjkQzTg\" alt=\"\"\/><\/figure>\n<\/div>\n<p>SUM\u2019s core technique is to mix answerable and unanswerable problems during training. Questions are modified to become ambiguous or unsolvable while maintaining plausibility. The training prompts instruct models to say \u201cI don\u2019t know\u201d for unanswerable inputs. By introducing only 10% of the SUM data into reinforcement finetuning, models begin to leverage inference-time reasoning to evaluate uncertainty. This structure allows them to refuse answers more appropriately without impairing their performance on solvable problems.<\/p>\n<p>Performance analysis shows significant improvements. After training with SUM, the Qwen2.5-7B model increased its refusal rate from 0.01 to 0.73 on the SUM benchmark and from 0.01 to 0.81 on the UMWP benchmark. On the SelfAware dataset, refusal accuracy rose dramatically from 0.01 to 0.94. Llama-3.1-8B-Instruct showed a similar trend, with refusal rates improving from 0.00 to 0.75 on SUM and from 0.01 to 0.79 on UMWP. Despite these gains in refusal behavior, accuracy on answerable datasets, such as GSM8K and MATH-500, remained stable, with most changes ranging from 0.00 to -0.05. The minimal drop indicates that refusal training can be introduced without major sacrifices in task performance.<\/p>\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXdqxsNoNw1PbFv-V5aoecodl-WgwzUuuLUcdBTXDQL6n5U6bUju6CwW9yIscpT5bHm7YFV33C-ap2-Sx-XM2L10wQY0LholjD7TN26PD6LtGFPiaogjDGgWthu5m0R04f4gqekXtA?key=emUTDo0-GigzW1TFjkQzTg\" alt=\"\"\/><\/figure>\n<\/div>\n<p>This study outlines a clear trade-off between improved reasoning and trustworthiness. Reinforcement finetuning, while powerful, tends to suppress cautious behavior. The SUM dataset corrects this by teaching models to recognize what they cannot solve. With only a small addition to training data, language models become better at identifying the boundaries of their knowledge. This approach marks a significant step in making AI systems not just smarter but also more careful and honest.<\/p>\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n<p><strong>Check out the\u00a0<a href=\"https:\/\/arxiv.org\/abs\/2505.13988\" target=\"_blank\" rel=\"noreferrer noopener\">Paper<\/a> and <a href=\"https:\/\/huggingface.co\/datasets\/lime-nlp\/Synthetic_Unanswerable_Math\" target=\"_blank\" rel=\"noreferrer noopener\">Dataset on Hugging Face<\/a><em>.<\/em><\/strong>\u00a0All credit for this research goes to the researchers of this project.<\/p>\n<p><img decoding=\"async\" src=\"https:\/\/s.w.org\/images\/core\/emoji\/15.1.0\/72x72\/1f195.png\" alt=\"\ud83c\udd95\" class=\"wp-smiley\"\/> <strong><mark>Did you know? Marktechpost is the fastest-growing AI media platform\u2014trusted by over 1 million monthly readers. <a href=\"https:\/\/promotion.marktechpost.com\/\">Book a strategy call to discuss your campaign goals<\/a>.<\/mark><\/strong> Also,\u00a0feel free to follow us on\u00a0<strong><a href=\"https:\/\/x.com\/intent\/follow?screen_name=marktechpost\" target=\"_blank\" rel=\"noreferrer noopener\"><mark>Twitter<\/mark><\/a><\/strong>\u00a0and don\u2019t forget to join our\u00a0<strong><a href=\"https:\/\/www.reddit.com\/r\/machinelearningnews\/\" target=\"_blank\" rel=\"noreferrer noopener\">95k+ ML SubReddit<\/a><\/strong>\u00a0and Subscribe to\u00a0<strong><a href=\"https:\/\/www.airesearchinsights.com\/subscribe\" target=\"_blank\" rel=\"noreferrer noopener\">our Newsletter<\/a><\/strong>.<\/p>\n<p>The post <a href=\"https:\/\/www.marktechpost.com\/2025\/06\/05\/usc-researchers-introduced-sum-synthetic-unanswerable-math-a-synthetic-dataset-to-reduce-hallucination-in-llms-via-reinforcement-fine-tuning\/\">Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning<\/a> appeared first on <a href=\"https:\/\/www.marktechpost.com\/\">MarkTechPost<\/a>.<\/p>","protected":false},"excerpt":{"rendered":"<p>Reinforcement finetuning uses reward signals to guide the large language model toward desirable behavior. This method sharpens the model\u2019s ability to produce logical and structured outputs by reinforcing correct responses. Yet, the challenge persists in ensuring that these models also know when not to respond\u2014particularly when faced with incomplete or misleading questions that don\u2019t have a definite answer. The problem arises when language models, after reinforcement finetuning, begin to lose their ability to refuse to answer unclear or ambiguous queries. Instead of signaling uncertainty, the models tend to produce confidently stated but incorrect responses. This phenomenon, identified in the paper as the \u201challucination tax,\u201d highlights a growing risk. As models are trained to perform better, they may also become more likely to hallucinate answers in situations where silence would be more appropriate. This is especially hazardous in domains that require high trust and precision. Tools currently used in training large language models often overlook the importance of refusal behavior. Reinforcement finetuning frameworks tend to reward only correct answers while penalizing incorrect ones, ignoring cases where a valid response should be no answer at all. The reward systems in use do not sufficiently reinforce refusal, resulting in overconfident models. For instance, the paper shows that refusal rates dropped to near zero across multiple models after standard RFT, demonstrating that current training fails to address hallucination properly. Researchers from the University of Southern California developed the Synthetic Unanswerable Math (SUM) dataset. SUM introduces implicitly unanswerable math problems by modifying existing questions through criteria such as missing key information or creating logical inconsistencies. The researchers used DeepScaleR as the base dataset and employed the o3-mini model to generate high-quality unanswerable questions. This synthetic dataset aims to teach models to recognize when a problem lacks sufficient information and respond accordingly. SUM\u2019s core technique is to mix answerable and unanswerable problems during training. Questions are modified to become ambiguous or unsolvable while maintaining plausibility. The training prompts instruct models to say \u201cI don\u2019t know\u201d for unanswerable inputs. By introducing only 10% of the SUM data into reinforcement finetuning, models begin to leverage inference-time reasoning to evaluate uncertainty. This structure allows them to refuse answers more appropriately without impairing their performance on solvable problems. Performance analysis shows significant improvements. After training with SUM, the Qwen2.5-7B model increased its refusal rate from 0.01 to 0.73 on the SUM benchmark and from 0.01 to 0.81 on the UMWP benchmark. On the SelfAware dataset, refusal accuracy rose dramatically from 0.01 to 0.94. Llama-3.1-8B-Instruct showed a similar trend, with refusal rates improving from 0.00 to 0.75 on SUM and from 0.01 to 0.79 on UMWP. Despite these gains in refusal behavior, accuracy on answerable datasets, such as GSM8K and MATH-500, remained stable, with most changes ranging from 0.00 to -0.05. The minimal drop indicates that refusal training can be introduced without major sacrifices in task performance. This study outlines a clear trade-off between improved reasoning and trustworthiness. Reinforcement finetuning, while powerful, tends to suppress cautious behavior. The SUM dataset corrects this by teaching models to recognize what they cannot solve. With only a small addition to training data, language models become better at identifying the boundaries of their knowledge. This approach marks a significant step in making AI systems not just smarter but also more careful and honest. Check out the\u00a0Paper and Dataset on Hugging Face.\u00a0All credit for this research goes to the researchers of this project. Did you know? Marktechpost is the fastest-growing AI media platform\u2014trusted by over 1 million monthly readers. Book a strategy call to discuss your campaign goals. Also,\u00a0feel free to follow us on\u00a0Twitter\u00a0and don\u2019t forget to join our\u00a095k+ ML SubReddit\u00a0and Subscribe to\u00a0our Newsletter. The post Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning appeared first on MarkTechPost.<\/p>","protected":false},"author":2,"featured_media":16906,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"pmpro_default_level":"","site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"default","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"var(--ast-global-color-4)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"_pvb_checkbox_block_on_post":false,"footnotes":""},"categories":[52,5,7,1],"tags":[],"class_list":["post-16905","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-club","category-committee","category-news","category-uncategorized","pmpro-has-access"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v25.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning - YouZum<\/title>\n<meta name=\"description\" content=\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/youzum.net\/zh\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/\" \/>\n<meta property=\"og:locale\" content=\"zh_CN\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning - YouZum\" \/>\n<meta property=\"og:description\" content=\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\" \/>\n<meta property=\"og:url\" content=\"https:\/\/youzum.net\/zh\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/\" \/>\n<meta property=\"og:site_name\" content=\"YouZum\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/DroneAssociationTH\/\" \/>\n<meta property=\"article:published_time\" content=\"2025-06-06T04:24:48+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png\" \/>\n\t<meta property=\"og:image:width\" content=\"796\" \/>\n\t<meta property=\"og:image:height\" content=\"1412\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"admin NU\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"\u4f5c\u8005\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin NU\" \/>\n\t<meta name=\"twitter:label2\" content=\"\u9884\u8ba1\u9605\u8bfb\u65f6\u95f4\" \/>\n\t<meta name=\"twitter:data2\" content=\"3 \u5206\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/\"},\"author\":{\"name\":\"admin NU\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c\"},\"headline\":\"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning\",\"datePublished\":\"2025-06-06T04:24:48+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/\"},\"wordCount\":654,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\"},\"image\":{\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png\",\"articleSection\":[\"AI\",\"Committee\",\"News\",\"Uncategorized\"],\"inLanguage\":\"zh-Hans\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/\",\"url\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/\",\"name\":\"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning - YouZum\",\"isPartOf\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png\",\"datePublished\":\"2025-06-06T04:24:48+00:00\",\"description\":\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\",\"breadcrumb\":{\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#breadcrumb\"},\"inLanguage\":\"zh-Hans\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"zh-Hans\",\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#primaryimage\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png\",\"width\":796,\"height\":1412},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/youzum.net\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/yousum.gpucore.co\/#website\",\"url\":\"https:\/\/yousum.gpucore.co\/\",\"name\":\"YouSum\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/yousum.gpucore.co\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"zh-Hans\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\",\"name\":\"Drone Association Thailand\",\"url\":\"https:\/\/yousum.gpucore.co\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"zh-Hans\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png\",\"width\":300,\"height\":300,\"caption\":\"Drone Association Thailand\"},\"image\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/DroneAssociationTH\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c\",\"name\":\"admin NU\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"zh-Hans\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png\",\"caption\":\"admin NU\"},\"url\":\"https:\/\/youzum.net\/zh\/members\/adminnu\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning - YouZum","description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/youzum.net\/zh\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/","og_locale":"zh_CN","og_type":"article","og_title":"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning - YouZum","og_description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","og_url":"https:\/\/youzum.net\/zh\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/","og_site_name":"YouZum","article_publisher":"https:\/\/www.facebook.com\/DroneAssociationTH\/","article_published_time":"2025-06-06T04:24:48+00:00","og_image":[{"width":796,"height":1412,"url":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png","type":"image\/png"}],"author":"admin NU","twitter_card":"summary_large_image","twitter_misc":{"\u4f5c\u8005":"admin NU","\u9884\u8ba1\u9605\u8bfb\u65f6\u95f4":"3 \u5206"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#article","isPartOf":{"@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/"},"author":{"name":"admin NU","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c"},"headline":"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning","datePublished":"2025-06-06T04:24:48+00:00","mainEntityOfPage":{"@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/"},"wordCount":654,"commentCount":0,"publisher":{"@id":"https:\/\/yousum.gpucore.co\/#organization"},"image":{"@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#primaryimage"},"thumbnailUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png","articleSection":["AI","Committee","News","Uncategorized"],"inLanguage":"zh-Hans","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/","url":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/","name":"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning - YouZum","isPartOf":{"@id":"https:\/\/yousum.gpucore.co\/#website"},"primaryImageOfPage":{"@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#primaryimage"},"image":{"@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#primaryimage"},"thumbnailUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png","datePublished":"2025-06-06T04:24:48+00:00","description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","breadcrumb":{"@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#breadcrumb"},"inLanguage":"zh-Hans","potentialAction":[{"@type":"ReadAction","target":["https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/"]}]},{"@type":"ImageObject","inLanguage":"zh-Hans","@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#primaryimage","url":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png","width":796,"height":1412},{"@type":"BreadcrumbList","@id":"https:\/\/youzum.net\/teaching-ai-to-say-i-dont-know-a-new-dataset-mitigates-hallucinations-from-reinforcement-finetuning\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/youzum.net\/"},{"@type":"ListItem","position":2,"name":"Teaching AI to Say \u2018I Don\u2019t Know\u2019: A New Dataset Mitigates Hallucinations from Reinforcement Finetuning"}]},{"@type":"WebSite","@id":"https:\/\/yousum.gpucore.co\/#website","url":"https:\/\/yousum.gpucore.co\/","name":"YouSum","description":"","publisher":{"@id":"https:\/\/yousum.gpucore.co\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/yousum.gpucore.co\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"zh-Hans"},{"@type":"Organization","@id":"https:\/\/yousum.gpucore.co\/#organization","name":"Drone Association Thailand","url":"https:\/\/yousum.gpucore.co\/","logo":{"@type":"ImageObject","inLanguage":"zh-Hans","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/","url":"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png","width":300,"height":300,"caption":"Drone Association Thailand"},"image":{"@id":"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/DroneAssociationTH\/"]},{"@type":"Person","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c","name":"admin NU","image":{"@type":"ImageObject","inLanguage":"zh-Hans","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/image\/","url":"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png","caption":"admin NU"},"url":"https:\/\/youzum.net\/zh\/members\/adminnu\/"}]}},"rttpg_featured_image_url":{"full":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png",796,1412,false],"landscape":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png",796,1412,false],"portraits":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png",796,1412,false],"thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ-150x150.png",150,150,true],"medium":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ-169x300.png",169,300,true],"large":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ-577x1024.png",577,1024,true],"1536x1536":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png",796,1412,false],"2048x2048":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ.png",796,1412,false],"trp-custom-language-flag":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ-7x12.png",7,12,true],"woocommerce_thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ-300x300.png",300,300,true],"woocommerce_single":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ-600x1064.png",600,1064,true],"woocommerce_gallery_thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/06\/AD_4nXdYPcM6xFZ63ytHfS6c6ZWGGTjHNnUt7zfXu8gT5PXaue2BNYuF6SHMO_ctJc0CNBHyDehspbunHjVUHHqWyHSCVvEjURszaVLSrbYV6BF0YLRtv1NV8bGiK5tRlgB3A9IIR8e0Jg-JzBBcQ-100x100.png",100,100,true]},"rttpg_author":{"display_name":"admin NU","author_link":"https:\/\/youzum.net\/zh\/members\/adminnu\/"},"rttpg_comment":0,"rttpg_category":"<a href=\"https:\/\/youzum.net\/zh\/category\/ai-club\/\" rel=\"category tag\">AI<\/a> <a href=\"https:\/\/youzum.net\/zh\/category\/committee\/\" rel=\"category tag\">Committee<\/a> <a href=\"https:\/\/youzum.net\/zh\/category\/news\/\" rel=\"category tag\">News<\/a> <a href=\"https:\/\/youzum.net\/zh\/category\/uncategorized\/\" rel=\"category tag\">Uncategorized<\/a>","rttpg_excerpt":"Reinforcement finetuning uses reward signals to guide the large language model toward desirable behavior. This method sharpens the model\u2019s ability to produce logical and structured outputs by reinforcing correct responses. Yet, the challenge persists in ensuring that these models also know when not to respond\u2014particularly when faced with incomplete or misleading questions that don\u2019t have&hellip;","_links":{"self":[{"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/posts\/16905","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/comments?post=16905"}],"version-history":[{"count":0,"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/posts\/16905\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/media\/16906"}],"wp:attachment":[{"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/media?parent=16905"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/categories?post=16905"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/youzum.net\/zh\/wp-json\/wp\/v2\/tags?post=16905"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}