{"id":30480,"date":"2025-08-09T05:58:33","date_gmt":"2025-08-09T05:58:33","guid":{"rendered":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/"},"modified":"2025-08-09T05:58:33","modified_gmt":"2025-08-09T05:58:33","slug":"alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models","status":"publish","type":"post","link":"https:\/\/youzum.net\/ja\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/","title":{"rendered":"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models"},"content":{"rendered":"<h3 class=\"wp-block-heading\"><strong>Smaller Models with Smarter Performance and 256K Context Support<\/strong><\/h3>\n<p>Alibaba\u2019s Qwen team has introduced two powerful additions to its small language model lineup: <strong>Qwen3-4B-Instruct-2507<\/strong> and <strong>Qwen3-4B-Thinking-2507<\/strong>. Despite having only 4 billion parameters, these models deliver exceptional capabilities across general-purpose and expert-level tasks while running efficiently on consumer-grade hardware. Both are designed with <strong>native 256K token context windows<\/strong>, meaning they can process extremely long inputs such as large codebases, multi-document archives, and extended dialogues without external modifications.<\/p>\n<h3 class=\"wp-block-heading\"><strong>Architecture and Core Design<\/strong><\/h3>\n<p>Both models feature <strong>4 billion total parameters<\/strong> (3.6B excluding embeddings) built across <strong>36 transformer layers<\/strong>. They use <strong>Grouped Query Attention (GQA)<\/strong> with <strong>32 query heads<\/strong> and <strong>8 key\/value heads<\/strong>, enhancing efficiency and memory management for very large contexts. They are <strong>dense transformer architectures<\/strong>\u2014not mixture-of-experts\u2014which ensures consistent task performance. Long-context support up to <strong>262,144 tokens<\/strong> is baked directly into the model architecture, and each model is pretrained extensively before undergoing <strong>alignment and safety post-training<\/strong> to ensure responsible, high-quality outputs.<\/p>\n<h3 class=\"wp-block-heading\"><strong>Qwen3-4B-Instruct-2507 \u2014 A Multilingual, Instruction-Following Generalist<\/strong><\/h3>\n<p>The <strong>Qwen3-4B-Instruct-2507<\/strong> model is optimized for speed, clarity, and user-aligned instruction following. It is designed to deliver direct answers without explicit step-by-step reasoning, making it perfect for scenarios where users want concise responses rather than detailed thought processes.<\/p>\n<p>Multilingual coverage spans <strong>over 100 languages<\/strong>, making it highly suitable for global deployments in chatbots, customer support, education, and cross-language search. Its <strong>native 256K context support<\/strong> enables it to handle tasks like analyzing large legal documents, processing multi-hour transcripts, or summarizing massive datasets without splitting the content.<\/p>\n<h4 class=\"wp-block-heading\"><strong>Performance Benchmarks:<\/strong><\/h4>\n<figure class=\"wp-block-table\">\n<table class=\"has-fixed-layout\">\n<thead>\n<tr>\n<th>Benchmark Task<\/th>\n<th>Score<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>General Knowledge (MMLU-Pro)<\/td>\n<td>69.6<\/td>\n<\/tr>\n<tr>\n<td>Reasoning (AIME25)<\/td>\n<td>47.4<\/td>\n<\/tr>\n<tr>\n<td>SuperGPQA (QA)<\/td>\n<td>42.8<\/td>\n<\/tr>\n<tr>\n<td>Coding (LiveCodeBench)<\/td>\n<td>35.1<\/td>\n<\/tr>\n<tr>\n<td>Creative Writing<\/td>\n<td>83.5<\/td>\n<\/tr>\n<tr>\n<td>Multilingual Comprehension (MultiIF)<\/td>\n<td>69.0<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<\/figure>\n<p>In practice, this means Qwen3-4B-Instruct-2507 can handle everything from <strong>language tutoring in multiple languages<\/strong> to <strong>generating rich narrative content<\/strong>, while still providing competent performance in reasoning, coding, and domain-specific knowledge.<\/p>\n<h3 class=\"wp-block-heading\"><strong>Qwen3-4B-Thinking-2507 \u2014 Expert-Level Chain-of-Thought Reasoning<\/strong><\/h3>\n<p>Where the Instruct model focuses on concise responsiveness, the <strong>Qwen3-4B-Thinking-2507<\/strong> model is engineered for <strong>deep reasoning and problem-solving<\/strong>. It automatically generates explicit <strong>chains of thought<\/strong> in its outputs, making its decision-making process transparent\u2014especially beneficial for complex domains like mathematics, science, and programming.<\/p>\n<p>This model excels at <strong>technical diagnostics<\/strong>, <strong>scientific data interpretation<\/strong>, and <strong>multi-step logical analysis<\/strong>. It\u2019s suited for advanced AI agents, research assistants, and coding companions that need to reason through problems before answering.<\/p>\n<h4 class=\"wp-block-heading\"><strong>Performance Benchmarks:<\/strong><\/h4>\n<figure class=\"wp-block-table\">\n<table class=\"has-fixed-layout\">\n<thead>\n<tr>\n<th>Benchmark Task<\/th>\n<th>Score<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>Math (AIME25)<\/td>\n<td>81.3%<\/td>\n<\/tr>\n<tr>\n<td>Science (HMMT25)<\/td>\n<td>55.5%<\/td>\n<\/tr>\n<tr>\n<td>General QA (GPQA)<\/td>\n<td>65.8%<\/td>\n<\/tr>\n<tr>\n<td>Coding (LiveCodeBench)<\/td>\n<td>55.2%<\/td>\n<\/tr>\n<tr>\n<td>Tool Usage (BFCL)<\/td>\n<td>71.2%<\/td>\n<\/tr>\n<tr>\n<td>Human Alignment<\/td>\n<td>87.4%<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<\/figure>\n<p>These scores demonstrate that Qwen3-4B-Thinking-2507 can match or even surpass much larger models in reasoning-heavy benchmarks, allowing more accurate and explainable results for mission-critical use cases.<\/p>\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img fetchpriority=\"high\" decoding=\"async\" width=\"1024\" height=\"576\" data-attachment-id=\"73416\" data-permalink=\"https:\/\/www.marktechpost.com\/2025\/08\/08\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/image-79\/\" data-orig-file=\"https:\/\/www.marktechpost.com\/wp-content\/uploads\/2025\/08\/image-3.png\" data-orig-size=\"1920,1080\" data-comments-opened=\"1\" data-image-meta='{\"aperture\":\"0\",\"credit\":\"\",\"camera\":\"\",\"caption\":\"\",\"created_timestamp\":\"0\",\"copyright\":\"\",\"focal_length\":\"0\",\"iso\":\"0\",\"shutter_speed\":\"0\",\"title\":\"\",\"orientation\":\"0\"}' data-image-title=\"image\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/www.marktechpost.com\/wp-content\/uploads\/2025\/08\/image-3-300x169.png\" data-large-file=\"https:\/\/www.marktechpost.com\/wp-content\/uploads\/2025\/08\/image-3-1024x576.png\" src=\"https:\/\/www.marktechpost.com\/wp-content\/uploads\/2025\/08\/image-3-1024x576.png\" alt=\"\" class=\"wp-image-73416\" \/><\/figure>\n<\/div>\n<h3 class=\"wp-block-heading\"><strong>Across Both Models<\/strong><\/h3>\n<p>Both the Instruct and Thinking variants share key advancements. The <strong>256K native context window<\/strong> allows for seamless work on extremely long inputs without external memory hacks. They also feature <strong>improved alignment<\/strong>, producing more natural, coherent, and context-aware responses in creative and multi-turn conversations. Furthermore, both are <strong>agent-ready<\/strong>, supporting API calling, multi-step reasoning, and workflow orchestration out-of-the-box.<\/p>\n<p>From a deployment perspective, they are highly efficient\u2014capable of running on <strong>mainstream consumer GPUs<\/strong> with quantization for lower memory usage, and fully compatible with modern inference frameworks. This means developers can <strong>run them locally or scale them in cloud environments<\/strong> without significant resource investment.<\/p>\n<h3 class=\"wp-block-heading\"><strong>Practical Deployment and Applications<\/strong><\/h3>\n<p>Deployment is straightforward, with <strong>broad framework compatibility<\/strong> enabling integration into any modern <a href=\"https:\/\/www.marktechpost.com\/2025\/01\/14\/what-is-machine-learning-ml\/\" target=\"_blank\">ML<\/a> pipeline. They can be used in edge devices, enterprise virtual assistants, research institutions, coding environments, and creative studios. Example scenarios include:<\/p>\n<ul class=\"wp-block-list\">\n<li><strong>Instruction-Following Mode<\/strong>: Customer support bots, multilingual educational assistants, real-time content generation.<\/li>\n<li><strong>Thinking Mode<\/strong>: Scientific research analysis, legal reasoning, advanced coding tools, and agentic automation.<\/li>\n<\/ul>\n<h3 class=\"wp-block-heading\"><strong>Conclusion <\/strong><\/h3>\n<p>The Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507 prove that <strong><a href=\"https:\/\/www.marktechpost.com\/2025\/01\/12\/what-are-small-language-models-slms\/\" target=\"_blank\">small language models<\/a> can rival and even outperform larger models in specific domains<\/strong> when engineered thoughtfully. Their blend of long-context handling, strong multilingual capabilities, deep reasoning (in Thinking mode), and alignment improvements makes them powerful tools for both everyday and specialist AI applications. With these releases, Alibaba has set a new benchmark in making <strong>256K-ready, high-performance AI models<\/strong> accessible to developers worldwide.<\/p>\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" \/>\n<p>Check out the\u00a0<strong><a href=\"https:\/\/huggingface.co\/Qwen\/Qwen3-4B-Instruct-2507\" target=\"_blank\" rel=\"noreferrer noopener\">Qwen3-4B-Instruct-2507 Model<\/a> <\/strong>and <strong><a href=\"https:\/\/huggingface.co\/Qwen\/Qwen3-4B-Thinking-2507\" target=\"_blank\" rel=\"noreferrer noopener\">Qwen3-4B-Thinking-2507 Model<\/a>.<\/strong>\u00a0Feel free to check out our\u00a0<strong><mark><a href=\"https:\/\/github.com\/Marktechpost\/AI-Tutorial-Codes-Included\" target=\"_blank\" rel=\"noreferrer noopener\">GitHub Page for Tutorials, Codes and Notebooks<\/a><\/mark><\/strong>.\u00a0Also,\u00a0feel free to follow us on\u00a0<strong><a href=\"https:\/\/x.com\/intent\/follow?screen_name=marktechpost\" target=\"_blank\" rel=\"noreferrer noopener\"><mark>Twitter<\/mark><\/a><\/strong>\u00a0and don\u2019t forget to Subscribe to\u00a0<strong><a href=\"https:\/\/www.aidevsignals.com\/\" target=\"_blank\" rel=\"noreferrer noopener\">our Newsletter<\/a><\/strong>.<\/p>\n\n<div class=\"wp-block-buttons is-horizontal is-content-justification-center is-layout-flex wp-container-core-buttons-is-layout-499968f5 wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button\"><a class=\"wp-block-button__link has-luminous-vivid-orange-background-color has-background wp-element-button\" href=\"https:\/\/news.ycombinator.com\/submitlink?u=https:\/\/www.marktechpost.com\/2025\/08\/08\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\" target=\"_blank\" rel=\"noreferrer noopener\"><img decoding=\"async\" src=\"https:\/\/s.w.org\/images\/core\/emoji\/16.0.1\/72x72\/1f1fe.png\" alt=\"\ud83c\uddfe\" class=\"wp-smiley\" \/> Discuss on Hacker News <\/a><\/div>\n<div class=\"wp-block-button\"><a class=\"wp-block-button__link has-vivid-cyan-blue-background-color has-background wp-element-button\" href=\"https:\/\/www.reddit.com\/r\/machinelearningnews\/\" target=\"_blank\" rel=\"noreferrer noopener\"> <img decoding=\"async\" src=\"https:\/\/s.w.org\/images\/core\/emoji\/16.0.1\/72x72\/1f1f7.png\" alt=\"\ud83c\uddf7\" class=\"wp-smiley\" \/> Join our ML Subreddit <\/a><\/div>\n<div class=\"wp-block-button\"><a class=\"wp-block-button__link has-vivid-green-cyan-background-color has-background wp-element-button\" href=\"https:\/\/promotion.marktechpost.com\/\" target=\"_blank\" rel=\"noreferrer noopener\"> <img decoding=\"async\" src=\"https:\/\/s.w.org\/images\/core\/emoji\/16.0.1\/72x72\/1f1f8.png\" alt=\"\ud83c\uddf8\" class=\"wp-smiley\" \/> Sponsor us <\/a><\/div>\n<\/div>\n<p>The post <a href=\"https:\/\/www.marktechpost.com\/2025\/08\/08\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\">Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models<\/a> appeared first on <a href=\"https:\/\/www.marktechpost.com\/\">MarkTechPost<\/a>.<\/p>","protected":false},"excerpt":{"rendered":"<p>Smaller Models with Smarter Performance and 256K Context Support Alibaba\u2019s Qwen team has introduced two powerful additions to its small language model lineup: Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507. Despite having only 4 billion parameters, these models deliver exceptional capabilities across general-purpose and expert-level tasks while running efficiently on consumer-grade hardware. Both are designed with native 256K token context windows, meaning they can process extremely long inputs such as large codebases, multi-document archives, and extended dialogues without external modifications. Architecture and Core Design Both models feature 4 billion total parameters (3.6B excluding embeddings) built across 36 transformer layers. They use Grouped Query Attention (GQA) with 32 query heads and 8 key\/value heads, enhancing efficiency and memory management for very large contexts. They are dense transformer architectures\u2014not mixture-of-experts\u2014which ensures consistent task performance. Long-context support up to 262,144 tokens is baked directly into the model architecture, and each model is pretrained extensively before undergoing alignment and safety post-training to ensure responsible, high-quality outputs. Qwen3-4B-Instruct-2507 \u2014 A Multilingual, Instruction-Following Generalist The Qwen3-4B-Instruct-2507 model is optimized for speed, clarity, and user-aligned instruction following. It is designed to deliver direct answers without explicit step-by-step reasoning, making it perfect for scenarios where users want concise responses rather than detailed thought processes. Multilingual coverage spans over 100 languages, making it highly suitable for global deployments in chatbots, customer support, education, and cross-language search. Its native 256K context support enables it to handle tasks like analyzing large legal documents, processing multi-hour transcripts, or summarizing massive datasets without splitting the content. Performance Benchmarks: Benchmark Task Score General Knowledge (MMLU-Pro) 69.6 Reasoning (AIME25) 47.4 SuperGPQA (QA) 42.8 Coding (LiveCodeBench) 35.1 Creative Writing 83.5 Multilingual Comprehension (MultiIF) 69.0 In practice, this means Qwen3-4B-Instruct-2507 can handle everything from language tutoring in multiple languages to generating rich narrative content, while still providing competent performance in reasoning, coding, and domain-specific knowledge. Qwen3-4B-Thinking-2507 \u2014 Expert-Level Chain-of-Thought Reasoning Where the Instruct model focuses on concise responsiveness, the Qwen3-4B-Thinking-2507 model is engineered for deep reasoning and problem-solving. It automatically generates explicit chains of thought in its outputs, making its decision-making process transparent\u2014especially beneficial for complex domains like mathematics, science, and programming. This model excels at technical diagnostics, scientific data interpretation, and multi-step logical analysis. It\u2019s suited for advanced AI agents, research assistants, and coding companions that need to reason through problems before answering. Performance Benchmarks: Benchmark Task Score Math (AIME25) 81.3% Science (HMMT25) 55.5% General QA (GPQA) 65.8% Coding (LiveCodeBench) 55.2% Tool Usage (BFCL) 71.2% Human Alignment 87.4% These scores demonstrate that Qwen3-4B-Thinking-2507 can match or even surpass much larger models in reasoning-heavy benchmarks, allowing more accurate and explainable results for mission-critical use cases. Across Both Models Both the Instruct and Thinking variants share key advancements. The 256K native context window allows for seamless work on extremely long inputs without external memory hacks. They also feature improved alignment, producing more natural, coherent, and context-aware responses in creative and multi-turn conversations. Furthermore, both are agent-ready, supporting API calling, multi-step reasoning, and workflow orchestration out-of-the-box. From a deployment perspective, they are highly efficient\u2014capable of running on mainstream consumer GPUs with quantization for lower memory usage, and fully compatible with modern inference frameworks. This means developers can run them locally or scale them in cloud environments without significant resource investment. Practical Deployment and Applications Deployment is straightforward, with broad framework compatibility enabling integration into any modern ML pipeline. They can be used in edge devices, enterprise virtual assistants, research institutions, coding environments, and creative studios. Example scenarios include: Instruction-Following Mode: Customer support bots, multilingual educational assistants, real-time content generation. Thinking Mode: Scientific research analysis, legal reasoning, advanced coding tools, and agentic automation. Conclusion The Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507 prove that small language models can rival and even outperform larger models in specific domains when engineered thoughtfully. Their blend of long-context handling, strong multilingual capabilities, deep reasoning (in Thinking mode), and alignment improvements makes them powerful tools for both everyday and specialist AI applications. With these releases, Alibaba has set a new benchmark in making 256K-ready, high-performance AI models accessible to developers worldwide. Check out the\u00a0Qwen3-4B-Instruct-2507 Model and Qwen3-4B-Thinking-2507 Model.\u00a0Feel free to check out our\u00a0GitHub Page for Tutorials, Codes and Notebooks.\u00a0Also,\u00a0feel free to follow us on\u00a0Twitter\u00a0and don\u2019t forget to Subscribe to\u00a0our Newsletter. Discuss on Hacker News Join our ML Subreddit Sponsor us The post Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models appeared first on MarkTechPost.<\/p>","protected":false},"author":2,"featured_media":30481,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"pmpro_default_level":"","site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"default","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"var(--ast-global-color-4)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"_pvb_checkbox_block_on_post":false,"footnotes":""},"categories":[52,5,7,1],"tags":[],"class_list":["post-30480","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-club","category-committee","category-news","category-uncategorized","pmpro-has-access"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v25.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models - YouZum<\/title>\n<meta name=\"description\" content=\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/youzum.net\/ja\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\" \/>\n<meta property=\"og:locale\" content=\"ja_JP\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models - YouZum\" \/>\n<meta property=\"og:description\" content=\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\" \/>\n<meta property=\"og:url\" content=\"https:\/\/youzum.net\/ja\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\" \/>\n<meta property=\"og:site_name\" content=\"YouZum\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/DroneAssociationTH\/\" \/>\n<meta property=\"article:published_time\" content=\"2025-08-09T05:58:33+00:00\" \/>\n<meta name=\"author\" content=\"admin NU\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"\u57f7\u7b46\u8005\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin NU\" \/>\n\t<meta name=\"twitter:label2\" content=\"\u63a8\u5b9a\u8aad\u307f\u53d6\u308a\u6642\u9593\" \/>\n\t<meta name=\"twitter:data2\" content=\"4\u5206\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\"},\"author\":{\"name\":\"admin NU\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c\"},\"headline\":\"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models\",\"datePublished\":\"2025-08-09T05:58:33+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\"},\"wordCount\":766,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\"},\"image\":{\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png\",\"articleSection\":[\"AI\",\"Committee\",\"News\",\"Uncategorized\"],\"inLanguage\":\"ja\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\",\"url\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\",\"name\":\"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models - YouZum\",\"isPartOf\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png\",\"datePublished\":\"2025-08-09T05:58:33+00:00\",\"description\":\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\",\"breadcrumb\":{\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#breadcrumb\"},\"inLanguage\":\"ja\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"ja\",\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#primaryimage\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png\",\"width\":1024,\"height\":576},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/youzum.net\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/yousum.gpucore.co\/#website\",\"url\":\"https:\/\/yousum.gpucore.co\/\",\"name\":\"YouSum\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/yousum.gpucore.co\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"ja\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\",\"name\":\"Drone Association Thailand\",\"url\":\"https:\/\/yousum.gpucore.co\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"ja\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png\",\"width\":300,\"height\":300,\"caption\":\"Drone Association Thailand\"},\"image\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/DroneAssociationTH\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c\",\"name\":\"admin NU\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"ja\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png\",\"caption\":\"admin NU\"},\"url\":\"https:\/\/youzum.net\/ja\/members\/adminnu\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models - YouZum","description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/youzum.net\/ja\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/","og_locale":"ja_JP","og_type":"article","og_title":"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models - YouZum","og_description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","og_url":"https:\/\/youzum.net\/ja\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/","og_site_name":"YouZum","article_publisher":"https:\/\/www.facebook.com\/DroneAssociationTH\/","article_published_time":"2025-08-09T05:58:33+00:00","author":"admin NU","twitter_card":"summary_large_image","twitter_misc":{"\u57f7\u7b46\u8005":"admin NU","\u63a8\u5b9a\u8aad\u307f\u53d6\u308a\u6642\u9593":"4\u5206"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#article","isPartOf":{"@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/"},"author":{"name":"admin NU","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c"},"headline":"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models","datePublished":"2025-08-09T05:58:33+00:00","mainEntityOfPage":{"@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/"},"wordCount":766,"commentCount":0,"publisher":{"@id":"https:\/\/yousum.gpucore.co\/#organization"},"image":{"@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#primaryimage"},"thumbnailUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png","articleSection":["AI","Committee","News","Uncategorized"],"inLanguage":"ja","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/","url":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/","name":"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models - YouZum","isPartOf":{"@id":"https:\/\/yousum.gpucore.co\/#website"},"primaryImageOfPage":{"@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#primaryimage"},"image":{"@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#primaryimage"},"thumbnailUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png","datePublished":"2025-08-09T05:58:33+00:00","description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","breadcrumb":{"@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#breadcrumb"},"inLanguage":"ja","potentialAction":[{"@type":"ReadAction","target":["https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/"]}]},{"@type":"ImageObject","inLanguage":"ja","@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#primaryimage","url":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png","width":1024,"height":576},{"@type":"BreadcrumbList","@id":"https:\/\/youzum.net\/alibaba-qwen-unveils-qwen3-4b-instruct-2507-and-qwen3-4b-thinking-2507-refreshing-the-importance-of-small-language-models\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/youzum.net\/"},{"@type":"ListItem","position":2,"name":"Alibaba Qwen Unveils Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507: Refreshing the Importance of Small Language Models"}]},{"@type":"WebSite","@id":"https:\/\/yousum.gpucore.co\/#website","url":"https:\/\/yousum.gpucore.co\/","name":"YouSum","description":"","publisher":{"@id":"https:\/\/yousum.gpucore.co\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/yousum.gpucore.co\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"ja"},{"@type":"Organization","@id":"https:\/\/yousum.gpucore.co\/#organization","name":"Drone Association Thailand","url":"https:\/\/yousum.gpucore.co\/","logo":{"@type":"ImageObject","inLanguage":"ja","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/","url":"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png","width":300,"height":300,"caption":"Drone Association Thailand"},"image":{"@id":"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/DroneAssociationTH\/"]},{"@type":"Person","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c","name":"admin NU","image":{"@type":"ImageObject","inLanguage":"ja","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/image\/","url":"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png","caption":"admin NU"},"url":"https:\/\/youzum.net\/ja\/members\/adminnu\/"}]}},"rttpg_featured_image_url":{"full":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png",1024,576,false],"landscape":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png",1024,576,false],"portraits":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png",1024,576,false],"thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey-150x150.png",150,150,true],"medium":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey-300x169.png",300,169,true],"large":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png",1024,576,false],"1536x1536":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png",1024,576,false],"2048x2048":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey.png",1024,576,false],"trp-custom-language-flag":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey-18x10.png",18,10,true],"woocommerce_thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey-300x300.png",300,300,true],"woocommerce_single":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey-600x338.png",600,338,true],"woocommerce_gallery_thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/08\/image-3-1024x576-wva1Ey-100x100.png",100,100,true]},"rttpg_author":{"display_name":"admin NU","author_link":"https:\/\/youzum.net\/ja\/members\/adminnu\/"},"rttpg_comment":0,"rttpg_category":"<a href=\"https:\/\/youzum.net\/ja\/category\/ai-club\/\" rel=\"category tag\">AI<\/a> <a href=\"https:\/\/youzum.net\/ja\/category\/committee\/\" rel=\"category tag\">Committee<\/a> <a href=\"https:\/\/youzum.net\/ja\/category\/news\/\" rel=\"category tag\">News<\/a> <a href=\"https:\/\/youzum.net\/ja\/category\/uncategorized\/\" rel=\"category tag\">Uncategorized<\/a>","rttpg_excerpt":"Smaller Models with Smarter Performance and 256K Context Support Alibaba\u2019s Qwen team has introduced two powerful additions to its small language model lineup: Qwen3-4B-Instruct-2507 and Qwen3-4B-Thinking-2507. Despite having only 4 billion parameters, these models deliver exceptional capabilities across general-purpose and expert-level tasks while running efficiently on consumer-grade hardware. Both are designed with native 256K token&hellip;","_links":{"self":[{"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/posts\/30480","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/comments?post=30480"}],"version-history":[{"count":0,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/posts\/30480\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/media\/30481"}],"wp:attachment":[{"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/media?parent=30480"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/categories?post=30480"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/youzum.net\/ja\/wp-json\/wp\/v2\/tags?post=30480"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}