{"id":14783,"date":"2025-05-25T03:05:02","date_gmt":"2025-05-25T03:05:02","guid":{"rendered":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/"},"modified":"2025-05-25T03:05:02","modified_gmt":"2025-05-25T03:05:02","slug":"optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers","status":"publish","type":"post","link":"https:\/\/youzum.net\/fr\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/","title":{"rendered":"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers"},"content":{"rendered":"<p>LLMs have shown impressive capabilities across various programming tasks, yet their potential for program optimization has not been fully explored. While some recent efforts have used LLMs to enhance performance in languages like C++ and Python, the broader application of LLMs to optimize code, especially in low-level programming contexts, remains limited. Existing LLM benchmarks largely focus on code generation from natural language or solving GitHub issues, as seen in HumanEval, MBPP, APPS, SWE-bench, and SWE-agent. Moreover, models such as Codex, AlphaCode, and Code Llama primarily aim to improve code generation quality rather than performance. However, select research has begun addressing optimization, including parallelization and code efficiency improvements, though many of these approaches are constrained by the need for formal verification, limiting scalability.<\/p>\n<p>In contrast, some newer methods embrace test-based validation, allowing optimization of more complex programs with loops. Learning-based strategies in compiler optimization\u2014like AutoPhase, which uses reinforcement learning for pass sequencing, and Coreset, which applies graph neural networks\u2014have shown promise in improving performance. Superoptimization techniques aim to find the most efficient version of a program but are typically restricted to small-scale problems. Additionally, frameworks like AutoTVM and Ansor have focused on optimizing GPU kernel code through statistical modeling and search. Recently, LLM-driven optimization has gained attention, with reinforcement learning approaches guiding LLMs using feedback from test cases. Techniques like CodeRL and PPOCoder leverage policy optimization methods to fine-tune models for better performance, even across resource-constrained programming languages like Verilog.\u00a0<\/p>\n<p>Stanford, UIUC, CMU, and Visa Research researchers explore using LLMs to optimize assembly code performance\u2014an area traditionally handled by compilers like GCC. They introduce a reinforcement learning framework using Proximal Policy Optimization (PPO), guided by a reward balancing correctness and speedup over the gcc -O3 baseline. Using a dataset of 8,072 real-world programs, their model, Qwen2.5-Coder-7B-PPO, achieves a 96.0% test pass rate and a 1.47\u00d7 average speedup, outperforming 20 other models, including Claude-3.7-sonnet. Their results show that with RL training, LLMs can effectively outperform conventional compiler optimizations.\u00a0<\/p>\n<p>The methodology involves optimizing compiled C programs for performance using an RL approach. Given a C program C, it is compiled to assembly P using gcc -O3. The goal is to generate a new assembly program P\u2019 that is functionally equivalent but faster. Correctness is verified using a test set, and speedup is measured by execution time improvement. Using CodeNet as the dataset, the authors apply PPO to train a language model that generates improved code. Two reward functions\u2014Correctness-Guided Speedup and Speedup-Only\u2014are used to guide training based on program validity, correctness, and performance gains.\u00a0<\/p>\n<p>The study evaluates various language models on optimizing assembly code, revealing that most models struggle with low test pass rates and minimal speedups. However, Qwen2.5-Coder-7B-PPO, trained with reinforcement learning, significantly outperforms others, achieving 96% accuracy and a 1.47\u00d7 average speedup. Ablation studies show that using gcc -O3 as a reference aids performance, while removing it leads to sharp declines. Notably, models like Claude-3.7-sonnet can surpass compilers by identifying hardware-specific optimizations, such as replacing loops with a single popcnt instruction, demonstrating their ability to perform semantic-level code transformations beyond traditional compiler capabilities.\u00a0<\/p>\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg?key=0LMdRrgIV9G7NDzk6S6Cfg\" alt=\"\" \/><\/figure>\n<\/div>\n<p>In conclusion, the study explores using LLMs to optimize assembly code, a domain where traditional compilers struggle due to the complexity of low-level performance tuning. The authors fine-tune Qwen2.5-Coder-7B using PPO, rewarding both correctness (via test cases) and speedup over gcc -O3. They introduce a benchmark of 8,072 real-world C programs to evaluate performance. The model achieves a 96.0% test pass rate and a 1.47\u00d7 average speedup, outperforming 20 other models, including Claude-3.7-sonnet. While effective, limitations include a lack of formal correctness guarantees and variability in hardware performance across systems.\u00a0<\/p>\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" \/>\n<p><strong>Check out the <a href=\"https:\/\/arxiv.org\/abs\/2505.11480\" target=\"_blank\" rel=\"noreferrer noopener\">Paper<\/a><\/strong>.\u00a0All credit for this research goes to the researchers of this project. Also,\u00a0feel free to follow us on\u00a0<strong><a href=\"https:\/\/x.com\/intent\/follow?screen_name=marktechpost\" target=\"_blank\" rel=\"noreferrer noopener\"><mark>Twitter<\/mark><\/a><\/strong>\u00a0and don\u2019t forget to join our\u00a0<strong><a href=\"https:\/\/www.reddit.com\/r\/machinelearningnews\/\" target=\"_blank\" rel=\"noreferrer noopener\">95k+ ML SubReddit<\/a><\/strong> and Subscribe to <strong><a href=\"https:\/\/www.airesearchinsights.com\/subscribe\" target=\"_blank\" rel=\"noreferrer noopener\">our Newsletter<\/a><\/strong>.<\/p>\n<p>The post <a href=\"https:\/\/www.marktechpost.com\/2025\/05\/24\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/\">Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers<\/a> appeared first on <a href=\"https:\/\/www.marktechpost.com\/\">MarkTechPost<\/a>.<\/p>","protected":false},"excerpt":{"rendered":"<p>LLMs have shown impressive capabilities across various programming tasks, yet their potential for program optimization has not been fully explored. While some recent efforts have used LLMs to enhance performance in languages like C++ and Python, the broader application of LLMs to optimize code, especially in low-level programming contexts, remains limited. Existing LLM benchmarks largely focus on code generation from natural language or solving GitHub issues, as seen in HumanEval, MBPP, APPS, SWE-bench, and SWE-agent. Moreover, models such as Codex, AlphaCode, and Code Llama primarily aim to improve code generation quality rather than performance. However, select research has begun addressing optimization, including parallelization and code efficiency improvements, though many of these approaches are constrained by the need for formal verification, limiting scalability. In contrast, some newer methods embrace test-based validation, allowing optimization of more complex programs with loops. Learning-based strategies in compiler optimization\u2014like AutoPhase, which uses reinforcement learning for pass sequencing, and Coreset, which applies graph neural networks\u2014have shown promise in improving performance. Superoptimization techniques aim to find the most efficient version of a program but are typically restricted to small-scale problems. Additionally, frameworks like AutoTVM and Ansor have focused on optimizing GPU kernel code through statistical modeling and search. Recently, LLM-driven optimization has gained attention, with reinforcement learning approaches guiding LLMs using feedback from test cases. Techniques like CodeRL and PPOCoder leverage policy optimization methods to fine-tune models for better performance, even across resource-constrained programming languages like Verilog.\u00a0 Stanford, UIUC, CMU, and Visa Research researchers explore using LLMs to optimize assembly code performance\u2014an area traditionally handled by compilers like GCC. They introduce a reinforcement learning framework using Proximal Policy Optimization (PPO), guided by a reward balancing correctness and speedup over the gcc -O3 baseline. Using a dataset of 8,072 real-world programs, their model, Qwen2.5-Coder-7B-PPO, achieves a 96.0% test pass rate and a 1.47\u00d7 average speedup, outperforming 20 other models, including Claude-3.7-sonnet. Their results show that with RL training, LLMs can effectively outperform conventional compiler optimizations.\u00a0 The methodology involves optimizing compiled C programs for performance using an RL approach. Given a C program C, it is compiled to assembly P using gcc -O3. The goal is to generate a new assembly program P\u2019 that is functionally equivalent but faster. Correctness is verified using a test set, and speedup is measured by execution time improvement. Using CodeNet as the dataset, the authors apply PPO to train a language model that generates improved code. Two reward functions\u2014Correctness-Guided Speedup and Speedup-Only\u2014are used to guide training based on program validity, correctness, and performance gains.\u00a0 The study evaluates various language models on optimizing assembly code, revealing that most models struggle with low test pass rates and minimal speedups. However, Qwen2.5-Coder-7B-PPO, trained with reinforcement learning, significantly outperforms others, achieving 96% accuracy and a 1.47\u00d7 average speedup. Ablation studies show that using gcc -O3 as a reference aids performance, while removing it leads to sharp declines. Notably, models like Claude-3.7-sonnet can surpass compilers by identifying hardware-specific optimizations, such as replacing loops with a single popcnt instruction, demonstrating their ability to perform semantic-level code transformations beyond traditional compiler capabilities.\u00a0 In conclusion, the study explores using LLMs to optimize assembly code, a domain where traditional compilers struggle due to the complexity of low-level performance tuning. The authors fine-tune Qwen2.5-Coder-7B using PPO, rewarding both correctness (via test cases) and speedup over gcc -O3. They introduce a benchmark of 8,072 real-world C programs to evaluate performance. The model achieves a 96.0% test pass rate and a 1.47\u00d7 average speedup, outperforming 20 other models, including Claude-3.7-sonnet. While effective, limitations include a lack of formal correctness guarantees and variability in hardware performance across systems.\u00a0 Check out the Paper.\u00a0All credit for this research goes to the researchers of this project. Also,\u00a0feel free to follow us on\u00a0Twitter\u00a0and don\u2019t forget to join our\u00a095k+ ML SubReddit and Subscribe to our Newsletter. The post Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers appeared first on MarkTechPost.<\/p>","protected":false},"author":2,"featured_media":14784,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"pmpro_default_level":"","site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"default","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"var(--ast-global-color-4)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"_pvb_checkbox_block_on_post":false,"footnotes":""},"categories":[52,5,7,1],"tags":[],"class_list":["post-14783","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-club","category-committee","category-news","category-uncategorized","pmpro-has-access"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v25.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers - YouZum<\/title>\n<meta name=\"description\" content=\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/youzum.net\/fr\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/\" \/>\n<meta property=\"og:locale\" content=\"fr_FR\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers - YouZum\" \/>\n<meta property=\"og:description\" content=\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\" \/>\n<meta property=\"og:url\" content=\"https:\/\/youzum.net\/fr\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/\" \/>\n<meta property=\"og:site_name\" content=\"YouZum\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/DroneAssociationTH\/\" \/>\n<meta property=\"article:published_time\" content=\"2025-05-25T03:05:02+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png\" \/>\n\t<meta property=\"og:image:width\" content=\"915\" \/>\n\t<meta property=\"og:image:height\" content=\"552\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"admin NU\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"\u00c9crit par\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin NU\" \/>\n\t<meta name=\"twitter:label2\" content=\"Dur\u00e9e de lecture estim\u00e9e\" \/>\n\t<meta name=\"twitter:data2\" content=\"3 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/\"},\"author\":{\"name\":\"admin NU\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c\"},\"headline\":\"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers\",\"datePublished\":\"2025-05-25T03:05:02+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/\"},\"wordCount\":670,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\"},\"image\":{\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png\",\"articleSection\":[\"AI\",\"Committee\",\"News\",\"Uncategorized\"],\"inLanguage\":\"fr-FR\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/\",\"url\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/\",\"name\":\"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers - YouZum\",\"isPartOf\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png\",\"datePublished\":\"2025-05-25T03:05:02+00:00\",\"description\":\"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19\",\"breadcrumb\":{\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#breadcrumb\"},\"inLanguage\":\"fr-FR\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"fr-FR\",\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#primaryimage\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png\",\"width\":915,\"height\":552},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/youzum.net\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/yousum.gpucore.co\/#website\",\"url\":\"https:\/\/yousum.gpucore.co\/\",\"name\":\"YouSum\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/yousum.gpucore.co\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"fr-FR\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/yousum.gpucore.co\/#organization\",\"name\":\"Drone Association Thailand\",\"url\":\"https:\/\/yousum.gpucore.co\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"fr-FR\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png\",\"width\":300,\"height\":300,\"caption\":\"Drone Association Thailand\"},\"image\":{\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/DroneAssociationTH\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c\",\"name\":\"admin NU\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"fr-FR\",\"@id\":\"https:\/\/yousum.gpucore.co\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png\",\"contentUrl\":\"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png\",\"caption\":\"admin NU\"},\"url\":\"https:\/\/youzum.net\/fr\/members\/adminnu\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers - YouZum","description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/youzum.net\/fr\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/","og_locale":"fr_FR","og_type":"article","og_title":"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers - YouZum","og_description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","og_url":"https:\/\/youzum.net\/fr\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/","og_site_name":"YouZum","article_publisher":"https:\/\/www.facebook.com\/DroneAssociationTH\/","article_published_time":"2025-05-25T03:05:02+00:00","og_image":[{"width":915,"height":552,"url":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png","type":"image\/png"}],"author":"admin NU","twitter_card":"summary_large_image","twitter_misc":{"\u00c9crit par":"admin NU","Dur\u00e9e de lecture estim\u00e9e":"3 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#article","isPartOf":{"@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/"},"author":{"name":"admin NU","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c"},"headline":"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers","datePublished":"2025-05-25T03:05:02+00:00","mainEntityOfPage":{"@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/"},"wordCount":670,"commentCount":0,"publisher":{"@id":"https:\/\/yousum.gpucore.co\/#organization"},"image":{"@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#primaryimage"},"thumbnailUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png","articleSection":["AI","Committee","News","Uncategorized"],"inLanguage":"fr-FR","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/","url":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/","name":"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers - YouZum","isPartOf":{"@id":"https:\/\/yousum.gpucore.co\/#website"},"primaryImageOfPage":{"@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#primaryimage"},"image":{"@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#primaryimage"},"thumbnailUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png","datePublished":"2025-05-25T03:05:02+00:00","description":"\u0e01\u0e34\u0e08\u0e01\u0e23\u0e23\u0e21\u0e40\u0e01\u0e35\u0e48\u0e22\u0e27\u0e01\u0e31\u0e1a\u0e42\u0e14\u0e23\u0e19","breadcrumb":{"@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#breadcrumb"},"inLanguage":"fr-FR","potentialAction":[{"@type":"ReadAction","target":["https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/"]}]},{"@type":"ImageObject","inLanguage":"fr-FR","@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#primaryimage","url":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png","width":915,"height":552},{"@type":"BreadcrumbList","@id":"https:\/\/youzum.net\/optimizing-assembly-code-with-llms-reinforcement-learning-outperforms-traditional-compilers\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/youzum.net\/"},{"@type":"ListItem","position":2,"name":"Optimizing Assembly Code with LLMs: Reinforcement Learning Outperforms Traditional Compilers"}]},{"@type":"WebSite","@id":"https:\/\/yousum.gpucore.co\/#website","url":"https:\/\/yousum.gpucore.co\/","name":"YouSum","description":"","publisher":{"@id":"https:\/\/yousum.gpucore.co\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/yousum.gpucore.co\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"fr-FR"},{"@type":"Organization","@id":"https:\/\/yousum.gpucore.co\/#organization","name":"Drone Association Thailand","url":"https:\/\/yousum.gpucore.co\/","logo":{"@type":"ImageObject","inLanguage":"fr-FR","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/","url":"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/2024\/11\/tranparent-logo.png","width":300,"height":300,"caption":"Drone Association Thailand"},"image":{"@id":"https:\/\/yousum.gpucore.co\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/DroneAssociationTH\/"]},{"@type":"Person","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/97fa48242daf3908e4d9a5f26f4a059c","name":"admin NU","image":{"@type":"ImageObject","inLanguage":"fr-FR","@id":"https:\/\/yousum.gpucore.co\/#\/schema\/person\/image\/","url":"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png","contentUrl":"https:\/\/youzum.net\/wp-content\/uploads\/avatars\/2\/1746849356-bpfull.png","caption":"admin NU"},"url":"https:\/\/youzum.net\/fr\/members\/adminnu\/"}]}},"rttpg_featured_image_url":{"full":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png",915,552,false],"landscape":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png",915,552,false],"portraits":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png",915,552,false],"thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw-150x150.png",150,150,true],"medium":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw-300x181.png",300,181,true],"large":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png",915,552,false],"1536x1536":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png",915,552,false],"2048x2048":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw.png",915,552,false],"trp-custom-language-flag":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw-18x12.png",18,12,true],"woocommerce_thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw-300x300.png",300,300,true],"woocommerce_single":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw-600x362.png",600,362,true],"woocommerce_gallery_thumbnail":["https:\/\/youzum.net\/wp-content\/uploads\/2025\/05\/AD_4nXdxqXsXoJLSlQVLVHCJG256NvHRYNb-_IAJBulneLnV5RoKCzbLfYSTltd68KSMEA4hK25q8lZqiKQGizdrI28Qrgd5XrkIhYO8OSR9vnv12vfU5dJ5FooIPuDQ_adZSTVzRZ7Emg-vfb6Kw-100x100.png",100,100,true]},"rttpg_author":{"display_name":"admin NU","author_link":"https:\/\/youzum.net\/fr\/members\/adminnu\/"},"rttpg_comment":0,"rttpg_category":"<a href=\"https:\/\/youzum.net\/fr\/category\/ai-club\/\" rel=\"category tag\">AI<\/a> <a href=\"https:\/\/youzum.net\/fr\/category\/committee\/\" rel=\"category tag\">Committee<\/a> <a href=\"https:\/\/youzum.net\/fr\/category\/news\/\" rel=\"category tag\">News<\/a> <a href=\"https:\/\/youzum.net\/fr\/category\/uncategorized\/\" rel=\"category tag\">Uncategorized<\/a>","rttpg_excerpt":"LLMs have shown impressive capabilities across various programming tasks, yet their potential for program optimization has not been fully explored. While some recent efforts have used LLMs to enhance performance in languages like C++ and Python, the broader application of LLMs to optimize code, especially in low-level programming contexts, remains limited. Existing LLM benchmarks largely\u2026","_links":{"self":[{"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/posts\/14783","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/comments?post=14783"}],"version-history":[{"count":0,"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/posts\/14783\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/media\/14784"}],"wp:attachment":[{"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/media?parent=14783"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/categories?post=14783"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/youzum.net\/fr\/wp-json\/wp\/v2\/tags?post=14783"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}