{"id":21584,"date":"2026-03-11T14:48:00","date_gmt":"2026-03-11T14:48:00","guid":{"rendered":"https:\/\/scannn.com\/expanding-metas-custom-silicon-to-power-our-ai-workloads\/"},"modified":"2026-03-11T14:48:00","modified_gmt":"2026-03-11T14:48:00","slug":"expanding-metas-custom-silicon-to-power-our-ai-workloads","status":"publish","type":"post","link":"https:\/\/scannn.com\/lv\/expanding-metas-custom-silicon-to-power-our-ai-workloads\/","title":{"rendered":"Expanding Meta\u2019s Custom Silicon to Power Our AI Workloads"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p><span style=\"font-weight: 400\">In 2023, we developed the <\/span><a href=\"https:\/\/ai.meta.com\/blog\/meta-training-inference-accelerator-AI-MTIA\/\"><span style=\"font-weight: 400\">Meta Training and Inference Accelerator (MTIA)<\/span><\/a><span style=\"font-weight: 400\">, a family of custom-built <\/span><a href=\"https:\/\/about.fb.com\/news\/2024\/04\/introducing-our-next-generation-infrastructure-for-ai\/\"><span style=\"font-weight: 400\">silicon chips<\/span><\/a><span style=\"font-weight: 400\"> to power our AI workloads efficiently. Now, we\u2019re developing and deploying four new generations of chips within the next two years \u2014 a much faster pace than typical chip cycles \u2014 to support ranking, recommendations, and GenAI workloads.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400\">As our current AI workloads continue to grow and evolve, we\u2019re taking a portfolio approach to scale our infrastructure capacity by sourcing silicon from a range of industry leaders, while keeping our own MTIA custom silicon at the center of our AI infrastructure strategy.\u00a0<\/span><\/p>\n<h2>A Custom Solution<\/h2>\n<p><span style=\"font-weight: 400\">We deploy hundreds of thousands of MTIA chips for inference workloads across both organic content and ads on our apps. These chips are specifically designed for our workloads, and are part of a custom full-stack solution, helping us create a highly optimized system that\u2019s tailored to our needs. This system achieves greater compute efficiency than general use chips for our intended purposes, making MTIA much more cost efficient.\u00a0<\/span><\/p>\n<h2>Four Chips in Two Years<span style=\"font-weight: 400\">\u00a0<\/span><\/h2>\n<p><span style=\"font-weight: 400\">We\u2019re continuing to advance the MTIA roadmap by developing four new generations of chips, each bringing significant improvements in compute, memory bandwidth, and efficiency. MTIA 300 will be used for ranking and recommendations training, and is already in production. MTIA 400, 450 and 500 will be capable of handling all workloads, but we will primarily use these chips to support GenAI inference production in the near future and into 2027.<\/span><\/p>\n<p><span style=\"font-weight: 400\">The modularity of our silicon allows these new chips to drop into existing rack system infrastructure, accelerating time-to-production.<\/span><\/p>\n<h2>Our MTIA Strategy<\/h2>\n<p><span style=\"font-weight: 400\">We\u2019ve developed a competitive strategy for MTIA by prioritizing rapid, iterative development, an inference-first focus, and frictionless adoption by building natively on industry standards.\u00a0<\/span><\/p>\n<p><i><span style=\"font-weight: 400\">Rapid, Iterative Development\u00a0<\/span><\/i><\/p>\n<p><span style=\"font-weight: 400\">While the industry typically launches a new AI chip every one to two years, we\u2019ve developed the capacity to release ours every six months or less by building on our modular, reusable designs. This accelerated pace enables us to quickly adapt to evolving AI techniques, adopt the latest hardware technologies, and minimize costs associated with developing and deploying new chip generations.\u00a0<\/span><\/p>\n<p><i><span style=\"font-weight: 400\">Inference-First Focus\u00a0<\/span><\/i><\/p>\n<p><span style=\"font-weight: 400\">Mainstream chips are typically built for the most demanding workload \u2014 large-scale GenAI pre-training \u2014 and then applied, often less cost-effectively, to other workloads like GenAI inference. We take the opposite approach: MTIA 450 and 500 are optimized first for GenAI inference, and they can then be used to support other workloads as needed, including ranking and recommendations training and inference, as well as GenAI training. This keeps MTIA well-tuned to the anticipated growth in GenAI inference demand.<\/span><\/p>\n<p><i><span style=\"font-weight: 400\">Building on Industry Standards<\/span><\/i><\/p>\n<p><span style=\"font-weight: 400\">MTIA is built on industry\u2011standard software and hardware ecosystems, like PyTorch, vLLM, Triton, and the Open Compute Project (OCP), from the beginning, enabling frictionless adoption of MTIA chips. Beyond industry-standard software, MTIA\u2019s system and rack solutions align with OCP standards, enabling MTIA to be seamlessly deployed in data centers.<\/span><\/p>\n<h2>Our Portfolio Approach<span style=\"font-weight: 400\">\u00a0<\/span><\/h2>\n<p><span style=\"font-weight: 400\">There is no single chip that can meet all the demands across our varying needs, which is why we\u2019re working to deploy a variety of chips that are optimized for each of our different workloads. We believe our portfolio approach will enable us to advance and innovate at an unmatched pace, bringing us closer to our goal of creating personal superintelligence for all.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400\">To learn more about the MTIA roadmap, head to the <\/span><a href=\"https:\/\/ai.meta.com\/blog\/meta-mtia-scale-ai-chips-for-billions\/\"><span style=\"font-weight: 400\">Meta AI blog<\/span><\/a><span style=\"font-weight: 400\">. <\/span><\/p>\n<\/p><\/div>\n<p><script async defer crossorigin=\"anonymous\" src=\"https:\/\/connect.facebook.net\/en_US\/sdk.js#xfbml=1&#038;version=v5.0\"><\/script><br \/>\n<br \/><br \/>\n<br \/><a href=\"https:\/\/about.fb.com\/news\/2026\/03\/expanding-metas-custom-silicon-to-power-our-ai-workloads\/\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>In 2023, we developed the Meta Training and Inference Accelerator (MTIA), a family of custom-built silicon chips to power our AI workloads efficiently. Now, we\u2019re developing and deploying four new generations of chips within the next two years \u2014 a much faster pace than typical chip cycles \u2014 to support ranking, recommendations, and GenAI workloads.\u00a0 [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":21585,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[123],"tags":[],"class_list":["post-21584","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-facebook"],"_links":{"self":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts\/21584","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/comments?post=21584"}],"version-history":[{"count":0,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts\/21584\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/media\/21585"}],"wp:attachment":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/media?parent=21584"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/categories?post=21584"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/tags?post=21584"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}