{"id":18510,"date":"2024-04-10T15:49:20","date_gmt":"2024-04-10T15:49:20","guid":{"rendered":"http:\/\/scannn.com\/introducing-our-next-generation-infrastructure-for-ai\/"},"modified":"2024-04-10T15:49:20","modified_gmt":"2024-04-10T15:49:20","slug":"introducing-our-next-generation-infrastructure-for-ai","status":"publish","type":"post","link":"https:\/\/scannn.com\/lv\/introducing-our-next-generation-infrastructure-for-ai\/","title":{"rendered":"Introducing Our Next Generation Infrastructure for AI"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p><span style=\"font-weight: 400;\">The next generation of Meta\u2019s large-scale <\/span><a href=\"https:\/\/about.fb.com\/news\/2023\/05\/metas-infrastructure-for-ai\/\"><span style=\"font-weight: 400;\">infrastructure is being built with AI in mind,<\/span><\/a><span style=\"font-weight: 400;\"> including supporting new generative AI products, recommendation systems and advanced AI research. It\u2019s an <\/span><a href=\"https:\/\/engineering.fb.com\/2024\/03\/12\/data-center-engineering\/building-metas-genai-infrastructure\/\"><span style=\"font-weight: 400;\">investment we expect will grow<\/span><\/a><span style=\"font-weight: 400;\"> in the years ahead, as the compute requirements to support AI models increase alongside the models\u2019 sophistication.<\/span><\/p>\n<p><a href=\"https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/01_Rotating-Chip.gif\"><img fetchpriority=\"high\" decoding=\"async\" class=\"alignnone size-large wp-image-40810\" src=\"https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/01_Rotating-Chip.gif?w=960&amp;resize=960%2C836\" alt=\"A GIF of a chip rotating. \" width=\"960\" height=\"836\" data-recalc-dims=\"1\"\/><\/a><\/p>\n<p><span style=\"font-weight: 400;\">Last year, we unveiled our<\/span><a href=\"https:\/\/ai.meta.com\/blog\/meta-training-inference-accelerator-AI-MTIA\/\"> <span style=\"font-weight: 400;\">Meta Training and Inference Accelerator (MTIA) v1<\/span><\/a><span style=\"font-weight: 400;\">, our first-generation AI inference accelerator that we designed in-house with Meta\u2019s AI workloads in mind. It was designed specifically for our deep learning recommendation models that are improving a variety of experiences across our apps and technologies.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">MTIA is a long-term bet to provide the most efficient architecture for Meta\u2019s unique workloads. As AI workloads become increasingly important to our products and services, this efficiency will be central to our ability to provide the best experiences for our users around the world. MTIA v1 was an important step in improving the compute efficiency of our infrastructure and better supporting our software developers as they build AI models that will facilitate new and better user experiences.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The next generation of MTIA is part of our broader full-stack development program for custom, domain-specific silicon that addresses our unique workloads and systems. This new version of MTIA more than doubles the compute and memory bandwidth of our previous solution while maintaining our close tie-in to our workloads. It is designed to efficiently serve the ranking and recommendation models that provide high-quality recommendations to users.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">This chip\u2019s architecture is fundamentally focused on providing the right balance of compute, memory bandwidth and memory capacity for serving ranking and recommendation models.\u00a0<\/span><\/p>\n<p><a href=\"https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/02_chips-architecture.gif\"><img loading=\"lazy\" decoding=\"async\" loading=\"lazy\" class=\"alignnone size-large wp-image-40811\" src=\"https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/02_chips-architecture.gif?w=960&amp;resize=960%2C836\" alt=\"A GIF showing the chip's architecture.\" width=\"960\" height=\"836\" data-recalc-dims=\"1\"\/><\/a><\/p>\n<p><span style=\"font-weight: 400;\">MTIA has been deployed in our data centers and is now serving models in production. We are already seeing the positive results of this program as it\u2019s allowing us to dedicate and invest in more compute power for our more intensive AI workloads.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The results so far show that this MTIA chip can handle both low complexity and high complexity ranking and recommendation models which are key components of Meta\u2019s products.\u00a0 Because we control the whole stack, we can achieve greater efficiency compared to commercially available GPUs (graphics processing units).\u00a0<\/span><\/p>\n<p class=\"jetpack-slideshow-noscript robots-nocontent\">This slideshow requires JavaScript.<\/p>\n<h2><span style=\"font-weight: 400;\">Meta\u2019s Ongoing Investment in Custom Silicon<\/span><\/h2>\n<p><span style=\"font-weight: 400;\">MTIA will be an important piece of our long-term roadmap to build and scale the most powerful and efficient infrastructure possible for Meta\u2019s unique AI workloads.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">We\u2019re designing our custom silicon to work in cooperation with our existing infrastructure as well as with new, more advanced hardware (including next-generation GPUs) that we may leverage in the future. Meeting our ambitions for our custom silicon means investing not only in compute silicon but also in memory bandwidth, networking and capacity, as well as other next-generation hardware systems.<\/span><\/p>\n<p><a href=\"https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg\"><img loading=\"lazy\" decoding=\"async\" loading=\"lazy\" class=\"alignnone size-large wp-image-40814\" src=\"https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg?w=960&amp;resize=960%2C836\" alt=\"An image to show our ongoing investment in custom silicon.\" width=\"960\" height=\"836\" srcset=\"https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg?w=1920 1920w, https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg?w=300 300w, https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg?w=768 768w, https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg?w=1024 1024w, https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg?w=1536 1536w, https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg?w=1240 1240w, https:\/\/about.fb.com\/wp-content\/uploads\/2024\/04\/04_Metas-Ongoing-Investment-in-Custom-Silicon.jpg?w=689 689w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" data-recalc-dims=\"1\"\/><\/a><\/p>\n<p><span style=\"font-weight: 400;\">We currently have several programs underway aimed at expanding the scope of MTIA, including support for GenAI workloads. And we\u2019re only at the beginning of this journey.<\/span><\/p>\n<\/p><\/div>\n<p><script async defer crossorigin=\"anonymous\" src=\"https:\/\/connect.facebook.net\/en_US\/sdk.js#xfbml=1&#038;version=v5.0\"><\/script><br \/>\n<br \/><br \/>\n<br \/><a href=\"https:\/\/about.fb.com\/news\/2024\/04\/introducing-our-next-generation-infrastructure-for-ai\/\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>The next generation of Meta\u2019s large-scale infrastructure is being built with AI in mind, including supporting new generative AI products, recommendation systems and advanced AI research. It\u2019s an investment we expect will grow in the years ahead, as the compute requirements to support AI models increase alongside the models\u2019 sophistication. Last year, we unveiled our [&hellip;]<\/p>\n","protected":false},"author":16,"featured_media":18511,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[123],"tags":[],"class_list":["post-18510","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-facebook"],"_links":{"self":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts\/18510","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/users\/16"}],"replies":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/comments?post=18510"}],"version-history":[{"count":0,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts\/18510\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/media\/18511"}],"wp:attachment":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/media?parent=18510"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/categories?post=18510"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/tags?post=18510"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}