{"id":324342,"date":"2026-04-25T13:34:52","date_gmt":"2026-04-25T08:04:52","guid":{"rendered":"https:\/\/ebiztoday.news\/?p=324342"},"modified":"2026-04-25T13:34:52","modified_gmt":"2026-04-25T08:04:52","slug":"metas-compute-grab-continues-with-agreement-to-deploy-tens-of-hundreds-of-thousands-of-aws-graviton-cores","status":"publish","type":"post","link":"https:\/\/ebiztoday.news\/index.php\/2026\/04\/25\/metas-compute-grab-continues-with-agreement-to-deploy-tens-of-hundreds-of-thousands-of-aws-graviton-cores\/","title":{"rendered":"Meta\u2019s compute grab continues with agreement to deploy tens of hundreds of thousands of AWS Graviton cores"},"content":{"rendered":"<div id=\"remove_no_follow\">\n<div class=\"grid grid--cols-10@md grid--cols-8@lg article-column\">\n<div class=\"col-12 col-10@md col-6@lg col-start-3@lg\">\n<div class=\"article-column__content\">\n<section class=\"wp-block-bigbite-multi-title\">\n<div class=\"container\"><\/div>\n<\/section>\n<p>Meta is constant its compute grab because the agentic AI race accelerates to a sprint.<\/p>\n<p>Today, the corporate announced a partnership with <a href=\"https:\/\/www.networkworld.com\/article\/4157477\/ai-demand-is-so-high-aws-customers-are-trying-to-buy-out-its-entire-capacity.html\" target=\"_blank\">Amazon Web Services<\/a> (AWS) that can bring \u201ctens of hundreds of thousands\u201d of AWS Graviton5 cores (one chip comprises 192 cores) into its compute portfolio, with the choice to expand as its AI capabilities grow. This may make the Llama builder one among the most important Graviton customers on this planet.<\/p>\n<p>The move builds on Meta\u2019s expansive partnerships with nearly every chip and compute provider within the business. It\u2019s working with Nvidia, Arm, and AMD, in addition to constructing its own internal training and inference accelerator chip.<\/p>\n<p>\u201cIt feels very difficult to maintain track of what Meta is doing, with all of those chip deals and announcements around in-house development,\u201d said <a href=\"https:\/\/moorinsightsstrategy.com\/team\/matt-kimball\/\" target=\"_blank\" rel=\"noreferrer noopener\">Matt Kimball<\/a>, VP and principal analyst at Moor Insights &#038; Strategy. This makes for \u201cexciting times that tell us just how incredibly priceless silicon is straight away.\u201d<\/p>\n<h2 class=\"wp-block-heading\" id=\"controlling-the-system-not-just-scale\">Controlling the system, not only scale<\/h2>\n<p>Graphics processing units (GPUs) are essential for giant language model (LLM) training, but agentic AI requires an entire recent workload capability. CPUs like Graviton5 are rising to this challenge, supporting intensive workloads like real-time reasoning, multi-step tasks, frontier model training, code generation, and deep research.<\/p>\n<p>AWS says Graviton5 has the flexibility to handle \u201cbillions of interactions\u201d and to coordinate complex, multi-stage agentic tasks. It&#8217;s built on the <a href=\"https:\/\/aws.amazon.com\/ec2\/nitro\/\" target=\"_blank\" rel=\"noreferrer noopener\">AWS Nitro System<\/a> to support high performance, availability, and security.<\/p>\n<p>\u201cThis is basically about control of the AI system, not only scale,\u201d said Kimball. As AI evolves toward persistent, agentic workloads, the role of the CPU becomes \u201cquite meaningful;\u201d it serves because the control plane, handling orchestration, managing memory, scheduling, and other intensive tasks across accelerators.<\/p>\n<p>\u201cThis is particularly true in agentic environments, where the workloads will probably be less linear and more stateful,\u201d he identified. So, ensuring a supply of those resources just is sensible.<\/p>\n<h2 class=\"wp-block-heading\" id=\"reflecting-metas-diversified-approach-to-hardware\">Reflecting Meta\u2019s diversified approach to hardware<\/h2>\n<p>The agreement builds on Meta\u2019s long-standing partnership with AWS, but in addition reflects what the corporate calls its \u201cdiversified approach\u201d to infrastructure. \u201cNo single chip architecture can efficiently serve every workload,\u201d the corporate <a href=\"https:\/\/about.fb.com\/news\/2026\/04\/meta-partners-with-aws-on-graviton-chips-to-power-agentic-ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">emphasized<\/a>.<\/p>\n<p>Proving the purpose, Meta recently <a href=\"https:\/\/www.networkworld.com\/article\/4145022\/meta-is-developing-more-ai-chips-for-itself.html\" target=\"_blank\">announced 4 recent generations<\/a> of its MTIA training and inference accelerator chip and signed a <a href=\"https:\/\/www.networkworld.com\/article\/4137299\/amd-strikes-massive-ai-chip-deal-with-meta.html\" target=\"_blank\">massive deal<\/a> with AMD to tap into 6GW value of CPUs and AI accelerators. It also entered right into a <a href=\"https:\/\/www.networkworld.com\/article\/4135325\/meta-scoops-up-more-of-nvidias-ai-chip-output.html\" target=\"_blank\">multi-year partnership<\/a> with Nvidia to access hundreds of thousands of Blackwell and Rubin GPUs and to integrate Nvidia Spectrum-X Ethernet switches into its platform, and was also one among Arm\u2019s first major CPU customers.<\/p>\n<p>Within the wake of all this, <a href=\"https:\/\/www.infotech.com\/profiles\/nabeel-sherif\" target=\"_blank\" rel=\"noreferrer noopener\">Nabeel Sherif<\/a>, a principal advisory director at Info-Tech Research Group, posed the burning query: \u201cWhat are they going to do with all this capability?\u201d<\/p>\n<p>Primarily it&#8217;s going to support Meta\u2019s internal experimentation and innovation, he said, but it surely also lays the groundwork and provides the capability for Meta to supply its own agentic AI services, as an example, its <a href=\"https:\/\/www.infoworld.com\/article\/3975132\/meta-will-offer-its-llama-ai-model-as-an-api-too.html\" target=\"_blank\">Llama AI model as an API<\/a>, to the market.<\/p>\n<p>\u201cWhat those [services] will appear like and what platforms and tools they\u2019ll use, in addition to what guardrails they\u2019ll provide to users, continues to be unclear, but it surely\u2019s going to be interesting to see it develop,\u201d said Sherif.<\/p>\n<p>The expanded capability will enable a diversity of use cases and experimentation across various architectures and platforms, he said. Meta could have many options, and access to provide in an environment currently characterised not only by a wide range of latest CPU approaches, but by significant supply chain constraints. The AWS deal ought to be viewed as a complement to its partnerships and investments in other platforms like ARM, Nvidia, and AMD.<\/p>\n<p>Kimball agreed that the move is \u201cmost definitely additive,\u201d not a substitute or substitution. Meta isn\u2019t moving off GPUs or accelerators, it\u2019s constructing around them. \u201cThat is about assembling a heterogeneous system, not picking a single winner,\u201d he said. \u201cIn actual fact, I believe for many, heterogeneity is critical to long run success.\u201d<\/p>\n<p>Nvidia still dominates training and loads of inference, while AMD is becoming \u201cincreasingly more relevant at scale,\u201d Kimball noted. Arm, meanwhile, whether through CPU, custom silicon or other efforts, gives Meta architectural control, and Graviton5 matches into that blend as a \u201ccost- and efficiency-optimized general-purpose compute layer.\u201d<\/p>\n<h2 class=\"wp-block-heading\" id=\"a-question-of-strategy\">A matter of strategy<\/h2>\n<p>The more interesting query is around strategy: Does this signal Meta is becoming a compute provider? Kimball doesn\u2019t think so, noting that it\u2019s likely the corporate isn\u2019t seeking to directly compete with hyperscalers as a general-purpose cloud. \u201cThat is more about vertical integration of their very own AI stack,\u201d he said.<\/p>\n<p>The move gives them the flexibility to support internal workloads more efficiently, in addition to providing the infrastructure foundation to reveal more of that capability externally, whether through APIs, partnerships, or other means, he said.<\/p>\n<p>And there\u2019s a price dynamic here, too, Kimball noted. As inference becomes persistent, especially with agentic systems, economics shift away from peak floating-point operations per second (FLOPS) (a measure of compute performance) and toward sustained efficiency and total cost of ownership (TCO).<\/p>\n<p>CPUs like Graviton5 are well positioned for the parts of that workload that don\u2019t require accelerators, but still have to run constantly. \u201cAt Meta\u2019s scale, even small efficiency gains per workload compound quickly,\u201d Kimball identified.<\/p>\n<p>For developers and enterprise IT, the signal is pretty clear, he noted: The AI stack is getting more heterogeneous, not less so. Enterprises are going to see tighter coupling between CPUs, GPUs, and specialized accelerators, with workloads increasingly split across them based on behavior (prefill versus decode, stateless versus stateful, burst versus persistent).<\/p>\n<p>\u201cThe implication is that infrastructure decisions should develop into more workload-aware,\u201d said Kimball. \u201cIt\u2019s less about \u2018which cloud?\u2019 and more about \u2018where does this specific a part of the applying run most efficiently?\u2019\u201d<\/p>\n<p><em>This text originally appeared on <a href=\"https:\/\/www.networkworld.com\/article\/4163379\/metas-compute-grab-continues-with-agreement-to-deploy-tens-of-millions-of-aws-graviton-cores.html\" target=\"_blank\">NetworkWorld<\/a>.<\/em><\/p>\n<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Meta is constant its compute grab because the agentic AI race accelerates to a sprint. Today, the corporate announced a partnership with Amazon Web Services (AWS) that can bring \u201ctens of hundreds of thousands\u201d of AWS Graviton5 cores (one chip comprises 192 cores) into its compute portfolio, with the choice to expand as its AI [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":324343,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[10],"tags":[11475,9669,7772,2254,51020,3682,11610,51019,5083,4947,10881],"class_list":["post-324342","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-technology","tag-agreement","tag-aws","tag-compute","tag-continues","tag-cores","tag-deploy","tag-grab","tag-graviton","tag-metas","tag-millions","tag-tens"],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/posts\/324342","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/comments?post=324342"}],"version-history":[{"count":2,"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/posts\/324342\/revisions"}],"predecessor-version":[{"id":324345,"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/posts\/324342\/revisions\/324345"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/media\/324343"}],"wp:attachment":[{"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/media?parent=324342"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/categories?post=324342"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/ebiztoday.news\/index.php\/wp-json\/wp\/v2\/tags?post=324342"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}