{"id":944397,"date":"2026-03-13T11:09:10","date_gmt":"2026-03-13T15:09:10","guid":{"rendered":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/"},"modified":"2026-03-13T11:09:10","modified_gmt":"2026-03-13T15:09:10","slug":"aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud","status":"publish","type":"post","link":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/","title":{"rendered":"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud"},"content":{"rendered":"<p>        <!--.bwalignc { text-align: center; list-style-position: inside }\n.bwlistdisc { list-style-type: disc }body {font:normal small Arial,Helvetica,sans-serif;color:#000;background-color:#fff;padding:24px;margin:0;} a img {border:0;} h3 {font-size:medium;color:#000;margin:0 0 1em 0; text-align:center;}-->  <\/p>\n<p class=\"bwalignc\"><b>AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud<\/b><\/p>\n<p class=\"bwalignc\">\nDeployed in AWS data centers and accessed through Amazon Bedrock, AWS Trainium + Cerebras CS-3 solution will accelerate inference speed<\/p>\n<p><b>Key Takeaways<\/b><\/p>\n<ul class=\"bwlistdisc\">\n<li>\nFastest inference coming soon: AWS and Cerebras are partnering to deliver the fastest AI inference available through Amazon Bedrock, launching in the next couple of months.<\/p>\n<\/li>\n<li>\nIndustry-leading speed and performance: With AWS Trainium optimized for prefill and Cerebras CS-3 optimized for decode, this innovative integrated system will provide unmatched performance and speed for AI inference.<\/p>\n<\/li>\n<li>\nPioneering cloud collaboration: AWS is the first cloud provider for Cerebras&#8217;s disaggregated inference solution, available exclusively through Amazon Bedrock.<\/p>\n<\/li>\n<\/ul>\n<p>SEATTLE &amp; SUNNYVALE, Calif.&#8211;(<a href=\"http:\/\/www.businesswire.com\">BUSINESS WIRE<\/a>)&#8211;<br \/>\nAmazon Web Services, Inc. (AWS), an Amazon.com, Inc. company (NASDAQ: AMZN), and Cerebras Systems today announced a collaboration that will, in the coming months, deliver the fastest AI inference solutions available for generative AI applications and LLM workloads. The solution, to be deployed on Amazon Bedrock in AWS data centers, combines AWS Trainium-powered servers, Cerebras CS-3 systems, and Elastic Fabric Adapter (EFA) networking. Later this year, AWS will also offer leading open-source LLMs and Amazon Nova using Cerebras hardware.<\/p>\n<p id=\"news-body-cta\">This press release features multimedia. View the full release here: <a href=\"https:\/\/www.businesswire.com\/news\/home\/20260313406341\/en\/\" rel=\"nofollow\">https:\/\/www.businesswire.com\/news\/home\/20260313406341\/en\/<\/a><\/p>\n<div id=\"bwbodyimg\" style=\"width: 480px;float:left;padding-left:0px;padding-right:20px;padding-top:0px;padding-bottom:0px\"><img decoding=\"async\" src=\"https:\/\/mms.businesswire.com\/media\/20260313406341\/en\/2746335\/4\/aws_cerebras_black_4.jpg\" alt=\"Amazon is deploying Cerebras Wafer Scale Engines in AWS datacenters\u200b. Ultra fast inference will be available through AWS Bedrock, bringing industry leading performance to the largest hyperscale cloud.\u200b\" \/><\/p>\n<p style=\"font-size:85%\">Amazon is deploying Cerebras Wafer Scale Engines in AWS datacenters\u200b. Ultra fast inference will be available through AWS Bedrock, bringing industry leading performance to the largest hyperscale cloud.\u200b<\/p>\n<\/div>\n<p>\n\u201cInference is where AI delivers real value to customers, but speed remains a critical bottleneck for demanding workloads like real-time coding assistance and interactive applications,\u201d said David Brown, Vice President, Compute &amp; ML Services, AWS. \u201cWhat we&#8217;re building with Cerebras solves that: by splitting the inference workload across Trainium and CS-3, and connecting them with Amazon\u2019s Elastic Fabric Adapter, each system does what it&#8217;s best at. The result will be inference that&#8217;s an order of magnitude faster and higher performance than what&#8217;s available today.&#8221;<\/p>\n<p>\n\u201cPartnering with AWS to build a disaggregated inference solution will bring the fastest inference to a global customer base,\u201d said Andrew Feldman, Founder and CEO of Cerebras Systems. \u201cEvery enterprise around the world will be able to benefit from blisteringly fast inference within their existing AWS environment.\u201d<\/p>\n<p><b>How It Works: Inference Disaggregation<\/b><\/p>\n<p>\nThe Trainium + CS-3 solution enables \u201cinference disaggregation,\u201d a technique which separates AI inference into two stages: prompt processing, or \u201cprefill,\u201d and output generation, or \u201cdecode.\u201d These two stages have profoundly different computational characteristics. Prefill is natively parallel, computationally intensive, and requires moderate memory bandwidth. Decode, on the other hand, is inherently serial, computationally light, and memory bandwidth intensive. Decode typically represents the majority of inference time in these scenarios because each output token must be generated sequentially.<\/p>\n<p>\nBecause each stage has a different computational challenge, they each benefit from different compute architectures and low-latency, high-bandwidth EFA networking between them. By strategically disaggregating the inference problem \u2014 with Trainium optimized for prefill and the Cerebras CS-3 optimized for decode \u2014 the two different computational challenges can be optimized in a specialized way.<\/p>\n<p>\nBuilt on the AWS Nitro System \u2014 the foundation of AWS&#8217;s secure, high-performance cloud infrastructure \u2014 the new solution will ensure that Cerebras CS-3 systems and Trainium-powered instances operate with the same security, isolation, and operational consistency customers expect from AWS.<\/p>\n<p><b>AWS Trainium for Prefill and Cerebras CS-3 for Decode<\/b><\/p>\n<p>\nTrainium is Amazon&#8217;s purpose-built AI chip, designed to deliver scalable performance and cost efficiency for training and inference across a broad range of generative AI workloads. Two of the world&#8217;s leading AI labs\u2014Anthropic and OpenAI\u2014are committed to Trainium. Anthropic has named AWS its primary training partner and is using Trainium to train and deploy its models, while OpenAI will consume 2 gigawatts of Trainium capacity through AWS infrastructure to support demand for Stateful Runtime Environment, frontier models, and other advanced workloads. Since its recent release, Trainium3 has seen strong customer adoption, with organizations across industries committing significant capacity.<\/p>\n<p>\nCerebras&#8217; CS-3 is the world&#8217;s fastest AI inference system. It delivers thousands of times greater memory bandwidth than the fastest GPU. As reasoning models now represent a majority of inference to compute and generate more tokens per request as they \u201cthink\u201d through problems, the need to accelerate this portion of the workflow has grown accordingly. OpenAI, Cognition, Mistral, and others use Cerebras to accelerate their most demanding workloads, especially agentic coding where developer productivity is constrained by inference speed.<\/p>\n<p>\nIn the disaggregated solution, CS-3 will be fully dedicated to decoding acceleration, enabling dramatically higher capacity for fast output tokens. With Trainium handling prefill, the CS-3 handling decode operations, and high-speed EFA networking connecting them, each processor will deliver maximum token capacity for its focused part of the workload.<\/p>\n<p><b>About Amazon Web Services<\/b><\/p>\n<p>\nAmazon Web Services (AWS) is guided by customer obsession, pace of innovation, commitment to operational excellence, and long-term thinking. By democratizing technology for nearly two decades and making cloud computing and generative AI accessible to organizations of every size and industry, AWS has built one of the fastest-growing enterprise technology businesses in history. Millions of customers trust AWS to accelerate innovation, transform their businesses, and shape the future. With the most comprehensive AI capabilities and global infrastructure footprint, AWS empowers builders to turn big ideas into reality. Learn more at aws.amazon.com and follow @AWSNewsroom.<\/p>\n<p><b>About Cerebras Systems<\/b><\/p>\n<p>\nCerebras Systems builds the fastest AI infrastructure in the world. We are a team of pioneering computer architects, computer scientists, AI researchers, and engineers of all types. We have come together to make AI blisteringly fast through innovation and invention because we believe that when AI is fast it will change the world. Our flagship technology, the Wafer Scale Engine 3 (WSE-3) is the world\u2019s largest and fastest AI processor. 56 times larger than the largest GPU, the WSE uses a fraction of the power per unit compute while delivering inference and training more than 20 times faster than the competition. Leading corporations, research institutes and governments on four continents chose Cerebras to run their AI workloads. Cerebras solutions are available on premise and in the cloud, for further information, visit cerebras.ai or follow us on LinkedIn, X and\/or Threads.<\/p>\n<p>\nThis press release contains forward-looking statements, including statements regarding the expected benefits of our products and the transaction described herein. These statements are subject to risks and uncertainties that could cause actual results to differ materially. Neither we nor any other person assumes responsibility for the accuracy and completeness of forward-looking statements. The forward-looking statements included in this press release relate only to events and information as of the date hereof. Cerebras undertakes no obligation to update or revise any forward-looking statement as a result of new information, future events or otherwise, except as otherwise required by law.<\/p>\n<p><img decoding=\"async\" alt=\"\" src=\"https:\/\/cts.businesswire.com\/ct\/CT?id=bwnews&amp;sty=20260313406341r1&amp;sid=flmnd&amp;distro=nx&amp;lang=en\" style=\"width:0;height:0\" \/><span class=\"bwct31415\" \/><\/p>\n<p id=\"mmgallerylink\"><span id=\"mmgallerylink-phrase\">View source version on businesswire.com: <\/span><span id=\"mmgallerylink-link\"><a href=\"https:\/\/www.businesswire.com\/news\/home\/20260313406341\/en\/\" rel=\"nofollow\">https:\/\/www.businesswire.com\/news\/home\/20260313406341\/en\/<\/a><\/span><\/p>\n<p><b>Media Contact<br \/>\n<\/b><br \/><a rel=\"nofollow\" href=\"mailto:pr@zmcommunications.com\">pr@zmcommunications.com<\/a><\/p>\n<p><b>KEYWORDS:<\/b> California Washington United States North America<\/p>\n<p><b>INDUSTRY KEYWORDS:<\/b> Data Management Technology Software Networks Artificial Intelligence Internet Hardware<\/p>\n<p><b>MEDIA:<\/b><\/p>\n<table cellpadding=\"3\" cellspacing=\"3\">\n<tr>\n<td><font face=\"Arial\" size=\"2\"><b>Photo<\/b><\/font><\/td>\n<\/tr>\n<tr>\n<td><img decoding=\"async\" src=\"https:\/\/mms.businesswire.com\/media\/20260313406341\/en\/2746335\/3\/aws_cerebras_black_4.jpg\" alt=\"Photo\" \/><\/td>\n<\/tr>\n<tr>\n<td><font face=\"Arial\" size=\"2\">Amazon is deploying Cerebras Wafer Scale Engines in AWS datacenters\u200b. Ultra fast inference will be available through AWS Bedrock, bringing industry leading performance to the largest hyperscale cloud.\u200b<\/font><\/td>\n<\/tr>\n<\/table>\n","protected":false},"excerpt":{"rendered":"<p>AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud Deployed in AWS data centers and accessed through Amazon Bedrock, AWS Trainium + Cerebras CS-3 solution will accelerate inference speed Key Takeaways Fastest inference coming soon: AWS and Cerebras are partnering to deliver the fastest AI inference available through Amazon Bedrock, launching in the next couple of months. Industry-leading speed and performance: With AWS Trainium optimized for prefill and Cerebras CS-3 optimized for decode, this innovative integrated system will provide unmatched performance and speed for AI inference. Pioneering cloud collaboration: AWS is the first cloud provider for Cerebras&#8217;s disaggregated inference solution, available exclusively through Amazon Bedrock. SEATTLE &amp; SUNNYVALE, Calif.&#8211;(BUSINESS WIRE)&#8211; &hellip; <\/p>\n<p class=\"link-more\"><a href=\"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/\" class=\"more-link\">Continue reading<span class=\"screen-reader-text\"> &#8220;AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud&#8221;<\/span><\/a><\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[],"tags":[],"class_list":["post-944397","post","type-post","status-publish","format-standard","hentry"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.5 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud - Market Newsdesk<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud - Market Newsdesk\" \/>\n<meta property=\"og:description\" content=\"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud Deployed in AWS data centers and accessed through Amazon Bedrock, AWS Trainium + Cerebras CS-3 solution will accelerate inference speed Key Takeaways Fastest inference coming soon: AWS and Cerebras are partnering to deliver the fastest AI inference available through Amazon Bedrock, launching in the next couple of months. Industry-leading speed and performance: With AWS Trainium optimized for prefill and Cerebras CS-3 optimized for decode, this innovative integrated system will provide unmatched performance and speed for AI inference. Pioneering cloud collaboration: AWS is the first cloud provider for Cerebras&#8217;s disaggregated inference solution, available exclusively through Amazon Bedrock. SEATTLE &amp; SUNNYVALE, Calif.&#8211;(BUSINESS WIRE)&#8211; &hellip; Continue reading &quot;AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud&quot;\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/\" \/>\n<meta property=\"og:site_name\" content=\"Market Newsdesk\" \/>\n<meta property=\"article:published_time\" content=\"2026-03-13T15:09:10+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/mms.businesswire.com\/media\/20260313406341\/en\/2746335\/4\/aws_cerebras_black_4.jpg\" \/>\n<meta name=\"author\" content=\"Newsdesk\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Newsdesk\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"6 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/\"},\"author\":{\"name\":\"Newsdesk\",\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/#\\\/schema\\\/person\\\/482f27a394d4fda80ecb5499e519d979\"},\"headline\":\"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud\",\"datePublished\":\"2026-03-13T15:09:10+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/\"},\"wordCount\":1214,\"image\":{\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/mms.businesswire.com\\\/media\\\/20260313406341\\\/en\\\/2746335\\\/4\\\/aws_cerebras_black_4.jpg\",\"inLanguage\":\"en-US\"},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/\",\"url\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/\",\"name\":\"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud - Market Newsdesk\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/mms.businesswire.com\\\/media\\\/20260313406341\\\/en\\\/2746335\\\/4\\\/aws_cerebras_black_4.jpg\",\"datePublished\":\"2026-03-13T15:09:10+00:00\",\"author\":{\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/#\\\/schema\\\/person\\\/482f27a394d4fda80ecb5499e519d979\"},\"breadcrumb\":{\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/#primaryimage\",\"url\":\"https:\\\/\\\/mms.businesswire.com\\\/media\\\/20260313406341\\\/en\\\/2746335\\\/4\\\/aws_cerebras_black_4.jpg\",\"contentUrl\":\"https:\\\/\\\/mms.businesswire.com\\\/media\\\/20260313406341\\\/en\\\/2746335\\\/4\\\/aws_cerebras_black_4.jpg\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/#website\",\"url\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/\",\"name\":\"Market Newsdesk\",\"description\":\"Latest Business News in Real Time\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/#\\\/schema\\\/person\\\/482f27a394d4fda80ecb5499e519d979\",\"name\":\"Newsdesk\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a0d0bd5b0f0ca12a265a459b13169dac35f33776d8501eda5e68844a366f2f46?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a0d0bd5b0f0ca12a265a459b13169dac35f33776d8501eda5e68844a366f2f46?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a0d0bd5b0f0ca12a265a459b13169dac35f33776d8501eda5e68844a366f2f46?s=96&d=mm&r=g\",\"caption\":\"Newsdesk\"},\"url\":\"https:\\\/\\\/www.marketnewsdesk.com\\\/index.php\\\/author\\\/newsdesk\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud - Market Newsdesk","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/","og_locale":"en_US","og_type":"article","og_title":"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud - Market Newsdesk","og_description":"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud Deployed in AWS data centers and accessed through Amazon Bedrock, AWS Trainium + Cerebras CS-3 solution will accelerate inference speed Key Takeaways Fastest inference coming soon: AWS and Cerebras are partnering to deliver the fastest AI inference available through Amazon Bedrock, launching in the next couple of months. Industry-leading speed and performance: With AWS Trainium optimized for prefill and Cerebras CS-3 optimized for decode, this innovative integrated system will provide unmatched performance and speed for AI inference. Pioneering cloud collaboration: AWS is the first cloud provider for Cerebras&#8217;s disaggregated inference solution, available exclusively through Amazon Bedrock. SEATTLE &amp; SUNNYVALE, Calif.&#8211;(BUSINESS WIRE)&#8211; &hellip; Continue reading \"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud\"","og_url":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/","og_site_name":"Market Newsdesk","article_published_time":"2026-03-13T15:09:10+00:00","og_image":[{"url":"https:\/\/mms.businesswire.com\/media\/20260313406341\/en\/2746335\/4\/aws_cerebras_black_4.jpg","type":"","width":"","height":""}],"author":"Newsdesk","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Newsdesk","Est. reading time":"6 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/#article","isPartOf":{"@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/"},"author":{"name":"Newsdesk","@id":"https:\/\/www.marketnewsdesk.com\/#\/schema\/person\/482f27a394d4fda80ecb5499e519d979"},"headline":"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud","datePublished":"2026-03-13T15:09:10+00:00","mainEntityOfPage":{"@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/"},"wordCount":1214,"image":{"@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/#primaryimage"},"thumbnailUrl":"https:\/\/mms.businesswire.com\/media\/20260313406341\/en\/2746335\/4\/aws_cerebras_black_4.jpg","inLanguage":"en-US"},{"@type":"WebPage","@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/","url":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/","name":"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud - Market Newsdesk","isPartOf":{"@id":"https:\/\/www.marketnewsdesk.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/#primaryimage"},"image":{"@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/#primaryimage"},"thumbnailUrl":"https:\/\/mms.businesswire.com\/media\/20260313406341\/en\/2746335\/4\/aws_cerebras_black_4.jpg","datePublished":"2026-03-13T15:09:10+00:00","author":{"@id":"https:\/\/www.marketnewsdesk.com\/#\/schema\/person\/482f27a394d4fda80ecb5499e519d979"},"breadcrumb":{"@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/#primaryimage","url":"https:\/\/mms.businesswire.com\/media\/20260313406341\/en\/2746335\/4\/aws_cerebras_black_4.jpg","contentUrl":"https:\/\/mms.businesswire.com\/media\/20260313406341\/en\/2746335\/4\/aws_cerebras_black_4.jpg"},{"@type":"BreadcrumbList","@id":"https:\/\/www.marketnewsdesk.com\/index.php\/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/www.marketnewsdesk.com\/"},{"@type":"ListItem","position":2,"name":"AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud"}]},{"@type":"WebSite","@id":"https:\/\/www.marketnewsdesk.com\/#website","url":"https:\/\/www.marketnewsdesk.com\/","name":"Market Newsdesk","description":"Latest Business News in Real Time","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.marketnewsdesk.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/www.marketnewsdesk.com\/#\/schema\/person\/482f27a394d4fda80ecb5499e519d979","name":"Newsdesk","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/a0d0bd5b0f0ca12a265a459b13169dac35f33776d8501eda5e68844a366f2f46?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/a0d0bd5b0f0ca12a265a459b13169dac35f33776d8501eda5e68844a366f2f46?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/a0d0bd5b0f0ca12a265a459b13169dac35f33776d8501eda5e68844a366f2f46?s=96&d=mm&r=g","caption":"Newsdesk"},"url":"https:\/\/www.marketnewsdesk.com\/index.php\/author\/newsdesk\/"}]}},"_links":{"self":[{"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/posts\/944397","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/comments?post=944397"}],"version-history":[{"count":0,"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/posts\/944397\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/media?parent=944397"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/categories?post=944397"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.marketnewsdesk.com\/index.php\/wp-json\/wp\/v2\/tags?post=944397"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}