{"id":7573,"date":"2025-05-17T02:16:06","date_gmt":"2025-05-16T17:16:06","guid":{"rendered":"https:\/\/aireviewirush.com\/?p=7573"},"modified":"2025-05-17T02:16:06","modified_gmt":"2025-05-16T17:16:06","slug":"new-amazon-ec2-p6-b200-cases-powered-by-nvidia-blackwell-gpus-to-speed-up-ai-improvements","status":"publish","type":"post","link":"https:\/\/aireviewirush.com\/?p=7573","title":{"rendered":"New Amazon EC2 P6-B200 cases powered by NVIDIA Blackwell GPUs to speed up AI improvements"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div id=\"\">\n<table id=\"amazon-polly-audio-table\">\n<tbody>\n<tr>\n<td id=\"amazon-polly-audio-tab\">\n<div id=\"amazon-polly-by-tab\">\n            <a href=\"https:\/\/aws.amazon.com\/polly\/\" target=\"_blank\" rel=\"noopener noreferrer\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/a0.awsstatic.com\/aws-blog\/images\/Voiced_by_Amazon_Polly_EN.png\" alt=\"Voiced by Polly\" width=\"554\" height=\"56\"\/><\/a>\n           <\/div>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>As we speak, we\u2019re saying the overall availability of <a href=\"https:\/\/aws.amazon.com\/ec2\" target=\"_blank\" rel=\"noopener\">Amazon Elastic Compute Cloud (Amazon EC2)<\/a> <a href=\"https:\/\/aws.amazon.com\/ec2\/instance-types\/p6\" target=\"_blank\" rel=\"noopener\">P6-B200 cases<\/a> powered by NVIDIA B200 to handle buyer wants for prime efficiency and scalability in <a href=\"https:\/\/aws.amazon.com\/ai\/\" target=\"_blank\" rel=\"noopener\">synthetic intelligence (AI)<\/a>, <a href=\"https:\/\/aws.amazon.com\/ai\/machine-learning\/\" target=\"_blank\" rel=\"noopener\">machine studying (ML)<\/a>, and <a href=\"https:\/\/aws.amazon.com\/hpc\/\" target=\"_blank\" rel=\"noopener\">excessive efficiency computing (HPC)<\/a> functions.<\/p>\n<p>Amazon EC2 P6-B200 cases speed up a broad vary of GPU-enabled workloads however are particularly well-suited for large-scale distributed AI coaching and inferencing for <a href=\"https:\/\/aws.amazon.com\/what-is\/foundation-models\/\" target=\"_blank\" rel=\"noopener\">basis fashions (FMs)<\/a> with reinforcement studying (RL) and distillation, multimodal coaching and inference, and HPC functions similar to local weather modeling, drug discovery, seismic evaluation, and insurance coverage threat modeling.<\/p>\n<p>When mixed with <a href=\"https:\/\/aws.amazon.com\/hpc\/efa\/\" target=\"_blank\" rel=\"noopener\">Elastic Material Adapter<\/a> (EFAv4) networking, hyperscale clustering by <a href=\"https:\/\/aws.amazon.com\/ec2\/ultraclusters\/\" target=\"_blank\" rel=\"noopener\">EC2 UltraClusters<\/a>, and superior virtualization and safety capabilities by <a href=\"https:\/\/aws.amazon.com\/ec2\/nitro\/\" target=\"_blank\" rel=\"noopener\">AWS Nitro System<\/a>, you possibly can prepare and serve FMs with elevated velocity, scale, and safety. These cases additionally ship as much as two instances the efficiency for AI coaching (time to coach) and inference (tokens\/sec) in comparison with <a href=\"https:\/\/aws.amazon.com\/ec2\/instance-types\/p5\/\" target=\"_blank\" rel=\"noopener\">EC2 P5en cases<\/a>.<\/p>\n<p>You&#8217;ll be able to speed up time-to-market for coaching FMs and ship sooner inference throughput, which lowers inference price and helps enhance adoption of generative AI functions in addition to elevated processing efficiency for HPC functions.<\/p>\n<p><strong><u>EC2 P6-B200 cases specs<\/u><\/strong><br \/>New EC2 P6-B200 cases present eight NVIDIA B200 GPUs with 1440 GB of excessive bandwidth GPU reminiscence, fifth Era Intel Xeon Scalable processors (Emerald Rapids), 2 TiB of system reminiscence, and 30 TB of native NVMe storage.<\/p>\n<p>Listed below are the specs for EC2 P6-B200 cases:<\/p>\n<table style=\"border: 2px solid black;border-collapse: collapse;margin-left: auto;margin-right: auto\">\n<tbody>\n<tr style=\"border-bottom: 1px solid black;background-color: #e0e0e0\">\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>Occasion dimension<\/strong><\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>GPUs (NVIDIA B200)<\/strong><\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>GPU<br \/>reminiscence (GB)<\/strong><\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>vCPUs<\/strong><\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>GPU Peer to see (GB\/s)<\/strong><\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>Occasion storage (TB)<\/strong><\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>Community bandwidth (Gbps)<\/strong><\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>EBS bandwidth (Gbps)<\/strong><\/td>\n<\/tr>\n<tr style=\"border-bottom: 1px solid black\">\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\"><strong>P6-b200.48xlarge<\/strong><\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\">8<\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\">1440 HBM3e<\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\">192<\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\">1800<\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\">8 x 3.84 NVMe SSD<\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\">8 x 400<\/td>\n<td style=\"border-right: 1px solid black;padding: 4px;text-align: center\">100<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>These cases characteristic as much as 125 % enchancment in GPU TFLOPs, 27 % enhance in GPU reminiscence dimension, and 60 % enhance in GPU reminiscence bandwidth in comparison with P5en cases.<\/p>\n<p><strong><u>P6-B200 cases in motion<\/u><\/strong><br \/>You need to use P6-B200 cases within the US West (Oregon) <a href=\"https:\/\/docs.aws.amazon.com\/glossary\/latest\/reference\/glos-chap.html#region\" target=\"_blank\" rel=\"noopener\">AWS Area<\/a> by way of <a href=\"https:\/\/aws.amazon.com\/ec2\/capacityblocks\/\" target=\"_blank\" rel=\"noopener\">EC2 Capability Blocks for ML<\/a>. To order your EC2 Capability Blocks, select <strong>Capability Reservations<\/strong> on the <a href=\"https:\/\/us-east-2.console.aws.amazon.com\/ec2\/home?region=us-east-2#CapacityReservations:\" target=\"_blank\" rel=\"noopener\">Amazon EC2 console<\/a>.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-96219 size-full\" style=\"border: solid 1px #ccc\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/da4b9237bacccdf19c0760cab7aec4a8359010b0\/2025\/05\/15\/2025-ec2-p6-b200-instance-capacity-block.jpg\" alt=\"\" width=\"2514\" height=\"1798\"><\/p>\n<p>Choose <strong>Buy Capability Blocks for ML<\/strong> after which select your complete capability and specify how lengthy you want the EC2 Capability Block for <strong>p6-b200.48xlarge<\/strong> cases. The entire variety of days you can reserve EC2 Capability Blocks is 1-14 days, 21 days, 28 days, or multiples of seven as much as 182 days. You&#8217;ll be able to select your earliest begin date for as much as 8 weeks upfront.<\/p>\n<p>Now, your EC2 Capability Block shall be scheduled efficiently. The entire worth of an EC2 Capability Block is charged up entrance, and the worth doesn\u2019t change after buy. The fee shall be billed to your account inside 12 hours after you buy the EC2 Capability Blocks. To be taught extra, go to <a href=\"https:\/\/docs.aws.amazon.com\/AWSEC2\/latest\/UserGuide\/ec2-capacity-blocks.html\" target=\"_blank\" rel=\"noopener\">Capability Blocks for ML<\/a> within the Amazon EC2 Person Information.<\/p>\n<p>When launching P6-B200 cases, you should utilize <a href=\"https:\/\/aws.amazon.com\/machine-learning\/amis\/\" target=\"_blank\" rel=\"noopener\">AWS Deep Studying AMIs<\/a> (DLAMI) to assist EC2 P6-B200 cases. DLAMI supplies ML practitioners and researchers with the infrastructure and instruments to rapidly construct scalable, safe, distributed ML functions in preconfigured environments.<\/p>\n<p>To run cases, you should utilize <a href=\"https:\/\/console.aws.amazon.com\/ec2\" target=\"_blank\" rel=\"noopener\">AWS Administration Console<\/a>, <a href=\"https:\/\/aws.amazon.com\/cli\/\" target=\"_blank\" rel=\"noopener\">AWS Command Line Interface (AWS CLI)<\/a> or <a href=\"http:\/\/docs.aws.amazon.com\/AWSJavaScriptSDK\/latest\/AWS\/EC2.html\" target=\"_blank\" rel=\"noopener\">AWS SDKs<\/a>.<\/p>\n<p>You&#8217;ll be able to combine EC2 P6-B200 cases seamlessly with varied AWS managed providers similar to <a href=\"https:\/\/aws.amazon.com\/eks\/\" target=\"_blank\" rel=\"noopener\">Amazon Elastic Kubernetes Companies (Amazon EKS)<\/a>, <a href=\"https:\/\/aws.amazon.com\/s3\" target=\"_blank\" rel=\"noopener\">Amazon Easy Storage Service (Amazon S3)<\/a>, and <a href=\"https:\/\/aws.amazon.com\/fsx\/lustre\/\" target=\"_blank\" rel=\"noopener\">Amazon FSx for Lustre<\/a>. Help for <a href=\"https:\/\/aws.amazon.com\/sagemaker-ai\/hyperpod\/\" target=\"_blank\" rel=\"noopener\">Amazon SageMaker HyperPod<\/a> can be coming quickly.<\/p>\n<p><strong><u>Now accessible<\/u><\/strong><br \/>Amazon EC2 P6-B200 cases can be found in the present day within the US West (Oregon) Area and could be bought as <a href=\"https:\/\/aws.amazon.com\/ec2\/capacityblocks\/\" target=\"_blank\" rel=\"noopener\">EC2 Capability blocks for ML<\/a>.<\/p>\n<p>Give Amazon EC2 P6-B200 cases a attempt within the <a href=\"https:\/\/console.aws.amazon.com\/ec2\/\" target=\"_blank\" rel=\"noopener\">Amazon EC2 console<\/a>. To be taught extra, discuss with the <a href=\"https:\/\/aws.amazon.com\/ec2\/instance-types\/p6\/\" target=\"_blank\" rel=\"noopener\">Amazon EC2 P6 occasion web page<\/a> and ship suggestions to <a href=\"https:\/\/repost.aws\/tags\/TAO-wqN9fYRoyrpdULLa5y7g\/amazon-ec-2\" target=\"_blank\" rel=\"noopener\">AWS re:Submit for EC2<\/a> or by way of your typical AWS Help contacts.<\/p>\n<p>\u2014 <a href=\"https:\/\/twitter.com\/channyun\" target=\"_blank\" rel=\"noopener\">Channy<\/a><\/p>\n<hr\/>\n<p>How is the Information Weblog doing? Take this <a href=\"https:\/\/amazonmr.au1.qualtrics.com\/jfe\/form\/SV_eyD5tC5xNGCdCmi\" target=\"_blank\" rel=\"noopener\">1 minute survey<\/a>!<\/p>\n<p><em>(This <a href=\"https:\/\/amazonmr.au1.qualtrics.com\/jfe\/form\/SV_eyD5tC5xNGCdCmi\" target=\"_blank\" rel=\"noopener\">survey<\/a> is hosted by an exterior firm. AWS handles your data as described within the <a href=\"https:\/\/aws.amazon.com\/privacy\/?trk=4b29643c-e00f-4ab6-ab9c-b1fb47aa1708&amp;sc_channel=blog\" target=\"_blank\" rel=\"noopener\">AWS Privateness Discover<\/a>. AWS will personal the information gathered by way of this survey and won&#8217;t share the knowledge collected with survey respondents.)<\/em><\/p>\n<p>       <!-- '\"` -->\n      <\/div>\n<p><script async src=\"\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><br \/>\n<br \/><\/p>\n","protected":false},"excerpt":{"rendered":"<p>As we speak, we\u2019re saying the overall availability of Amazon Elastic Compute Cloud (Amazon EC2) P6-B200 cases powered by NVIDIA B200 to handle buyer wants for prime efficiency and scalability in synthetic intelligence (AI), machine studying (ML), and excessive efficiency computing (HPC) functions. Amazon EC2 P6-B200 cases speed up a broad vary of GPU-enabled workloads [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":7575,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[],"class_list":{"0":"post-7573","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-cloud-computing"},"_links":{"self":[{"href":"https:\/\/aireviewirush.com\/index.php?rest_route=\/wp\/v2\/posts\/7573","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/aireviewirush.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/aireviewirush.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/aireviewirush.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/aireviewirush.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=7573"}],"version-history":[{"count":1,"href":"https:\/\/aireviewirush.com\/index.php?rest_route=\/wp\/v2\/posts\/7573\/revisions"}],"predecessor-version":[{"id":7574,"href":"https:\/\/aireviewirush.com\/index.php?rest_route=\/wp\/v2\/posts\/7573\/revisions\/7574"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/aireviewirush.com\/index.php?rest_route=\/wp\/v2\/media\/7575"}],"wp:attachment":[{"href":"https:\/\/aireviewirush.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=7573"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/aireviewirush.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=7573"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/aireviewirush.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=7573"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}