ZawyaZawya

PRESSR: New AWS AI Factories transform customers’ existing infrastructure into high performance AI environments

Refinitiv7 min de leitura
  • AWS AI Factories deliver dedicated infrastructure combining the latest NVIDIA accelerated computing platform, Trainium chips, AWS high-speed, low-latency networking, and AWS AI services.
  • Customers can leverage their existing data center space, network connectivity, and power while AWS handles the complexity of deployment and management of the integrated infrastructure.
  • AWS AI Factories help enterprise and public sector organizations meet their data sovereignty and regulatory requirements, with accelerated deployment timelines.

Dubai, UAE - As governments and large organizations seek to scale AI projects, some are turning to the concept of an “AI factory” to address their unique sovereignty and compliance needs. But, building a high-performance AI factory requires a comprehensive set of management, database, storage, and security services—complexity that few customers want to take on themselves. To address this need, today we announced AWS AI Factories, a new offering that provides enterprises and governments with dedicated AWS AI infrastructure deployed in their own data centers. AWS AI Factories combine the latest AI accelerators including cutting-edge NVIDIA AI computing and Trainium chips, AWS high-speed, low-latency networking, high-performance storage and databases, security, and energy-efficient infrastructure together with comprehensive AI services like Amazon Bedrock and Amazon SageMaker so customers can rapidly develop and deploy AI applications at scale. 

Organizations in regulated industries and the public sector face a critical AI infrastructure challenge in getting their large-scale AI projects deployed. Building their own AI capabilities requires massive capital investments in GPUs, data centers, and power, plus navigating complex procurement cycles, selecting the right AI model for their use case, and licensing models from different AI providers. This creates multi-year timelines and operational complexity that diverts focus from their core business goals.

AWS AI Factories address this challenge by deploying dedicated AWS AI infrastructure in customers’ own data centers, operated exclusively for them. AWS AI Factories operate like a private AWS Region that gives secure, low-latency access to compute, storage, database, and AI services. This approach lets you leverage existing data center space and power capacity you’ve already acquired, and gives access to AWS AI infrastructure and services—from the latest AI chips for training and inference to tools for building, training, and deploying AI models. It also provides managed services that offer access to leading foundation models without having to negotiate separate contracts with model providers—all while helping you meet security, data sovereignty, and regulatory requirements for where data is processed and stored. Leveraging nearly two decades of cloud leadership and unmatched experience in architecting large-scale AI systems, we are able to deploy secure, reliable AI infrastructure faster than most organizations can on their own, saving years of buildout effort and managing operational complexity.

AWS and NVIDIA expand collaboration to accelerate customer AI infrastructure deployments 

The relationship between AWS and NVIDIA goes back 15 years, to when we launched the world’s first GPU cloud instance, and today we offer the widest range of GPU solutions for customers. Building on our longstanding collaboration to deliver advanced AI infrastructure, AWS and NVIDIA make it possible for customers to build and run large language models faster, at scale, and more securely than anywhere else—now in your own data centers. With the NVIDIA-AWS AI Factories integration, AWS customers have seamless access to the NVIDIA accelerated computing platform, full stack NVIDIA AI software, and thousands of GPU-accelerated applications to deliver high performance, efficiency, and scalability for building next-generation AI solutions. We continue to bring the best of our technologies together. The AWS Nitro System, Elastic Fabric Adapter (EFA) petabit scale networking, and Amazon EC2 UltraClusters support the latest NVIDIA Grace Blackwell and the next-generation NVIDIA Vera Rubin platforms. In the future, AWS will support NVIDIA NVLink Fusion high-speed chip interconnect technology in next-generation Trainium4 and Graviton chips, and in the Nitro System. This integration makes it possible for customers to accelerate time to market and achieve better performance.

“Large-scale AI requires a full-stack approach—from advanced GPUs and networking to software and services that optimize every layer of the data center. Together with AWS, we’re delivering all of this directly into customers’ environments,” said Ian Buck, vice president and general manager of Hyperscale and HPC at NVIDIA. “By combining NVIDIA’s latest Grace Blackwell and Vera Rubin architectures with AWS’s secure, high-performance infrastructure and AI software stack, AWS AI Factories allow organizations to stand up powerful AI capabilities in a fraction of the time and focus entirely on innovation instead of integration.” 

Helping the public sector accelerate AI adoption 

AWS AI Factories are built to meet AWS's rigorous security standards of providing governments with the confidence to run their most sensitive workloads across all classification levels: Unclassified, Sensitive, Secret, and Top Secret. AWS AI Factories will also provide governments around the world with the availability, reliability, security, and control they need to help their own economies advance and take advantage of the benefits of AI technologies.

AWS and NVIDIA are collaborating on a strategic partnership with HUMAIN, the global company based in Saudi Arabia building full-stack AI capabilities, with AWS building a first-of-its-kind "AI Zone" in Saudi Arabia featuring up to 150,000 AI chips including GB300 GPUs, dedicated AWS AI infrastructure, and AWS AI services, all within a HUMAIN purpose-built data center. “The AI factory AWS is building in our new AI Zone represents the beginning of a multi-gigawatt journey for HUMAIN and AWS. From inception, this infrastructure has been engineered to serve both the accelerating local and global demand for AI compute,” said Tareq Amin, CEO of HUMAIN. “What truly sets this partnership apart is the scale of our ambition and the innovation in how we work together. We chose AWS because of their experience building infrastructure at scale, enterprise grade reliability, breadth of AI capabilities, and depth of commitment to the region. Through a shared commitment to global market expansion, we are creating an ecosystem that will shape the future of how AI ideas can be built, deployed, and scaled for the whole world.” 

For more details on AWS AI Factories, visit the product page.

AI Factories FAQs

1. What are AWS AI Factories?

AWS AI Factories provide customers with dedicated deployments that combine the latest AI hardware accelerators, including next-generation NVIDIA GPUs and AWS Trainium AI chips, specialized networking, and high-performance storage. Each AWS AI Factory is a physically separate deployment available exclusively for the customer or their designated trusted community, allowing customers to rapidly build their AI-powered applications while meeting their data residency and sovereignty requirements. AWS builds these AI Factories using customer-provided data centers and power infrastructure, while also giving customers the flexibility to use their own GPUs, meeting them wherever they are in their AI journey.

With a broad set of AWS AI services like Amazon Bedrock and Amazon SageMaker, customers have immediate access to leading foundation models without needing to negotiate separate contracts with individual model providers. At the same time, these AI Factories seamlessly integrate with the broader set of AWS services, allowing customers to benefit from the unparalleled security, reliability, and capabilities of the AWS Cloud.

2. Who are AWS AI Factories for?

AWS AI Factories are ideal for customers that require dedicated or sovereign AI infrastructure to power their mission-critical AI workloads. The primary target customers are: 1) government entities looking to leverage advanced AI capabilities to drive digital AI transformation, and 2) large enterprises often in regulated industries across sectors like energy, healthcare, manufacturing, and finance that require high-performance AI infrastructure. AWS AI Factories enable customers to rapidly accelerate AI adoption by using AI services like Amazon Bedrock and SageMaker in their data centers, with the flexibility to use their GPUs and network connectivity in a secure, tailored environment.

3. In what facilities does AWS deploy AWS AI Factories?

AWS deploys AI Factories exclusively within customer-owned data centers. We work closely with you to implement and manage these environments in your facilities, while maintaining AWS's stringent security standards.

4. Can AWS use customers’ NVIDIA GPUs in AWS AI Factories?

Yes, AWS can use customers’ NVIDIA GPUs in AWS AI Factories. AWS works with customers and NVIDIA to redirect customers' purchased NVIDIA GPUs to AWS and integrate them into the dedicated AWS AI Factories. This allows customers to leverage their GPU investments while also accessing the advanced AWS AI services, networking, and storage provided by AWS AI Factories.

5. Can AWS provision and manage GPUs for customers?

Yes. AWS provisions and manages a variety of GPUs within AWS AI Factories, including instances powered by the latest NVIDIA GPUs. Customers can access these GPUs as part of their AWS AI Factories deployment without needing to manage the underlying hardware infrastructure.

6. In which countries can AWS AI factories be deployed?

AWS AI Factories can be deployed in countries where Regions and Local Zones have been announced. To find the specific list of countries, you can visit the AWS Regions page and the AWS Local Zones page in the AWS Documentation.

Send us your press releases to pressrelease.zawya@lseg.com

Disclaimer: The contents of this press release was provided from an external third party provider. This website is not responsible for, and does not control, such external content. This content is provided on an “as is” and “as available” basis and has not been edited in any way. Neither this website nor our affiliates guarantee the accuracy of or endorse the views or opinions expressed in this press release.

The press release is provided for informational purposes only. The content does not provide tax, legal or investment advice or opinion regarding the suitability, value or profitability of any particular security, portfolio or investment strategy. Neither this website nor our affiliates shall be liable for any errors or inaccuracies in the content, or for any actions taken by you in reliance thereon. You expressly agree that your use of the information within this article is at your sole risk.

To the fullest extent permitted by applicable law, this website, its parent company, its subsidiaries, its affiliates and the respective shareholders, directors, officers, employees, agents, advertisers, content providers and licensors will not be liable (jointly or severally) to you for any direct, indirect, consequential, special, incidental, punitive or exemplary damages, including without limitation, lost profits, lost savings and lost revenues, whether in negligence, tort, contract or any other theory of liability, even if the parties have been advised of the possibility or could have foreseen any such damages.

Entrar ou criar uma conta gratuita para ler essa notícia