Latest news

NVIDIA and AWS deepen technical cooperation to standardize global AI infrastructures

NVIDIA and AWS are expanding their technical collaboration through deeper integration of hardware, interconnect technologies and software layers to provide a unified and scalable platform for future AI needs. The collaboration includes AWS cloud infrastructure as well as custom chips Trainium4, Graviton and the Nitro system. At the heart of this is the implementation of NVIDIA NVLink Fusion, a technology for coupling customized AI systems that AWS integrates into its own silicon solutions. The combination of NVLink Scale-up Interconnect and the NVIDIA MGX rack design creates a platform designed for low power consumption, high bandwidth and consistent latency. The integration is part of a multi-year roadmap aimed at accelerating the development of new cloud scaling architectures.

AWS is using Trainium4 in combination with NVLink Fusion for the first time, increasing the efficiency of inference workloads and agentic AI models while reducing management overhead. Already active MGX racks with NVIDIA GPUs are extended by NVLink Fusion with a unified system structure. Additionally, AWS leverages NVLink Fusion’s supplier ecosystem, which provides components such as chassis, power and cooling solutions to simplify rack-level deployment. Support for Elastic Fabric Adapter and Nitro enables the upcoming NVIDIA Vera Rubin architecture to be fully embedded into the existing AWS network fabric.

In the area of international AI infrastructure, AWS is planning the global expansion of sovereign AI clouds, which will take stricter regulatory frameworks into account. These structures are based on NVIDIA Blackwell GPUs such as HGX B300 and GB300 NVL72 as well as the new RTX PRO 6000 Blackwell Server Edition for visual applications. The systems form the foundation of the AI Factories operated by AWS, which provide companies with dedicated AI data centers without violating data sovereignty requirements. Public sector organizations in particular are expected to benefit from the combination of AWS cloud services and NVIDIA computing architectures, as it allows for extensive model training with complete control over the most sensitive data.

The collaboration will also be expanded at software level. The open NVIDIA Nemotron models are integrated into Amazon Bedrock and enable the development of generative applications and multimodal agents without infrastructural management overhead. Immediate deployment via Bedrock’s serverless model lowers the barriers to entry for companies. The first users from the security and finance sectors are already using this integration productively. In parallel, Amazon OpenSearch Service receives GPU-based acceleration based on NVIDIA cuVS, allowing vector indexes to be built faster. Early test results indicate a significant reduction in latency with simultaneous cost savings, which is particularly relevant for systems with retrieval augmented generation.

For the operation of agent-based systems, NVIDIA and AWS provide a coordinated software chain consisting of Strands Agents, the NVIDIA NeMo Agent Toolkit and Bedrock AgentCore. This combination covers the steps from development to performance tuning and scalable execution. The cooperation also includes the simulation of physical AI systems. NVIDIA Cosmos world models are available via NIM microservices in Amazon EKS and enable real-time robotics applications. The same models can be used for extensive data synthesis via AWS Batch. The scenarios generated in this way are suitable for validation in Isaac Sim and Isaac Lab. Companies in the robotics sector use this infrastructure to collect, process and train robot-generated data.

The deepened partnership marks a further step in the long-standing technological relationship between the two companies. As part of this, NVIDIA was recognized as AWS Global GenAI Infrastructure and Data Partner of the Year. Both partners are increasingly focusing their architecture on global, scalable and data-sovereign AI infrastructures that are designed to provide complex models securely and efficiently.

Conclusion

The expansion of the collaboration between NVIDIA and AWS leads to a unification of hardware, interconnects and software services, increasing the scalability and performance of AI infrastructures. The integration of NVLink Fusion into AWS silicon represents a structural change in the design of cloud-based AI systems. Complementary software optimizations and the integration of sovereign AI clouds expand the offering towards fully controllable high-performance environments for enterprises and public institutions.

Sources

Source Key statement Link to
NVIDIA press release Expanded strategic collaboration with AWS, integration of NVLink Fusion, expansion of sovereign AI clouds https://www.nvidia.com/en-us/news/amazon-aws-nvidia-blackwell-nvlink-fusion-ai-factories-2024
AWS Newsroom Information on AI Factories, Blackwell GPUs and software-side integrations https://www.aboutamazon.com/news/innovation-at-amazon/amazon-and-nvidia-expand-collaboration-to-power-next-generation-of-ai
Amazon Bedrock documentation Description of Nemotron integration and deployment via Bedrock https://aws.amazon.com/bedrock/nemotron
NVIDIA Developer Technical basics of cuVS, NIM, NeMo, Isaac Sim and Cosmos WFMs https://developer.nvidia.com/cuda-toolkit-ai-frameworks-neural-models-simulation-cuvs-nim-nemo-isaac-cosmos

Bisher keine Kommentare

Kommentar

Lade neue Kommentare

Redaktion

Artikel-Butler

3,723 Kommentare 14,335 Likes

NVIDIA und AWS erweitern ihre technische Zusammenarbeit durch eine tiefere Integration von Hardware, Interconnect-Technologien und Softwareebenen, um eine einheitliche und skalierbare Plattform für zukünftige KI-Anforderungen bereitzustellen. Die Kooperation umfasst sowohl die Cloud-Infrastruktur von AWS als auch die kundenspezifischen Chips Trainium4, Graviton und das Nitro-System. Im Mittelpunkt steht die Implementierung von NVIDIA NVLink Fusion, einer Technologie zur Kopplung von maßgeschneiderten KI-Systemen, die AWS in seine eigenen Siliziumlösungen integriert. Durch die Verbindung aus NVLink Scale-up-Interconnect und dem NVIDIA MGX-Rackdesign entsteht eine Plattform, die auf geringen Energieverbrauch, hohe Bandbreite und konsistente Latenzen ausgelegt ist. Die Integration ist Teil einer mehrjährigen Roadmap, die darauf […] (read full article...)

Antwort Gefällt mir

Danke für die Spende



Du fandest, der Beitrag war interessant und möchtest uns unterstützen? Klasse!

Hier erfährst Du, wie: Hier spenden.

Hier kannst Du per PayPal spenden.

About the author

Karsten Rabeneck-Ketme

Werbung

Werbung