HomeWinBuzzer NewsMicrosoft Cuts Nvidia GB200 Orders, Prioritizes GB300 Amid Production Delays

Microsoft Cuts Nvidia GB200 Orders, Prioritizes GB300 Amid Production Delays

Microsoft reduces Nvidia GB200 GPU orders by 40% due to production delays, shifting focus to the upcoming GB300 series to sustain its AI advancements.

-

Microsoft has reduced its Nvidia GB200 GPU orders by 40%, citing ongoing production delays and unresolved technical challenges, according to Taiwanese news outlet Commercial Times.

Nvidia’s struggles with its Blackwell architecture reportedly have pushed the GB200’s release timeline into 2025, forcing Microsoft to pivot to the forthcoming GB300 chips for its Azure cloud services.

This decision underscores the mounting pressure on Nvidia as it balances its dominant market position with production hurdles and increasing competition from Google and Amazon Web Services (AWS).

Microsoft’s Shift to GB300: A Calculated Response

Microsoft’s decision to scale back GB200 orders reflects both its reliance on Nvidia and the urgency of adapting to the evolving AI hardware landscape. The GB200, a critical component of Nvidia’s Blackwell architecture, was initially slated for a mid-2024 release. However, production delays caused by low-yield cartridge connectors and unresolved overheating issues have pushed availability to March 2025.

According to the Commercial Times, “Nvidia’s production of Blackwell GPUs faces delays due to cartridge connector issues, which have exhibited low yield rates during critical testing phases. This has caused supply disruptions and forced Nvidia to reconsider its production strategy.”

Additionally, thermal management remains a significant concern. In November, reports emerged that Nvidia’s GPUs face overheating issues in high-density data center racks, requiring multiple design revisions. These challenges have disrupted deployment plans for major tech clients, including Microsoft and Google.

Faced with these issues, Microsoft has redirected its focus to Nvidia’s GB300, the next-generation GPU designed to overcome many of the GB200’s limitations. Expected to launch in mid-2025, the GB300 is said to feature a socketed design for easier installation, 12 layers of HBM3e memory, and liquid cooling for improved energy efficiency.

Related: DeepX DX-M1 AI Chip Hits 90% Yield with Samsung’s 5nm Technology

Nvidia’s Blackwell Successes and Ongoing Challenges

Despite the setbacks, Nvidia remains a leader in AI hardware. Its GB200 GPUs demonstrated exceptional performance in the MLPerf Training v4.1 benchmarks, doubling the training speed for GPT-3 compared to the H100 and delivering 62% gains in image generation tasks.

During Nvidia’s Q2 FY2025 earnings call, CEO Jensen Huang acknowledged challenges but emphasized the company’s commitment to addressing them.

These benchmarks underscore Nvidia’s continued dominance, but the production delays have created openings for competitors like Google. Google’s sixth-generation Trillium TPU achieved a 3.8x improvement in training tasks compared to its predecessor, narrowing the gap with Nvidia in certain benchmarks.

Related: Nvidia Q3 2024 Up To $35.1 billion, 94% Year-Over-Year: Jensen Huang Praises Blackwell Adoption

Microsoft and Nvidia: A Strategic Partnership

Microsoft’s early adoption of Nvidia’s Blackwell GPUs highlights its strategic reliance on Nvidia for AI infrastructure. In October 2024, Azure was named the first cloud provider to deploy GB200 GPUs, leveraging their capabilities to optimize energy efficiency and performance.

However, unlike competitors Google and AWS, which have developed proprietary AI chips, Microsoft’s exclusive reliance on Nvidia has its risks. While Nvidia’s hardware remains unmatched in many respects, its production challenges could disrupt Microsoft’s plans if delays persist.

Related: Amazon Develops Custom AI Chips to Compete Directly with NVIDIA

Technical Highlights of Blackwell GPUs

The GB200 and GB300 GPUs are built on TSMC’s CoWoS-L packaging technology, which integrates multiple chiplets on a single substrate to achieve data transfer speeds of up to 10 TB/s. This design, while innovative, has faced challenges related to thermal expansion and heat dissipation.

The GPUs also feature HBM3e memory, enabling up to 8 TBps of bandwidth and 192GB of memory capacity. These specifications are critical for large-scale AI workloads, such as training large language models (LLMs) and generative AI systems.

Nvidia’s advancements in floating-point precision, particularly 4-bit FP, have further enhanced its GPUs’ efficiency for AI applications. This approach reduces computational overhead while maintaining the accuracy required for complex tasks like GPT-4 training and Stable Diffusion.

Last Updated on December 5, 2024 9:47 pm CET

Markus Kasanmascheff
Markus Kasanmascheff
Markus has been covering the tech industry for more than 15 years. He is holding a Master´s degree in International Economics and is the founder and managing editor of Winbuzzer.com.
0 0 votes
Article Rating
Subscribe
Notify of
guest
0 Comments
Newest
Oldest Most Voted
Inline Feedbacks
View all comments
0
We would love to hear your opinion! Please comment below.x
()
x