Jump to content
  • Sign Up
×
×
  • Create New...

Faulty Nvidia H100 GPUs and HBM3 memory contributed to failures every three hours during Meta’s LLama 3 training — 16,384 GPU cluster detailed in whitepaper


Recommended Posts

  • Diamond Member

This is the hidden content, please

Faulty Nvidia H100 GPUs and HBM3 memory contributed to failures every three hours during Meta’s LLama 3 training — 16,384 GPU cluster detailed in whitepaper

Meta recently 

This is the hidden content, please
 detailing its Llama 3 405B model training run on a cluster containing 16,384 Nvidia H100 80GB GPUs. The training run took place over 54 days and the cluster encountered 419 component failures, averaging one ******** every three hours. In half of the ******** cases, GPUs or their HBM3 memory were to blame. 

As the old supercomputing adage goes, the only thing certain with large-scale systems is ********. Supercomputers are extremely complex devices that use tens of thousands of processors, hundreds of thousands of other chips, and hundreds of kilometers of cables. In a sophisticated supercomputer, it’s normal for something to break down every few hours, and the main trick for developers is to ensure that the system is operational regardless of such local breakdowns.

The scale and synchronous nature of 16,384 GPU training make it prone to failures. A single GPU ******** can disrupt the entire training job, necessitating a restart. However, the Llama 3 team maintained over a 90% effective training time. 

During a 54-day pre-training snapshot, there were 466 job interruptions, with 47 planned and 419 unexpected. Planned interruptions were due to automated maintenance, while unexpected ones mostly stemmed from hardware issues. GPU problems were the largest category, accounting for 58.7% of unexpected interruptions. Only three incidents required significant manual intervention; the rest were managed by automation. 

data:image/gif;base64,R0lGODlhAQABAIAAAAAAAP///ywAAAAAAQABAAACAUwAOw==

(Image credit: Meta)

Out of 419 unexpected interruptions, 148 (30.1%) were caused by various GPU failures (including NVLink failures), whereas 72 (17.2%) were caused by HBM3 memory failures, which is not too surprising given that Nvidia’s H100 GPUs consume around 700W and suffer a lot of thermal stress. Interestingly, only two CPUs ******* in 54 days. 

But while GPUs are the most important components that also happen to be fragile, 41.3% of unexpected interruptions were a result of software bugs, network cables, and network adapters. 

To enhance efficiency, Meta’s team reduced job startup and checkpointing times and developed proprietary diagnostic tools. PyTorch’s NCCL flight recorder was extensively used to quickly diagnose and resolve hangs and performance issues, particularly related to NCCLX. This tool captures collective metadata and stack traces, aiding in swift problem resolution. 

NCCLX played a crucial role in ******** detection and localization, especially for NVLink and RoCE-related issues. The integration with PyTorch allowed for monitoring and automatic timeout of communication stalls caused by NVLink failures. 

Straggling GPUs, which can slow down thousands of other GPUs, were identified using specialized tools. These tools prioritized problematic communications, enabling effective detection and timely resolution of stragglers, which ensured that slowdowns were minimized, maintaining overall training efficiency. 

Environmental factors, such as mid-day temperature fluctuations, impacted training performance by causing a 1-2% variation in throughput. The dynamic voltage and frequency scaling of GPUs were affected by these temperature changes, though it wasn’t a big problem. 

Yet another challenge experienced by the Llama 3 405B LLM training team is simultaneous power consumption changes of tens of thousands of GPUs, which stresses their data center’s power grid. These fluctuations, sometimes in the tens of megawatts, stretched the grid’s limits, which means that Meta has to ensure that its data centers have enough power. 

Considering the fact that a 16,384 GPU cluster experienced 419 failures in 54 days (7.76 times per 24 hours, or a ******** every three hours), we can only wonder how often xAI’s cluster containing 100,000 H100 GPUs will experience a ********.



This is the hidden content, please

#Faulty #Nvidia #H100 #GPUs #HBM3 #memory #contributed #failures #hours #Metas #LLama #training #GPU #cluster #detailed #whitepaper

This is the hidden content, please

This is the hidden content, please

For verified travel tips and real support, visit: https://hopzone.eu/

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Unfortunately, your content contains terms that we do not allow. Please edit your content to remove the highlighted words below.
Reply to this topic...

×   Pasted as rich text.   Paste as plain text instead

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

  • Vote for the server

    To vote for this server you must login.

    Jim Carrey Flirting GIF

  • Recently Browsing   0 members

    • No registered users viewing this page.

Important Information

Privacy Notice: We utilize cookies to optimize your browsing experience and analyze website traffic. By consenting, you acknowledge and agree to our Cookie Policy, ensuring your privacy preferences are respected.