New Discussion
Home
Community
0
0
Share

Intel's new Gaudi 3 accelerators massively undercut Nvidia GPUs as AI race heats up

1 question · 1 view
D
David Taylor
Jun 09, 2024

Can you detail the performance benchmarks provided by Intel for the Gaudi 3 in comparison to Nvidia's H100?

socratic-gpt
Socrates

Intel provided several benchmarks to compare the performance of its Gaudi 3 AI accelerator against Nvidia's H100 GPU. According to Intel, the Gaudi 3 is capable of delivering up to 40% faster training times than the H100 in large 8,192-chip clusters. In a smaller 64-chip setup, the Gaudi 3 is said to offer 15% higher throughput than the H100 on the popular LLaMA 2 language model.

For AI inference tasks, Intel claims that the Gaudi 3 is twice as fast as the H100 on models like LLaMA and Mistral. These benchmarks suggest that the Gaudi 3 can offer competitive performance to the H100 in various AI workloads while potentially undercutting Nvidia's pricing significantly. However, it's important to note that the Gaudi chips don't have optimizations for Nvidia's CUDA platform, which most AI software relies on today. This could pose a challenge for Intel in convincing enterprises to refactor their code for Gaudi.

Like
Dislike
Sources
Copy
Betty😄⭐
Betty😄⭐
Jun 09, 2024

How does the cost of Intel's Gaudi 3 accelerator compare to Nvidia's H100 data center GPU?

socratic-gpt
Socrates

The cost of Intel's Gaudi 3 accelerator is significantly lower than that of Nvidia's H100 data center GPU. The Gaudi 3 is priced at around $15,000 per unit, while the H100 costs approximately $30,000 per unit. This makes the Gaudi 3 50% cheaper than the H100, offering a more cost-effective solution for AI acceleration.

Like
Dislike
Sources
Copy
More
Mary009
Mary009
Jun 09, 2024

What are the key features of Intel's Gaudi 3 AI accelerator chips that were unveiled at Computex?

socratic-gpt
Socrates

Intel's Gaudi 3 AI accelerator chips, unveiled at Computex, boast several key features:

  1. Performance: The Gaudi 3 is designed to deliver similar performance to Nvidia's H100 GPU but at a significantly lower cost1. Intel claims that the Gaudi 3 either keeps pace with or outperforms the H100 across a variety of important AI training and inference workloads.

  2. Pricing: The flagship Gaudi 3 accelerator will cost around $15,000 per unit when purchased individually, which is 50% cheaper than Nvidia's competing H100 data center GPU. For the Gaudi 3, an 8-accelerator kit configuration costs $125,000. Intel insists it's two-thirds cheaper than alternative solutions at that high-end performance tier.

  3. Open Standards: The Gaudi chips leverage open standards like Ethernet for easier deployment. This could potentially make them more appealing to enterprises looking for flexibility and freedom from proprietary lock-in.

  4. Partnerships: To drive adoption, Intel says it has lined up at least 10 major server vendors – including new Gaudi 3 partners like Asus, Foxconn, Gigabyte, Inventec, Quanta, and Wistron. Familiar names like Dell, HPE, Lenovo, and Supermicro are also on board.

  5. Benchmarks: Intel's benchmarks show the Gaudi 3 delivering up to 40 percent faster training times than the H100 in large 8,192-chip clusters. Even a smaller 64-chip Gaudi 3 setup offers 15 percent higher throughput than the H100 on the popular LLaMA 2 language model. For AI inference, Intel claims a 2x speed advantage over the H100 on models like LLaMA and Mistral.

  6. Architecture: The Gaudi 3 is architected for efficient large-scale AI compute and is manufactured on a 5-nanometer (nm) process. It is designed to allow activation of all engines in parallel, including the Matrix Multiplication Engine (MME), Tensor Processor Cores (TPCs), and Networking Interface Cards (NICs).

  7. Memory: The Gaudi 3 accelerator features 128 gigabytes (GB) of HBMe2 memory, providing ample memory for processing large GenAI datasets.

  8. Networking: Twenty-four 200 gigabit (Gb) Ethernet ports are integrated into every Intel Gaudi 3 accelerator, providing flexible and open-standard networking. They enable efficient scaling to support large compute clusters and eliminate vendor lock-in from proprietary networking fabrics.

  9. Software: Intel Gaudi software integrates the PyTorch framework and provides optimized Hugging Face community-based models. This allows GenAI developers to operate at a high abstraction level for ease of use and productivity and ease of model porting across hardware types.

  10. Availability: The Intel Gaudi 3 accelerators are expected to be available in Q3 2024.

Like
Dislike
Sources
Copy
More
Socrates may produce inaccurate information. Verify important details.
0 New Question