Intel Gaudi, Xeon, and AI PC Accelerate Meta Llama 3 GenAI Workloads 

  • Intel’s AI products, including Gaudi, Xeon, Core Ultra processors, and Arc graphics, are optimized for Meta Llama 3, a next-gen large language model.
  • Collaboration between Intel and Meta ensures performance and simplicity in AI solutions.
  • Intel’s investments in software and AI ecosystems drive readiness for dynamic AI innovations.
  • Intel Gaudi 2 accelerators and Xeon processors exhibit optimized performance for Llama 2 and 3 models.
  • Intel Core Ultra processors and Arc Graphics deliver impressive performance for Llama 3.
  • Meta anticipates introducing new capabilities and enhanced performance for Llama 3.

Main AI News:

Empowering the AI landscape, Intel’s suite of AI products, spanning from Gaudi and Xeon in data centers to AI PCs, stands poised to catapult Meta Llama 3, its cutting-edge large language model, to new heights. Meta’s launch of Llama 3, the next iteration in large language models, has prompted Intel to validate its AI product lineup for the inaugural Llama 3 8B and 70B models across various platforms including Intel Gaudi accelerators, Xeon processors, Core Ultra processors, and Intel Arc graphics.

Intel actively engages with AI software leaders to deliver solutions that seamlessly blend performance and ease of use. Meta Llama 3 signifies a significant leap forward in large language models for AI. As a key provider of AI hardware and software, Intel is honored to collaborate with Meta to leverage models like Llama 3, empowering the ecosystem to innovate cutting-edge AI applications.” – Wei Li, Intel Vice President and General Manager of AI Software Engineering.

Aligned with its mission to democratize AI, Intel invests in software and AI ecosystems to ensure its products stay at the forefront of dynamic AI innovations. In data centers, Intel Gaudi and Xeon processors, equipped with Intel Advanced Matrix Extension (Intel AMX) acceleration, offer customers versatile options to meet diverse and evolving needs.

Moreover, Intel Core Ultra processors and Intel Arc graphics solutions serve as both local development platforms and deployment solutions across myriad devices, supporting a range of comprehensive software frameworks and tools such as PyTorch and Intel® Extension for PyTorch, alongside OpenVINO toolkit for model development and inference.

Intel’s rigorous testing and performance evaluations for Llama 3 8B and 70B models utilize open-source software, including PyTorch, DeepSpeed, Intel Optimum Habana library, and Intel Extension for PyTorch, ensuring the latest in software optimizations. For detailed performance insights, readers are directed to the Intel Developer Blog.

  • Intel Gaudi 2 accelerators exhibit optimized performance for Llama 2 models – spanning 7B, 13B, and 70B parameters – and now boast initial performance metrics for the new Llama 3 model. With the maturation of Intel Gaudi software, running the new Llama 3 model and generating results for inference and fine-tuning becomes a seamless process. Additionally, Llama 3 is supported on the recently unveiled Intel Gaudi 3 accelerator.
  • Addressing demanding end-to-end AI workflows, Intel Xeon processors undergo rigorous optimization to minimize latency in LLM operations. Intel® Xeon® 6 processors featuring Performance-cores (codenamed Granite Rapids) showcase a 2x improvement in Llama 3 8B inference latency compared to 4th Gen Intel Xeon processors, enabling the execution of larger language models like Llama 3 70B in under 100ms per generated token.
  • Intel Core Ultra processors and Intel Arc Graphics demonstrate remarkable prowess in handling Llama 3 workloads. Initial tests reveal that Intel Core Ultra processors already outpace typical human reading speeds, while the Intel Arc A770 GPU equipped with Xe Matrix eXtensions (XMX) AI acceleration and 16GB of dedicated memory delivers exceptional performance for LLM tasks.

In the near future, Meta anticipates unveiling additional capabilities, expanded model sizes, and enhanced performance. Intel remains committed to optimizing its AI product lineup to bolster support for this burgeoning large language model paradigm.

Conclusion:

Intel’s strategic alignment with Meta’s Llama 3 model underscores its commitment to driving AI innovation across diverse applications. With optimized performance across its product portfolio, Intel is poised to capture significant market share in the rapidly expanding AI landscape. This collaboration solidifies Intel’s position as a leading provider of AI hardware and software solutions, catering to the evolving needs of AI developers and enterprises alike.

Source