Intel launches Xeon 6 with Performance-core (P-cores) and Gaudi 3 AI accelerators. This showcases Intel’s commitment to delivering powerful AI systems with optimal performance per watt and lower total cost of ownership (TCO).
“Demand for AI is leading to a massive transformation in the data center, and the industry is asking for choice in hardware, software, and developer tools,” said Justin Hotard, Intel executive vice president and general manager of the Data Center and Artificial Intelligence Group. “With our launch of Xeon 6 with P-cores and Gaudi 3 AI accelerators, Intel is enabling an open ecosystem that allows our customers to implement all their workloads with greater performance, efficiency, and security.”
These are two significant AI infrastructure updates within its data center portfolio. The Intel Xeon 6 processor, featuring P-cores, delivers twice the performance of its predecessor with increased core count, double the memory bandwidth, and AI acceleration in every core, addressing the performance needs of AI across edge, data center, and cloud environments. Meanwhile, the Intel® Gaudi 3 AI Accelerator, optimized for large-scale generative AI, offers 64 Tensor processor cores, eight matrix multiplication engines, 128 GB of HBM2e memory, and 24 200 Gb Ethernet ports for scalable networking. Seamlessly compatible with PyTorch, Hugging Face transformer, and diffuser models, Gaudi 3 is also part of Intel’s collaboration with IBM, aimed at
Deploying AI at scale involves considerations such as flexible deployment options, competitive price-performance ratios, and accessible AI technologies. Intel’s robust x86 infrastructure and extensive open ecosystem position it to support enterprises in building high-value AI systems with an optimal TCO and performance per watt. Notably, 73% of GPU-accelerated servers use Intel Xeon as the host CPU3.
Intel partners with leading OEMs including Dell Technologies and Supermicro to develop co-engineered systems tailored to specific customer needs for effective AI deployments. Dell Technologies is currently co-engineering RAG-based solutions leveraging Gaudi 3 and Xeon 6.
Transitioning generative AI (Gen AI) solutions from prototypes to production-ready systems presents challenges in real-time monitoring, error handling, logging, security, and scalability. Intel addresses these challenges through co-engineering efforts with OEMs and partners to deliver production-ready retrieval-augmented generation (RAG) solutions.
These solutions, built on the Open Platform Enterprise AI (OPEA) platform, integrate OPEA-based microservices into a scalable RAG system, optimized for Xeon and Gaudi AI systems, designed to allow customers to easily integrate applications from Kubernetes, Red Hat OpenShift AI, and Red Hat Enterprise Linux AI.
Intel’s Tiber portfolio offers business solutions to tackle challenges such as access, cost, complexity, security, efficiency, and scalability across AI, cloud, and edge environments. The Intel Tiber Developer Cloud now provides preview systems of Intel Xeon 6 for tech evaluation and testing. Additionally, select customers will gain early access to Intel Gaudi 3 for validating AI model deployments, with Gaudi 3 clusters to begin rolling out next quarter for large-scale production deployments.
New service offerings include SeekrFlow, an end-to-end AI platform from Seekr for developing trusted AI applications. The latest updates feature Intel Gaudi software’s newest release and Jupyter notebooks loaded with PyTorch 2.4 and Intel one API and AI tools 2024.2, which include new AI acceleration capabilities and support for Xeon 6 processors.