Intel Unleashes Enterprise AI with Gaudi 3, AI Open Systems Strategy and New Customer Wins
At Vision 2024,
NEWS HIGHLIGHTS
-
Intel unveiled a comprehensive AI strategy for enterprises, with open, scalable systems that work across all AI segments. - Introduced the Intel® Gaudi® 3 AI accelerator, delivering 50% on average better inference1 and 40% on average better power efficiency2 than Nvidia H100 – at a fraction of the cost.
-
Intel announced Gaudi 3 availability to original equipment manufacturers (OEMs) – including Dell Technologies, HPE, Lenovo and Supermicro – broadening the AI data center market offerings for enterprises. -
Announced new
Intel Gaudi accelerator customers and partners, including Bharti Airtel, Bosch, CtrlS, IBM, IFF, Landing AI, Ola, NAVER, NielsenIQ, Roboflow and Seekr. -
Intel announced the intention to create an open platform for enterprise AI together with SAP, RedHat, VMware and other industry leaders to accelerate deployment of secure generative AI (GenAI) systems, enabled by retrieval-augmented generation (RAG). -
Through the
Ultra Ethernet Consortium (UEC),Intel is leading open Ethernet networking for AI fabric. The company introduced an array of AI-optimized Ethernet solutions, including the AI NIC (network interface card) and AI connectivity chiplets.
This press release features multimedia. View the full release here: https://www.businesswire.com/news/home/20240409077438/en/
“Innovation is advancing at an unprecedented pace, all enabled by silicon – and every company is quickly becoming an AI company,” said
More:
Enterprises are looking to scale GenAI from pilot to production. To do so, they need readily available solutions, built on performant and cost- and energy-efficient processors like the
Introducing Gaudi 3 for AI Training and Inference
The
In comparison to Nvidia H100,
Read more at “
Generating Value for Customers with
-
NAVER
: To develop a powerful large language model (LLM) for the deployment of advanced AI services globally, from cloud to on-device. NAVER has confirmed
Intel Gaudi’s foundational capability in executing compute operations for large-scale transformer models with outstanding performance per watt. - Bosch : To explore further opportunities for smart manufacturing, including foundational models generating synthetic datasets of manufacturing anomalies to provide robust, evenly-distributed training sets (e.g., automated optical inspection).
-
IBM
: Using 5th Gen Intel® Xeon® processors for its watsonx.data™ data store and working closely with
Intel to validate the watsonx™ platformforIntel Gaudi accelerators. -
Ola/Krutrim
: To pre-train and fine-tune its first
India foundational model with generative capabilities in 10 languages, producing industry-leading price/performance versus market solutions. Krutrim is now pre-training a larger foundational model on an Intel® Gaudi® 2 cluster. -
NielsenIQ
, an
Advent International portfolio company: To enhance its GenAI capabilities by training domain-specific LLMs on the world’s largest consumer buying behavior database, enhancing its client service offerings while adhering to rigorous privacy standards. -
Seekr
: Leader in trustworthy AI runs production workloads on
Intel Gaudi 2, Intel® Data Center GPU Max Series and Intel® Xeon® processors in the Intel® Tiber™ Developer Cloud for LLM development and production deployment support. - IFF : Global leader in food, beverage, scent and biosciences will leverage GenAI and digital twin technology to establish an integrated digital biology workflow for advanced enzyme design and fermentation process optimization.
-
CtrlS Group : Collaborating to build an AI supercomputer forIndia -based customers and scaling CtrlS cloud services forIndia with additional Gaudi clusters. - Bharti Airtel : Embracing the power of Intel’s cutting-edge technology, Airtel plans to leverage its rich telecom data to enhance its AI capabilities and turbo charge the experiences of its customers. The deployments will be in line with Airtel’s commitment to stay at the forefront of technological innovation and help drive new revenue streams in a rapidly evolving digital landscape.
- Landing AI : Fine-tuned domain-specific large vision model for use in segmenting cells and detecting cancer.
- Roboflow : Running production workloads of YOLOv5, YOLOv8, CLIP, SAM and ViT models for its end-to-end computer vision platform.
-
Infosys
: Global leader in next-generation digital services and consulting announced a strategic collaboration to bring
Intel technologies including 4th and 5th GenIntel Xeon processors,Intel Gaudi 2 AI accelerators and Intel® Core™ Ultra to Infosys Topaz – an AI-first set of services, solutions and platforms that accelerate business value using generative AI technologies.
Ecosystem Rallies to Develop Open Platform for Enterprise AI
In collaboration with Anyscale, Articul8,
As initial steps in this effort,
In addition to the
New Intel® Xeon® 6 Processors:
-
Intel Xeon 6 processors with E-cores (code-namedSierra Forest ):- 4x performance per watt improvement4 and 2.7x better rack density5 compared with 2nd Gen Intel® Xeon® processors.
- Customers can replace older systems at a ratio of nearly 3-to-1, drastically lowering energy consumption and helping meet sustainability goals6.
-
Intel Xeon 6 processors with P-cores (code-named Granite Rapids):- Incorporate software support for the MXFP4 data format, which reduces next token latency by up to 6.5x versus 4th Gen Intel® Xeon® processors using FP16, with the ability to run 70 billion parameter Llama-2 models7.
Client, Edge and Connectivity:
-
Intel® Core™ Ultra processors are powering new capabilities for productivity, security and content creation, providing a great motivation for businesses to refresh their PC fleets.
Intel expects expect to ship 40 million AI PCs in 2024, with more than 230 designs, from ultra-thin PCs to handheld gaming devices. -
Next-generation
Intel Core Ultra client processor family (code-namedLunar Lake ), launching in 2024, will have more than 100 platform tera operations per second (TOPS) and more than 45 neural processing unit (NPU) TOPS for next-generation AI PCs. -
Intel announced new edge silicon across theIntel Core Ultra, Intel® Core™ and Intel® Atom processor and Intel® Arc™ graphics processing unit (GPU) families of products, targeting key markets including retail, industrial manufacturing and healthcare. All new additions to Intel’s edge AI portfolio will be available this quarter and will be supported by the Intel® Tiber™ Edge Platform this year. -
Through the
Ultra Ethernet Consortium (UEC),Intel is leading open Ethernet networking for AI fabrics, introducing an array of AI-optimized Ethernet solutions. Designed to transform large scale-up and scale-out AI fabrics, these innovations enable training and inferencing for increasingly vast models, with sizes expanding by an order of magnitude in each generation. The lineup includes theIntel AI NIC, AI connectivity chiplets for integration into XPUs, Gaudi-based systems, and a range of soft and hard reference AI interconnect designs forIntel Foundry.
A unified experience makes it easier for enterprise customers and developers to find solutions that fit their needs, accelerate innovation and unlock value without compromising on security, compliance or performance. Customers can begin exploring the
For more information on
Forward-Looking Statements
This release contains forward-looking statements, including with respect to:
- our business plans and strategy and anticipated benefits therefrom;
- our AI strategy and AI accelerators;
- our open platforms approach and ecosystem support with respect to AI; and
- other characterizations of future events or circumstances.
Such statements involve many risks and uncertainties that could cause our actual results to differ materially from those expressed or implied, including those associated with:
- the high level of competition and rapid technological change in our industry;
- the significant long-term and inherently risky investments we are making in R&D and manufacturing facilities that may not realize a favorable return;
- the complexities and uncertainties in developing and implementing new semiconductor products and manufacturing process technologies;
- our ability to time and scale our capital investments appropriately and successfully secure favorable alternative financing arrangements and government grants;
- implementing new business strategies and investing in new businesses and technologies;
- changes in demand for our products;
-
macroeconomic conditions and geopolitical tensions and conflicts, including geopolitical and trade tensions between the US and
China , the impacts ofRussia's war onUkraine , tensions and conflict affectingIsrael , and rising tensions between mainlandChina andTaiwan ; - the evolving market for products with AI capabilities;
- our complex global supply chain, including from disruptions, delays, trade tensions and conflicts, or shortages;
- product defects, errata and other product issues, particularly as we develop next-generation products and implement next-generation manufacturing process technologies;
- potential security vulnerabilities in our products;
- increasing and evolving cybersecurity threats and privacy risks;
- IP risks including related litigation and regulatory proceedings;
- the need to attract, retain, and motivate key talent;
- strategic transactions and investments;
- sales-related risks, including customer concentration and the use of distributors and other third parties;
- our significantly reduced return of capital in recent years;
- our debt obligations and our ability to access sources of capital;
- complex and evolving laws and regulations across many jurisdictions;
- fluctuations in currency exchange rates;
- changes in our effective tax rate;
- catastrophic events;
- environmental, health, safety, and product regulations;
- our initiatives and new legal requirements with respect to corporate responsibility matters; and
-
other risks and uncertainties described in this release, our most recent Annual Report on Form 10-K and our other filings with the
U.S. Securities and Exchange Commission (SEC).
All information in this release reflects management's expectations as of the date of this release, unless an earlier date is specified. We do not undertake, and expressly disclaim any duty, to update such statements, whether as a result of new information, new developments, or otherwise, except to the extent that disclosure may be required by law.
About
1 NV H100 comparison based on https://nvidia.github.io/TensorRT-LLM/performance.html#h100-gpus-fp8 ,
2 NV H100 comparison based on https://nvidia.github.io/TensorRT-LLM/performance.html#h100-gpus-fp8 ,
3 NV H100 comparison based on: https://developer.nvidia.com/deep-learning-performance-training-inference/training,
4 Based on architectural projections as of
5 Based on architectural projections as of
6 Based on architectural projections as of
7 See Vision 2024 section of intel.com/performanceindex for workloads and configurations. Results may vary.
©
View source version on businesswire.com: https://www.businesswire.com/news/home/20240409077438/en/
1-973-997-1154
danielle.mann@intel.com
Source: