As the world experiences a generational shift to artificial intelligence, each of us is participating in a new era of global expansion enabled by silicon. It’s the “Siliconomy,” where systems powered by AI are imbued with autonomy and agency, assisting us across both knowledge-based and physical-based tasks as part of our everyday environments.
At Intel Innovation, the company unveiled technologies to bring AI everywhere and to make it more accessible across all workloads – from client and edge to network and cloud. These include easy access to AI solutions in the cloud, better price performance for Intel data center AI accelerators than the competition offers, tens of millions of new AI-enabled Intel PCs shipping in 2024 and tools for securely powering AI deployments at the edge.
AI requires a broad range of solutions developed with openness and security in mind to speed innovation. Intel’s portfolio of AI-enabling hardware and software – from CPUs, GPUs and accelerators to the oneAPI programming model, OpenVINO developer toolkit and libraries that empower the AI ecosystem – provides competitive, high-performance, open-standards solutions for customers to quickly deploy AI at scale.
More: Intel Innovation 2023 (Press Kit)
Intel Developer Cloud Reaches General Availability
Intel announced general availability of the Intel® Developer Cloud, which gives developers an easy path to test and deploy AI and high performance computing applications and solutions across the latest Intel CPUs, GPUs and AI accelerators. Developers can also take advantage of cutting-edge tools to enable advanced AI and performance. The details:
- The Intel Developer Cloud is built on a foundation of advanced central processing units (CPUs) that are purpose-built for AI, graphics processing units (GPUs), and Intel® Gaudi®2 processors for Deep Learning, along with open software and tools. The cloud development environment also provides access to the latest Intel hardware platforms, such as 5th Gen Intel® Xeon® Scalable processors (code-named Emerald Rapids), which will become available in the Intel® Development Cloud in the next few weeks and launch on Dec. 14, and Intel® Data Center GPU Max Series 1100 and 1550.
- Developers can use the Intel Developer Cloud to build, test and optimize AI and high performance computing applications. They can also run small- to large-scale AI training, model optimization and inference workloads that deploy with performance and efficiency. Based on an open software foundation with oneAPI – the open multiarchitecture, multivendor programming model – Intel Developer Cloud provides hardware choice and freedom from proprietary programming models to support accelerated computing and code reuse and portability.
More: Intel Developer Cloud
Customer and Performance Momentum in the Data Center
Intel announced AI performance updates and industry momentum for its data center and artificial intelligence product portfolio, including Intel Gaudi2 and 3; 4th Gen Intel® Xeon®; 5th Gen Intel Xeon; and future-generation Xeon processors code-named Sierra Forest and Granite Rapids. The details:
- Intel announced a large AI supercomputer will be built entirely on Intel Xeon processors and 4,000 Intel Gaudi2 AI hardware accelerators, with Stability AI as the anchor customer.
- Dell Technologies and Intel are collaborating to offer AI solutions to meet customers wherever they are on their AI journey. PowerEdge systems with Xeon and Gaudi will support AI workloads ranging from large-scale training to base-level inferencing.
- Alibaba Cloud has reported 4th Gen Xeon as a viable solution for real-time large language model (LLM) inference in its model-serving platform DashScope, with 4th Gen Xeon achieving a 3x acceleration in response time because of its built-in Intel® Advanced Matrix Extensions (Intel® AMX) accelerators and other software optimizations.1
- Granite Rapids will include industry-leading Performance-cores (P-cores), offering better AI performance than any other CPU, and a 2x to 3x boost over 4th Gen Xeon for AI workloads.2
New AI Experiences Powered by Intel Core Ultra Processors
Intel will usher in the age of the AI PC with the upcoming Intel Core Ultra processors, code-named Meteor Lake, featuring Intel’s first integrated neural processing unit, or NPU, for power-efficient AI acceleration and local inference on the PC. Intel confirmed Core Ultra will launch Dec. 14. The details:
- Core Ultra delivers low-latency AI compute that is connectivity-independent with stronger data privacy.
- Core Ultra integrates an NPU into client silicon for the first time. The NPU is built to enable low power and high quality and provide entirely new PC experiences. It is ideal for workloads migrating from the CPU that need higher quality or efficiency, or for workloads that would typically run in the cloud due to lack of efficient client compute.
- Core Ultra represents an inflection point in Intel’s client processor roadmap: It’s the first client chiplet design enabled by Foveros packaging technology. In addition to the NPU and major advances in power-efficient performance thanks to Intel 4 process technology, the new processor brings discrete-level graphics performance with onboard Intel® Arc™ graphics.
- Core Ultra’s disaggregated architecture delivers a balance of performance and power across AI-driven tasks:
- The GPU has performance parallelism and throughput, ideal for AI infused in media, 3D applications and the render pipeline.
- The NPU is a dedicated low-power AI engine for sustained AI and AI offload.
- The CPU has a fast response ideal for lightweight, single-inference low-latency AI tasks.
- Intel highlighted a collaboration with Acer to bring AI to its upcoming Core Ultra systems showcasing how the new “Acer Parallax” software feature uses the NPU to add a 3D look and feel to user images.
Powering AI at the Edge
The opportunity for edge computing is immense, fueled by the demand for automating systems and analyzing data through AI. OpenVINO is Intel’s AI inferencing and deployment runtime of choice for developers on client and edge platforms. With the OpenVINO developer toolkit, Intel is making AI at the edge even more accessible. Developer downloads of the OpenVINO toolkit have seen a 90% year-over-year increase in the past year alone. The details:
- OpenVINO 2023.1, powered by oneAPI, makes generative AI more accessible for real-world scenarios, enabling developers to write once and deploy across a broad range of devices and AI applications.
- The newest release – available for download on OpenVINO.ai – brings Intel closer to the vision of any model on any hardware anywhere.
- OpenVINO 2023.1 enables developers to optimize standard PyTorch, TensorFlow or ONNX models and offers full support for the forthcoming Core Ultra processors. It also provides more model compression techniques, improved GPU support and memory consumption for dynamic shapes, as well as more portability and performance to run across the entire compute continuum: cloud, client and edge.
- During the Innovation Day 1 keynote Intel demonstrated Fit:match, an AI solution improving today’s retail fitting-room experience. Fit:match’s 3D Concierge experience uses Intel® RealSense™ Depth Cameras with lidar sensors, Intel Core processors and OpenVINO. With a focus on security and privacy, the solution can scan and match thousands of products to ensure an optimal fit for the customer, which increases purchasing conversions and reduces return rates.
This release contains forward-looking statements, including with respect to Intel’s business plans and strategy, process and product roadmaps, and current and future technologies, as well as the anticipated benefits therefrom. Such statements involve many risks and uncertainties that could cause our actual results to differ materially from those expressed or implied, including: changes in demand for our products; changes in product mix; the complexity and fixed cost nature of our manufacturing operations; the high level of competition and rapid technological change in our industry; the significant upfront investments in R&D and our business, products, technologies, and manufacturing capabilities; vulnerability to new product development and manufacturing-related risks, including product defects or errata, particularly as we develop next generation products and implement next generation process technologies; risks associated with a highly complex global supply chain, including from disruptions, delays, trade tensions, or shortages; sales-related risks, including customer concentration and the use of distributors and other third parties; potential security vulnerabilities in our products; cybersecurity and privacy risks; investment and transaction risk; intellectual property risks and risks associated with litigation and regulatory proceedings; evolving regulatory and legal requirements across many jurisdictions; geopolitical and international trade conditions; our debt obligations; risks of large scale global operations; macroeconomic conditions; impacts of the COVID 19 or similar such pandemic; and other risks and uncertainties described in our earnings release dated July 27, 2023, our most recent Annual Report on Form 10-K and our other filings with the U.S. Securities and Exchange Commission. All information in this press release reflects Intel management views as of the date hereof unless an earlier date is specified. Intel does not undertake, and expressly disclaims any duty, to update such statements, whether as a result of new information, new developments, or otherwise, except to the extent that disclosure may be required by law.
1 Intel does not control or audit third party data. You should consult other sources to evaluate accuracy.
2 Based on architectural projections as of Aug. 21, 2023, relative to 4th Gen Intel Xeon processors. Your results may vary.