At the 2025 OCP Global Summit, Intel unveiled Crescent Island, a new GPU built for the age of real-time, everywhere AI. As inference workloads eclipse training, Intel’s chip—powered by Xe3P architecture and 160 GB of LPDDR5X—promises efficient, scalable performance in air-cooled servers. Sampling in 2026, it anchors Intel’s unified AI stack spanning PCs to data centers.

At the 2025 OCP Global Summit, Intel revealed a major addition to its AI accelerator lineup: the Intel Data Center GPU code-named Crescent Island. Designed to handle the explosive growth in AI inference workloads, the chip will deliver high memory capacity and energy-efficient performance for next-generation deployments.
“AI is shifting from static training to real-time, everywhere inference—driven by agentic AI,” said Sachin Katti, Intel CTO. “Scaling these complex workloads requires heterogeneous systems that match the right silicon to the right task, powered by an open software stack. Intel’s Xe architecture data center GPU will give customers the efficient headroom and flexibility they need as token volumes surge.”
AI workloads are rapidly transitioning from massive training runs to real-time inference at scale. Performance alone no longer defines success—systems-level design does. Efficient inference depends on a workload-centric, open architecture that combines diverse compute types under a unified, developer-friendly software ecosystem.
Intel is positioning itself to meet that challenge across the full spectrum—from the AI PC to the data center and the industrial edge—anchored by its Intel Xeon 6 processors and Intel GPUs.
Through the co-design of hardware, software, and orchestration layers, and close collaboration with the Open Compute Project (OCP) community, Intel aims to make scalable AI inference practical, power-efficient, and deployable wherever it’s needed most.
About Crescent Island
The Crescent Island GPU targets air-cooled enterprise servers, optimized for power, cost, and inference-specific performance. Its design centers on delivering ample on-board memory bandwidth and capacity—critical for handling ever-larger model contexts and continuous token streams.
Key specifications include:
- Xe3P microarchitecture tuned for high performance-per-watt efficiency
- 160 GB of LPDDR5X memory, offering strong capacity and throughput
- Compute fabric: Data-parallel execution across multiple SIMT-style clusters, tuned for transformer inference, recommendation systems, and large-context token handling.
- System design: Power- and cost-optimized for air-cooled enterprise servers, eliminating liquid-cooling overhead while maintaining sustained throughput.
- Support for a wide range of data types, enabling “tokens-as-a-service” providers and real-time inference applications
The stack includes:
- A heterogeneous runtime linking Xe GPUs and Xeon 6 processors through shared memory semantics and common orchestration APIs.
- Open compiler and graph-optimization layers for workload portability across CPUs, GPUs, and custom accelerators.
- Compatibility with container-based deployment and orchestration frameworks for inference-at-scale environments.
Intel’s open, unified AI software stack—currently being developed and validated on Arc Pro B-Series GPUs—will extend seamlessly to Crescent Island, allowing developers to tune and deploy workloads early.
Customer sampling for Crescent Island is expected in the second half of 2026, marking the next phase in Intel’s roadmap to deliver efficient, scalable AI inference solutions across every compute domain.
What do we think?
As inference overtakes training as the dominant AI workload, compute architecture must balance performance, power, and system cost. Intel’s co-design strategy—linking silicon, system engineering, and open-source software—aims to reduce integration friction and accelerate deployment.
Crescent Island represents this next step: a data-parallel GPU engineered for scalable inference, high utilization at small batch sizes, and efficient operation within standard air-cooled infrastructure.
If you’d like to know more about the AIP landscape, we just happen to have a new 356-page report on the 137 AIP vendors chasing the $387 billion processor market across five market segments. https://www.jonpeddie.com/store/annual-ai-processors-market-development-report/