Arm has unveiled the AGI CPU, a cutting-edge family of data center processors featuring up to 136 cores, marking a significant shift in the company’s approach. This innovation is built on TSMC's advanced 3nm technology and integrates Neoverse V3 cores, developed in collaboration with Meta. For the first time in its 35-year history, Arm is delivering its own production processors rather than licensing its intellectual property to other partners.
Designed for what Arm describes as "agentic AI infrastructure," this CPU is optimized for the critical task of orchestrating accelerators and efficiently managing data movement in expansive AI systems.
With a robust architecture, the AGI CPU can host up to 136 Neoverse V3 cores, each capable of reaching speeds of up to 3.2 GHz under full load and peaking at 3.7 GHz. This performance is achieved within a thermal design power (TDP) of 300 watts. The processor supports a substantial 12 channels of DDR5 memory, operating at speeds up to 8800 MT/s, resulting in over 800 GB/s of total memory bandwidth—equivalent to 6GB/s per core—while attempting to maintain latency below 100 nanoseconds. Additionally, it features 96 PCIe Gen6 lanes and integrates support for CXL 3.0, facilitating memory expansion and pooling.
Arm's reference design includes a 10U dual-node server conforming to the Open Compute Project’s DC-MHS specifications, where two AGI CPUs can be housed within a blade. Each standard air-cooled 36kW rack can accommodate 30 blades, yielding a staggering total of 8,160 cores. Moreover, Arm has teamed up with Supermicro to offer a high-performance liquid-cooled configuration that can support 336 chips and exceed 45,000 cores.
According to Arm, the AGI CPU surpasses the latest x86 platforms in performance per rack by more than double. However, it’s worth noting that this claim relies on Arm's internal assessments, and independent benchmarks are yet to validate these figures.
While GPUs have traditionally dominated discussions around AI hardware, there is an increasing need for more powerful general-purpose computing solutions as systems like OpenClaw gain traction. Arm aims to address this growing demand, potentially catering to a segment often overlooked by major players like Nvidia and Micron, who have heavily focused on AI-centric applications.
Among the early adopters of the AGI CPU is Meta, which plans to integrate the processor with its proprietary MTIA accelerators. Santosh Janardhan, Meta's head of infrastructure, emphasized their collaborative effort and commitment to a long-term development roadmap.
Arm has also secured commercial interest from companies such as Cerebras, Cloudflare, F5, OpenAI, Positron, Rebellions, SAP, and SK Telecom. Sachin Katti, leading industrial compute at OpenAI, remarked that the AGI CPU is set to enhance OpenAI's infrastructure, specifically within the orchestration layer essential for managing extensive AI tasks.
Historically, Arm has functioned as an IP licensing enterprise, allowing partners like Apple, Nvidia, and AWS to craft customized chips using Arm's architectural designs. With the introduction of the AGI CPU, Arm expands its offerings to include production-ready silicon that clients can deploy directly, in addition to its IP licensing and Compute Subsystems programs.
Arm indicated that the AGI CPU lineup will coexist with the existing Arm Neoverse CSS product offerings, and they have already committed to future products. The company is keen to clarify that this move complements rather than competes with its current licensees. However, it remains to be seen how Arm will navigate this landscape while providing chips to the same data centers that house competing products from Nvidia, AWS, Google, and Microsoft.