There’s a new CPU maker.
Chapters13
An introduction noting that most viewers are using ARM-based devices, and clarifying that ARM itself doesn’t manufacture CPUs.
ARM unveils the AGI CPU with 136 Neoverse V3 cores on a 3nm process, promising best-in-class performance-per-watt for data centers.
Summary
Linus Tech Tips’ coverage of ARM’s ARM Everywhere event centers on the new AGI CPU, a processor designed for performance, scale, and efficiency. The chip uses up to 136 Neoverse V3 cores, each paired with 2 MB of L2 cache, all built on TSMC’s 3nm node and capable of 3.6 GHz. Linus highlights that ARM’s design chooses SMT multi-threading and stable performance over aggressive single-core boosts to keep power use predictable. A 12-channel DDR5 memory controller feeds every core at a steady 6 GB/s, a choice Linus says helps data centers avoid buffering for peak demand. The AGI CPU includes 96 PCIe Gen 6 lanes and CXL 3.0, enabling massive shared memory pools across PCIe. ARM showcased rack-level designs that could fit up to two CPUs per motherboard, and a single 36 kW OCP rack holds 300 W per AGI CPU, which enables dense configurations with impressive power efficiency. In an 200 kW OCP rack, ARM claims 42 eight-node systems, totaling 45,696 cores and over a petabyte of RAM, with roughly half the power budget of competing x86 systems. The conversation shifts from hardware specs to practical economics, arguing that power efficiency, reduced legacy bottlenecks, and favorable memory latency today translate into strong long-term performance-per-watt. ARM explains the branding around AI readiness, noting CPUs are still the coordination layer even as accelerators handle the tensor work, and that headroom for higher core counts per watt could improve once AI workloads scale. Linus wraps by noting the live demos — including real-time H.264 to H.65 encoding with concurrent computer vision — emphasized software porting support and ARM’s multi-model roadmap for IP licensing, compute subsystems, and CPUs. If you’re curious about roadmaps and commercial implications, Linus suggests watching ARM’s longer-term plans, including a second CPU expected next year. Cap it off with a nod to a CES ARM feature that mirrored the event, and you’ve got a snapshot of how ARM is aiming to redefine data-center economics for next-gen workloads.
Key Takeaways
- ARM’s AGI CPU can host up to 136 Neoverse V3 cores on TSMC 3nm with 3.6 GHz tops and 2 MB L2 cache per unit.
- A 12-channel DDR5 memory controller delivers 6 GB/s per core, supporting stable power and performance in data-center racks.
- The design emphasizes SMT, consistent power draw, and reduced legacy bottlenecks to improve performance-per-watt versus traditional x86 CPUs.
- In a 200 kW OCP rack, ARM projects 45,696 cores and over a petabyte of RAM across 42 eight-node systems, at about half the power budget of competitors.
- ARM frames this as a multi-model business: licensing IP, licensing compute subsystems, and selling CPUs, offering flexible paths for customers.
- Live demos showed ARM handling mixed workloads (video encoding plus computer vision) on hardware brought to the showroom floor, highlighting software porting support.
- The first CPU is positioned as a safe first step with a second CPU on the horizon for next year, signaling ongoing ARM expansion into data-center CPUs.”],
Who Is This For?
Essential viewing for data-center planners, hyperscale developers, and hardware enthusiasts curious about ARM’s push into CPU design and high-density racks. It outlines both the architectural choices and the commercial strategy behind ARM’s AGI CPU.
Notable Quotes
"up to 136 ARM Neoverse V3 cores with 2 MGB of level 2 cache each built on TSMC's 3nmter process node and it can run at up to 3.6 GHz"
—Core hardware specs highlighted at launch.
"By issuing SMT multi-threading and the highly variable power consumption that's associated with constantly fluctuating clock speeds, not to mention designing a 12 channel DDR5 memory controller that can feed every individual core with a consistent 6 GB per second of bandwidth."
—Describes the design philosophy for stable performance.
"Every AGI CPU draws just 300 W"
—Key power efficiency figure for rack calculations.
"This is just a safe first attempt. The best is yet to come with our second CPU due next year."
—ARM’s roadmap and optimism about future CPUs.
"ARM figures they can pack 42 8 node one piece systems for a grand total of 45,696 cores and over a pabyte of RAM"
—Rack-level capacity projection.
Questions This Video Answers
- How many cores does ARM's AGI CPU have and what is the memory config?
- What makes ARM AGI’s power efficiency claims credible for data-center use?
- How does ARM plan to position AGI CPUs among IP licensing, compute subsystems, and physical CPUs?
- What are the implications of PCIe Gen6 and CXL 3.0 support for ARM in hyperscale environments?
- When is the second ARM CPU expected and what might it imply for the ARM ecosystem?
ARM AGI CPUNeoverse V3TSMC 3nmDDR5PCIe Gen 6CXL 3.0OCP rackdata center power efficiencyARM licensing modelsLinus Tech Tips
Full Transcript
If you're watching this right now, you're probably using at least one ARM CPU to do it. Or well, not an ARM CPU because ARM doesn't actually make CPUs. Or do they? That's the big news that they sponsored us down here to their ARM everywhere event to announce. Behind me and in my hand is the ARM AGI CPU built for performance, scale, and as always, efficiency. up to 136 ARM Neoverse V3 cores with 2 MGB of level 2 cache each built on TSMC's 3nmter process node and it can run at up to 3.6 GHz which right out of the gate raises some questions doesn't it just 3.6 GHz can it like dynamically boost a single core way higher or something?
No. And according to ARM, that's actually a key feature, not a bug. By issuing SMT multi-threading and the highly variable power consumption that's associated with constantly fluctuating clock speeds, not to mention designing a 12 channel DDR5 memory controller that can feed every individual core with a consistent 6 GB per second of bandwidth. ARM is ensuring that every core in the CPU will perform its best at all times and keep power consumption more consistent, which will allow data centers to design to how much power their racks will consistently consume rather than having to build in a buffer for how much they might consume at peak.
And that's huge considering that cooling and especially power are just about the hottest commodities in a world that is rapidly scaling data center infrastructure. Each AGI CPU has 96 lanes of PCI Express Gen 6 with support for CXL 3.0 for deploying massive shared memory pools over PCIe. And ARM showed off node designs with their hardware partners that deployed up to two of these CPUs on a single motherboard. Super cool, but not exactly world changing yet. To see the vision that led ARM to spend the last few years bringing this to life, you got to zoom out and look beyond the individual node to the rack level.
This rack contains 30 two node one piece servers. So for those keeping count at home, that's 8,160 CPU cores. Okay, still not that big of a deal. I mean, dense CPU racks are already a thing. Well, here comes the big reveal. This sick error message hoodie is now available from ltstore.com. JK. Okay, I mean it is, but that's not the big reveal. The big reveal is that everything that I just told you fits in a standard OCP 36 kW air cooled rack. Each AGI CPU draws just 300 W, a significant reduction compared to flagship x86 CPUs.
So when you throw liquid cooling at them, the numbers get frankly kind of ridiculous. In an OCP 200 kW rack, ARM figures they can pack 42 8 node one piece systems for a grand total of 45,696 cores and over a pabyte of RAM. All while consuming only about half of that total available power budget. They are pegging the bottom line performance per watt in the neighborhood of double compared to x86. And this is largely thanks to carrying less legacy croft, but also thanks to architectural choices like using fewer chiplets to keep memory latency down along with ARM's traditional strength in instructions per clock and taking just a no silicon wasted approach to their design.
With the cost and scarcity of power, that's a number that is going to perk up a lot of ears. But why though? Everybody knows that CPUs aren't good at AI compared to GPUs or application specific neural processors. So, uh, what's with the branding? ARM met that question head on. While GPUs and neural accelerators get all the attention, CPUs are still chugging along in the background coordinating tasks with ARM estimating that a typical deployment today is going to have about 30 million cores per gawatt. But here's the thing, that's with humans handling most of the token requests.
AI agents push requests much faster and um don't sleep, meaning that your expensive AI accelerators can end up sitting around because the CPU coordinators can't keep up with all of those requests. So, ARM figures that that 30 million cores per gawatt number could go up a lot in the head node next to the accelerator rack as high as about four times as many. But, uh here's the thing. when these are doing all the actual AI work. Nobody's going to want to spend more power budget on all of those CPUs. Well, that's where ARM comes in with their famously power efficient designs.
Let's go to Nick from the lab to see this thing in action. Many of the demos were focused on the ease of porting software to ARM and the support they're building for developers, which makes a lot of sense, but isn't very visual. So, let's check out this one instead, where they're encoding a 1080p video from H.264 264 to H65 while running computer vision at the same time on the same CPU. Let's go take a look at the man behind the curtain. That's not a video recording. ARM actually had the stones to do it live, bringing an actual server running the actual hardware here to the show floor.
But um awkward question. Doesn't all of this put ARM kind of in direct competition with their own customers? you know, the ones who licensed their IP and their compute subsystems, the guys who got them where they are today. Well, on paper, yes. Um, absolutely. But from ARM's perspective, this is actually something that many of their customers were asking for. Expanding on that, ARM laid out how their road map and their policies account for how all three of their business models are going to go forward. And they're positioning this as a choice between IP licensing, compute subsystems licensing, and physical CPUs.
Or hey, why not some combination of all three? They'll gladly take your money any way you want to give it to them. Contact your local sales representative. Will is here. He can be reached afterwards. And it seems like that's the plan for the long haul. In a move that I don't think I've ever seen before, ARMS stood up on stage and said the quiet part out loud. This is just a safe first attempt. The best is yet to come with our second CPU due next year. Like obviously given the timelines of silicon development, but you almost never hear that from a company who probably wants you to buy the hardware they have today from partners like, you know, for example, Super Micro.
Pretty wild. If you guys enjoyed this video, you might enjoy the one that we did at CES, also in partnership with ARM, highlighting some of the unexpected places that you can find ARM technology.
More from Linus Tech Tips
Get daily recaps from
Linus Tech Tips
AI-powered summaries delivered to your inbox. Save hours every week while staying fully informed.



