Array
(
    [content] => 
    [params] => Array
        (
            [0] => /forum/threads/ex-intel-engineers-are-developing-the-biggest-baddest-cpu-in-the-world-by-targeting-ipc-not-clockspeed-or-core-counts.23027/
        )

    [addOns] => Array
        (
            [DL6/MLTP] => 13
            [Hampel/TimeZoneDebug] => 1000070
            [SV/ChangePostDate] => 2010200
            [SemiWiki/Newsletter] => 1000010
            [SemiWiki/WPMenu] => 1000010
            [SemiWiki/XPressExtend] => 1000010
            [ThemeHouse/XLink] => 1000970
            [ThemeHouse/XPress] => 1010570
            [XF] => 2021770
            [XFI] => 1050270
        )

    [wordpress] => /var/www/html
)

Ex-Intel engineers are developing the 'biggest, baddest CPU in the world' by targeting IPC, not clockspeed or core counts

Daniel Nenni

Admin
Staff member
 AheadComputing chip illustration.

Credit: AheadComputing

Intel, CPUs, and the concept of "badness" aren't necessarily things you'd want to shout about, what with numerous well-documented issues afflicting Intel's recent processors. But a new Oregon-based startup called AheadComputing is leaning hard on the Intel provenance of its founders while claiming that it is creating, "the biggest, baddest CPU in the world." And it's going to do it via IPC or instructions per clock, not cranking up the operating frequencies or throwing in more cores.

That is some statement. All four of AheadComputing's founders had long careers at Intel, dating all the way back to ye olde 386 processor through to the latest Intel Core-branded chips. What's more, AheadComputing also appointed CPU design legend Jim Keller to its board in March. That's at least a vote of confidence, even if it seems unlikely Keller will be involved in the design of AheadComputing's CPUs.

The company is very young, having launched in July last year with a plan to, "develop and license breakthrough, high-performance 64-bit RISC-V processor cores." RISC-V, of course, is an open-source instruction set that exists to present a more modern and cost effective alternative to the proprietary x86 and Arm standards.

Currently, RISC-V chips tend to be found in embedded applications and commercial devices. RISC-V has yet to make much of an impact in PCs or phones, for instance.

Exactly how AheadComputing is going to deliver on that promise of the "biggest, baddest CPU in the world" isn't totally clear beyond the focus in IPC. It's a fabless startup, which means it won't manufacture chips itself. But then the likes of AMD and Nvidia are fabless, too. It's really only Intel that designs and manufactures its own chips, and that business is coming under increasing pressure.

According to AheadComputing's CEO Debbie Marr, "the x86 ecosystem is fiercely defending its territory but is destined to lose in the end." As for Arm, she says, "we anticipate that the ARM ecosystem will experience considerable strain in the coming years. If ARM's current customers are pressured excessively, they will consider transitioning to an alternative architecture like RISC-V."

In response, AheadComputing claims it will, "demonstrate leadership in CPU performance and performance per watt in a very short timeframe and start building the second generation of products that will demonstrate our commitment to a roadmap with large gains in performance generation over generation."

AheadComputing says it will achieve that via IPC, or instructions processed per clock, as opposed to operating frequency or adding cores. "If the performance and efficiency from the multi-core scaling era are slowing down, then it's time for the CPU designers to find a different way to use the additional gates from new process technologies. CPU designers must look towards IPC. This will require increasing the functions for each core rather than increasing the number of cores. If we do this intelligently, AheadComputing will provide performance improvements regardless of workload parallelism," says co-founder Jonathan Pearce.

That latter point could be critical. When Intel's plans for 10 GHz-plus computing hit the wall towards the latter end of the 2000's, the company dramatically changed tack in favour of multi-core computing as a way to add performance in the absence of substantial clockspeed improvements.

Intel Core i9-14900K being installed into a motherboard CPU socket

Adding lots of cores isn't always the best way to improve performance. | Credit: Intel

The problem with adding cores is that it relies on multi-threaded workloads. That's fine for many tasks, like 3D rendering. But it's not a magic bullet for every computational task. Indeed, that's why AMD's eight-core Ryzen 7 9800X3D is the weapon of choice for PC gaming, currently. Adding another eight cores in the form of the Ryzen 9 9950X3D typically doesn't do a whole lot for gaming performance.

Whatever, aside from that focus on IPC as opposed to adding cores, AheadComputing isn't going into any detail. For sure, it will be years before the company's CPU core designs have any chance of showing up in a device you can actually buy.

Personally, if you offered me a CPU with either 50% more IPC or 50% more cores, I'd take the IPC every time. That will deliver in almost any circumstance, while multi-core CPUs can be a bit more hit and miss. Aiming for improved IPC also tends to make for better efficiency, which is great for mobile PCs.


Anywho, for now we'll have to chalk AheadComputing down as a slow burn. The company has strong provenance, but it's anyone's guess as to whether it will, in reality, make an impact. My best guess is that if it manages to come up with an interesting core design, it'll get snapped up by one of the big boys, just as the startup Nuvia was bought by Qualcomm and its Oryon CPU cores ended up in the new Snapdragon X chips.

And all of that is before you even begin to ponder the odds of any RiSC-V chip making an impact on the PC. Industry watchers have been predicting Arm chips would take over the PC for decades. That still hasn't happened.

 
Daniel writes:
> if you offered me a CPU with either 50% more IPC or 50% more cores, I'd take the IPC every time.

Set the way-back machine to the mid-90s. A group of ex-Amdahl* CPU design engineers went to work for a startup company called Rise. They pulled in a few other design engineers (I was ex-Amdahl design engineer from the system storage unit) and some really sharp college grads. In 1997, iirc, we released the Rise mp6 (Pentium II class) CPU, later followed by mp6 II. Intel flat our said no one could do DVD decode at less than 400 MHz. Our mp6 chip could do DVD decode at 266 MHz.

The Rise CPU had a third pipeline and it exceeded Intel's IPC by a significant amount (less than 50% though, as it was not always 100% full). Transmeta made a splash after Rise, but it was still a small ripple in the pond.

Intel had educated the market that only clock speed mattered and we could not undo the narrative... I wish these folks luck.

* - For those not familiar with Amdahl, the company was found by Gene Amdahl from IBM. It designed and manufactured IBM compatible business mainframes - System 360/370/390 and storage systems.
 
Amphere is ex Intel as well. In fact, you will be hard pressed to find a CPU start-up without some Intel Inside.

I love the Amdahl reference. I trained on minicomputers in college but we did have a CDC mainframe somewhere on campus for "big jobs". Mini computers killed the mainframes and workstations (SUN) killed the minicomputers, and now we are queuing up big jobs in the cloud. Simply amazing.
 
Set the way-back machine to the mid-90s. A group of ex-Amdahl* CPU design engineers went to work for a startup company called Rise. They pulled in a few other design engineers (I was ex-Amdahl design engineer from the system storage unit) and some really sharp college grads. In 1997, iirc, we released the Rise mp6 (Pentium II class) CPU, later followed by mp6 II. Intel flat our said no one could do DVD decode at less than 400 MHz. Our mp6 chip could do DVD decode at 266 MHz.

The Rise CPU had a third pipeline and it exceeded Intel's IPC by a significant amount (less than 50% though, as it was not always 100% full). Transmeta made a splash after Rise, but it was still a small ripple in the pond.
Very cool you worked on that! I remember being excited by the Rise chip at the time - powerful MMX at lower clock speeds, and a good upgrade for older Socket 7 boards (with BIOS/mods). The mp6 design lived on for a while longer for embedded and industrial purposes: https://en.wikipedia.org/wiki/Vortex86

BTW - here's a "retro benchmark compilation" including the Rise mp6 thread: https://www.vogons.org/viewtopic.php?t=34666

And specifically raw ALU and FPU performance of "686" class chips: https://www.vogons.org/download/file.php?id=11657

For ALU, the Rise mp6 is a bit faster per clock than Pentium II, FPU looks to be a lot slower though, but better than other contemporaries like the Cyrix 6x86MX, and competitive with K6.
 
Amphere is ex Intel as well. In fact, you will be hard pressed to find a CPU start-up without some Intel Inside.

I love the Amdahl reference. I trained on minicomputers in college but we did have a CDC mainframe somewhere on campus for "big jobs". Mini computers killed the mainframes and workstations (SUN) killed the minicomputers, and now we are queuing up big jobs in the cloud. Simply amazing.
Even Apple has Intel inside 🤣🤣
 
" And it's going to do it via IPC or instructions per clock, not cranking up the operating frequencies or throwing in more cores."

The fact of the matter is that you need to do all three to be competitive.

and in fact, if there is a fourth lever to pull, any good pragmatic computer architect will be remiss not to take full advantage of it.
 
The only "high volume" application I can think of which is so dependent on single threaded performance is PC gaming. IDC says the unit volume in 2024 was about 11 million units. A lot of that market will be x86 dependent, at least for the foreseeable future. This doesn't look promising as a TAM.

I'd rather have many cores, since I'm not a gamer.
 
The only "high volume" application I can think of which is so dependent on single threaded performance is PC gaming. IDC says the unit volume in 2024 was about 11 million units. A lot of that market will be x86 dependent, at least for the foreseeable future. This doesn't look promising as a TAM.

I'd rather have many cores, since I'm not a gamer.
Treat for you since Nova Lake is supposed to be 52 Core Desktop CPU

1750174333912.png
 
The only "high volume" application I can think of which is so dependent on single threaded performance is PC gaming. IDC says the unit volume in 2024 was about 11 million units. A lot of that market will be x86 dependent, at least for the foreseeable future. This doesn't look promising as a TAM.

I'd rather have many cores, since I'm not a gamer.
I misread IDC's numbers. Senility must be getting worse. The 11 million units was a quarterly number. 2025 annual projection for desktop + laptop gaming is about 45 million units, but they'll still need to find OEMs willing to build and distribute PCs and laptops with these CPUs. I'm still a skeptic, but I should at least get the data right.

 
The only "high volume" application I can think of which is so dependent on single threaded performance is PC gaming. IDC says the unit volume in 2024 was about 11 million units. A lot of that market will be x86 dependent, at least for the foreseeable future. This doesn't look promising as a TAM.

I'd rather have many cores, since I'm not a gamer.
I think there is a category of "lightly threaded" apps that are "strong single core dependent". i.e. Even if they ultimately use 8+ cores, they're still heavily gated by the fastest ~ 1-4 cores on a processor. I suspect this is the majority of user applications*.

Of course other apps are memory bandwidth and/or latency dependent. Think usually FPU heavy applications for the bandwidth (data compression, AI), or games for the latency. Architectures like Zen make this point where some applications speed up with dual CCD Ryzen, not because of the # of cores, but because the bus allows for more usage of memory bandwidth with 2 CCDs than a single CCD.

A very high performance CPU architecture I think could win a lot of benchmarks; if it were such that you had ~ 4 "extra big" cores, 8 "big" cores, and then an army of efficient/little cores vs. "big little" that we have today". I suspect though this new "very high IPC" chip will fall flat enough on frequency it won't move the needle much if at all.


*A data point: https://www.techpowerup.com/review/intel-core-i9-12900k-e-cores-only-performance/2.html
Compare the 12600K (6 big / 4 little cores) vs. the 12900K e-cores only (8 little). The 12600K has 74% higher overall application performance. 10400F is another good comparison - 6 significantly older P-cores (Skylake arch) is 9.5% faster than 8 e-cores.
 
Treat for you since Nova Lake is supposed to be 52 Core Desktop CPU

View attachment 3284

I'm torn on this.

On one hand, I think reviews are going to basically point out that there's literally no point to go above Core Ultra 5 for 99+% of all users, and Core Ultra 3 for 90+% of users. OTOH that's possibly true for Ryzen's lineup today.

I'd be surprised if they offered this many cores on desktop though.. (with only dual channel memory. )
 
I think there is a category of "lightly threaded" apps that are "strong single core dependent". i.e. Even if they ultimately use 8+ cores, they're still heavily gated by the fastest ~ 1-4 cores on a processor. I suspect this is the majority of user applications*.
I largely agree, but on most clients there are numerous background processes and threads running which are doing relatively intense functions at random times, and I think they're better served, and the foreground apps are better served, by multiple cores. For example, synchronizing shared photo and video albums, various cloud storage functions (like back-up), real-time encrypt/decrypt if your storage is entirely encrypted (highly recommended), etc. Also, if you run Chrome... what a mess of threads, processes, and helper threads that thing is. Putting Chrome on one core is probably going to cause a lot of cache thrashing, just guessing. Safari and Edge look less messy lately, but Chrome is the leader.
Of course other apps are memory bandwidth and/or latency dependent. Think usually FPU heavy applications for the bandwidth (data compression, AI), or games for the latency. Architectures like Zen make this point where some applications speed up with dual CCD Ryzen, not because of the # of cores, but because the bus allows for more usage of memory bandwidth with 2 CCDs than a single CCD.
I agree.
A very high performance CPU architecture I think could win a lot of benchmarks; if it were such that you had ~ 4 "extra big" cores, 8 "big" cores, and then an army of efficient/little cores vs. "big little" that we have today". I suspect though this new "very high IPC" chip will fall flat enough on frequency it won't move the needle much if at all
I think these ex-Intel folks are possibly too much focused on single-thread CPU performance. Say that it does run games a little better than the best over-clocked Intel CPU, unless the retail system is significantly cheaper I can't believe it'll steal a lot of market share. And the game vendors will hate it because being RISC-V it'll require a port and a huge verification effort for each game. I'm still finding it tough to get excited.
 
Last edited:
I think these ex-Intel folks are possibly too much focused on single-thread CPU performance. Say that it does run games a little better than the best over-clocked Intel CPU, unless the retail system is significantly cheaper I can't believe it'll steal a lot of market share. And the game vendors will hate it because being RISC-V it'll require a port and a huge verification effort for each game. I'm still finding it tough to get excited.
100% agree; It would need to be something like literally twice as fast to get people genuinely interested, and I don't see that happening :)

Honestly, this feels a lot like the car company Lucid. on paper, their EVs are superior technically in many respects to the rest of the market, but they have no real path to owning the market on their own.. (and similar $B costs to build market share that a new chip company has).. But they would potentially be a good set of IP to pick up and build from..
 
Back
Top