Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
ARM’s Neoverse N2: Cortex A710 for Servers (chipsandcheese.com)
100 points by zdw on Aug 18, 2023 | hide | past | favorite | 97 comments


Do you people honestly believe we can have a real discussion about CPUs on HN? These posts are always hijacked into an Apple discussion (currently 21 comments out of 23 are in the Apple thread)

back to the topic: ARM and AMD CPUs have about the same L1 instruction cache size. But don't they have a very different instruction density, giving one a noticeable advantage over the other?


> These posts are always hijacked into an Apple discussion

Or occasionally by RISC-V "enthusiasts" who claim that none of it matters since ARM is dead and everybody is going to switch to RISC-V in the very near (but undetermined) future..


>ARM is dead

Not dead (yet). Just a legacy ISA.

>everybody is going to switch to RISC-V

RISC-V is inevitable.


> RISC-V is inevitable.

Exactly, I find this sort of fanaticism about an instruction set (of all things) so puzzling.

Many of the comments which mention RISC-V have anything constructive to say or any meaningful arguments. All the fanboyism makes me feel a bit like I'm on reddit (or worse)...


There's a bunch of such select phrases the use of which is suggested by the top of the organization[0].

I recognize the value of some level of coordination in marketing efforts, and as I'm no expert I just follow their recommendations, like "RISC-V is inevitable".

0. https://www.youtube.com/watch?v=r9DEM216748


I never realized that it's actually RISC-five and not RISC-/v/.


ARM is dead not because I am a RISCV enthusiast, but because ARM is a terrible company that is taking forever to enter the server market.

Neoverse N2 chips should have been available years ago. Once Jim Keller releases his Ascalon server RISC-V CPUs that are competitive with Zen x86 server CPUs, there is basically no room for ARM servers left since ARM couldn't manage to get their feet into the server market soon enough. So many ARM server vendors canceled their CPU projects, to the point that the safest prediction you can make about any new ARM server CPU is that it will be canceled. The thing about ARM is that it is already dead in the server market.


ARM servers have a better history of mass production than server CPU startups have of making it to 5 years without being acquired. But then that's probably the goal.


Regardless of the merit of your points, why would you willingly walk into the caricature that was so obviously painted?


Amazon thinks otherwise.


The uops are similar so if x86 has a density advantage, it will be paid back during micro-op expansion. Whether this is really a win or not depends on where your bottlenecks are,


Isn't L1 I$ at ISA level? There have been some attempts to cache microops instead but I don't think current CPUs do that.


You mean a µop cache of size approaching that of a L1 instruction cache?

The A710 has a 1536-entry µop cache. That is 3/16 of the L1 cache (counting each instr as four bytes)


A710 and Zen4 both have post-decode uop caches, up to 6750 ops/core in Zen4 and 1536 ops/core in A710.

(Because the uop caches must fill lines aligned to real L1 cache, utilization cannot be perfect, and x86 has a somewhat harder time to achieve full utilization, while the uop cache on arm is easier to utilize because instructions are more regular, only "losing" slots when instructions are fused.)


> ARM and AMD CPUs have about the same L1 instruction cache size. But don't they have a very different instruction density, giving one a noticeable advantage over the other?

Instruction cache sizes are not set by what you want, but by what you can get away with. The fetch path has both a lot of distance to travel and transistors to switch, and is typically a very critical path for clock speed, and while adding more pipeline stages for it helps with clocks, it also directly hurts performance near branches.

So in the end you have to limit the size depending on what clocks you target and how many stages you are willing to give the fetch path.


> Do you people honestly believe we can have a real discussion about CPUs on HN?

It's not just CPUs, feels like many discussions related to personal computing in general get hijacked into Apple discussions, likely because there's many people here desperate to rationalize their purchase of mediocre hardware at a 10x markup.


> desperate to rationalize their purchase of mediocre hardware at a 10x markup

Are silly claims like this necessary? Yes, Apple massively overcharges on memory and storage upgrades because their users obviously have no other options...

But besides Macs are pretty good value if you're prioritizing battery life, slimness and fan sound.

I guess latest XPS 15/17 are fairly competitive with MBP 14/16. Especially if you don't care about battery life and heat that much, the 13700H/ 13900H seem to be on par with the M2 Pro/Max. The base price is up to $500 lower and you get replaceable RAM (also slower) and 2 m2 slots.

However I don't think there is anything that's close the the M2 Air on the market atm?


> I don't think there is anything that's close the the M2 Air on the market atm?

I will be "that guy"; the original M1 Macbook Air struggled to compete with the Ryzen 4800u at launch. Current Ryzen mobile chips put up very well against the M2, considering the architecture gap:

https://browser.geekbench.com/macs/macbook-air-2022

https://browser.geekbench.com/v6/cpu/1604691

This is the sort of discussion the grandparent comment was lamenting though. Honest-to-god, none of this matters; dev shops will use whatever they need at the end of the day. Infecting the discussion with pedestrian "muh ARM revolution" comments always ruins it though. I like ARM for reading e-books as much as the next guy, but server use is a whole different discussion. Even mentioning Apple's consumer offerings is a punch-line like saying you want to beat OpenAI using consumer-grade GPUs. They exist in the ARM space, but they could not be any less-present in the server one. Their relationship to the topic at-hand is fraught and near non-existent if it weren't for their history of failure with Xserve.


> I guess latest XPS 15/17 are fairly competitive with MBP 14/16.

Nah, there's been a history of poor QA and just general poor build issues. Have they fixed the coil whine yet? It's still there last I checked.


Agree that Apple fanboys hijacking all CPU threads are indeed annoying (along with the TSMC and ASML chip hipsters) and I wish moderators would handle this so we could enjoy on-topic discussion without the usual "my M1 is so good, every other non-M1 CPU is now pointless" spam, but I think it's also that people would like to have Apple's M1 level of performance and efficiency from other vendors of more open systems and not be conditioned of always buying into the Apple ecosystem for it, which we know isn't gonna happen, as nobody else is as vertically integrated and has the market share and profit margins of Apple to pull off a X86 to ARM transition for consumers like Apple did, nor is there any standardized ARM platform yet, like the trinational IBM PC clones were, where you can just plug in an extern GPU.

The transition to ARM will happen first in the server/cloud space, but consumer will have to wait much longer for it to reach PCs, if it ever does, as X86 is also making progress.


> X86 to ARM transition for consumers Apple did. > ARM will happen first in the server/cloud space, but consumer will have to wait much longer.

It's not obvious to me that this is somehow inevitable. Yeah power efficiency wise Apple is ahead currently. But are there any reasons AMD/Intel why wouldn't be able to catch up and surpass Apple in a generation or two?


Not while Apple is the main investor in TSMCs leading edge nodes, plus, X86 still has a lot of backwards compatibility cruft keeping it back.

Eventually the gap will narrow as Apple already picked the lowest hanging fruits with the M1 design, and as we saw the jump to M2 was no longer as impressive, so we just have to wait.


> Not while Apple is the main investor in TSMCs

It's not inconceivable that Intel might catch up and surpass TSMC. I mean they were in a somewhat similar situation they are in now back in the early to mid 2000s.


Maybe, maybe not, maybe Samsung ups their game, who knows, nobody can predict what will happen with semi fab players and their processes in the distant future, it's all just speculation at this point.


If you believe Intel, the "distant future" is 2025. Thats when they say they will take fab process leadership. Of course, you should take what any company says about future timelines with some degree of skepticism, but Intel isnt that far behind currently and the first product on their next node (Intel 4) is being launched next month.


Are Dell or HP selling ARM servers to the general public, or still only to the cloud providers?


HP, at least, doesn't seem to want to sell any type of server to anyone.

I recently tried to configure a Proliant DL385 Gen11, in order to give my sales rep a coherent list of "this is what I want" in terms of CPUs, memory and disks. And let me tell you, it's a sales experience that might just as well be a Monty Python sketch, it's so bad.

I mean literally, one of their subdomains had a certificate expired so that I had to add a security exception just to view the information. The page in question had select boxes that couldn't be scrolled, anything I did would just scroll the main page instead. Their other configurator (because ofc they need at least two different ones) was completely unusable, when you try to select a drive or something you get an explosion of recursive "to select that you also need to select this cable kit, one of those 4 cages and deselect that other thing" warnings. Like, if you know what to fix, just fix it!!


This is beyond frustrating, as Ampere servers (low power) are tailor made for on-prem compute. I like because I want my devs and data scientists doing green field stuff to have $0 marginal cost for compute. premature optimization being the root of all evil, and all that.


HP is


Indeed. It's an Ampere chip inside a ProLiant Gen11 https://www.theregister.com/2022/06/28/hpe_arm_proliant_serv...


Worth having a look on the Lenovo x13s as well.


Samsung, Dell and HP are also selling Qualcomm Windows laptops. Why (besides the novelty factor) would anyone get one instead of an Intel/AMD/Apple laptops is beyond me though..


Battery life is a major factor.

Windows is not really my thing for many years, so Apple is on the table in the battery life requirement, but I don’t think I’m coming back to Intel/AMD soon.


Apparently those ARM based applications don't write themselves.


The point about 52bit vs 48bit address width is interesting. It seems to reflect the fact that both intel and amd are going after the supercomputer market. I’m curious if there are other use cases where such a large address width affects traditional server computing. For instance does it prevent using bits for tagging pointers?


I sense a degree of disappointment, both in the article and in some of the comments here, that Arm hasn’t made more rapid progress in getting competitive server CPUs to market.

It’s clearly not the instruction set architecture. I wonder if the problem is that Arm’s business model makes it hard to justify the investment needed given the licensing fees they feel they can charge.

My reading between the lines on Nuvia was that Arm was looking to make more on Nuvia cores and providing Nuvia with substantial help. But Qualcomm broke into the party.

Happy to be told I’m wrong though.


Qualcomm helped save Nuvia from the Apple lawsuit though.


ROB was never defined in that technical article. It is one of the basic of technical writing that it should be defined on first use.


They also didn't define DTLB, ALU, or CPU for that matter.

I thought it was a good choice. They were assuming deep familiarity with modern CPU architectures through the whole text, so a one-sentence definition of all the terms would either have been too much, or not enough, depending on the reader.


It depends entirely on your audience. There is too much to define in these kinds of articles for it to be practical, it would way more than double the length, which is why it's better to just refer to a computer architecture course.


Dear Tim Cook,

I have an idea for Apple's next great product. Bring back Xserves, but with M-series processors in them this time. Then hire the Asahi Linux developers and get Linux running very solidly on them. Make it so you can buy them with either macOS or Linux pre-installed.

Sincerely,

lockhouse


It's likely they already have a bare bones version of linux running solidly in house.


I've heard solid rumors that they do have their own internal linux distro, specifically so the hardware teams can do bringup without a dependency on the XNU team(s). That Linux distro is also supposed to be kind of a shit show held together with bubblegum and paperclips, in contrast to asahi which is bending over backwards to do things right (gods bless them).


To be fair, if it's only for hardware bring up then I would kind of expect it to to be hacked together without any regard for things you care about in a "real" distro. (Heavily patched kernel? Who cares, we're not going to update it. Can't self-host? Whatever, cross compile the whole thing. Software leaks memory? No problem, just reboot the test rig hourly. Broken package manager? We can build+flash a new image every time we want to change the installed software.)


There’s a slight problem with this approach - you can’t trust your test results in such environment… if you manage to complete your test suites at all before the next revision drops.


> in contrast to asahi which is bending over backwards to do things right

Not having to reverse engineer drivers for pretty much every single component would probably require a fraction of the effort though. And it's not like Apple would really need to actually maintain their own distro, releasing the drivers (which they are more than capabable of doing) would be perfectly sufficient.


Dear lockhouse,

What does any of this have to do with Neoverse?


Well apple chips are Arm but better so...


A little late now that Ampere and Nvidia have their entries in the game. Pushing for Apple hardware in the server rack would probably end the same way it did the first time. It would take a cultural shift for Apple to align themselves with the server market.


Given that there are several brands of Mac mini server rack mount kits on the market, clearly there is a demand for Apple hardware in the server room. If nothing else they would be a nice alternative to rack mount Mac Pros for a build farm. I would think the performance per watt would be very competitive with other server vendors on the market as well.


They fill a niche, and so did Xserve. It had an accompanying software suite that was fully fleshed out to an Apple user's likings.

The biggest gap right now (afaik) is Apple's virtualization limits. Regardless of the cores in your system, any Mac can host no more than 2 instances of MacOS. It's a huge kneecap to how Macs can be used as a server, and a death-blow to any serious hosting company that wants to provision them as a virtual resource.

> I would think the performance per watt would be very competitive with other server vendors on the market as well.

I can go get a 4 core Ampere ARM instance for free. The market goes so much cheaper than Apple's CPUs it's not even funny. Performance-wise, Apple kinda got their lunch ate by Nvidia's Grace server. There isn't really a demand to fill for high-performance ARM computing, and the cheap end goes so cheap that Apple can't even touch them.

I see two ways for Apple to escape this:

1. They can double-down on their ecosystem investments to create a "CUDA-killer" and accompanying hardware, spending tens-of-billions of dollars across multiple years to obsolete Nvidia's industry ties.

2. Initiate or join an Open Source pact to write a CUDA replacement or translator a-la Vulkan/DXVK that gives vendors no choice but to abandon Nvidia.

Of course, the easier play is to just pretend all of this is beneath you and continue building your current ecosystem. It's what I would do, and I don't see a rational reason why Tim Cook would read your pitch and greenlight a project like that.


> I don't see a rational reason why Tim Cook would read your pitch and greenlight a project like that.

Why not? Even just Apple's internal expense could justify it - for the same reason Google, Microsoft, Amazon etc all dog food their own cloud. Apple currently pays AWS, Akamai, Cloudflare, etc lots of $$$.

As the current focus is to expand on services and managed offerings you would focus on controlling such a cost (as is typical Apple).

> I can go get a 4 core Ampere ARM instance for free. The market goes so much cheaper than Apple's CPUs it's not even funny.

Is the market full of people that only use 4 cores? If that was really the case AWS would have closed shop already instead of being the market leader. It's not like Oracle doesn't charge you for anything else.

> Apple kinda got their lunch ate by Nvidia's Grace server

And where is Grace? Which cloud currently? It doesn't work that way. If all you think about a server and its ecosystem is hardware then lots of Android phones have better hardware than iphone e.g. in screens, battery life, etc. Was that really the deciding factor? No.

I mean by your logic, AMD's EPYC has been winning Intel Xeon's by years. Has Xeon been decimated? No. There's still room for both. AMD can't even produce enough. Considering Apple can exclusive lock TSMC down for a year on bleeding edge, who can really compete if Apple was in?


> Is the market full of people that only use 4 cores

Ampere CPUs on Oracle, Hetzner and (I assume?) Azure are still remarkable cheap compared to Intel/AMD if they are sufficient for your workflows.

I'm not sure if the margins in this market are high enough for Apple to enter it and they would probably still have issues scaling production capacity if they are serious about it.


> Ampere CPUs on Oracle, Hetzner and (I assume?) Azure are still remarkable cheap compared to Intel/AMD if they are sufficient for your workflows.

Hetzner has always been there. It didn't impact Intel/AMD use cases. There are lots of cheap providers (since the start of time) and I doubt that's the space Apple would ever get into.


It's just not a market Apple can infiltrate. If Apple wants to compete on the high-end, they have to dethrone Nvidia and their current high-performance ARM servers (H100 clusters and whatnot). That is a suicide mission for Apple or any company really, regardless of the money they have on-hand.

So... who would they be selling to, now? I genuinely think such a "server Mac" would echo the Xserve without a fundamental shift in the way Apple treats the software community.


> It's just not a market Apple can infiltrate. If Apple wants to compete on the high-end, they have to dethrone Nvidia and their current high-performance ARM servers (H100 clusters and whatnot).

Why? And why is it only about hardware? So AWS, GCP and Azure all run Intel/AMD, who do they need to dethrone?

For all it matters Apple "cloud" could even use these nVidia servers themselves. Not like AWS only uses Graviton. So, so?

> So... who would they be selling to, now? I genuinely think such a "server Mac" would echo the Xserve without a fundamental shift in the way Apple treats the software community.

Right, why would they only do hardware? As is typical Apple the $ is in the ecosystem and services. The competition would be to offer cloud services that has a better experience just like they do for phones and others. The macbook pro was always more expensive and yet people still purchase them.


> The competition would be to offer cloud services that has a better experience just like they do for phones and others. The macbook pro was always more expensive and yet people still purchase them.

Which is both less relevant and quite different for enterprise services than consumer products. It's also not that obvious that Apple has a lot of expertise in this area and could provide something better than AWS/Azure/etc.. and Apple is not exactly the company which is willing to compete on lower margins.


> The competition would be to offer cloud services that has a better experience just like they do for phones and others

You are just back around to describing Xserve again. Regardless of whether it is software or hardware-based, an Apple Cloud has no upsell in today's market. That's why Xserve died, and it's why it would die again.


The WAS multi-billion contract ends this or next year. Apple cloud is imo one of the easiest ways for them to grow. It's not coincidental that they're working hard on VirtualizationKit etc. Biggest problem would be production of chips


> Biggest problem would be production of chips

I would actually argue that is the easy part. Apple already has access to cutting-edge nodes if they want to pay for them. The hard part is designing a server chip from Apple Silicon, a decidedly mobile SOC.

For instance; what if Apple wanted to ship a system that competes with a DGX GH200? Where would you even start with something like the M2 Ultra?

- You would either need more cores per SOC, or more SOCs per chipset to match a DGX system

- Custom memory controllers are a must (running them at regular speeds depends on how much bandwidth is left)

- PCI would need to exposed on-hardware and supported in MacOS, Linux and third-party software

- An multi-instance interconnect like Infiniband would need to be working and attractive (as well as a CUDA translator, if they want any hope)

And that's just what they would need to match the high-performance ARM industry right now. If they wanted to compete or knock Nvidia's block off, it would take a lot more investment, time and attitude-changing.


You don’t need that to run most cloud services, and where most of the profits come from at this point for aws/azure


Where do you get a 4 core ampere for free?


Not OP, but I presume he is referring to Oracle Cloud, as they offer "free-forever" 4 core (24 Gb RAM) instances.

Fairly good deal, the only problem being that it's Oracle.

ps. if you do decide to give them a try, be sure to upgrade to "pay-as-you-go" (and never go beyond the free offering) otherwise they will close your account without any warning (ask me how I know).


Thanks! that's good to know. As much as I like free things I may stay away from that one.


Suit yourself. It's great for the stuff I don't want to host on my homeserver, like AI Discord bots and big remote builds.

That being said, I wouldn't trust the server as far as I could throw it. It's a bad deal unless you have ephemeral compute requirements that don't matter too much.


or just offer the hardware exclusively on Apple Cloud.


You're evil, but going places.


It’d be additional services revenue, and they’re all about services.


> You're evil, but going places.

Just on par with AWS graviton...


I look forward to the day when AKS is assumed to be Apple and not Azure.


Better get a chair to wait on.


Xserve and A/UX hardly had anything to offer over other UNIX on the server.

Likewise Apple also had their own Linux distribution that went nowhere, MkLinux.


> Likewise Apple also had their own Linux distribution that went nowhere, MkLinux.

MkLinux (MicroKernel Linux) was not a distribution but a modified Linux kernel to be loaded as a Mach server to run under CMU (later OSF) Mach 3. It abstracted idiosyncrasises of the Apple NuBus and peripherals away from the Linux kernel that did not (or could not altogether) support – the NuBus itself and the NuBus attached devices.

The distribution was a stock RedHat v4 or v5.

Amusingly, despite CMU Mach 3 being a µ-kernel, there was nothing «micro» about the size of its own binary, and – off the top of my head – the Mach binary size was larger than that of the vmlinuz at the time.


It was still a distribution, the Linux kernel packaged alongside whatever else is necessary to make it work in an usable operating system.

If anything, it was actually pity it didn't took off, as its design was quite interesting.


Risc-v is already coming for the data center and Apple have their hands full.


RISC-V is at-best 10-15 years away from being in the position to where ARM is now.


That isn't how CPU design works. The microarchitecture is what matters and you can use the same techniques as x86 uses today to get good performance.

Just because there is a new ISA does not mean a company cannot design a modern CPU.

For example, the Vision Star Five is uses an in order CPU core that is competitive with the out of order A72 core. The performance gap on day one is much smaller. It also has UEFI support and opensource GPU drivers for Power VR are in development by imagination, so the software support will also quickly catch up or surpass ARM SBCs.

The same will happen in the server market. Nobody who works on a RISCV server chip has to wait for ARM to deliver a CPU core that doesn't suck.


Likewise, because there is a modern CPU, doesn't mean the relevant partners that actually matter in the industry care enough about its existence.

Or that they show up in computer shops for regular buyers, instead of computer enthusiasts, used to install GNU/Linux into any kind of motherboard with a CPU on it.


I have no insights about this but Jim Keller disagrees with that statement


Jim Keller is someone to pay attention to, but as a sibling says, JK has a commercial IP interest, which means the touted openness advantage could easily be lost. This insight comes from people who have consulted for organisations like Tenstorrent and SambaNova and both companies focussed on designing for HPC and AI. General-purpose consumer computing platform (i.e. servers, desktops and laptops) is yet to be forthcoming from anyone in the space, and they are a way off.


> which means the touted openness advantage could easily be lost

RISC-V openness has nothing to do with processor implementation, it's solely in the instruction set.

The advantage it has is that it provides compatibility across vendors and anyone can become a vendor of RISC-V IP. Comercial IP doesn't affect that, and if anything makes multiple IP vendors more likely.

The availability of server processors depends on the market rather than any technical limitation. If someone decides to take the risk of investing in the development of a strong server product then we'll get it. It depends on some company believing there is a market for it.


> which means the touted openness advantage could easily be lost

I don't expect them to open source their processors, but they do definitely contribute back to the community. E.g. they open sourced their first risc-v vector implementation (ocelot) on top of BOOM. It's in order and they used it as a reference for their later out of order commercial designs, but they also could've keep it to them self's.


JK got vested interested in such a wishful thinking statement. I take it as marketing and not engineering.


You can have a vested interest in something and still say the truth. Its not mutually exclusive.


Couldn't it be the other way around? That JK got into RISC-V because of its promising future?


In the overall market, probably, but in the data center?


Yes, and soon we will have PIC powered and AVR powered servers in the data center, joining our ARM and RISC V servers.


Don't forget the 6502.xxlarge instance coming soon to AWS :) /s


I'm betting the shop on 68k.24xlarge


You can have whatever you want, via FPGA instances (like F1).


There are ARM and RISC-V servers. The ARM ones are even available in major clouds. Why are you comparing to microcontrollers?


> RISC-V servers

There are?


I'm counting the HiFive; I suppose it depends how you define your terms. The point is that RISC-V can scale down to compete with PIC/AVR, but it can also scale up to "real" compute so it's a bit nonsense to dismiss it as only being for embedded.


> I'm counting the HiFive

Isn't it years behind ARM/x86 performance wise?

> nonsense to dismiss it as only being for embedded

They unlikely to be competitive anywhere else without significant investment.

> but it can also scale up to "real" compute

Did anyone doubt that you could theoretically scale up ARM to be competitive with x86 back in ~2000? It still took them several decades to achieve that..


> but it can also scale up to "real" compute so it's a bit nonsense to dismiss it as only being for embedded.

Especially after being selected for EU super computer effort.


I'm still betting on Intel bringing back Itanium.

Cause why not.. having more architectures (for no apparent reasons) sounds cool.


Keep dreaming.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: