this post was submitted on 22 Sep 2023
9 points (54.6% liked)

Technology

35127 readers
201 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
 

tr:dr; he says "x86 took over the server market" because it was the same architecture developers in companies had on their machines thus it made it very easy to develop applications on their machines to then ship to the servers.

Now this, among others he made, are very good points on how and why it is hard for ARM to get mainstream on the datacenter, however I also feel like he kind lost touch with reality on this one...

He's comparing two very different situations, more specifically eras. Developers aren't so tied anymore like they used to be to the underlaying hardware. The software development market evolved from C to very high language languages such as Javascript/Typescript and the majority of stuff developed is done or will be done in those languages thus the CPU architecture becomes irrelevant.

Obviously very big companies such as Google, Microsoft and Amazon are more than happy to pay the little "tax" to ensure Javascript runs fine on ARM than to pay the big bucks they pay for x86..

What are your thoughts?

all 41 comments
sorted by: hot top controversial new old
[–] [email protected] 53 points 1 year ago* (last edited 1 year ago) (3 children)

He has a strong opinion, but he hasn't lost the plot. It's very reasonable to say you need to develop on the architecture you wanted to deploy to. If you want to be efficient, so most companies are going to deploy to architecture they have locally.

But you're taking comments from 2019. Nowadays lots of Mac developers develop directly on arm. So by his own argument, those Mac developers would be more comfortable deploying to an arm-based architecture cuz the running on an arm-based architecture.

So broadly I agree with him, or his past comments from 2019, you're going to need local developer environments, before you're going to get efficient server software

[–] [email protected] 8 points 1 year ago* (last edited 1 year ago)

ARM on Mac isn't nearly as helpful for workloads on an ARM server as x86 PC for an x86 server. The differences in hardware behavior between the two x86 parts is small because the platforms are standardized way beyond the instruction set. The ARM server on the other hand has nothing to do with the Mac beyond the instruction set. Something runs great on your Mac because of the on-SoC ridiculously fast RAM. You throw it on an ARM server with completely different ARM CPUs, slotted RAM and a bottleneck shows up.

[–] [email protected] 6 points 1 year ago

I hope we get there soon with RISC-V.

[–] [email protected] 2 points 1 year ago (3 children)

I hate my M2 Mac because I hate Macs and Docker doesn't always work correctly.

[–] [email protected] 1 points 1 year ago

I'd be surprised if Docker worked at all on an M2, because it doesn't work worth a shit on an x86 Mac.

[–] [email protected] 1 points 1 year ago (1 children)

If you run an ARM system inside docker, it works much better!

Many pre-baked images may be x86 only. However, thanks to M processors there’s a real demand for more than Raspberry Pi, so this will get better too.

[–] [email protected] 1 points 1 year ago (1 children)

Unfortunately I was trying to build WebRTC, which is supported on Linux only.

[–] [email protected] 1 points 1 year ago (1 children)

There’s aarch64 version of Linux.

[–] [email protected] 1 points 1 year ago

Not all the dependencies are supported on aarch64 unfortunately.

[–] [email protected] 0 points 1 year ago (1 children)
[–] [email protected] 1 points 1 year ago

Have job

Get paid to suffer

[–] [email protected] 34 points 1 year ago (1 children)

As someone dealing with enterprise software for living, what he's saying absolutely makes sense, and I deal mostly in web applications (where I never really have to worry about the low level stuff).

Just because the top layer seems to be the same, doesn't mean the underlying ones are. There's a reason why perfect bug compatibility is a thing (or maybe, was, in RHEL ecosystem?).

Things that looks like slam dunks in theories are never such in practice. Weird bugs pop up from time to time; and believe me, they will!

It might be rare, you may only see it once or twice in a project; but when it happens, you're gonna want to be ready, or people will question your ability to do your job.

[–] [email protected] 3 points 1 year ago (1 children)

The cross-compiling point makes sense but, since this is a 4.5 year old message, the state of ARM in the cloud has changed. Now developers do actually have ARM-based machines because of Apple. AWS has Graviton2 instances now and they are a lot cheaper than similarly specced x86_64 instances. ARM is a viable consideration that can be made.

[–] [email protected] 2 points 1 year ago

While its true that having ARM ecosystem is more feasible now, there's not many companies that's willing to equip their whole team with very specific model of laptop, with almost no servicable parts for no perceivable benefit. No, Pinebooks as well as Raspberry Pi laptops and cyberdecks are not feasible for industry.

Most companies are not looking for gimmicks for work, even when they make some for living; so no, looking cool is not a benefit that defeats all that cost.

Meanwhile, most people in the industry, such as myself, and my current bosses & colleagues, and my previous bosses & colleagues, and probably all my future bosses & colleagues are fine running x86 for production servers. It got everything we'd need, including upgradable RAM and decades worth of collective experience, which I cannot say ARM has.

At the same time, I have some hope for RISC-V. It won't take over the industry anytime soon, but it's been showing some promise for long term.

[–] [email protected] 21 points 1 year ago

I’ve got an ARM Mac. I’ve got ARM VPSes from Hetzner, and I’m compiling native code for the server.

It’s definitely easier to develop, build, and test on the same architecture, than to deal with cross-compilation and emulation.

So I think Linus is right.

[–] [email protected] 20 points 1 year ago (2 children)

He is sort of right, back in 2019. Even then, IBM PowerPC mainframe are still thriving.

Now, new language with cross compilation with some maturity are here. Major cloud providers now have ARM base machines ready, even designing to their own need.

ARM is in the datacenter market and become a trend.

The only thing I worried about, is the architecture of ARM are too fractured. AWS Graviton might behave differently than Ampere Altra, despite both have the ARM ISA.

[–] [email protected] 3 points 1 year ago

Different cores, different topologies, different interconnects, different memory throughputs... fahgedabouddit.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (1 children)

[This comment has been deleted by an automated system]

[–] [email protected] 1 points 1 year ago (1 children)

With x86, there are AMD and Intel. With ARM, how many designers are here? With more designers, the smaller the potential common ground is, and more code paths to optimize, thus cost more to build.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago)

[This comment has been deleted by an automated system]

[–] [email protected] 19 points 1 year ago (5 children)

The luxuries you have to not know a thing about enterprise grade servers because your world is JavaScript was made possible, and continues to be made possible, by people working on layers that do require familiarity with the underlying hardware.

[–] [email protected] 4 points 1 year ago* (last edited 1 year ago)

Right, whenever someone like Linus talks about developers he's probably not referring to your run-of-the-mill code monkey making simple web apps.

[–] [email protected] 3 points 1 year ago

And that underlying stuff doesn't run the same on x86 and dog knows who's ARM implementation.

[–] [email protected] 2 points 1 year ago* (last edited 1 year ago) (1 children)

[This comment has been deleted by an automated system]

[–] [email protected] 1 points 1 year ago

I think people underestimate the challenges involved when building software systems tightly coupled to the underlying hardware (like if you are a team tasked with building a next gen server).

Successful companies in the space don't underestimate it though, the engineers who do the work don't underestimate it, and Linus doesn't underestimate it either.

The domain knowledge in your org required to mitigate the business risk isn't trivial. The value proposition always needs to be pretty juicy to overcome the inertia present caused by institutional familiarity. Like, can we save a few million on silicon? Sure. Do we think we understand the challenges well enough to keep our hardware release schedules without taking shortcuts that will result in reputational impact? Do we think we have the right people in place to oversee the switch?

Over and over again, it comes back to "is it worth it", and it's much more complex of a question to offer than just picking the cheaper chips.

I imagine at this point there is probably a metric fuckton of enterprise software what strictly dictate that it must be run on X86. Even if it doesn't have to. If you stray from the vendor hardware requirements, bullshit or not, you'll lose your support. There is likely friction on some consumer segments as well on the uptake.

[–] [email protected] 2 points 1 year ago

Yeah but you have to write Javascript. :-D

[–] [email protected] 0 points 1 year ago* (last edited 1 year ago)

The luxuries you have to not know a thing about enterprise grade servers because your world is JavaScript was made possible, (...) by people working on layers that do require familiarity with the underlying hardware.

That's kind my point... Since everyone is or will be coding on Javascript (or other languages that run on virtual machines / "layers") general developers won't have a problem running on ARM datacenters anymore. Big cloud providers will take the opportunity to move to ARM as it is cheaper for them.

And btw, the people making JS fast and stable on ARM are, most likely, not that familiar with server grade hardware. They're optimizing for phones and whatever where ARM was born.

[–] [email protected] 18 points 1 year ago* (last edited 1 year ago) (1 children)

JavaScript and TS are script languages with little to nothing to do with threading

[–] [email protected] 1 points 1 year ago (1 children)

It can do multithreading using worker threads, IIRC.

[–] [email protected] 1 points 1 year ago

TS and JS are completely single threaded unless you have multiple instances of node running at the same time.

[–] [email protected] 16 points 1 year ago

X86 and AMD64 based stuff is fairly standard in terms of a motherboard with a BIOS/UEFI and peripheral busses. ARM has for a long time been kind of a mess in this regard, and there are several varieties of ARM architecture that don't play nicely with code compiled for others.

Don't get me wrong. ARM can be great for certain types of workloads. It's typically more efficient at lower power than X86, and better at various types of math. That's why we DO see it available on ARM for certain stuff like Lambda functions, but you probably won't be running full VM environments on it.

Last: notice how it's been hard to find certain varieties of Pi and various other stuff running ARM? There's shortages all over the place but I'm general Intel and AMD have been able to apply demand for their CPU's.

Yes, devs aren't tied to hardware, but there are efficiencies of scale to consider

[–] [email protected] 14 points 1 year ago (1 children)

It's tough to debug issues when you can't run on the same hardware directly.

There's a reason that arm support in open source software has exploded in the past few years, and it's because of apple silicon.

I'll agree that it's easier now, with most developers using higher level runtimes, but someone's got to get those runtimes working, and it's much easier to develop if you have a laptop running that hardware.

[–] [email protected] 21 points 1 year ago

Raspberry Pi also brought arm64 to a lot more people.

[–] [email protected] 11 points 1 year ago

The linked message is from 2019, i.e. per-M1 Apple laptops and at a time when arm in datacenter was just starting out.

Tbh, I feel like it's kinda pointless to discuss a comment made by someone over 4-years ago. Both the environment and the person itself can change a lot in that time.

[–] [email protected] 7 points 1 year ago* (last edited 1 year ago) (1 children)

"The software development market evolved from C to very high language languages such as Javascript/Typescript and the majority of stuff developed is done or will be done in those languages thus the CPU architecture becomes irrelevant."

I saw someone else make a similar comment about C. People track these things, and C has been in the top 2 most widely used languages for more than 2 decades. Not knowing this should probably make you wonder why your background has resulted in such a narrow experience.

https://en.m.wikipedia.org/wiki/TIOBE_index#

[–] [email protected] -4 points 1 year ago* (last edited 1 year ago) (1 children)

Look, I'm not saying C is important nor that people aren't using it but... Let me ask you one thing, if you look at the majority of the web (not specific cases) you'll find that 76% of it is PHP. Furthermore if you think that everyone is moving to mobile apps you'll get a mix of Java/Kotlin, Swift and a very strong move to towards cross-platform stuff that is, in most cases, based on Javascript. To make things worse bootcamps for wanna be devs have been teaching node as a valid backend solution for quite a while now. We see startups going that route and things going perfectly well.

Since we've that huge market for higher level that run perfectly well on ARM do you really thing that stuff made in C really dictates the future of the market? The "issue" I see with the link you've provided is simple: nobody is developing "run of the mill" solutions with C anymore like we used to and those are the solutions that have the numbers to move the market. Nowadays C is operating systems, libraries for higher level languages, engines such as the JS V8, a ton of IoT devices (that ironically are ARM), low level electronics, industrial automation and financial use cases where performance is really important.

C is going to stay on specific places but nobody develops websites, desktop and mobile applications with hence my simplistic "the software development market evolved from C to very high language languages such as Javascript/Typescript" conclusion.

The market is moved by the large masses and the masses use technologies that are not bound anymore to architectures like other used to be.

[–] [email protected] 5 points 1 year ago* (last edited 1 year ago)

It's odd that you're saying you shouldn't consider the specific cases where C excels and then narrowing down things to the Web, where languages like php excel. So now you probably have some idea why your experience is so narrow. There's a lot more to programming than the Web, and there's always going to be.

[–] [email protected] 5 points 1 year ago* (last edited 1 year ago) (1 children)

From what I learned at university:
CISC instruction set (x86) was developed to adress the technical reality of its time - time costly CPU operation and fast read from storage. Not long after that the situation has changed - storage reads became slower in comparison to computing time (putting it simply it's faster to read an archive and unpack it than to read unpacked thing). But in the meantime the PC boom has happened. In a way backward compatibility and market inertia locked us with instruction set that is not the best optimised for our tech, despite the fact that RISC (for example ARM) was conceived earlier.

In a way software (compilers and interpreters too) is like a muscle. The more/wider it's used, the better it becomes. You can be writing in python but if your interpreter has some missed optimization opportunities, your code will be running faster on architecture with a better optimized interpreter available.

From personal observations:
The biggest cost of software is not to write something super efficient. It's maintainability (readability and debugging), ease of use (onboarding/training time) and versatility ("let's add the feature that is missing to what we have, instead of reinventing the wheel and maintaining two toolsets").

The new languages are not created because they can do something faster than assembler (they can't, btw). If assembly code is written as optimal as possible, high level languages can at best be as fast. Writing such assembly is a problem behind the keyboard, not a technical limitation. The only thing high-level languages do better is how much time it takes a human to work with it.
I would not be surprised to learn that bigger part of these big bucks you mention go not into optimization but rather into "how can we work around that difference so the high-level interface stays the same as for more widely used x86?"

In the end it all boils down to machine code - it's the only thing that really exists when it comes to executing code. If your "human to bits translator" produces unoptimized binaries, it doesn't matter how high-level your code was written in.
And sometime in the meantime we've arrived at a level when even a few behemoths like Google or Microsoft throwing money into research (not that I believe they are doing so when it comes to optimization) is enough.
It's the field use that from time to time provides a use-case that helps finding edge-case where optimization can be made.
To purposefully find it? Dumping your datacenter in liquid nitrogen might be cheaper and probably more predictable.

So yeah, I mostly agree with him.
Maybe the times have changed a little, the thing that gave RISCs the most kick were smartphones, then one board computers, so not long ago. The improvements are always bigger at the beginning.
But the fact that some companies are trying to get RISC back into userland in my opinion means that the computer world has only started to heal itself after the effects of PC boom. There's around 20 year difference where x86 was the main thing and RISC was a niche

[–] [email protected] 2 points 1 year ago* (last edited 1 year ago)

[This comment has been deleted by an automated system]