208
submitted 2 months ago by misk@sopuli.xyz to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] jqubed@lemmy.world 13 points 2 months ago

I don’t really know if ARM adds benefits I’d really notice as an end user, but it’ll be interesting to see if this really goes through and upends the dominant architecture we’ve seen for really 40+ years.

[-] SMillerNL@lemmy.world 42 points 2 months ago

As an ARM Mac user, I wouldn’t trade all this new battery life for an x86 processor

[-] aniki@lemm.ee 14 points 2 months ago* (last edited 2 months ago)

Second this. Not to mention INSTANT resume from hibernation! It's fucking crazy. I can use this thing ALL DAY doing webGL CAD work and Orca Slicer and barely scratch 50%.

[-] catloaf@lemm.ee 4 points 2 months ago

With a modern system, I honestly don't think there's a noticeable difference between suspend to ram and suspend to disk. They've gotten the boot times down so much that it's lightning-fast. My work laptop's default is suspend to disk, and I don't notice a difference except when it prompts for the bitlocker password.

S0 standby is borderline unusable on many PCs. On Apple silicon macs it's damn near flawless.

My current laptop is probably the last machine to support S3 standby and I do not look forward to replacing it and being forced back into a laptop that overheats and crashes in my backpack in less than 15 minutes. On my basic T14 it works ok for the most part, but my full fat Thinkpad P1 with an i9 is in S0 standby for longer than a few minutes, and sometimes uses more power than when it was fully on. Maybe Meteor lake with it's LP E cores will fix this but I doubt it.

[-] vox@sopuli.xyz 2 points 2 months ago* (last edited 2 months ago)

tbh it has been nearly flawless on win11 for me with an amd cpu

(just make sure to disable automatic windows/defender updates unless you want to get woken up by jet turbine sounds in the middle of the night)

[-] pycorax@lemmy.world 12 points 2 months ago

There's nothing stopping x86-64 processors from being power efficient. This article is pretty technical but does a really good explanation of why that's the case: https://chipsandcheese.com/2024/03/27/why-x86-doesnt-need-to-die/

It's just that traditionally Intel and AMD earn most of their money from the server and enterprise sectors where high performance is more important than super low power usage. And even with that, AMD's Z1 Extreme also gets within striking distance of the M3 at a similar power draw. It also helps that Apple is generally one node ahead.

[-] SquiffSquiff@lemmy.world 5 points 2 months ago

If there's 'nothing stopping' it then why has nobody done it? Apple moved from x86 to ARM. Mobile is all ARM. All the big cloud providers are doing their own ARM chips. Intel killed off much of the architectural competition with Itanic in the early 2000's. Why stop?

[-] pycorax@lemmy.world 1 points 2 months ago

Their primary money makers are what's stopping them I reckon. Apple's move to ARM is because they already had a ton of experience with building their own in house processors for their mobile devices and ARM licenses stock chip designs, making it easier for other companies to come up with their own custom chips whereas there really isn't any equivalent for x86-64. There were some disagreements between Intel and AMD over patents on the x86 instruction set too.

[-] QuaternionsRock@lemmy.world 1 points 2 months ago

This article fails to mention the single biggest differentiator between x86 and ARM: their memory models. Considering the sheer amount of everyday software that is going multithreaded, this is a huge issue, and the reason why ARM drastically outperforms x86 running software like modern web browsers.

[-] pycorax@lemmy.world 2 points 2 months ago

Do you mind elaborating what is it about the difference on their memory models that makes a difference?

[-] QuaternionsRock@lemmy.world 3 points 2 months ago

Here is a great article on the topic. Basically, x86 spends a comparatively enormous amount of energy ensuring that its strong memory guarantees are not violated, even in cases where such violations would not affect program behavior. As it turns out, the majority of modern multithreaded programs only occasionally rely on these guarantees, and including special (expensive) instructions to provide these guarantees when necessary is still beneficial for performance/efficiency in the long run.

For additional context, the special sauce behind Apple’s Rosetta 2 is that the M family of SoCs actually implement an x86 memory model mode that is selectively enabled when executing dynamically translated multithreaded x86 programs.

[-] pycorax@lemmy.world 1 points 2 months ago

Thanks for the links, they're really informative. That said, it doesn't seem to be entirely certain that the extra work done by the x86 arch would incur a comparatively huge difference in energy consumption. Granted, that isn't really the point of the article. I would love to hear from someone who's more well versed in CPU design on the impact of it's memory model. The paper is more interesting with regards to performance but I don't find it very conclusive since it's comparing ARM vs TSO on an ARM processor. It does link this paper which seems more relevant to our discussion but a shame that it's paywalled.

[-] sunbeam60@lemmy.one 2 points 2 months ago

On the x86 architecture, RAM is used by the CPU and the GPU has a huge penalty when accessing main RAM. It therefore has onboard graphics memory.

On ARM this is unified so GPU and CPU can both access the same memory, at the same penalty. This means a huge class of embarrassingly parallel problems can be solved quicker on this architecture.

[-] pycorax@lemmy.world 1 points 2 months ago

Do x86 CPUs with iGPUs not already use unified memory? I'm not exactly sure what you mean but are you referring to the overhead of having to do data copying over from CPU to GPU memory on discrete graphics cards when performing GPU calculations?

[-] sunbeam60@lemmy.one 1 points 2 months ago

Yes unified and extremely slow compared to an ARM architecture’s unified memory, as the GPU sort of acts as if it was discrete.

[-] pycorax@lemmy.world 1 points 2 months ago

Do you have any sources for this? Can't seem to find anything specific describing the behaviour. It's quite surprising to me since the Xbox and PS5 uses unified memory on x86-64 and would be strange if it is extremely slow for such a use case.

[-] sunbeam60@lemmy.one 1 points 2 months ago

It’s been a while since I’ve coded on the Xbox, but at least in the 360, the memory wasn’t really unified as such. You had 10 MB of EDRAM that formed your render target and then there was specialised functions to copy the EDRAM output to DRAM. So it was still separated and while you could create buffers in main memory that you access in the shaders, at some penalty.

It’s not that unified memory can’t be created, but it’s not the architecture of a PC, where peripheral cards communicate over the PCI bus, with great penalties to touch RAM.

[-] pycorax@lemmy.world 1 points 2 months ago

Well for the current generation consoles they're both x86-64 CPUs with only a single set of GDDR6 memory shared across the CPU and GPU so I'm not sure if you have such a penalty anymore

It’s not that unified memory can’t be created, but it’s not the architecture of a PC, where peripheral cards communicate over the PCI bus, with great penalties to touch RAM.

Are there any tests showing the difference in memory access of x86-64 CPUs with iGPUs compared to ARM chips?

[-] QuaternionsRock@lemmy.world 1 points 2 months ago* (last edited 2 months ago)

That’s actually not what I was referring to, although the unified memory architecture is certainly more power efficient for mixed-intensive workloads. The cost of transferring to/from dedicated GPU memory is (unsurprisingly) quite large.

[-] SMillerNL@lemmy.world 1 points 2 months ago

So there is something stopping them. The manufacturers.

load more comments (32 replies)
this post was submitted on 20 May 2024
208 points (94.4% liked)

Technology

57175 readers
3843 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS