20 % of shit is still shit.
Luckily my Intel GPU is lacking so many features that it barely has to do anything at all.
All things related to technology hardware, with a focus on computing hardware.
Rules (Click to Expand):
Follow the Lemmy.world Rules - https://mastodon.world/about
Be kind. No bullying, harassment, racism, sexism etc. against other users.
No Spam, illegal content, or NSFW content.
Please stay on topic, adjacent topics (e.g. software) are fine if they are strongly relevant to technology hardware. Another example would be business news for hardware-focused companies.
Please try and post original sources when possible (as opposed to summaries).
If posting an archived version of the article, please include a URL link to the original article in the body of the post.
Some other hardware communities across Lemmy:
Icon by "icon lauk" under CC BY 3.0
20 % of shit is still shit.
Luckily my Intel GPU is lacking so many features that it barely has to do anything at all.
"After discussion between Intel and Canonical’s security teams, we are in agreement that Spectre no longer needs to be mitigated for the GPU at the Compute Runtime level. At this point, Spectre has been mitigated in the kernel, and a clear warning from the Compute Runtime build serves as a notification for those running modified kernels without those patches. For these reasons, we feel that Spectre mitigations in Compute Runtime no longer offer enough security impact to justify the current performance tradeoff."
Seems like the Spectre mitigations were applied for GPU compute as well.
1 to 5 percentage points is not great, but in and around the margin of error. 20% is a huge performance delta.