390
you are viewing a single comment's thread
view the rest of the comments
[-] Alphane_Moon@lemmy.world 16 points 1 month ago

They honestly seem to be done with high-end "enthusiast" GPUs. There is probably more money/potential for iGPUs and low/middle level products optimized for laptops.

[-] Technus@lemmy.zip 14 points 1 month ago

Their last few generations of flagship GPUs have been pretty underwhelming but at least they existed. I'd been hoping for a while that they'd actually come up with something to give Nvidia's xx80 Ti/xx90 a run for their money. I wasn't really interested in switching teams just to be capped at the equivalent performance of a xx70 for $100-200 more.

[-] TheGrandNagus@lemmy.world 29 points 1 month ago

The 6900XT/6950XT were great.

They briefly beat Nvidia until Nvidia came out with the 3090 Ti. Even then, it was so close you couldn't tell them apart with the naked eye.

Both the 6000 and 7000 series have had cards that compete with the 80-class cards, too.

The reality is that people just buy Nvidia no matter what. Even the disastrous GTX 480 outsold ATI/AMD's cards in most markets.

The $500 R9 290X was faster than the $1000 Titan, with the R9 290 being just 5% slower and $400, and yet AMD lost a huge amount of money on it.

AMD has literally made cards faster than Nvidia's for half the price and lost money on them.

It's simply not viable for AMD to spend a fortune creating a top-tier GPU only to have it not sell well because Nvidia's mindshare is arguably even better than Apple's.

Nvidia's market share is over 80%. And it's not because their cards are the rational choice at the price points most consumers are buying at. It really cannot be stressed enough how much of a marketing win Nvidia is.

[-] sugar_in_your_tea@sh.itjust.works 11 points 1 month ago* (last edited 1 month ago)

Yup, it's the classic name-brand tax. That, and Nvidia also wins on features, like RTX and AI/compute.

But most people don't actually use those features, so most people seem to be buying Nvidia due to brand recognition. AMD has dethroned Intel on performance and price, yet somehow Intel remains dominant on consumer PCs, though the lead is a lot smaller than before.

If AMD wants to take over Nvidia, they'll need consistently faster GPUs and lower prices with no compromises on features. They'd have to invest a ton to get there, and even then Nvidia would probably sell better than AMD on name recognition alone. Screw that! It makes far more sense for them to stay competitive and suck up a bunch of the mid-range market and transition the low-end market to APUs. Intel can play at the low-mid range markets, and AMD will slot themselves as a bit better than Intel, and a better value than Nvidia.

That said, I think AMD needs to go harder on the datacenter for compute, because that's where the real money is, and it's all going to Nvidia. If they can leverage their processors to provide a better overall solution for datacenter compute, they could translate that into prosumer compute devices. High end gaming is cool, but it's not nearly as lucrative as datacenter. I would hesitate to make AI-specific chips, but instead make high quality general compute chips so they can take advantage of whatever comes after the current wave of AI.

I think AMD should also get back into ARM and low-power devices. The snapdragon laptops have made a big splash, and that market could explode once the software is refined, and AMD should be poised to dominate it. They already have ARM products, they just need to make low-power, high performance products for the laptop market.

[-] pycorax@lemmy.world 3 points 1 month ago

I think AMD should also get back into ARM and low-power devices. The snapdragon laptops have made a big splash, and that market could explode once the software is refined, and AMD should be poised to dominate it. They already have ARM products, they just need to make low-power, high performance products for the laptop market.

They don't need to go with ARM. There's nothing inherently wrong with the x86 instruction set that prevents them from making low power processors, it's just that it doesn't make sense for them to build an architecture for that market since the margins for servers are much higher. Even then, the Z1 Extreme got pretty close to Apple's M2 processors.

Lunar Lake has also shown that x86 can match or beat Qualcomm's ARM chips while maintaining full compatibility with all x86 applications.

[-] sugar_in_your_tea@sh.itjust.works 1 points 1 month ago* (last edited 1 month ago)

it’s just that it doesn’t make sense for them to build an architecture for that market since the margins for servers are much higher

Hence ARM. ARM already has designs for low power, high performance chips for smaller devices like laptops. Intel is chasing that market, and AMD could easily get a foot in the door by slapping their label on some designs, perhaps with a few tweaks (might be cool to integrate their graphics cores?). They already have ARM cores for datacenter workloads, so it probably wouldn't be too crazy to try it out on business laptops.

The 6000 series from AMD were so great because they picked the correct process node. Nvidia went with the far inferior Samsung 8nm node over TSMCs 7. Yet Nvidia still kept up with AMD in most areas (ignoring ray tracing).

Even the disastrous GTX 480 outsold ATI/AMD's cards in most markets.

The “disastrous” Fermi cards were also compute monsters. Even after the 600 series came out people were buying the 500 series over them because they performed so much better for the money. Instead of picking up a Kepler Quadra card in order to get double precision you could get a regular ass GTX 580 and do the same thing.

[-] pycorax@lemmy.world 13 points 1 month ago

Were the 6000 series not competitive? I got a 6950 XT for less than half the price of the equivalent 3090. It's an amazing card.

[-] vithigar@lemmy.ca 9 points 1 month ago

Yes, they were, and that highlights the problem really. Nvidia's grip on mind share is so strong that AMD releasing cards that matched or exceeded at the top end didn't actually matter and you still have people saying things like the comment you responded to.

It's actually incredible how quickly the discourse shifted from ray tracing being a performance hogging gimmick and DLSS being a crutch to them suddenly being important as soon as AMD had cards that could beat Nvidia's raster performance.

[-] JaY_III@lemmy.ca 2 points 1 month ago

The 6000 series is faster in Raster but slower in Ray Tracing.

Reviews have been primarily pushing cards based on RT since it has become available. nVidia has a much larger marketing budget than AMD, and ever since they have been able to leverage the fact they have the fastest Ray Tracing, AMD share has been noise diving.

[-] pycorax@lemmy.world 4 points 1 month ago

I mean I guess? But the question here was about value and no way is RT worth double the price.

[-] sugar_in_your_tea@sh.itjust.works 3 points 1 month ago

It is if that's the main thing you care about.

[-] pycorax@lemmy.world 3 points 1 month ago

Wouldn't be the first time they did this though, I wouldn't be surprised if they jump back into the high end once they're ready.

this post was submitted on 19 Oct 2024
390 points (99.0% liked)

Technology

59681 readers
4879 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS