Most active commenters
  • senko(3)
  • npteljes(3)
  • tankenmate(3)
  • homebrewer(3)
  • MegaDeKay(3)

←back to thread

Nvidia won, we all lost

(blog.sebin-nyshkim.net)
977 points todsacerdoti | 53 comments | | HN request time: 0.822s | source | bottom
Show context
__turbobrew__ ◴[] No.44468824[source]
> With over 90% of the PC market running on NVIDIA tech, they’re the clear winner of the GPU race. The losers are every single one of us.

I have been rocking AMD GPU ever since the drivers were upstreamed into the linux kernel. No regrets.

I have also realized that there is a lot out there in the world besides video games, and getting all in a huff about it isn’t worth my time or energy. But consumer gotta consoooooom and then cry and outrage when they are exploited instead of just walking away and doing something else.

Same with magic the gathering, the game went to shit and so many people got outraged and in a big huff but they still spend thousands on the hobby. I just stopped playing mtg.

replies(22): >>44468885 #>>44468985 #>>44469036 #>>44469146 #>>44469164 #>>44470357 #>>44470480 #>>44470607 #>>44471458 #>>44471685 #>>44471784 #>>44471811 #>>44472146 #>>44472400 #>>44473527 #>>44473828 #>>44473856 #>>44476633 #>>44485501 #>>44487391 #>>44489487 #>>44493815 #
1. darkoob12 ◴[] No.44470357[source]
I am not a gamer and don't why AMD GPUs aren't good enough. It's weird since both Xbox and PlayStation are using AMD GPUs.

I guess there games that you can only play on PC with Nvidia graphics. That begs the question why someone create a game and ignore large console market.

replies(9): >>44470371 #>>44470393 #>>44470614 #>>44470709 #>>44471156 #>>44472918 #>>44473655 #>>44476308 #>>44479276 #
2. PoshBreeze ◴[] No.44470371[source]
Nvidia is the high end, AMD is the mid segment and Intel is the low end. In reality I am playing 4K on HellDivers with 50-60FPS on a 6800XT.

Traditionally the NVIDIA drivers have been more stable on Windows than the AMD drivers. I choose an AMD card because I wanted a hassle free experience on Linux (well as much as you can).

3. ErrorNoBrain ◴[] No.44470393[source]
ive used an amd card for a couple years

its been great. flawless in fact.

replies(1): >>44472615 #
4. senko ◴[] No.44470614[source]
> AMD GPUs aren't good enough.

Software. AMD has traditionally been really bad at their drivers. (They also missed the AI train and are trying to catch up).

I use Linux and have learned not to touch AMD GPUs (and to a lesser extent CPUs due to chipset quality/support) a long time ago. Even if they are better now, (I feel) Intel integrated (if no special GPU perf needed) or NVidia are less risky choices.

replies(3): >>44470640 #>>44470887 #>>44471900 #
5. rewgs ◴[] No.44470640[source]
> I use Linux and have learned not to touch AMD GPUs (and to a lesser extent CPUs due to chipset quality/support) a long time ago. Even if they are better now, (I feel) Intel integrated (if no special GPU perf needed) or NVidia are less risky choices.

Err, what? While you're right about Intel integrated GPUs being a safe choice, AMD has long since been the GPU of choice for Linux -- it just works. Whereas Nvidia on Linux has been flaky for as long as I can remember.

replies(3): >>44470651 #>>44470851 #>>44470902 #
6. michaelmrose ◴[] No.44470651{3}[source]
They have never been flaky on the x11 desktop
7. datagram ◴[] No.44470709[source]
AMD cards are fine from a raw performance perspective, but Nvidia has built themselves a moat of software/hardware features like ray-tracing, video encoding, CUDA, DLSS, etc where AMD's equivalents have simply not been as good.

With their current generation of cards AMD has caught up on all of those things except CUDA, and Intel is in a similar spot now that they've had time to improve their drivers, so it's pretty easy now to buy a non-Nvidia card without feeling like you're giving anything up.

replies(2): >>44470751 #>>44470848 #
8. SSLy ◴[] No.44470751[source]
AMD RT is still slower than Nvidia's.
9. jezze ◴[] No.44470848[source]
I have no experience of using it so I might be wrong but AMD has ROCm which has something called HIP that should be comparable to CUDA. I think it also has a way to automatically translate CUDA calls into HIP as well so it should work without the need to modify your code.
replies(4): >>44470886 #>>44471442 #>>44472280 #>>44472309 #
10. senko ◴[] No.44470851{3}[source]
Had major problems with xinerama, suspend/resume, vsync, probably a bunch of other stuff.

That said, I've been avoiding AMD in general for so long the ecosystem might have really improved in the meantime, as there was no incentive for me to try and switch.

Recently I've been dabbling in AI where AMD GPUs (well, sw ecosystem, really) are lagging behind. Just wasn't worth the hassle.

NVidia hw, once I set it up (which may be a bit involved), has been pretty stable for me.

replies(2): >>44471472 #>>44472519 #
11. StochasticLi ◴[] No.44470886{3}[source]
it's mostly about AI training at this point. the software for this only supports CUDA well.
12. ho_schi ◴[] No.44470887[source]
This is wrong. For 14 years the recommendation on Linux is:

    * Purchase always AMD.      
    * Purchase never Nvidia.
    * Intel is also okay.
Because the AMD drivers are good and open-source. And AMD cares about bug reports. The one from Nvidia can and will create issues because they’re closed-source and avoided for years to support Wayland. Now Nvidia published source-code and refuses to merge it into Linux and Mesa facepalm

While Nvidia comes up with proprietary stuff AMD brought us Vulkan, FreeSync, supported Wayland well already with Implicit-Sync (like Intel) and used the regular Video-Acceleration APIs for long time.

Meanwhile Nvidia:

https://registry.khronos.org/OpenGL/extensions/NV/NV_robustn...

    It’s not a bug, it’s a feature!

Their bad drivers still don’t handle simple actions like a VT-Switch or Suspend/Resume. If a developer doesn’t know about that extension the users suffer for years.

Okay. But that is probably only a short term solution? It is Nvidias short term solution since 2016!

https://www.phoronix.com/news/NVIDIA-Ubuntu-2025-SnR

replies(3): >>44471817 #>>44472122 #>>44472448 #
13. simion314 ◴[] No.44470902{3}[source]
>Err, what? While you're right about Intel integrated GPUs being a safe choice, AMD has long since been the GPU of choice for Linux -- it just works. Whereas Nvidia on Linux has been flaky for as long as I can remember.

Not OP, I had same experience in the past with AMD,I bought a new laptop and in 6 months the AMD decided that my card is obsolete and no longer provided drivers forcing me to be stuck with older kernel/X11 , so I switched to NVIDIA and after 2 PC changes I still use NVIDIA since the official drivers work great, I really hope AMD this time is putting the effort to keep older generations of cards working on latest kernels/X11 maybe next card will be AMD.

But this is an explanations why us some older Linux users have bad memories with AMD and we had good reason to switch over to NVIDIA and no good reason to switch back to AMD

14. npteljes ◴[] No.44471156[source]
What I experienced is that AI is a nightmare on AMD in Linux. There is a myriad of custom things that one needs to do, and even that just breaks after a while. Happened so much on my current setup (6600 XT) that I don't bother with local AI anymore, because the time investment is just not worth it.

It's not that I can't live like this, I still have the same card, but if I were looking to do anything AI locally with a new card, for sure it wouldn't be an AMD one.

replies(3): >>44471354 #>>44473307 #>>44473336 #
15. eden-u4 ◴[] No.44471354[source]
I don't have much experience with ROCm for large trainings, but NVIDIA is still shit with driver+cuda version+other things. The only simplification is due to ubuntu and other distros that already do the heavy lift by installing all required components, without much configuration.
replies(2): >>44471743 #>>44474569 #
16. tankenmate ◴[] No.44471442{3}[source]
`I think it also has a way to automatically translate CUDA calls`

I suspect the thing you're referring to is ZLUDA[0], it allows you to run CUDA code on a range of non NVidia hardware (for some value of "run").

[0] https://github.com/vosen/ZLUDA

replies(1): >>44473140 #
17. tankenmate ◴[] No.44471472{4}[source]
I run llama.cpp using Vulkan and AMD CPUs, no need to install any drivers (or management software for that matter, nor any need to taint the kernel meaning if I have an issue it's easy to get support). For example the other day when a Mesa update had an issue I had a fix in less than 36 hours (without any support contract or fees) and `apt-mark hold` did a perfect job until there was a fix. Performance for me is within a couple of % points, and with under-volting I get better joules per token.
18. npteljes ◴[] No.44471743{3}[source]
Oh I'm sure. The thing is that with AMD I have the same luxury, and the wretched thing still doesn't work, or has regressions.
19. josephg ◴[] No.44471817{3}[source]
I've been using a 4090 on my linux workstation for a few years now. Its mostly fine - with the occasional bad driver version randomly messing things up. I'm using linux mint. Mint uses X11, which, while silly, means suspend / resume works fine.

NVIDIA's drivers also recently completely changed how they worked. Hopefully that'll result in a lot of these long term issues getting fixed. As I understand it, the change is this: The nvidia drivers contain a huge amount of proprietary, closed source code. This code used to be shipped as a closed source binary blob which needed to run on your CPU. And that caused all sorts of problems - because its linux and you can't recompile their binary blob. Earlier this year, they moved all the secret, proprietary parts into a firmware image instead which runs on a coprocessor within the GPU itself. This then allowed them to - at last - opensource (most? all?) of their remaining linux driver code. And that means we can patch and change and recompile that part of the driver. And that should mean the wayland & kernel teams can start fixing these issues.

In theory, users shouldn't notice any changes at all. But I suspect all the nvidia driver problems people have been running into lately have been fallout from this change.

replies(2): >>44472157 #>>44472472 #
20. jorams ◴[] No.44471900[source]
> I use Linux and have learned not to touch AMD GPUs

The situation completely changed with the introduction of the AMDGPU drivers integrated into the kernel. This was like 10 years ago.

Before then the AMD driver situation on Linux was atrocious. The open source drivers performed so bad you'd get better performance out of Intel integrated graphics than an expensive AMD GPU, and their closed source drivers were so poorly updated you'd have to downgrade the entire world for the rest of your software to be compatible. At that time Nvidia was clearly ahead, even though the driver needs to be updated separately and they invented their own versions of some stuff.

With the introduction of AMDGPU and the years after that everything changed. AMD GPUs now worked great without any effort, while Nvidia's tendency to invent their own things really started grating. Much of the world started moving to Wayland, but Nvidia refused to support some important common standards. Those that really wanted their stuff to work on Nvidia had to introduce entirely separate code paths for it, while other parts of the landscape refused to do so. This started improving again a few years ago, but I'm not aware of the current state because I now only use Intel and AMD hardware.

replies(2): >>44472911 #>>44475472 #
21. quicksilver03 ◴[] No.44472122{3}[source]
The AMD drivers are open source, but they definitely are not good. Have a look at the Fedora discussion forums (for example https://discussion.fedoraproject.org/t/fedora-does-not-boot-... ) to see what happens about each month.

I have no NVIDIA hardware, but I understand that the drivers are even worse than AMD's.

Intel seems to be, at the moment, the least worse compromise between performance and stability,

replies(1): >>44473052 #
22. nirv ◴[] No.44472157{4}[source]
No browser on Linux supports any other backend for video acceleration except VAAPI, as far as I know. AMD and Intel use VAAPI, while Nvidia uses VDPAU, which is not supported anywhere. This single fact means that with Nvidia graphics cards on Linux, there isn't even such a simple and important feature for users as video decoding acceleration in the browser. Every silly YouTube video will use CPU (not iGPU, but CPU) to decode video, consuming resources and power.

Yes, there are translation layers[1] which you have to know about and understand how to install correctly, which partially solve the problem by translating from VAAPI to NVDEC, but this is certainly not for the average user.

Hopefully, in the future browsers will add support for the new Vulkan Video standard, but for now, unfortunately, one has to hardcode the browser launch parameters in order to use the integrated graphics chip's driver (custom XDG-application file for AMD APU in my case: ~/.local/share/applications/Firefox-amdgpu.desktop): `Exec=env LIBVA_DRIVER_NAME=radeonsi DRI_PRIME=0 MOZ_ENABLE_WAYLAND=1 __NV_PRIME_RENDER_OFFLOAD=0 __GLX_VENDOR_LIBRARY_NAME=radeons i /usr/bin/firefox-beta %u`.

[1] https://github.com/elFarto/nvidia-vaapi-driver/

replies(2): >>44472332 #>>44475494 #
23. Almondsetat ◴[] No.44472280{3}[source]
AMD "has" ROCm just like Intel "has" AVX-512
24. whatevaa ◴[] No.44472309{3}[source]
Consumer card ROCm support is straight up garbage. CUDA support project was also killed.

AMD doesn't care about consumers anymore either. All the money in AI.

replies(1): >>44472705 #
25. whatevaa ◴[] No.44472332{5}[source]
VAAPI support in browsers is also bad and oftenly requires some forcing.

On my Steam deck, I have to use vulkan. AV1 decoder is straight up buggy, have to disable it with config or extensions.

26. homebrewer ◴[] No.44472448{3}[source]
I have zero sympathy for Nvidia and haven't used their hardware for about two decades, but amdgpu is the sole reason I stick to linux-lts kernels. They introduce massive regressions into every mainline release, even if I delay kernel updates by several minor versions (to something like x.y.5), it's still often buggy and crashy.

They do care about but reports, and their drivers — when given time to stabilize — provide the best experience across all operating systems (easy updates, etc), but IME mainline kernels should be treated as alpha-to-beta material.

27. homebrewer ◴[] No.44472472{4}[source]
They opened a tiny kernel level sliver of their driver, everything else (including OpenGL stack et al) is and will still be closed.

Sadly, a couple of years ago someone seriously misunderstood the news about "open sourcing" their drivers and spread that misunderstanding widely; many people now think their whole driver stack is open, when in reality it's like 1% of the code — the barest minimum they could get away with (I'm excluding GSP code here).

The real FOSS driver is Nova, and it's driven by the community with zero help from Nvidia, as always.

replies(1): >>44479520 #
28. homebrewer ◴[] No.44472519{4}[source]
> I've been avoiding AMD in general

I have no opinion on GPUs (I don't play anything released later than about 2008), but Intel CPUs have had more problems over the last five years than AMD, including disabling the already limited support for AVX-512 after release and simply burning themselves to the ground to get an easy win in initial benchmarks.

I fear your perception of their products is seriously out of date.

replies(2): >>44472738 #>>44476819 #
29. sfn42 ◴[] No.44472615[source]
Same. Bought a 6950xt for like $800ish or something like that a few years ago and it's been perfect. Runs any game I want to play on ultra 1440p with good fps. No issues.

Maybe there's a difference for the people who buy the absolute top end cards but I don't. I look for best value and when I looked into it amd looked better to me. Also got an amd CPU which has aso been great.

30. MangoToupe ◴[] No.44472705{4}[source]
> AMD doesn't care about consumers anymore either. All the money in AI.

I mean, this also describes the quality of NVIDIA cards. And their drivers have been broken for the last two decades if you're not using windows.

31. senko ◴[] No.44472738{5}[source]
> I fear your perception of their products is seriously out of date.

How's the chipset+linux story these days? That was the main reason for not choosing AMD CPU for me the last few times I was in the market.

32. MegaDeKay ◴[] No.44472911{3}[source]
I use the amdgpu driver and my luck has not been as good as yours. Can't sleep my PC without having it wake up to fill my logs with spam [0] and eventually crash.

Then there is the (in)famous AMD reset bug that makes AMD a real headache to use with GPU passthrough. The card can't be properly reset when the VM shuts down so you have to reboot the PC to start the VM a second time. There are workarounds but they only work on some cards & scenarios [1] [2]. This problem goes back to around the 390 series cards so they've had forever to properly implement reset according to the pci spec but haven't. nvidia handles this flawlessly

[0] https://gitlab.freedesktop.org/drm/amd/-/issues/3911

[1] https://github.com/gnif/vendor-reset

[2] https://github.com/inga-lovinde/RadeonResetBugFix

replies(2): >>44474660 #>>44505976 #
33. Cthulhu_ ◴[] No.44472918[source]
AMD GPU's are fine, but nvidia's marketing (overt and covert / word-of-mouth) is better. "RTX On" is a meme where people get convinced the graphics are over 9000x "better"; it's a meaningless marketing expression but a naive generation of fairly new PC gamers are eating it up.

And... they don't need to. Most of the most played video games on PC are all years old [0]. They're online multiplayer games that are optimized for average spec computers (and mobile) to capture as big a chunk of the potential market as possible.

It's flexing for clout, nothing else to it. And yet, I can't say it's anything new, people have been bragging, boasting and comparing their graphics cards for decades.

[0] https://activeplayer.io/top-15-most-popular-pc-games-of-2022...

replies(1): >>44473432 #
34. roenxi ◴[] No.44473052{4}[source]
Although you get to set your own standards "A bug was discovered after upgrading software" isn't very illuminating vis a vis quality. That does happen from time to time in most software.

In my experience an AMD card on linux is a great experience unless you want to do something AI related, in which case there will be random kernel panics (which, in all fairness, may one day go away - then I'll be back on AMD cards because their software support on Linux was otherwise much better than Nvidia's). There might be some kernel upgrades that should be skipped, but using an older kernel is no problem.

35. smallmancontrov ◴[] No.44473140{4}[source]
For an extremely flexible value of "run" that you would be extremely unwise to allow anywhere near a project whose success you have a stake in.
replies(1): >>44473933 #
36. phronimos ◴[] No.44473307[source]
Are you referring to AI training, prediction/inference, or both? Could you give some examples for what had to be done and why? Thanks in advance.
replies(1): >>44473522 #
37. FredPret ◴[] No.44473336[source]
I set up a deep learning station probably 5-10 years ago and ran into the exact same issue. After a week of pulling out my hair, I just bought an Nvidia card.
38. keyringlight ◴[] No.44473432[source]
One thing I wonder about is whether PC gaming is splitting into two distinct tiers, high end for those with thousands to spend on their rig and studios who are pathfinders (id, Remedy, 4A, etc) in graphics, then the wider market for cheaper/older systems and studios going for broad appeal. I know the market isn't going to be neatly divided and more of a blurry ugly continuum.

The past few years (2018 with the introduction of RT and upscaling reconstruction seems as good a milestone as any) feel like a transition period we're not out of yet, similar to the tail end of the DX9/Playstation3/Xbox360 era when some studios were moving to 64bit and DX11 as optional modes, almost like PC was their prototyping platform for when they made completed the jump with PS4/Xbox one and more mature PC implementations. It wouldn't surprise me if it takes more years and titles built targeting the next generation consoles before it's all settled.

replies(1): >>44476131 #
39. npteljes ◴[] No.44473522{3}[source]
Sure! I'm referring to setting up a1111's stable diffusion webui, and setting up Open WebUI.

Wrt/ a1, it worked at one point (a year ago) after 2-3 hours of tinkering, then regressed to not working at all, not even from fresh installs on new, different Linuxes. I tried the main branch and the AMD specific fork as well.

Wrt/ Open WebUI, it works, but the thing uses my CPU.

40. wredcoll ◴[] No.44473655[source]
A significant part of the vocal "gamers" is about being "the best" which translates into gpu benchmarking.

You don't get headlines and hype by being an affordable way to play games at a decent frame rate, you achieve it by setting New Fps Records.

41. tankenmate ◴[] No.44473933{5}[source]
To quote "The Dude"; "Well ... ummm ... that's ... ahh ... just your opinion man". There are people who are successfully running it in production, but of course depending on your code, YMMV.
42. int_19h ◴[] No.44474569{3}[source]
On Ubuntu, in my experience, installing the .deb version of the CUDA toolkit pretty much "just works".
43. eptcyka ◴[] No.44474660{4}[source]
I was under the impression that nvidia just didn't let consumer cards do GPU passthrough.
replies(1): >>44515820 #
44. pjmlp ◴[] No.44475472{3}[source]
The open source driver for the Netboooks APU was never as good as either the Windows version, or the closed source that predated it.

Lesser OpenGL version, and I never managed to have hardware accelerated video until it died last year.

45. pjmlp ◴[] No.44475494{5}[source]
I never managed to get it working on my Netbook APU.
46. phatfish ◴[] No.44476131{3}[source]
Once the "path tracing" that the current top end Nvidia cards can pull off reaches mainstream it will settle down. The PS6 isn't going to be doing path tracing because the hardware for that is being decided now. I'd guess PS7 time frame. It will take console level hardware pricing to bring the gaming GPU prices down.

I understand the reason for moving to real time ray-tracing. It is much easier for development, and apparently the data for baked/pre-rendered lighting in these big open worlds was getting out of hand. Especially with multiple time-of-day passes.

But, it is only the "path tracing" that top end Nvidia GPUs can do that matches baked lighting detail.

The standard ray-tracing in the latest Doom for instance has a very limited number of entities that actually emit light in a scene. I guess there is the main global illumination source, but many of the extra lighting details in the scene don't emit light. This is a step backward compared to baked lighting.

Even shots from the plasma weapon don't cast any light into the scene with the standard ray-tracing, which Quake 3 was doing.

47. AngryData ◴[] No.44476308[source]
I am a gamer, and I don't understand why everyone flocks to Nvidia either unless they are buying the newest flagship card. Maybe just because "the best card" is from Nvidia so many assume Nvidia must be the best for everyone? For multiple generations ive gotten better card for my dollar at any "mid-tier" gaming level with AMD, and have had zero complaints.
48. michaelmrose ◴[] No.44476819{5}[source]
I believe this is correct. Linux drivers and support duration were garbage at least 2003-2015. AMD fanboys feveretly expressed opinions notwithstanding. Especially so when AMD started the process of open sourcing their drivers even though many chips already existing didn't qualify for the new upcoming open source drivers. 2015-2018 drivers were acceptable but performance was poorer than Nvidia and wayland support wasn't a notable for most parties.

Now wayland support is an important factor and AMD is a perfectly acceptable and indeed economical choice.

Basically 15 years inertia is hard to counter.

49. Bratmon ◴[] No.44479276[source]
AMD GPUs are 5 years behind Nvidia. But that logically means that if you thought Nvidia graphics looked fine in 2020, you'll think AMD graphics look fine now.
50. xorcist ◴[] No.44479520{5}[source]
Just recently Alex Courbot with an @nvidia address have become co-maintainer of Nova. Apparently he has pushed open source inside nVidia before, so there's hope for the future.
51. greenmartian ◴[] No.44505976{4}[source]
I just read your comment today. And the good news is, someone just found the commit that triggered the sleep issue literally a couple hours ago. Fingers crossed a fix is incoming.
replies(1): >>44515810 #
52. MegaDeKay ◴[] No.44515810{5}[source]
Yes, I saw that! No response from the devs yet. Hopefully this gets into the kernel soon as a bugfix.
53. MegaDeKay ◴[] No.44515820{5}[source]
For the longest time they had something in their driver that threw an error and bailed if it detected the GPU was being passed through to a VM but that was easily worked around in qemu or libvirt. nvidia must have realized that that check in the driver was pointless and removed it.