Most active commenters

    ←back to thread

    1045 points mfiguiere | 11 comments | | HN request time: 0.902s | source | bottom
    Show context
    Cu3PO42 ◴[] No.39346489[source]
    I'm really rooting for AMD to break the CUDA monopoly. To this end, I genuinely don't know whether a translation layer is a good thing or not. On the upside it makes the hardware much more viable instantly and will boost adoption, on the downside you run the risk that devs will never support ROCm, because you can just use the translation layer.

    I think this is essentially the same situation as Proton+DXVK for Linux gaming. I think that that is a net positive for Linux, but I'm less sure about this. Getting good performance out of GPU compute requires much more tuning to the concrete architecture, which I'm afraid devs just won't do for AMD GPUs through this layer, always leaving them behind their Nvidia counterparts.

    However, AMD desperately needs to do something. Story time:

    On the weekend I wanted to play around with Stable Diffusion. Why pay for cloud compute, when I have a powerful GPU at home, I thought. Said GPU is a 7900 XTX, i.e. the most powerful consumer card from AMD at this time. Only very few AMD GPUs are supported by ROCm at this time, but mine is, thankfully.

    So, how hard could it possibly to get Stable Diffusion running on my GPU? Hard. I don't think my problems were actually caused by AMD: I had ROCm installed and my card recognized by rocminfo in a matter of minutes. But the whole ML world is so focused on Nvidia that it took me ages to get a working installation of pytorch and friends. The InvokeAI installer, for example, asks if you want to use CUDA or ROCm, but then always installs the CUDA variant whatever you answer. Ultimately, I did get a model to load, but the software crashed my graphical session before generating a single image.

    The whole experience left me frustrated and wanting to buy an Nvidia GPU again...

    replies(10): >>39346714 #>>39347956 #>>39348258 #>>39349464 #>>39349658 #>>39350019 #>>39350273 #>>39351237 #>>39354496 #>>39433413 #
    1. whywhywhywhy ◴[] No.39347956[source]
    > I'm really rooting for AMD to break the CUDA monopoly

    Personally I want Nvidia to break the x86-64 monopoly, with how amazing properly spec'd Nvidia cards are to work with I can only dream of a world where Nvidia is my CPU too.

    replies(4): >>39348006 #>>39348977 #>>39351000 #>>39352323 #
    2. kuschkufan ◴[] No.39348006[source]
    apt username
    3. smcleod ◴[] No.39348977[source]
    That’s already been done with ARM.
    4. weebull ◴[] No.39351000[source]
    > Personally I want Nvidia to break the x86-64 monopoly

    The one supplied by two companies?

    replies(2): >>39351725 #>>39353087 #
    5. Keyframe ◴[] No.39351725[source]
    Maybe he meant homogeneity which Nvidia did try and tries with Arm.. but, on the other hand how wild would it be for Nvidia to enter x86-64 as well? It's probably never going to happen due to licensing if nothing else, lest we remember nForce chipset ordeal with intel legal.
    replies(2): >>39353105 #>>39375351 #
    6. mickael-kerjean ◴[] No.39352323[source]
    How would this a good idea? I am not very familiar with GPU programming but the small amount I've tried was nothing but pain a few years ago on linux, it was so bad that Torvald publicly used the f word in a very public event. That aside, CUDA seem like a great way to lock people in even further like AWS does with absolutely everything
    replies(2): >>39354180 #>>39356819 #
    7. paulmd ◴[] No.39353087[source]
    "minor spelling/terminology mistake, activate the post-o-tron"
    8. paulmd ◴[] No.39353105{3}[source]
    https://en.wikipedia.org/wiki/Project_Denver#History
    9. Qwertious ◴[] No.39354180[source]
    >I am not very familiar with GPU programming but the small amount I've tried was nothing but pain a few years ago on linux, it was so bad that Torvald publicly used the f word in a very public event.

    I'm pretty sure Torvalds was giving the finger over the subject of GPU drivers (which run on the CPU), not programming on the Nvidia GPU itself. Particularly, they namedropped Bumblebee (and maybe Optimus?) which was more about power-management and making Nvidia cooperate with a non-Nvidia integrated GPU than it was about the Nvidia GPU itself.

    10. whywhywhywhy ◴[] No.39356819[source]
    >CUDA seem like a great way to lock people in even further like AWS does with absolutely everything

    Lock people in to something that didn’t exist in a way any user could use before it existed? I get people hate CUDAs dominance but no one else was pushing this before CUDA and Apple+AMD completely fumbled OpenCL.

    Can’t hate on something good just because it’s successful and I can’t be angry the talent behind the success wanting to profit.

    11. weebull ◴[] No.39375351{3}[source]
    Indeed, but I think people forget that the reason AMD have a license in the first place was because Intel's customers in the early days required a second source for it's processors.

    Who owns the Cyrix x86 license these days?