←back to thread

183 points spacebanana7 | 2 comments | | HN request time: 0.48s | source

I appreciate developing ROCm into something competitive with CUDA would require a lot of work, both internally within AMD and with external contributions to the relevant open source libraries.

However the amount of resources at stake is incredible. The delta between NVIDIA's value and AMD's is bigger than the annual GDP of Spain. Even if they needed to hire a few thousand engineers at a few million in comp each, it'd still be a good investment.

Show context
johnnyjeans ◴[] No.43547510[source]
> The delta between NVIDIA's value and AMD's is bigger than the annual GDP of Spain.

Nvidia is massively overvalued right now. AI has rocketed them into absolute absurdity, and it's not sustainable. Put aside the actual technology for a second and realize that public image of AI is at rock bottom. Every single time a company puts out AI-generated materials, they receive immense public backlash. That's not going away any time soon and it's only likely to get worse.

Speaking as someone that's not even remotely anti-AI, I wouldn't touch the shit with a 10 foot pole because of how bad the public image is. The moment that capital realizes this, that bubble is going to pop and it's going to pop hard.

replies(5): >>43547559 #>>43547590 #>>43547630 #>>43547711 #>>43548927 #
spacebanana7 ◴[] No.43547711[source]
> AI has rocketed them into absolute absurdity, and it's not sustainable

Why isn't it sustainable? Their biggest customers all have strong finances and legitimate demand. Google and Facebook would happily run every piece of user generated content through an LLM if they had enough GPUs. Same with Microsoft and every enterprise document.

The VC backed companies and Open AI are more fragile, but they're comparatively small customers.

replies(3): >>43547801 #>>43548050 #>>43548097 #
1. fancyfredbot ◴[] No.43548097[source]
Google may well want to run more of their content through an LLM, but they will not be using Nvidia hardware to do it, they'll be using their TPUs.

Amazon are on their third generation of in-house AI chips and Anthropic will be using those chips to train the next generation of Claude.

In other words, their biggest customers are looking for cheaper alternatives and are already succeeding in finding them.

replies(1): >>43549546 #
2. a5ehren ◴[] No.43549546[source]
Google and Amazon still have to buy tons of Nvidia HW to provide in their clouds. No one writes to their custom chips besides internal teams because the software stack doesn't exist.