Most active commenters
  • lolinder(3)

←back to thread

623 points magicalhippo | 16 comments | | HN request time: 0.001s | source | bottom
1. derbaum ◴[] No.42620643[source]
I'm a bit surprised by the amount of comments comparing the cost to (often cheap) cloud solutions. Nvidia's value proposition is completely different in my opinion. Say I have a startup in the EU that handles personal data or some company secrets and wants to use an LLM to analyse it (like using RAG). Having that data never leave your basement sure can be worth more than $3000 if performance is not a bottleneck.
replies(6): >>42621036 #>>42621592 #>>42622470 #>>42622485 #>>42622500 #>>42622740 #
2. sensesp ◴[] No.42621036[source]
100% I see many SMEs not willing to send their data to some cloud black box.
replies(1): >>42624496 #
3. originalvichy ◴[] No.42621592[source]
Even for established companies this is great. A tech company can have a few of these locally hosted and users can poll the company LLM with sensitive data.
4. lolinder ◴[] No.42622470[source]
Heck, I'm willing to pay $3000 for one of these to get a good model that runs my requests locally. It's probably just my stupid ape brain trying to do finance, but I'm infinitely more likely to run dumb experiments with LLMs on hardware I own than I am while paying per token (to the point where I currently spend way more time with small local llamas than with Claude), and even though I don't do anything sensitive I'm still leery of shipping all my data to one of these companies.

This isn't competing with cloud, it's competing with Mac Minis and beefy GPUs. And $3000 is a very attractive price point in that market.

replies(2): >>42623584 #>>42624306 #
5. diggan ◴[] No.42622485[source]
The price seems relatively competitive even compared to other local alternatives like "build your own PC". I'd definitely buy one of this (or even two if it works really well) for developing/training/using models that currently run on cobbled together hardware I got left after upgrading my desktop.
6. btbuildem ◴[] No.42622500[source]
Yeah that's cheaper than many prosumer GPUs on the market right now
7. 627467 ◴[] No.42622740[source]
> Having that data never leave your basement sure can be worth more than $3000 if performance is not a bottleneck

I get what you're saying, but there are also regulations (and your own business interest) that expects data redundancy/protection which keeping everything on-site doesnt seem to cover

8. ynniv ◴[] No.42623584[source]
I'm pretty frugal, but my first thought is to get two to run 405B models. Building out 128GB of VRAM isn't easy, and will likely cost twice this.
replies(1): >>42625092 #
9. logankeenan ◴[] No.42624306[source]
Have you been to the localLlama subreddit? It’s a great resource for running models locally. It’s what got me started.

https://www.reddit.com/r/LocalLLaMA/

replies(1): >>42624609 #
10. jckahn ◴[] No.42624496[source]
Exactly this. I would happily give $3k to NVIDIA to avoid giving 1 cent to OpenAI/Anthropic.
11. lolinder ◴[] No.42624609{3}[source]
Yep! I don't spend much time there because I got pretty comfortable with llama before that subreddit really got started, but it's definitely turned up some helpful answers about parameter tuning from time to time!
12. rsanek ◴[] No.42625092{3}[source]
You can get a M4 Max MBP with 128GB for $1k less than two of these single-use devices.
replies(4): >>42625963 #>>42625978 #>>42626307 #>>42627334 #
13. lolinder ◴[] No.42625963{4}[source]
Don't these devices provide 128GB each? So you'd need to price in two Macs to be a fair comparison to two Digits.
14. ynniv ◴[] No.42625978{4}[source]
These are 128GB each. Also, Nvidias inference speed is much higher than Apple's.

I do appreciate that my MBP can run models though!

15. layer8 ◴[] No.42626307{4}[source]
But then you have to use macOS.
16. ganoushoreilly ◴[] No.42627334{4}[source]
I read the Nvidia units are 250 Tflops vs the M4 Pro 27 Tflops. If they perform as advertised i'm in for two.