Most active commenters
  • foobiekr(4)
  • slt2021(3)
  • AnthonyMouse(3)

←back to thread

255 points tbruckner | 26 comments | | HN request time: 1.723s | source | bottom
Show context
adam_arthur ◴[] No.37420461[source]
Even a linear growth rate of average RAM capacity would obviate the need to run current SOTA LLMs remotely in short order.

Historically average RAM has grown far faster than linear, and there really hasn't been anything pressing manufacturers to push the envelope here in the past few years... until now.

It could be that LLM model sizes keep increasing such that we continue to require cloud consumption, but I suspect the sizes will not increase as quickly as hardware for inference.

Given how useful GPT-4 is already. Maybe one more iteration would unlock the vast majority of practical use cases.

I think people will be surprised that consumers ultimately end up benefitting far more from LLMs than the providers. There's not going to be much moat or differentiation to defend margins... more of a race to the bottom on pricing

replies(8): >>37420537 #>>37420948 #>>37421196 #>>37421214 #>>37421497 #>>37421862 #>>37421945 #>>37424918 #
1. MuffinFlavored ◴[] No.37421214[source]
> Given how useful GPT-4 is already. Maybe one more iteration would unlock the vast majority of practical use cases.

Unless I'm misunderstanding, doesn't OpenAI have a very vested interest to keep making their products so good/so complex/so large that consumer hobbyists can't just `git clone` an alternative that's 95% as good running locally?

replies(3): >>37421454 #>>37421498 #>>37421783 #
2. Frannyies ◴[] No.37421454[source]
They have a huge cost incentive to optimize it for runtime.

The magic of openai is their training data and architecture.

There is a real risk that a model gets leaked.

replies(1): >>37421998 #
3. chongli ◴[] No.37421498[source]
What is OpenAI's moat? Loads of people outside the company are working on alternative models. They may have a lead right now but will it last a few years? Will it even last 6 months?
replies(4): >>37421647 #>>37421649 #>>37421665 #>>37422380 #
4. yumraj ◴[] No.37421647[source]
> What is OpenAI's moat?

There’s none. Which is why Sam Altman has been crying wolf, in hope of regulatory barriers which can provide it the moat.

5. MuffinFlavored ◴[] No.37421649[source]
> What is OpenAI's moat?

From what I understand, if you take the absolute best cutting edge LLM with the most parameters and the most up to date model from GitHub/HuggingFace/whatever, it's very far off from the output you get from GPT-3.5 / GPT-4

aka full of hallucinations, not very useful

I don't know if this is the right way to look at it but if what George Hotz said about GPT-4 simply being "8 220B parameter models glued together by something called a mixture-of-experts", from what I understand, OpenAI's moat is:

their access/subsidiized cost to GPUs/infrastructure with Microsoft

the 8 220B models they have are really good/I don't think anything open source matches them/nobody can download "all of Reddit/Twitter/Wikipedia/StackOverflow/whatever else they trained on" anymore like they could given how everybody wants to protect/monetize their content now

and then the "router" / "MoE" piece seems to be something missing from open source offerings as well

replies(3): >>37421962 #>>37422981 #>>37426850 #
6. ben_w ◴[] No.37421665[source]
OpenAI's "moat" is basically the same as Adobe's or Microsoft's, give or take a metaphor, for Photoshop or Office.

Although see last week for previous responses: https://news.ycombinator.com/item?id=37333747

7. reckless ◴[] No.37421783[source]
Indeed they do, however companies like Meta (altruistically or not) are preventing OpenAI from building 'moats' by releasing models and architecture details in a very public way.
replies(2): >>37422263 #>>37422288 #
8. easygenes ◴[] No.37421962{3}[source]
Depending on the task, the best open models will outperform GPT-3.5, but would be more expensive to run at comparable speed. GPT-4 is in a league of its own.
9. slt2021 ◴[] No.37421998[source]
it is not really a moat if one engineer can leave openai with all the secret sauce in his head and replicate it elsewhere (anthropic?)
replies(2): >>37422647 #>>37423076 #
10. runjake ◴[] No.37422263[source]
I think it's a safe bet to say it's not altruistic. And, if Meta were to wrestle away OpenAI's moat, they'd eagerly create their own, given the opportunity.
replies(3): >>37422875 #>>37423084 #>>37426473 #
11. foobiekr ◴[] No.37422288[source]
Commoditize your complement strategies can just as likely put a market into a zombie state in the long run.
12. foobiekr ◴[] No.37422380[source]
Adoption and a mass of human feedback collected which is not available in the gleaned data sets.

Here’s another way to think about it. Why does ISA matter in CPUs? There are minor issues around efficiencies of various kinds, but the real advantage of any mainstream ISA is, in part, the availability of tooling (hence this was a correct and heavy early focus for the RISCV effort) but also a lot of ecosystem things you don’t see: for example, Intel and Arm have truly mammoth test and verification suites that represent thousands++ of man years of investment.

OpenAI almost certainly has a massive invisible accumulated value at this point.

The actual models themselves are the output in the same way that a packaged CPU is the output. How you got there matters almost as much or more.

replies(1): >>37426572 #
13. foobiekr ◴[] No.37422647{3}[source]
Name one software-based tech company where this isn’t true.
replies(1): >>37423025 #
14. passion__desire ◴[] No.37422875{3}[source]
Meta doesn't interact with its users in very obvious ways which MS, Google do. All its models magic happen behind the scenes. Meta can continue to release 2nd best models to undercut others and them going far too ahead. And Open Source community will take it from there. Dall-E is dead.
replies(2): >>37423063 #>>37427098 #
15. passion__desire ◴[] No.37422981{3}[source]
What if specialized smaller models is the best way ahead for community. I don't care if I am interacting with one big model which can do everything or I have to go to different websites to access specific models. All model sizes will be useful. Smaller models will be frequently used. Bigger less so.
16. slt2021 ◴[] No.37423025{4}[source]
microsoft? gogel? FB?
17. bugglebeetle ◴[] No.37423063{4}[source]
And if all open source extends their models, they can accrue those benefits back to themselves. This is already how they’ve become such a huge player in machine learning (open sourcing amazing stuff).
18. Frannyies ◴[] No.37423076{3}[source]
I only meant the trained model.

You would need to steal it all over again as soon as the next model is trained.

replies(1): >>37424110 #
19. sangnoir ◴[] No.37423084{3}[source]
> And, if Meta were to wrestle away OpenAI's moat, they'd eagerly create their own

Meta is already capable of monetizing content generated by the models: these models complement their business and they could not care less which model you're using to earn them advertising dollars, as long as you keep the (preferably high quality) content coming.

20. slt2021 ◴[] No.37424110{4}[source]
no need to steal the model if training process can be reliably replicated/adopted in clean room implementation with additional optimisations.

startup as legal entity has close to 0 value, most value is in intellectual property which is stored and transmitted by meatbags.

21. AnthonyMouse ◴[] No.37426473{3}[source]
> And, if Meta were to wrestle away OpenAI's moat, they'd eagerly create their own, given the opportunity.

At which point the new underdogs would have an interest in doing to them what they're doing to OpenAI.

Assuming progress for LLMs continues at a rapid pace for an extended period of time. It's not implausible that they'll get to a certain level past which non-trivial progress is hard, and if there is an open source model at that level there isn't going to be a moat.

22. AnthonyMouse ◴[] No.37426572{3}[source]
> Here’s another way to think about it. Why does ISA matter in CPUs?

Honestly the answer is that it mostly doesn't.

An ISA isn't viable without tooling, but that's why it's the first thing they all get. The only ISA with any significant moat is x86, and that's because there is so much legacy closed source software for it that people still need but would have to be emulated on any other architecture. And even that only works as long as x86 processors are competitive; if they fell behind then customers would just eat the emulation overhead on something else.

Other ISAs don't even have that. Would anybody actually be surprised if RISC-V took a huge chunk of out ARM's market share in the not too distant future?

replies(1): >>37429430 #
23. nmfisher ◴[] No.37426850{3}[source]
This isn’t really true (or at least, doesn’t apply across the board). Qwen (Alibaba’s open source model) outperforms GPT4 on Chinese language tasks, and I can further finetune it for my own tasks (which I’ve done, and I confirm it’s produces more natural output than GPT4).

Other benchmarks/anecdotes suggest fine-tuned code models are outperforming GPT4 too. The trend seems to be that smaller, fine-tuned task specific models outperform larger generalised models. It requires a lot of resources to pretrain the base model, but as we’ve seen, there’s no shortage of companies who are willing and able to do that.

Not to mention, all those other companies are already profitable, whereas OpenAI is already burning investor cash.

24. astrange ◴[] No.37427098{4}[source]
I think Dall-E isn't actually dead, but was merely renamed Bing Image Creator.
25. foobiekr ◴[] No.37429430{4}[source]
That's literally my point. The problem is that there's a massive amount of hidden infrastructure behind those that you don't see and that "oh look everyone has a big model" isn't as impressive as it sounds.
replies(1): >>37430771 #
26. AnthonyMouse ◴[] No.37430771{5}[source]
But the open source infrastructure is getting built too. And the infrastructure is mostly independent of the model. This is Falcon 180B running using the code from llama.cpp.