Most active commenters
  • aurareturn(16)
  • tim333(6)
  • rsynnott(5)
  • HarHarVeryFunny(5)
  • johnnyanmac(5)
  • (4)
  • atomic128(4)
  • danielmarkbruce(4)
  • ViewTrick1002(4)
  • ben_w(3)

←back to thread

The AI Investment Boom

(www.apricitas.io)
271 points m-hodges | 172 comments | | HN request time: 1.296s | source | bottom
1. hn_throwaway_99 ◴[] No.41896346[source]
Reading this makes me willing to bet that this capital intensive investment boom will be similar to other enormous capital investment booms in US history, such as the laying of the railroads in the 1800s, the proliferation of car companies in the early 1900s, and the telecom fiber boom in the late 1900s. In all of these cases there was an enormous infrastructure (over) build out, followed by a crash where nearly all the companies in the industry ended up in bankruptcy, but then that original infrastructure build out had huge benefits for the economy and society as that infrastructure was "soaked up" in the subsequent years. E.g. think of all the telecom investment and subsequent bankruptcies in the late 90s/early 00s, but then all that dark fiber that was laid was eventually lit up and allowed for the explosion of high quality multimedia growth (e.g. Netflix and the like).

I think that will happen here. I think your average investor who's currently paying for all these advanced chips, data centers and energy supplies will walk away sorely disappointed, but this investment will yield huge dividends down the road. Heck, I think the energy investment alone will end up accelerating the switch away from fossil fuels, despite AI often being portrayed as a giant climate warming energy hog (which I'm not really disputing, but now that renewables are the cheapest form of energy, I believe this huge, well-funded demand will accelerate the growth of non-carbon energy sources).

replies(21): >>41896376 #>>41896426 #>>41896447 #>>41896726 #>>41898086 #>>41898206 #>>41898291 #>>41898436 #>>41898540 #>>41899659 #>>41900309 #>>41900633 #>>41903200 #>>41903363 #>>41903416 #>>41903838 #>>41903917 #>>41904566 #>>41905630 #>>41905809 #>>41906189 #
2. candiddevmike ◴[] No.41896376[source]
What will be the advantage of having a bunch of obsolete hardware? All I see is more e-waste.
replies(8): >>41896440 #>>41896450 #>>41896455 #>>41896456 #>>41896471 #>>41896532 #>>41896645 #>>41899790 #
3. bwanab ◴[] No.41896426[source]
While I agree with your essential conclusions, I don't think the automobile companies really fit. Many of the early 1900s companies (e.g. Ford, GM, Mercedes, even Chrysler) are still among the largest auto companies in the world.
replies(5): >>41896441 #>>41896540 #>>41896563 #>>41896580 #>>41897155 #
4. wslh ◴[] No.41896440[source]
I don't think the parent was specifically referring to hardware alone. The 'rails' in this context are also the AI algorithms and the underlying software. New research and development could lead to breakthroughs that allow us to use significantly less hardware than we currently do. Just as the dot-com crash wasn’t solely about the physical infrastructure but also about protocols like HTTP, I believe the AI boom will involve advancements beyond just hardware. There may be short-term excess, but the long-term benefits, particularly on the software side, could be immense.
replies(1): >>41903794 #
5. throwaway20222 ◴[] No.41896441[source]
There were hundreds of failed automotive companies and parts suppliers though. I think the argument is that many will die, some will survive and take all (most)
replies(2): >>41896761 #>>41905766 #
6. aurareturn ◴[] No.41896447[source]
I'm sure you are right. At some point, the bubble will crash.

The question remains is when the bubble will crash. We could be in the 1995 equivalent of the dotcom boom and not 1999. If so, we have 4 more years of high growth and even after the crash, the market will still be much bigger in 2029 than in 2024. Cisco was still 4x bigger in 2001 than in 1995.

One thing that is slightly different from past bubbles is that the more compute you have, the smarter and more capable AI.

One gauge I use to determine if we are still at the beginning of the boom is this: Does Slack sell an LLM chatbot solution that is able to give me reliable answers to business/technical decisions made over the last 2 years in chat? We don't have this yet - most likely because it's probably still too expensive to do this much inference with such high context window. We still need a lot more compute and better models.

Because of the above, I'm in the camp that believe we are actually closer to the beginning of the bubble than at the end.

Another thing I would watch closely to see when the bubble might pop is if LLM scaling laws are quickly breaking down and that more compute no longer yields more intelligence in an economical way. If so, I think the bubble would pop. All eyes are on GPT5-class models for signs.

replies(8): >>41896552 #>>41896790 #>>41898712 #>>41899018 #>>41899201 #>>41903550 #>>41904788 #>>41905320 #
7. iTokio ◴[] No.41896450[source]
Well, at least they are paving the way to more efficient hardware, GPU are way, way more energy efficient than CPU and the parallel architectures are the only remaining way to scale compute.

But yes, a lot of energy wasted in the growing phase.

replies(2): >>41896551 #>>41897866 #
8. goda90 ◴[] No.41896455[source]
Even if the hardware quickly becomes outdated, I'm not sure it'll become worthless so quickly. And there's also the infrastructure of the data center and new electricity generation to power them. Another thing that might survive a crash and carry on to help the future is all the code used to support valuable use cases.
9. almost_usual ◴[] No.41896456[source]
In the case of dark fiber the hardware was fine, wavelength division multiplexing was created which increased capacity by 100x in some cases crashing demand for new fiber.

I think OP is suggesting AI algorithms and training methods will be improve resulting in enormous performance gains with existing hardware causing a similar surplus of infrastructure and crash in demand.

replies(1): >>41896626 #
10. CamperBob2 ◴[] No.41896471[source]
Do you expect better hardware to suddenly start appearing on the market, fully-formed from the brow of Zeus?
11. bee_rider ◴[] No.41896532[source]
Maybe the bust will be so rough that TSMC will go out of business, and then these graphics cards will not go obsolete for quite a while.

Like Intel and Samsung might make a handful of better chips or whatever, but neither of their business models really involve being TSMC. So if the bubble pop took out TSMC, there wouldn’t be a new TSMC for a while.

12. cloud_hacker ◴[] No.41896540[source]
> While I agree with your essential conclusions, I don't think the automobile companies really fit. Many of the early 1900s companies (e.g. Ford, GM, Mercedes, even Chrysler) are still among the largest auto companies in the world.

American Automative filled for Bankruptcy multiple times.

American Government had to step in to back them up and bail them out.

replies(2): >>41903883 #>>41905774 #
13. dartos ◴[] No.41896551{3}[source]
GPUs are different than CPUs.

They’re way more efficient at matmuls, but start throwing branching logic at them and they slow down a lot.

Literally a percentage of their cores will noop while others are executing a branch, since all cores are lockstep.

14. vladgur ◴[] No.41896552[source]
Re: Slack chat:

Glean.com does it for the enterprise I work at: It consumes all of our knowledge sources including Slack, Google docs, wiki, source code and provides answers to complex specific questions in a way that’s downright magical.

I was converted into a believer when I described an issue to it, pointers to a source file in online git repo and it pointed me to another repository that my team did not own that controlled DNS configs that we were not aware about. These configs were the reason our code did not behave as we expected.

replies(4): >>41896575 #>>41896658 #>>41899040 #>>41901466 #
15. nemo44x ◴[] No.41896563[source]
That phase is called consolidation. It’s part of the cycle. The speculative over leveraged and mismanaged companies get merged into the winners or disappear if they have nothing of value.
16. aurareturn ◴[] No.41896575{3}[source]
Thanks. I didn't know that existed. But does it scale? Would it still work if large companies with many millions of Slack messages?

I suppose one reason Slack doesn't have a solution yet is because they're having a hard time getting it to work for large companies.

replies(2): >>41896647 #>>41896714 #
17. ◴[] No.41896580[source]
18. llamaimperative ◴[] No.41896626{3}[source]
How much of current venture spending is going into reusable R&D that can be moved forward in time the way that physical infrastructure in their examples were able to be used in the future?
replies(1): >>41900123 #
19. aurareturn ◴[] No.41896645[source]
>What will be the advantage of having a bunch of obsolete hardware? All I see is more e-waste.

The energy build out, data centers are not wasted. You can swap out A100 GPUs for BH200 GPUs in the same datacenter. A100s will be 5 years old when Blackwell is out - which is just about right for how long datacenter chips are expected to last.

I do, however, think that the industry will move to newer hardware faster to try to squeeze as much efficiency as possible due to the energy bottleneck. Therefore, I expect TSMC's N2 nodes to have huge demand. In fact, TSMC themselves have said designs for N2 far outnumber N3 at the same stage of the node. This is most likely due to AI companies that want to increase efficiency due to the lack of electricity.

20. ◴[] No.41896647{4}[source]
21. _huayra_ ◴[] No.41896658{3}[source]
This is the main "killer feature" I've personally experienced from GPT things: a much better contextual "search engine-ish" tool for combing through and correlating different internal data sources (slack, wiki, jira, github branches, etc).

AI code assistants have been a net neutral for me (they get enough idioms in C++ slightly incorrect that I have to spend a lot of time just reading the generated code thoroughly), but being able to say "tell me what the timeline for feature X is" and have it comb through a bunch of internal docs / tickets / git commit messages, etc, and give me a coherent answer with links is amazing.

replies(3): >>41896682 #>>41898324 #>>41905687 #
22. aurareturn ◴[] No.41896682{4}[source]
This is partly why I believe OS makers, Apple, Microsoft, Google, have a huge advantage in the future when it comes to LLMs.

They control the OS so they can combine and feed all your digital information to an LLM in a seamless way. However, in the very long term, I think their advantage will go away because at some point, LLMs could get so good that you don't need an OS like iOS anymore. An LLM could simply become standalone - and function without a traditional OS.

Therefore, I think the advantage for iOS, Android, Windows will increase in the next few years, but less powerful after that.

replies(3): >>41898863 #>>41902626 #>>41904320 #
23. hn_throwaway_99 ◴[] No.41896714{4}[source]
Yeah, Glean does this and there are a bunch of other competitors that do it as well.

I think you may be confused about the length of the context window. These tools don't pull all of your Slack history into the context window. They use a RAG approach to index all of your content into a vector DB, then when you make a query only the relevant document snippets are pulled into the context window. It's similar for example to how Cursor implements repository-wide AI queries.

replies(1): >>41896734 #
24. HarHarVeryFunny ◴[] No.41896726[source]
Similarly, I like to compare AI (more specifially LLM) "investment" to the cost of building the channel tunnel between UK and France. The original investors lost their shirt, but once built it is profitable to operate.
replies(1): >>41898651 #
25. aurareturn ◴[] No.41896734{5}[source]
I'm aware that one can't feed millions of messages into an LLM all at once. The only way to do this now is to use a RAG approach. But RAG approach has pros and cons and can miss crucial information. I think context window still matters a lot. The bigger the window, the more information you can feed in and the quality of answer should increase.

The point I'm trying to make is that increase context window will require more compute. Hence, we could still just be in the beginning of the compute/AI boom.

replies(1): >>41898924 #
26. aurareturn ◴[] No.41896761{3}[source]
But that happens in every bubble. Over investment, consolidation, huge winners in the end, and maybe eventually a single monopoly.
replies(1): >>41898544 #
27. HarHarVeryFunny ◴[] No.41896790[source]
> the more compute you have, the smarter and more capable AI

Well, this is taken on faith by OpenAI/etc, but obviously the curve has to flatten at some point, and appears to already be doing so. OpenAI are now experimenting with scaling inference-time compute (GPT-O1), but have said that it takes exponential increases in compute to produce linear gains in performance, so it remains to be seen if customers find this a worthwhile value.

replies(1): >>41896900 #
28. aurareturn ◴[] No.41896900{3}[source]
GPT-o1 does demonstrate my point: the more compute you have, the smarter the AI.

If you run chain of thoughts on an 8B model, it becomes a lot smarter too.

GPT-o1 isn't GPT5 though. I think OpenAI will have a chain of thoughts model for GPT5-class models as well. They're separate from normal models.

replies(1): >>41896980 #
29. HarHarVeryFunny ◴[] No.41896980{4}[source]
There is only so much that an approach like O1 can do, but anyways in terms of AI boom/bust the relevant question is whether this is a viable product. All sorts of consumer products could be improved by making them a lot more expensive, but there are cost/benefit limits to everything.

GPT-5 and Claude-4 will be interesting, assuming these are both pure transformer models (not COT), as they will be a measure how much benefit remains to be had from training set scaling. I'd expect gains will be more against narrow benchmarks, than in the overall feel of intelligence (LLM arena score?) one gets from the model.

replies(1): >>41899221 #
30. grecy ◴[] No.41897155[source]
A couple of them went bankrupt and got bailouts.
31. aurareturn ◴[] No.41897866{3}[source]
>But yes, a lot of energy wasted in the growing phase.

Why exactly is energy wasted during this phase?

Are you expecting hardware to become obsolete much faster? But that only depends on TSMC's node cadence, which is still 2-3 years. Therefore, AI hardware will still be bound to TSMC's cadence.

32. ben_w ◴[] No.41898086[source]
> but now that renewables are the cheapest form of energy, I believe this huge, well-funded demand will accelerate the growth of non-carbon energy sources

I think the renewables would have been built at the same rate anyway precisely because they're so cheap; but nuclear power, being expensive, would not be built if this bubble had not happened, and somehow nuclear does seem to be getting some of this money.

replies(3): >>41898253 #>>41898260 #>>41903672 #
33. from-nibly ◴[] No.41898206[source]
The problem is that all that mal investment will get bailed about by us regular shmucks. Get ready for the hamster wheel to start spinning faster.
34. atomic128 ◴[] No.41898253[source]
I want to point out to anyone who's interested in the nuclear angle that even before the AI data center demand story arrived, the uranium market was facing a persistent undersupply for the first time in its many decades of history. As a result, the (long-term contract) price of uranium has been steadily rising for years: https://www.cameco.com/invest/markets/uranium-price

After Fukushima (https://news.ycombinator.com/item?id=41768726), Japanese reactors were shut down and there was a glut of uranium available in the spot market. Simultaneously, Kazatomprom flooded the market with cheap ISR uranium. The price of uranium fell far below the cost of production and the mining companies were obliterated. The few miners that survived via their long-term contracts (primarily Cameco) put their less efficient mines into care and maintenance.

Now we're seeing the uranium mining business wake up. But after a decade of bear-market conditions the miners cannot serve the demand: they've underinvested, they've lost skilled labor, they've shrunk. The rebound in uranium supply will be slow, much slower than the rebound in demand. This is because uranium mining is an extremely difficult process. Look at how long NexGen Energy's Rook 1 Arrow mine has taken to develop, and that's prime ore (https://s28.q4cdn.com/891672792/files/doc_downloads/2022/03/...). Look at Kazatomprom's slowing growth rate (https://world-nuclear-news.org/Articles/Kazatomprom-lowers-2...), look at the incredible complexity of Cameco's mining operations: https://www.petersenproducts.com/articles/an-inflatable-tunn...

Here is a discussion of the uranium mining situtation: https://news.ycombinator.com/item?id=41661768 (including a very risky method of profiting from the undersupply of uranium, stock ticker SRUUF, not recommended). Note that Numerco's uranium spot price was put behind a paywall last week. You can still get the intra-day spot uranium price for free here: https://www.yellowcakeplc.com/

replies(1): >>41898979 #
35. synergy20 ◴[] No.41898260[source]
based on my reading nuclear power is much cheaper overall compared to wind solar etc?
replies(2): >>41898377 #>>41898565 #
36. kjkjadksj ◴[] No.41898291[source]
Not all of that infrastructure gets soaked up, plenty is abandoned. Look at the state of american passenger rail for example and how quickly the bottom of that industry dropped out. Many old rail right of ways sit abandoned today. Likewise with telecoms, e.g. the microwave network that also sits abandoned today.
replies(1): >>41900207 #
37. aaronblohowiak ◴[] No.41898324{4}[source]
>they get enough idioms in C++ slightly incorrect

this is part of why I stay in python when doing ai-assisted programming; there's so much training information out there for python and I _generally_ don't care about if its slightly off-idiom, its still probably fine.

replies(1): >>41900097 #
38. atomic128 ◴[] No.41898377{3}[source]
Yes, that's right. See the recent discussion here:

https://news.ycombinator.com/item?id=41860341

Basically, nuclear fission is clean baseload power. Wind and solar are not baseload power sources. They don't really compete. See discussion here: https://news.ycombinator.com/item?id=41858892

Furthermore, we're seeing interest (from Google and Amazon and Dow Chemical) in expensive but completely safe TRISO (HALEU) reactors (https://www.energy.gov/ne/articles/triso-particles-most-robu...). These companies want clean baseload power, with no risk of meltdown, and they're willing to pay for it. Here's what Amazon has chosen: https://x-energy.com/fuel/triso-x

TRISO (HALEU) reactors use more than 1.5 times the natural uranium per unit of energy produced because the higher burnup is offset by higher enrichment inputs (see page 11 at https://fuelcycleoptions.inl.gov/SiteAssets/SitePages/Home/1...), and the fuel is even more expensive to manufacture, but they are completely safe. This is a technology from the 1960's but it's attractive now because so much money is chasing clean baseload nuclear fission for data centers.

These "impossible to melt down" TRISO small modular nuclear fission reactors are what Elon Musk was talking about on the campaign trail last week, when he said:

  ELON MUSK: "The dangers of nuclear power are greatly
  overstated. You can make a nuclear reactor that is
  literally impossible to melt down even if you tried to
  melt it down. You could try to bomb the place, and it
  still wouldn't melt down. There should be no regulatory
  issues with that. There should be significant nuclear 
  reform."
https://x.com/AutismCapital/status/1847452008502219111
replies(1): >>41898598 #
39. jacobgorm ◴[] No.41898436[source]
Railroads and computer networks create network effects, I am not sure the same is true for data centers full of hardware that becomes outdated very quickly.
replies(2): >>41898465 #>>41903847 #
40. CSMastermind ◴[] No.41898465[source]
If they're building new power plants to support all those data centers than that power generation capacity might be put to good use doing something else.
replies(2): >>41903765 #>>41903844 #
41. jiggawatts ◴[] No.41898540[source]
Something people forget is that a training cluster with tens of thousands of GPUs is a general purpose supercomputer also! They can be used for all sorts of numerical modelling codes, not just AI. Protein folding, topology optimisation, route planning, satellite image processing, etc…

We bought a lot of shovels. Even if we don’t find more gold, we can dig holes for industry elsewhere.

replies(1): >>41899552 #
42. danielmarkbruce ◴[] No.41898544{4}[source]
There isn't a rule as to how it plays out. No huge winners in cars, no huge winners in rail. Lots of huge winners in internet.
replies(1): >>41899300 #
43. ViewTrick1002 ◴[] No.41898565{3}[source]
Not at all. Old paid of nuclear plants are competitive but new builds are insanely expensive leading to $140-220/MWh prices for the ratepayers before factoring in grid stability and transmission costs.[1]

The US has zero commercial reactors under construction and this is for one reason: economics.

The recent announcements from the hyperscalers are PPAs. If the company building the reactor can provide power at the agreed price they will take it off their hands. Thus creating a more stable financial environment to get funding.

They are not investing anything on their own. For a recent example NuScale another SMR developer essentially collapsed when their Utah deal fell through when nice renders and PowerPoints met real world costs and deadlines. [2]

[1]: https://www.lazard.com/media/gjyffoqd/lazards-lcoeplus-june-...

[2]: https://iceberg-research.com/2023/10/19/nuscale-power-smr-a-...

replies(2): >>41898819 #>>41899855 #
44. ViewTrick1002 ◴[] No.41898598{4}[source]
> Basically, nuclear fission is clean baseload power. Wind and solar are not baseload power sources. They don't really compete.

This means you don't understand how the grid works. California's baseload is ~15 GW while it peaks at 50 GW.

New built nuclear power is wholly unsuitable for load following duty due to the economics. It is an insane prospect when running at 100% 24/7, and even worse when it has to adapt.

Both nuclear power and renewables need storage, flexibility or other measures to match their inflexibility to the grid.

See the recent study where it was found that nuclear power needs to come down 85% in cost to be competitive with renewables, due to both options requiring dispatchable power to meet the grid load.

> The study finds that investments in flexibility in the electricity supply are needed in both systems due to the constant production pattern of nuclear and the variability of renewable energy sources. However, the scenario with high nuclear implementation is 1.2 billion EUR more expensive annually compared to a scenario only based on renewables, with all systems completely balancing supply and demand across all energy sectors in every hour. For nuclear power to be cost competitive with renewables an investment cost of 1.55 MEUR/MW must be achieved, which is substantially below any cost projection for nuclear power.

https://www.sciencedirect.com/science/article/pii/S030626192...

> These companies want clean baseload power, with no risk of meltdown, and they're willing to pay for it. Here's what Amazon has chosen

The recent announcements from the hyperscalers are PPAs. If the company building the reactor can provide power at the agreed price they will take it off their hands. Thus creating a more stable financial environment to get funding.

They are not investing anything on their own. For a recent example NuScale another SMR developer essentially collapsed when their Utah deal fell through when nice renders and PowerPoints met real world costs and deadlines.

https://iceberg-research.com/2023/10/19/nuscale-power-smr-a-...

> with no risk of meltdown

Then we should be able to remove the enormous subsidy the Price Anderson act adds to the industry right? Let all new reactors buy insurance for a Fukushima level accident in the open market.

Nuclear powerplants are currently insured for ~0.05% of the cost of a Fukushima style accident and pooled together the entire US industry covers less than 5%.

https://en.wikipedia.org/wiki/Price%E2%80%93Anderson_Nuclear...

45. tim333 ◴[] No.41898651[source]
I was an original investor. I still have shirts and shares in it but they could have done better.
46. jackcosgrove ◴[] No.41898712[source]
> One gauge I use to determine if we are still at the beginning of the boom is this

Has your barber/hairdresser recommended you buy NVDA?

replies(2): >>41898939 #>>41903766 #
47. floren ◴[] No.41898819{4}[source]
> leading to $140-220/MWh prices for the ratepayers

I'm on PG&E, I wish I could get my electricity for only $0.14/kWh

replies(1): >>41898926 #
48. thwarted ◴[] No.41898863{5}[source]
An LLM is an application that runs on an operating system like any other application. That the vendor of the operating system has tied it to the operating system is purely a marketing/force-it-onto-your-device/force-it-in-front-of-your-face play. It's forced bundling, just like Microsoft did with Internet Explorer 20 years ago.
replies(1): >>41899134 #
49. reissbaker ◴[] No.41898924{6}[source]
We might be even earlier — the 90s was a famous boom with a fast bust, but to me this feels closer to the dawn of the personal computer in the late 70s and early 80s: we can automate things now that were impossible to automate before. We might have a long time before seeing diminishing returns.
50. ViewTrick1002 ◴[] No.41898926{5}[source]
That cost is excluding grid stability and transmission costs.

From what I’ve understood PG&E’s largest problem is the massive payouts and infrastructure upgrades needed from the wildfires, not the cost of the electricity itself.

replies(1): >>41911465 #
51. arach ◴[] No.41898939{3}[source]
There was an NVDA earnings watch party in NY this summer and Jensen signed some boobs earlier this year. There are some signs but still room to run
52. ben_w ◴[] No.41898979{3}[source]
Uranium, at least the un-enriched kind you can just buy, was never the problem.

Even the peak of that graph (136… er, USD per lb?) is essentially a rounding error compared to everything else.

0.00191 USD/kWh? Something like that, depends on the type of reactor it goes in.

replies(1): >>41899042 #
53. tim333 ◴[] No.41899018[source]
You can never really tell, though following some market tea leaf readers they seem to think a few months from now, after a bit of a run up in the market. Here's one random datapoint, on mentions of "soft landing" in Bloomberg https://x.com/bravosresearch/status/1848047330794385494
replies(1): >>41899100 #
54. mvdtnz ◴[] No.41899040{3}[source]
My workpalce uses Glean and since it was connected to Slack it has become significantly worse. It routinely gives incorrect or VERY incomplete information, misattributes work to developers who may have casually mentioned a project at some time and worst of all presents jokes or sarcastic responses as fact.

Not only is it an extremely poor source of information, it has ruined the company's Slack culture as people are no longer willing to (for lack of a better term) shitpost knowing that their goofy sarcasm will now be presented to Glean users as fact.

replies(2): >>41899457 #>>41906216 #
55. atomic128 ◴[] No.41899042{4}[source]
You are correct. This is one of the advantages of nuclear power.

The fuel is a tiny fraction of the cost of running the plant. See discussion here, contrasting with natural gas: https://news.ycombinator.com/item?id=41858892

It is also important that the fuel is physically small so you can (and typically, do) store years of fuel on-site at the reactor. Nuclear is "secure" in the sense that it can provide "energy security".

replies(1): >>41899398 #
56. aurareturn ◴[] No.41899100{3}[source]
I read through a few pages of tweets from this author and it looks just like another perpetual doomsday pundit akin to Zerohedge.
replies(1): >>41899171 #
57. aurareturn ◴[] No.41899134{6}[source]
I predict that OpenAI will try to circumvent iOS and Android by making their own device. I think it will be similar to Rabbit R1, but not a scam, and a lot more capable.

They recently hired Jony Ive on a project - it could be this.

I think it'll be a long term goal - maybe in 3-4 years, a device similar to the Rabbit R1 would be viable. It's far too early right now.

replies(5): >>41900088 #>>41900923 #>>41900926 #>>41902275 #>>41903291 #
58. tim333 ◴[] No.41899171{4}[source]
Well yeah there may be a bit of that. I find them quite interesting for the data they bring up like the linked tweet but I don't really have an opinion as to whether they are any good at predicting things.

I was thinking re the data in the tweet, that there were a lot of mentions of "soft landing" before the dot com crash, before the 2006 property crash and now, it is quite likely there was an easy money policy preceding them and then government policy mostly focuses on consumer price inflation and unemployment, so they relax when those are both low and then hit the brakes when inflation goes up and then it moderates and things look good similar to now. But that ignores that easy money can also inflate asset prices, eg dot com stocks, houses in 06, or money losing AI companies like now. And then at some point that ends and the speculative asset prices go down rather than up, leaving people thinking oh dear we've borrowed to put loads of money into that dotcom/house/ai thing and now it's not worth much and we still have the debts...

At least that's my guess.

replies(2): >>41901413 #>>41903649 #
59. Terr_ ◴[] No.41899201[source]
> Does Slack sell an LLM chatbot solution that is able to give me reliable answers to business/technical decisions made over the last 2 years in chat?

Note that the presence of such a feature isn't the same as whether it's secure enough for normal use.

In particular, anything anyone said in the last 2 years in chat could poison the LLM into exfiltrating your data or giving false results chosen by the attacker, because of the fundamental problems of LLMs.

https://promptarmor.substack.com/p/data-exfiltration-from-sl...

60. aurareturn ◴[] No.41899221{5}[source]
I think OpenAI has already proven that it's a viable product. Their gross margins must be decent. I doubt they're making a loss for every token they inference.
replies(1): >>41899461 #
61. aurareturn ◴[] No.41899300{5}[source]
There were huge winners in cars. Ford and GM have historically been huge companies. Then oil companies became the biggest companies in the world mostly due to cars.
replies(1): >>41900792 #
62. ben_w ◴[] No.41899398{5}[source]
It would only be an advantage if everything in the power plant else wasn't so expensive.

And I'm saying that as someone who finds all this stuff cool and would like to see it used in international shipping.

replies(1): >>41899493 #
63. dcsan ◴[] No.41899457{4}[source]
Maybe have some off limits to glean shit posting channels?
64. HarHarVeryFunny ◴[] No.41899461{6}[source]
I don't think they've broken out O1 revenue, but it must be very small at the moment since it was only just introduced. Their O1-preview pricing doesn't seem to reflect the exponential compute cost, so perhaps it is not currently priced to be profitable. Overall, across all models and revenue streams, their revenue does exceed inference costs ($4B vs $2B), but they still are projected to lose $5B this year, $14B next year, and not make a profit until 2029 (and only then if they've increased revenue by 100x ...).

Training costs are killing them, and it's obviously not sustainable to keep spending more on research and training than the revenue generated. Training costs are expected to keep growing fast, while revenue per token in/out is plummeting - they need massive inference volume to turn this into a profitable business, and need to pray that this doesn't turn into a commodity business where they are not the low cost producer.

https://x.com/ayooshveda/status/1847352974831489321

https://x.com/Gloraaa_/status/1847872986260341224

replies(1): >>41900349 #
65. atomic128 ◴[] No.41899493{6}[source]
Discussed at length here: https://news.ycombinator.com/item?id=41863388

I already linked this above, twice. I know it's a hassle to read, it's Sunday afternoon, so don't worry about it.

It's not important whether you as an individual get this right or not, as long as society reaches the correct conclusion. Thankfully, we're seeing that happen, a worldwide shift toward the adoption of nuclear power.

Have a pleasant evening!

66. fhdsgbbcaA ◴[] No.41899552[source]
I think there is an LLM bubble for sure, but I’m very bullish on the ease with which one can generate new specialized models for various tasks that are not LLM.

For example, there’s a ton of room for developing all kinds of low latency, highly reliable, embedded classifiers in a number of domains.

It’s not as gee-whiz/sci-fi as an LLM demo, but I think potentially much bigger impact over time.

replies(2): >>41900350 #>>41901612 #
67. fsndz ◴[] No.41899659[source]
Building robust LLM-based applications is token-intensive. You often have to plan for the parsing and digestion of a lot of tokens for summarization or even retrieval augmented generation. Even the mere generation of marketing blogposts consumes a lot of output tokens in most cases. Not to mention that all robust cognitive architectures often rely on the generation of several samples for each prompt, custom retry logics, feedback loops, and reasoning tokens to achieve state of the art performance, all solutions powerfully token-intensive.

Luckily, the cost of intelligence is quickly dropping. GPT-4, one of OpenAI’s most capable models, is now priced at $2.5 per million input tokens and $10 per million output tokens. At its initial release in March 2023, the cost was respectively $10/1M input tokens and $30/1M for output tokens. That’s a huge $7.5/1M input tokens and $20/1M output tokens reduction in price. https://www.lycee.ai/blog/drop-o1-preview-try-this-alternati...

68. Mengkudulangsat ◴[] No.41899790[source]
All those future spare GPUs will make video game streaming dirt cheap.

Even poor people can enjoy 8k gaming on a phone soon.

replies(1): >>41900711 #
69. synergy20 ◴[] No.41899855{4}[source]
Thanks! I always thought it is due to people's safety concerns here instead of economic reasons. After all, nuclear plant is quite 'popular' in Europe, and China too these days.
replies(2): >>41900266 #>>41903154 #
70. marcus_holmes ◴[] No.41900088{7}[source]
Even if this is true (and I'm not saying it's not), they probably won't create their own OS. They'd be smarter to do what Apple did and clone a BSD (or similar) rather than start afresh.
replies(2): >>41901274 #>>41903781 #
71. ryandrake ◴[] No.41900097{5}[source]
Yea, I was thumbs-down on ai-assisted programming because when I tested it out, I tried it by adding things to my existing C and C++ projects, and its suggestions were... kind of wild. Then, a few months later I gave it another chance when I was writing some Python and was impressed. Finally, I used it on a new-from-blank-text-file Rust project and was pretty much blown away.
replies(4): >>41900253 #>>41900255 #>>41900878 #>>41901107 #
72. Eisenstein ◴[] No.41900123{4}[source]
Considering that models have been getting more powerful for the same number of parameters -- all of it.
replies(1): >>41903853 #
73. yowayb ◴[] No.41900207[source]
That may be true now, but Europe has had a resurgence in rail usage growth recently, so I'm not sure it's true forever.
74. ffujdefvjg ◴[] No.41900253{6}[source]
As someone who doesn't generally program, it was pretty good at getting me an init.lua set up for nvim with a bunch of plugins and some functions that would have taken me ages to do by hand. That said...it still took a day or two of working with it and troubleshooting everything, and while it's been reliable so far, I worry that it's not exactly idiomatic. I don't know enough to really say.

What it's really good at is taking my description of something and pointing me in the right direction to do my own research.

(two things that helped me with getting decent code were to describe the problem and desired solution, followed by a "Does that make sense?". This seems to get it to restate the problem itself and produce better solutions. The other thing was to copy the output into a fresh session, ask for a description of what the code does and what improvements could be made)

replies(2): >>41900331 #>>41900420 #
75. rayxi271828 ◴[] No.41900255{6}[source]
Wouldn't AI be worse at Rust than at C++ given the amount of code available in the respective languages?
replies(1): >>41900497 #
76. dalyons ◴[] No.41900266{5}[source]
It’s not popular in Europe at all.
replies(1): >>41900320 #
77. datavirtue ◴[] No.41900309[source]
I'm investing in undervalued businesses who are selling the shovels.
replies(1): >>41900537 #
78. synergy20 ◴[] No.41900320{6}[source]
France derives about 70% of its electricity from nuclear energy.

For Europe overall is 22%.

replies(1): >>41903810 #
79. skydhash ◴[] No.41900331{7}[source]
Not saying that it’s a better way, but I started with vim by copying someone conf (on Github), removing all extraneous stuff, then slowly familiarizing myself with the rest. Then it was a matter of reading the docs when I wanted some configuration. I believe the first part is faster than dealing with an LLM, especially when dealing with an unfamiliar software.
replies(1): >>41900535 #
80. nl ◴[] No.41900349{7}[source]
The thing is that OpenAI can choose to spend less on training at any time.

We've seen this before, with for example Amazon where they made a deliberate effort to avoid profitability by spending as much as possible on infrastructure until the revenue became some much they couldn't spend it.

Being in a position where you are highly cash-flow positive and it's strategic investment that is the cost seems like a good position.

replies(1): >>41903153 #
81. datavirtue ◴[] No.41900350{3}[source]
Spot on.
82. komali2 ◴[] No.41900420{7}[source]
The downside of this nvim solution is the same downside as both pasting big blobs of ai code into a repo, and, pasting big vim configs you find online into your vimrc: inability to explain the pasted code.

When you need something fast for whatever reason sure, but later when you want to tweak or add something, you'll have to finally sit down and learn basically the whole thing or at least a major part of it to do so anyway. Imo it's better to do that from the start but sometimes that's not always ideal.

replies(1): >>41901480 #
83. reverius42 ◴[] No.41900497{7}[source]
Maybe this is a case where more training data isn’t better. There is probably a lot of bad/old C++ out there in addition to new/modern C++, compared to Rust which is relatively all modern.
replies(1): >>41904660 #
84. ffujdefvjg ◴[] No.41900535{8}[source]
I agree with this approach generally, but I needed to use some lua plugins to do something specific fairly quickly, and didn't feel like messing around with it for weeks on end to get it just right.
85. 0xDEAFBEAD ◴[] No.41900537[source]
That's the funny thing about the AI boom. There's way more hype about shovel sellers than shovel buyers.

No one is getting excited about "AI slop", just the models that generate it. Funny situation.

replies(2): >>41902203 #>>41908970 #
86. lelanthran ◴[] No.41900633[source]
> Reading this makes me willing to bet that this capital intensive investment boom will be similar to other enormous capital investment booms in US history, such as the laying of the railroads in the 1800s, the proliferation of car companies in the early 1900s, and the telecom fiber boom in the late 1900s. In all of these cases there was an enormous infrastructure (over) build out, followed by a crash where nearly all the companies in the industry ended up in bankruptcy, but then that original infrastructure build out had huge benefits for the economy and society as that infrastructure was "soaked up" in the subsequent years.

> I think that will happen here.

Why? The rail network, road network and fiber network that was laid could be used for decades after their original investors went bust.

The current datacenters full of AI compute can't really be used for anything else if AI companies go bust.

That's the problem with investing in compute infrastructure - you need to have a plan to use it all up in the next 5 years, because after that you wouldn't even be able to give it away.

replies(3): >>41900917 #>>41902633 #>>41903793 #
87. shaklee3 ◴[] No.41900711{3}[source]
Most data center GPUs do not have game rendering hardware in them.
88. danielmarkbruce ◴[] No.41900792{6}[source]
GM went bankrupt. Ford would have without government intervention. Each have had periods of profitability but they weren't ever anything like microsoft/google etc. Ford has underperformed the stock market average since it went public like 70 odd years ago. GM got so big in the first place via acquisitions, not because the business of cars lent itself to a dominant player.

Huge by itself isn't the same as huge winner.

replies(2): >>41902450 #>>41903849 #
89. _huayra_ ◴[] No.41900878{6}[source]
The best I have ever seen were obscure languages with very strong type safety. Some researcher at a sibling org to my own told me to try it with the Lean language, and it basically gave flawless suggestions.

I'm guessing this is because the only training material was blogs from uber-nerdy CS researchers on a language where "mistakes" are basically impossible to write, and not a bunch of people flailing on forums asking about hello world-ish stuff and segfaulting examples.

90. margalabargala ◴[] No.41900917[source]
The renewable power infrastructure for those datacenters will still exist.

People will be able to buy those used GPUs cheap and run small local LLMs perhaps. A 10 year old computer today won't do state of the art games or run models, but is entirely acceptable for moderate computing use.

replies(3): >>41901582 #>>41902523 #>>41903756 #
91. ◴[] No.41900923{7}[source]
92. tightbookkeeper ◴[] No.41900926{7}[source]
I’m not even sure if they can make a website that takes text input to an executable and dumps the output.
93. fragmede ◴[] No.41901107{6}[source]
My data science friend tells me it's really good at writing bad pandas code because it's seen so much bad pandas code.

At the end of the day, it depends where you are in the hierarchy. Having it write code for me on a hobby project in react that's bad but works is one thing. I'm having a lot of fun with that. Having it write bad code for me professionally is another thing though. Either way, there's no going back to before ChatGPT, just like there's no going back to before Stack Overflow or Google. Or the Internet.

94. aurareturn ◴[] No.41901274{8}[source]
The LLM would become the OS.
replies(2): >>41901726 #>>41902566 #
95. throwaway2037 ◴[] No.41901413{5}[source]

    > 2006 property crash
Assuming you are talking about the US financial crisis, do you mean 2008, instead of 2006? As I recall, easy money (via mortgages) was still sloshing about, well into 2007.
replies(1): >>41901652 #
96. sofixa ◴[] No.41901466{3}[source]
> Glean.com does it for the enterprise I work at: It consumes all of our knowledge sources including Slack, Google docs, wiki, source code and provides answers to complex specific questions in a way that’s downright magical

There are a few other companies in this space (and it's not something that complex to DIY either); the issue is data quality. If your Google Docs and wikis contain obsolete information (because nobody updated them), it's just going to be shit in, shit out. Curating the input data is the challenging part.

97. shwaj ◴[] No.41901480{8}[source]
When I’ve used AI for writing shell scripts it used a lot of syntax that I couldn’t understand. So then I took the time to ask it to walk me through the parts that I didn’t understand. This took longer than blindly pasting what it generated, but still less time than it would have using search to learn to write my own script. With search, a lot of time is spent guessing the right search term. With chat, assuming it generated a reasonable answer (I know: a big assumption!), my follow-up questions can directly reference aspects of the generated code.
replies(1): >>41902311 #
98. flakeoil ◴[] No.41901582{3}[source]
But in terms of compute/watt those 10 year old data centers are going to suck and that is what counts for a data center.
99. jiggawatts ◴[] No.41901612{3}[source]
Agreed! One thing I noticed is that the LLM craze seems to have triggered some other developments in only vaguely related fields.

My favourite example is the astonishing pace with which reverse-rendering technology has progressed. It started with a paper by NVIDIA showing projections of 2D photos being "fitted" into a 3D volume of differentiable hashtables, and then the whole thing exploded when Guassian Splats were invented. I full expect this niche all by itself to generate a huge variety of practical applications. Computer games and movie special effects, obviously, but also AR/VR, industrial uses, mapping, drone navigation, etc...

100. tim333 ◴[] No.41901652{6}[source]
Yeah that one.
101. marcus_holmes ◴[] No.41901726{9}[source]
An LLM cannot "become" an OS. It can have an OS added to it, for sure, but that's a different thing. LLMs run on top of a software stack that runs on top of an OS. Incorporating that whole stack into a single binary does not mean it "becomes" an OS.

And the point stands: you would not write a new OS, even to incorporate it into your LLM. You'd clone a BSD (or similar) and start there.

replies(1): >>41902470 #
102. tim333 ◴[] No.41902203{3}[source]
The hypothesis is the AI slop will improve. A bit like 1990s internet where there was a lot of futzing with dial up modems to eventually get a fairly crappy web page. But you could tell it would get better.
103. vrighter ◴[] No.41902275{7}[source]
even then, the llm cannot possibly be a standalone os. For one thing, it cannot execute loops. So even something as simple as enumerating hardware at startup is impossible.
104. vrighter ◴[] No.41902311{9}[source]
having something explained to me has never helped me retain the information. That only happens if i spend the time actually figuring out stuff myself.
105. aurareturn ◴[] No.41902450{7}[source]
That's recent. Ford was founded in 1903. GM in 1908.

GM was America's largest employer as recently as the 90s.

replies(1): >>41906474 #
106. aurareturn ◴[] No.41902470{10}[source]
I don't think you're getting the main point. The only application that this physical device would run is ChatGPT (or some successor). You won't be able to install other apps on it like a normal OS. Everything you do is inside this LLM.

Underneath, it can be Linux, BSD, Unix, or nothing at all, whatever. It doesn't matter. That's not important.

OS was just a convenient phrase to describe this idea.

replies(2): >>41903237 #>>41910102 #
107. lelanthran ◴[] No.41902523{3}[source]
> People will be able to buy those used GPUs cheap and run small local LLMs perhaps.

Maybe; I find it unlikely though, because unlike CPUs, there's a large difference in compute/watt in subsequent generations of GPUs.[1]

I would imagine that, from an economics PoV, the payback for using a newer generation GPU over a previous generation GPU in terms of energy usage is going to be on the order of months, not years, so anyone needing compute for more than a month or two would save money by buying a new one at knockdown prices (because the market collapsed) than by getting old ones for free (because the market collapsed).

[1] Or maybe I am wrong about this - maybe each new generation is only slightly better than the previous one

108. glimshe ◴[] No.41902566{9}[source]
The LLM can't abstract PCI, USB, SATA etc from itself.
replies(1): >>41904964 #
109. matthewdgreen ◴[] No.41902626{5}[source]
I cannot tell you how much this echoes what people were saying during the dot com days :) Of course back then it was browsers and not LLMs. Looking back, people were both correct about this, yet we’re still having the same conversation about replacing the OS cartel.
110. HarHarVeryFunny ◴[] No.41903153{8}[source]
I don't know how you can compare Amazon vs OpenAI on the fundamentals of the two businesses. It's the difference in fundamentals that made Amazon a buy at absurd P/Es, as well as some degree of luck in AWS becoming so profitable, while OpenAI IMO seems much more of a dodgy value proposition.

Amazon were reinvesting and building scale, breadth and efficiency that has become an effective moat. How do you compete with Amazon Prime free delivery without your own delivery fleet, and how do you build that without the scale of operations?

OpenAI don't appear to have any moat, don't own their own datacenters, and the datacenters they are using are running on expensive NVIDIA chips. Compare to Google with their own datacenters and TPUs, Amazon with own datacenters and chips (Graviton), Meta with own datacenters (providing value to their core business) and chips - and giving away the product for free despite spending billions on it ... If this turns into the commodity business that it appears it may (all frontier models converging in performance), then OpenAI would seem to be in trouble.

Of course OpenAI could stop training at any time, but to extent that there is further performance to be had from further scaling and training, then they will be left behind by the likes of Meta who have a thriving core business to fund continued investment and are not dependent on revenue directly from AI.

111. ViewTrick1002 ◴[] No.41903154{5}[source]
We built a lot of nuclear back in the 70s and 80s which we still rely on with long term operation upgrades.

For modern nuclear power the only nuclear reactor under construction in France is Flamanville 3 which is 6x over budget and 12 years late on a 6 year construction timeline.

Hinkley Point C in the UK is in a similar quagmire and Olkiluoto 3 finally got finished last year after a near 20 year construction timeline.

Politically there's some noise from conservative politicians who can't hold a climate change denial position anymore, but still need to be contrarians.

The problem is the horrendous economics.

112. James_K ◴[] No.41903200[source]
My thoughts exactly. One of the best ways to develop countries is to just invest in a bunch of infrastructure. It probably won't be optimal, but it's better than not investing. It's interesting that the private bubbles in this case form a simulacrum of public investment. Instead of the government raising money to invest through taxes, capitalists just throw it away on fads. Perhaps it even helps to address inequality in the long run.

That said, I'm not sure the effect of digital infrastructure will be the same as physical infrastructure. A road has a clear material impact on all businesses in the area and their capacity to produce physical goods. But do data centres have the same effect? An extra lane on the road means you can get a greater volume of goods in and out to broaden operations to a larger area, but I don't see what positive effect two data centres could have on the average business. For as great as the internet is, I don't know how much value is created here. The question of what to do with a railroad is quite easily answered, but I'm not really sure what you can do with a datacentre. I guess whoever works it out will be decently rich.

But I feel we already have enough computing power, and the bottleneck in the whole process is making software that efficiently uses it (or knowledge of how to operate such software), rather than the power of devices themselves. Though perhaps as the bubble bursts, the price of programmers will also decrease significantly and the software issue will be resolved.

113. guitarlimeo ◴[] No.41903237{11}[source]
I got your main point from the first message, but still don't like redefining terminology like OS to mean what you did.
replies(2): >>41904202 #>>41904933 #
114. simonh ◴[] No.41903291{7}[source]
This is a similar situation to the view that the web would replace operating systems. All we'd need is a browser.

I don't think AI is ultimately even an application, it's a feature we will use in applications.

replies(1): >>41904880 #
115. btbuildem ◴[] No.41903363[source]
Of course it's similar: they're all instances of a capitalist "boom -> boost -> quit" cycle. At least this frontier is mostly digital in terms of resources being destroyed.

Maybe we will get a nuclear energy renaissance out of this, who knows.

116. wkat4242 ◴[] No.41903416[source]
Yeah this is the thing. Investors are only looking for the next bitcoin these days. Something that will pay back 10.000-fold in 2 years. They're cheering each other on and working themselves up drooling over the imagined profits. They no longer have any long-term vision.

If it doesn't meet those sky-high expectations it's a flop.

The same happened with metaverse, blockchain etc. Those technologies are kinda shitcanned now which is unfair too because they have excellent usecases where they add value. It was never going to be for everyone, and no we weren't all going to run around with an oculus quest 24/7.

I think these investors break it more than they do good.

117. rsynnott ◴[] No.41903550[source]
> Does Slack sell an LLM chatbot solution that is able to give me reliable answers to business/technical decisions made over the last 2 years in chat? We don't have this yet - most likely because it's probably still too expensive to do this much inference with such high context window.

So, your problem there is 'reliable'. LLMs, fairly fundamentally, cannot do 'reliable'. If you're looking for reliable, you likely are looking at a different tech entirely.

replies(1): >>41903843 #
118. rsynnott ◴[] No.41903649{5}[source]
> I was thinking re the data in the tweet, that there were a lot of mentions of "soft landing" before the dot com crash, before the 2006 property crash and now

There's a confirmation bias there, though. Economists, particularly pop economists, have predicted all 20 of the last two recessions; if you just say "the world is going to end" every year, then occasionally it kinda will, and certain people will think you're a visionary.

replies(1): >>41908463 #
119. rsynnott ◴[] No.41903672[source]
> but nuclear power, being expensive, would not be built if this bubble had not happened, and somehow nuclear does seem to be getting some of this money

Eh, it's generally SMRs, which remain kinda vapourware-y. I'd be a little surprised if anything concrete comes of it, tbh; I suspect that it is mostly PR cover for reactivating coal plants and the like. (The one possibly-real thing might be the Three Mile Island restart, but that in itself isn't particularly consequential.)

120. johnnyanmac ◴[] No.41903756{3}[source]
>People will be able to buy those used GPUs cheap and run small local LLMs perhaps.

That's not really how SaaS works these days. It will be a "cheap" subscription or expensive and focused on enterprise. Both those require maintance costs, which ruin the point of "cheap and small run LLM's".

And they sure aren't going to sell local copies. They'd rather go down with their ship than risk hackers dissecting the black box.

replies(1): >>41905815 #
121. johnnyanmac ◴[] No.41903765{3}[source]
Lots of ifs going on here. I haven't been so optimistic of tech this decade compared to the early '10's.
122. JKCalhoun ◴[] No.41903766{3}[source]
Doesn't that say more about Nvidia than it does about AI in general?

But I see your point. And yes, I think it has trickled down to the mainstream.

(By that metric, I guess Bitcoin crashed a few years ago.)

123. whywhywhywhy ◴[] No.41903781{8}[source]
Would be extremely surprising if it were anything other than an Android fork. The differentiator is gonna be the LLM, always on listening and the physical interface to it.

You're just burning money bothering to rewrite the rest of the stack when off the shelf will save you years.

124. JKCalhoun ◴[] No.41903793[source]
> The current datacenters full of AI compute can't really be used for anything else if AI companies go bust.

That's hard to know from this vantage point in the present.

Who knows what ideas will spring forth when there are all these AI-capable data-centers sitting out there on the cheap.

replies(1): >>41907055 #
125. johnnyanmac ◴[] No.41903794{3}[source]
That's my big worry. The Internet was made with an idea of being a common. LLM's are very much built with mentalities of trade secrets, from their data aquisition to the algorithms. I don't think such techniques will proliferate for commercial use as easily when the bubble bursts.
126. rsynnott ◴[] No.41903810{7}[source]
That French capacity was largely built a long time ago, though. Only a couple of nuclear plants have been built on Europe in the last decade, and they've generally overrun _horribly_ on costs.
127. bamboozled ◴[] No.41903838[source]
I was thinking about it today, it's absolutely wild we're building nuclear to fuel this boom alone. If it doesn't pan out as we expect, what is to happen to all this nuclear investment ? Sounds positive to me.
128. williamcotton ◴[] No.41903843{3}[source]
If an LLM is tasked with translating a full text to a summary of that text then it is very reliable.

This is akin to an analytic statement, eg, "all bachelors are married". The truth is completely within the definition of the statement. Compare this to a synthetic statement such as "it is raining outside". In this case the truth is contingent on facts outside of the statement itself.

When LLMs are faced with an analytic statement they are more reliable. When they are faced with a synthetic statement they are prone to confabulate and are unreliable.

replies(1): >>41904075 #
129. gloflo ◴[] No.41903844{3}[source]
As global society we must strive to reduce energy consumption, not to find new use cases for burning more energy. Our planet has limited resources.
130. SJC_Hacker ◴[] No.41903847[source]
The data centers themselves with all the supporting infrastructure (telecom/power), as well as all the chip fabs needed to build the hardware, even if the hardware itself becomes obsolete / breaks down on a 4-5 year time scale.

In the same way the rights of way obtained with all the railroads were, even if the rails / engines themselves had to be replaced every decade or so

But some hardware does last quite a long while. Fiber laid from 25 years ago is still pretty useful.

replies(1): >>41903858 #
131. johnnyanmac ◴[] No.41903849{7}[source]
>GM went bankrupt.

I'd call an 80 year run pretty damn good. having my company not just survive me and my children, but dominate an industry for that time seems like a good deal. It shows it wasn't my fault it failed.

>the business of cars lent itself to a dominant player.

I'd rather measure my business by impact, not stock numbers. That mentality is exactly why GM fell very slowly through the 70's (defying a while bunch of strategies Ford implemented to beat out the compeition. like knowledge retention ) and crashed by the 90's.

Money to keep operating is important too, but I don't think Ford lived a life of Picasso here.

replies(1): >>41906460 #
132. llamaimperative ◴[] No.41903853{5}[source]
That... is not relevant. The question is what percentage of R&D spend gets "encoded" into something that can survive the dissolution of its holding company and how much does a transfer to a new owner depreciate it.

I'd be shocked if more than like 20% of the VC money going into it would come out the other side during such an event.

133. ◴[] No.41903858{3}[source]
134. johnnyanmac ◴[] No.41903883{3}[source]
Every sector has its story like that. bankruptcy for a huge business isn't the same as an individual doing it.

And yea, it will vary. Amazon crashed hard on stocks through the 2000's. Google completely thrived. they are still considered on the same standing today as a trillionaire tech company

135. irunmyownemail ◴[] No.41903917[source]
A few thoughts, Netflix works fine over an old, slow DSL connection. Fossil fuels aren't going away this century, especially not with power hungry AI (setting aside discussion on whether AI is truly worth it).
replies(1): >>41904484 #
136. rsynnott ◴[] No.41904075{4}[source]
> If an LLM is tasked with translating a full text to a summary of that text then it is very reliable.

Hrm. I've found LLM summaries to be of... dubious reliability. When someone posts an article on this here orange website, these days someone will sometimes 'helpfully' post a summary generated by a magic robot. Have a look at these, sometime. They _often_ leave out key details, and sometimes outright make shit up.

Interesting article on someones' experiences with this recently: https://ea.rna.nl/2024/05/27/when-chatgpt-summarises-it-actu...

replies(1): >>41904564 #
137. aurareturn ◴[] No.41904202{12}[source]
Think of iOS and everything that it does such as downloading apps, opening apps, etc. Replace all of that with ChatGPT.

No need to get to the technicals such as whether it's UNIX or Linux talking to the hardware.

Just from a pure user experience standpoint, OpenAI would become iOS.

138. dash2 ◴[] No.41904320{5}[source]
Good comment. From Apple's point of view, AI could be a disruptive innovation: they've spent billions making extremely user-friendly interfaces, but that could become irrelevant if I can just ask my device questions.

But I think there will be a long period when people want both the traditional UI with buttons and sliders, and the AI that can do what you ask. (Analogy with phone keyboards where you can either speech-to-text, or slide to type, or type individual letters, or mix all three.)

139. baby_souffle ◴[] No.41904484[source]
> A few thoughts, Netflix works fine over an old, slow DSL connection.

debatable depending on what "fine" means. In any case, DSL really doesn't go far.

The old version from the early 2000's could work out to a few miles / km but only with absolutely perfect condition copper. The newer DSL versions are limited to much less distance even with good quality cable.

Each neighborhood has a head-end that does the copper <-> fiber transition. Unless you lived _really_ close to the Central Office, your DSL service was probably copper only for a few blocks before it transitioned to fiber going from TelCo central office to all the individual DSLAMs scattered about.

140. williamcotton ◴[] No.41904564{5}[source]
Sure, anecdotal evidence. Here's another anecdote:

Original article: https://osa1.net/posts/2024-10-09-oop-good.html

LLM (ChatGPT o1-preview) results: https://chatgpt.com/share/67166301-00d8-8013-9cf5-e8a980aca7...

LGTM!

I'd like to know which model is used in the article you've referenced as well as the prompt. I also suspect that 50 pages is pushing up to the limits of the context window and has an impact on the results.

---

As for the article itself... Use a language like F# or OCaml and you get a functional-first language that also supports OOP!

141. 7thaccount ◴[] No.41904566[source]
The amount of power needed for data centers over the next decade is estimated at like 80+ GW of growth by 2030 which is insane.

It'll either prompt serious investment in small modular reactors and rescuing older nuke plants about to retire, or we'll see a massive build out in gas. These companies want the power to be carbon free, so they're trying to do the former, but we'll see how practical that is. Small modular reactors are still pretty new and nobody knows how successful that will be.

At the end of the day, I feel like this will all crash and burn, but we may end up with some kind of nuclear renaissance. We're also expanding the transmission grid and building more wind, solar, and storage. However, I don't think that alone is going to satisfy the needs of these data centers that want to run nearly 24/7.

replies(1): >>41904853 #
142. ryandrake ◴[] No.41904660{8}[source]
Yes, I think that's it. There is a lot of horrible C++ code out there, especially on StackOverflow where "this compiled for me" sometimes ends up being the accepted answer. There are also a lot of ways to use C++ poorly/wrong without even knowing it.
143. n_ary ◴[] No.41904788[source]
I believe, what we will see in new few years is the complete(or nearly) abolision of all human friendly customer support and everything will be ChatBot or voice-chat-bot based support to reduce cost of service.

We will also get some nice things, like more intelligent IDE at affordable cost, think CursorAi costs $20/month(240/year), while whole JetBrain's package costs only 25/month(290/year).

However, I am a bit worried about all these data center and AI and energy use/scaling. While consumers are being pushed to more and more efficient energy usage and energy prices are definitely high(to what I would expect with massive renewable energy production), large corps and such will continue scaling energy usage higher and higher.

Also, the AI fad will eventually spook out a lot of free knowledge sharing on the open and everything will get behind paywall, so random poor kid in some poor country will no longer have access to some nice tutorial or documentations online to learn cool stuff because in some countries, what we call a price of "morning coffee" is actually could be a day's earning of an adult but not for non-privileged people. Without ability to pay for AI services, no more access to knowledge. Search engines will eventually drown in slop, I mean even google now frequently gives me "no resoult found" page and I need to use ddg/brave/bing to fish out some results still.

144. Kon-Peki ◴[] No.41904853[source]
Many data center tax breaks have carbon-free energy requirements (to varying degrees). If the pace of building data centers exceeds the ability to provide carbon-free energy, you may see a shift in the location of data centers, away from locations with good incentives and toward locations with the availability of energy regardless of its source.
145. gpderetta ◴[] No.41904880{8}[source]
> This is a similar situation to the view that the web would replace operating systems. All we'd need is a browser.

well, that's not a false statement. As much as I might dislike it, the raise of the web and web applications have made the OS themselves irrelevant for a significant number for tasks.

146. ogogmad ◴[] No.41904933{12}[source]
I don't think "OS" means anything definitive. It's not 1960. Nowadays, it's a thousand separate things stuck together.
147. ogogmad ◴[] No.41904964{10}[source]
What counts as an OS is subjective. The concept has always been a growing snowball.
148. sixhobbits ◴[] No.41905320[source]
I've been calling a crash for far too long so take with a pinch of salt BUT I think another four years of this is very unlikely.

1996 - Cisco was 23.4B or 0.3% of US GDP

2000 - Cisco peaked at 536B or 5.2% of US GDP

2020 - Nvidia was 144B or 0.7% of US GDP

2024 - Nvidia is 3.4T or 11.9% of US GDP

Numbers very rough and from different sources, but I'd be surprised if Nvidia doesn't pop within 1-2 years at most.

replies(3): >>41905411 #>>41905869 #>>41905954 #
149. dragontamer ◴[] No.41905411{3}[source]
Rumor is that there just isn't enough power to turn on all those fancy AI accelerators or Datacenters.

There's a reason Microsoft just outright purchased the entire output of 3-mile island (a full sized nuclear power plant).

At some point, people will stop buying GPUs because we've simply run out of power.

My only hope is that we don't suffer some kind of ill effect (ex: double the cost of consumer electricity or something, or local municipalities going bankrupt due to rising energy costs). The AI boom has so much money in it we need to account for tail wags dog effects.

replies(1): >>41905600 #
150. shikon7 ◴[] No.41905600{4}[source]
In a working market, we won't run out of power, but power becomes so expensive that it's no longer viable to use most of the GPUs. The boom shifts to power generation instead, and we will have a similar article "The power investment boom", where people will debate that we will stop building power plants because we've simply run out of GPUs to use that power.
replies(1): >>41905650 #
151. SoftTalker ◴[] No.41905630[source]
Except railroads, manufacturing plants, telecom fiber and those prior build-outs were for infrastructures that have useful lifetimes measured in decades.

Computing infrastrucuture that's even one decade old is essentially obsolete. Even desktop PCs are often life-cycled in 5 years, servers often the same.

If it takes AI a decade to find its way, most of today's investment won't be useful at that point.

replies(1): >>41905694 #
152. SoftTalker ◴[] No.41905650{5}[source]
Fortunately we're also electrifying transportation so there will be no shortage of demand for electrical power generation.
153. SoftTalker ◴[] No.41905687{4}[source]
Companies are going to have to do a lot less gatekeeping and siloing of data for this to really work. The companies that are totally transparent even internally are few and far between in my experience.
154. fizx ◴[] No.41905694[source]
Don't think of an H100. Think of the factories, the tooling, the datacenters and the power supply needed to light one up.
155. TeaBrain ◴[] No.41905766{3}[source]
>There were hundreds of failed automotive companies

What companies are you referring to?

156. TeaBrain ◴[] No.41905774{3}[source]
That's one automobile company. The parent mentioned "hundreds".
157. jbs789 ◴[] No.41905809[source]
I think this is the reasoned answer.

It’s interesting to observe who is making the counterpoint - it’s often very vocal fundraisers.

Of course you can argue they are raising because they believe, and I don’t (necessarily) doubt that in all cases.

158. margalabargala ◴[] No.41905815{4}[source]
Exactly. SaaS does not enter into it.

People will locally run the open models which are freely released, just like they do today with Llama and Whisper.

Most of the AI SaaS companies won't be around to have anything to say about it, because they will be casualties of the bust that will follow the boom. There will be a few survivors with really excellent models, and some people will pay for those, while many others simply use the good-enough freely available ones.

159. syndicatedjelly ◴[] No.41905869{3}[source]
Comparing a company’s market cap to the US GDP makes no sense to me. The former is the product of shares and stock price. The latter is total financial output of a country. What intuition is that supposed to provide?
replies(1): >>41906259 #
160. pirate787 ◴[] No.41905954{3}[source]
This comparison is silly. First of all, Cisco's scale was assembled through acquisitions, and hardware is a commodity business. Nvidia has largely grown organically and has CUDA software as a unique differentiator.

More importantly, Cisco's margins and PE were much higher than Nvidia's today.

You should use actual financial measures and not GDP national accounts which have zero bearing on business valuation.

replies(1): >>41906134 #
161. TacticalCoder ◴[] No.41906134{4}[source]
I don't think GP's comparison is as silly as you think. People thinking about "money" take many different numbers, from a shitload of source, into account.

There's a relation between P/E and future actual revenues of a company.

Imagine that a similar comparison would imply that it's projected that in a few years NVidia's revenues shall represent 10% of the US's GDP: do we really believe that's going to happen?

The Mag 7 + Broadcom have a market cap that is now 60% of the US's GDP. I know you think it's silly but... Doesn't that say something about the expect revenues of these companies in a few years?

Do we really think the Mag 7 + Broadcom (just an example) are really to represent the % of the actual US's GDP that that implies?

Just to be clear: I'm not saying it implies the percentage of the US GDP of these 8 companies alone is going to be 60% but there is a relation between the P/E of a company and its expected revenues. And revenues of companies do participate in the GDP computation.

I don't think it's as silly as several here think.

I also don't think GP should be downvoted: if we disagree, we can discuss it.

162. jordanb ◴[] No.41906189[source]
The only asset from the telcom bubble that was still valuable was fiber in the ground. All the endpoints were obsolete and had to be replaced within a few years. The fiber could be reused and installing it was expensive, so that was the main asset.

What asset from the AI bubble will still be valuable 5 years later? Probably not any warehouses full of 5 year old GPUs. Maybe nuclear power plants?

replies(1): >>41911471 #
163. rendang ◴[] No.41906216{4}[source]
Interesting. I still find it to be a net positive, but it is amusing when I ask it about a project and the source cited is a Slack thread I wrote 2 days prior
164. rendang ◴[] No.41906259{4}[source]
Comparing to total household wealth would be better (about $140T now, about $40T in 2000)
165. danielmarkbruce ◴[] No.41906460{8}[source]
Yup, it's pretty good.

It's just not a huge winner. Many industries don't work that way, there are no "huge winners" even if there are some companies that are huge. Oil & gas doesn't really have "huge winners". The huge companies are a result of huge amounts of capital being put to work.

166. danielmarkbruce ◴[] No.41906474{8}[source]
Largest employer is a strange way to describe a huge winner.
167. tivert ◴[] No.41907055{3}[source]
> Who knows what ideas will spring forth when there are all these AI-capable data-centers sitting out there on the cheap.

You still have to pay for power to run them. A lot of power. It won't be that cheap.

168. tim333 ◴[] No.41908463{6}[source]
True. Market/economic forecasting is quite unreliable, partly because you're trying to predict human behaviour which is changeable.
169. shadowmanifold ◴[] No.41908970{3}[source]
It is rather comical.

Everyone knows that panning for gold is a fools game so we have a gold pan/shovel bubble.

It is like having a massive lumber bubble and calling it a real estate bubble because someday we might actually build those houses.

170. marcus_holmes ◴[] No.41910102{11}[source]
I think what you mean is "Desktop" not "OS". You're just replacing all the windows, menus and buttons with a chat interface.
171. RF_Savage ◴[] No.41911465{6}[source]
Ain't that a direct result of them not investing in their infra for decades and all that technical debt catching up to them?
172. RF_Savage ◴[] No.41911471[source]
I doubt they will manage to do a significant nuclear buildout in mere five years. But future renevables will potentially benefit from the stronger grid.