https://www.anthropic.com/news/updates-to-our-consumer-terms
That was true when the tech leadership was an open question and it seemed like any one of the big players could make a breakthrough at any moment that would propel them to the top. Nowadays it has pattered out and the market is all about sustainable user growth. In that sense Anthropic is pretty overvalued, at least if you think that OpenAI's valuation is legit. And if you think OpenAI is overvalued, then Anthropic would be a no-go zone as an investor.
As the years go by, I'm finding myself being able to rely on those less and less, because every time I do, I eventually get disappointed by them working against their user base.
Meta downloaded copyrighted content and trained their models on it, OpenAI did the same.
Uber developed Greyball to cheat the officials and break the law.
Tesla deletes accident data and reports to the authorities they don't have it.
So forgive me I have zero trust in whatever these companies say.
I don't think we should be so quick to dismiss the holes LLMs are fulfilling as unnecessary. The only thing "necessary" is food water and shelter by some measures.
None. And even if it's the nicest goody two shoes company in the history of capitalism, the NSA will have your data and then there'll be a breach and then Russian cyber criminals will have it too.
At this point I'm with you on the zero trust: we should be shouting loud and clear to everyone, if you put data into a web browser or app, that data will at some point be sold for profit without any say so from you.
For me this been a pretty fundamental shift, where before I either had to figure out another way so I can move on, or had to spend weeks writing one function after learning the needed math, and now it can take me 10-30 minutes to nail perfectly.
The solution is to break up monopolies....
If you don’t take companies at their word, you need to be consistent about it.
Where did these companies claim they didn’t do this?
Even websites can be covered by copyright. It has always been known that they trained on copyrighted content. The output is considered derivative and therefore it’s not illegal.
And the underrated comparison was more towards the fact that I couldn't believe scaleAi's questionable accquisition by facebook and I still remember the conversation me and my brother were having which was, why doesn't facebook pay 2x, 3x the price of anthropic but buy anthropic instead of scaleAI itself
well I think the answer my brother told was that meta could buy it but anthropic is just not selling it
I don't own a car and only take public transit or bike. I fill my transit card with cash. I buy food in cash from the farmer's morning market. My tv isn't connected to the Internet, it's connected to a raspberry pi which is connected to my home lab running jellyfin and a YouTube archiving software. I de Googled and use an old used phone and foss apps.
It's all happened so gradually I didn't even realize how far I'd gone!
If your threat model is to unconditionally not trust the companies, what they're saying is irrelevant. Which is fair enough, you probably should not be using a service you don't trust at all. But there's not much of a discussion to be had when you can just assert that everything they say is a lie.
> Meta downloaded copyrighted content and trained their models on it, OpenAI did the same.
> Uber developed Greyball to cheat the officials and break the law.
These seem like randomly chosen generic grievances, not examples of companies making promises in their privacy policy (or similar) and breaking them. Am I missing some connection?
1. Anthropic reverses privacy stance, will train on Claude chats
3. Gun Maker Sig Sauer Citing National Security to Keep Documents from Public
4. Tesla said it didn't have key data in a fatal crash. Then a hacker found it
6. Meta might be secretly scanning your phone's camera roll
7. If you have a Claude account, they're going to train on your data moving forward
8. Ask HN: The government of my country blocked VPN access. What should I use?
> If you’re an existing user, you have until September 28, 2025 to accept the updated Consumer Terms and make your decision. If you choose to accept the new policies now, they will go into effect immediately. These updates will apply only to new or resumed chats and coding sessions. After September 28, you’ll need to make your selection on the model training setting in order to continue using Claude. You can change your choice in your Privacy Settings at any time.
Doesn’t say clearly it applies to all the prompts from the past.
https://www.anthropic.com/news/updates-to-our-consumer-terms
It has always been like this. Sites like Reddit, HN, and Digg and Boing Boing (when they were more popular) have always had a lot of stories under the category of online rights, privacy, and anger at big companies.
Once they admitted they are going to have to take money from folks who chop up journalists that made them feel sad, they proved the current pre token LLM based business model doesn't work. They haven't pulled the ads lever yet but the writing is on the wall.
Which means sadly only business with other revenue streams like M$, the Google, or Amazon can really afford it long term. I'm was rooting for Anthropic but it doesn't look good.
> Previous chats with no additional activity will not be used for model training.
seriously, the idea we need this is a joke. people need it to pretend they can do their job. the rest of us enjoy having quick help from it. and we have done without it for a very long time already..
Merely selling data is extremely low value compared to also having the surface monopoly to monetize it in a very high engagement and decisioning space.
I feel like you don’t understand the fundamental mechanics of the ad world. Ultimately, the big 4 own such immense decisions surface area it may be a while before any AI model company can create a product the get there.
That I don't know, and probably no one else, way too early to tell. I only responded to a comment stating "LLMs aren't a fundamental or core technology, they're an amusing party trick", which obviously I disagree with as for me they've been a fundamental shift in what I'm able to do.
> From the energy consumption impact on climate change alone I would say the answer is clearly no.
Ok, I guess that's fair enough. So if someone happens to use local models at home, in a home that is powered by solar power, then you'd feel LLM starting to be a net positive for humanity?
> And tons of other issues like how Big Tech is behaving.
This is such a big thing in general (that I agree with) but it has nothing to do with LLMs as a technology. Big Tech acts like they own the world and can do whatever they want with it, regardless if there are LLMs or not, so not sure why anyone would expect anything else.
An LLM can give you a hazy picture, but it's your job to focus it.
This is already happening.
> Wait until smart glasses or watches with AI overtake cellphones
Smartphones are crystalized perfection. It's such a peak design. The size, form factor, sensors, input/output modalities, and generalization are perfect. The reason companies are trying to supplant it is that they need to get out from under Google and Apple's control. It's not that anything is wrong with the smartphone.
VR has a long way to go in terms of hardware problems.
XR/AR is ridiculous. It's creepy, unstylish, and the utility is highly questionable. Nobody is going to want to be a walking ad.
Just think about the type of code these things are trained on and the fact you’re clearly some random non-specialist.
not if you have to constantly expend enormous sums to stay ahead of your competition or otherwise you lose your edge. It's not the best coding model because they got some mystical treasure in their basement. It's so rapidly becoming a commodity that at some point Microsoft or Google will just offer just as good a model for free and like search they'll just start milking people with ads.
That's likely one of the reasons for the shifting privacy stances, not just for training but because monetization of the product itself is probably looking pretty dim in the long run.
I can understand becoming reliant on a technology -- I expect most programmers today would be pretty lost with punch cards or line editors -- but LLM coding seems too new for true reliance to have formed yet...?
I am expecting AI companies to start using ads, it's inevitable as they need to make money at some point and $20 a month won't do it.
For ads the number of users is the main thing - the more users you have the bigger the market and more money you could earn. Google desperately needs to be in this space, that's why they are throwing a ton of money on AI.
Time will tell, but to me they feel like desktops did 20 years ago. The process of enshitification has turned simple tasks complicated and everyone wants a different, privacy destroying, frustrating to use "app", each of which has a slightly different UI paradigm, a mandatory subscription I've forgotten to cancel for two years straight, and a confusing name to remember. I now have something like 90 apps installed on my iphone, and I can only remember what something like 40 of them do. My damn cat box has an app, and instead of naming it something sensible like "Shitbox 2000" they named it "Whisker".
Was it "Foober Eats that had taco bell, or Instafart, maybe it was Dine-N-Dash? Where's the back button on this thing and why is it different from every other app? Is this an ad or content, does it even matter anymore? Why do I need another login, what happened to SSO? Why won't my password vault work for this one app? Did I register for this one with my google account or apple? Who took my pills? Stay off my lawn!"
When the day comes that I can just tell my device what to do, and let it get it done I'll be very happy to dump that cognitive load onto someone/something else.
That's just a misunderstanding, I'm not "vibing" anything. The tests are written by me, the API interfaces are written by me, the usages are written by me, and the implementation of these functions are written by an LLM, but reviewed to be up to code standards/quality by me.
If a function gives me the right output for the inputs I have in mind, does anything beyond that really matter?
This is because apps were never allowed to be installed like desktop software or as easy to access as websites. Developers had to cram in as much as possible and take as many permissions as possible because of how difficult Apple and Google made it.
If you could just search the web for an app, click a link, and have it instantly start working natively (sandboxed, with permissions), the world would be an amazing place.
Sure, that would make a difference, but it's not gonna happen anytime soon, other than hacker hobbyists, because no one is making money off of that.
> This is such a big thing in general (that I agree with) but it has nothing to do with LLMs as a technology.
Correct -- I don't have any issue with the technology itself, but rather how the technology is implemented and used, and the resources put towards its use. And BigTech are putting hundreds of $B into this -- for what end exactly besides potentially making tons of money off of consumer subscribers or ads a-la-Meta or Google? If BigTech was putting the same amount of money into technology that could actually benefit humanity (you know, like actually saving the world from potential future destruction by climate change), I'd have a much kinder view of them.
I disagree. Almost all of it should just be relatively standard API's designed for the AI to use, and we should all just use the AI as the standard interface. Many companies would collapse, because their entire anti-consumer business models would topple over, but that would be a good thing.
My point is that whenever we send our data to a third party, we can assume it could be abused, either unintentionally (by a hack, mistake etc.) or intentionally, because these companies are corrupted to the core and have a very relaxed attitude to obeying the law in general as these random examples show.
Well, this is what they claim. In practice, this is untrue on several levels. First, earlier OpenAI models were able to quote verbatim, and they were maimed later not to do that. Second, there were several lawsuits against OpenAI and not all of them ended. And finally, assuming that courts decide what they did was legal would mean everyone can legally download and use a copy of Libgen (part of "Books3") whereas the courts around the world are doing the opposite and are blocking access to Libgen country by country. So unless you set double standards, something is not right here. Even Meta employees torrenting Lingen knew that so let's not pretend we buy this rhetoric.