←back to thread

Claude Code now supports hooks

(docs.anthropic.com)
381 points ramoz | 1 comments | | HN request time: 0.284s | source
Show context
bionhoward ◴[] No.44429497[source]
Given the Anthropic legal terms forbid competing with them, what are we actually allowed to do with this? Seems confusing what is allowed.

No machine learning work? That would compete.

No writing stuff I would train AI on. Except I own the stuff it writes, but I can’t use it.

Can we build websites with it? What websites don’t compete with Anthropic?

Terminal games? No, Claude code is a terminal game, if you make a terminal game it competes with Claude?

Can their “trust and safety team” humans read everyone’s stuff just to check if we’re competing with LLMs (funny joke) and steal business ideas and use them at Anthropic?

Feels like the dirty secret of AI services is, every possible use case violates the terms, and we just have to accept we’re using something their legal team told us not to use? How is that logically consistent? Any safety concerns? This doesn’t seem like a law Asimov would appreciate.

It would be cool if the set of allowed use cases wasn’t empty. That might make Anthropic seem more intelligent

replies(5): >>44429509 #>>44429552 #>>44429556 #>>44429665 #>>44429697 #
ethan_smith ◴[] No.44429697[source]
Anthropic's terms typically restrict training competing AI models with their outputs, not building standard applications or websites that simply use their API as a tool.
replies(2): >>44431937 #>>44436367 #
1. bionhoward ◴[] No.44436367[source]
it reads, “competing products or services” — not “training competing AI models” with their outputs

Also, who the fuck wants a codebase they can’t train models on