←back to thread

170 points bookofjoe | 7 comments | | HN request time: 0.814s | source | bottom
1. janalsncm ◴[] No.43646748[source]
> Isaac Asimov describes artificial intelligence as “a phrase that we use for any device that does things which, in the past, we have associated only with human intelligence.”

This is a pretty good definition, honestly. It explains the AI Effect quite well: calculators aren’t “AI” because it’s been a while since humans were the only ones who could do arithmetic. At one point they were, though.

replies(2): >>43646765 #>>43647350 #
2. azinman2 ◴[] No.43646765[source]
Although calculators can now do things almost no humans can do, or at least in any reasonable time. But most (now) wouldn’t call it AI. It’s a tool, with a very limited domain
replies(1): >>43646878 #
3. janalsncm ◴[] No.43646878[source]
That’s my point, it’s not AI now. It used to be.
replies(1): >>43647290 #
4. hinkley ◴[] No.43647290{3}[source]
Similarly, we esteem performance optimizations so aggressively that a lot of things that used to be called performance work are now called architecture, good design. We just keep moving the goal posts to make things more comfortable.
5. saalweachter ◴[] No.43647350[source]
I mean, at one point "calculator" was a job title.
replies(2): >>43647593 #>>43650155 #
6. timewizard ◴[] No.43647593[source]
The abacus has existed for thousands of years. Those who had the job of "calculator" also used pencil and paper to manage larger calculations which they would have struggled to do without any tools.

That's humanity. We're tool users above anything else. This gets lost.

7. musicale ◴[] No.43650155[source]
And "computer".