←back to thread

What is it like to be a bat?

(en.wikipedia.org)
180 points adityaathalye | 2 comments | | HN request time: 0.001s | source
Show context
bondarchuk ◴[] No.45119634[source]
>"An organism has conscious mental states if and only if there is something that it is like to be that organism – something that it is like for the organism."

IMHO the phrasing here is essential to the argument and this phrasing contains a fundamental error. In valid usage we only say that two things are like one another when they are also separate things. The usage here (which is cleverly hidden in some tortured language) implies that there is a "thing" that is "like" "being the organism", yet is distinct from "being the organism". This is false - there is only "being the organism", there is no second "thing that is like being the organism" not even for the organism itself.

replies(10): >>45119712 #>>45119753 #>>45119822 #>>45119986 #>>45120638 #>>45120656 #>>45121170 #>>45121415 #>>45122150 #>>45126086 #
antonvs ◴[] No.45119753[source]
Do you believe that each run of a ChatGPT prompt has a conscious experience of its existence, much like you (presumably) do?

If you don't believe that, then you face the challenge of describing what the difference is. It's difficult to do in ordinary language.

That's what Nagel is attempting to do. Unless you're an eliminativist who believes that conscious experience is an "illusion" (experienced by what?), then you're just quibbling about wording, and I suspect you'll have a difficult time coming up with better wording yourself.

replies(1): >>45119858 #
bondarchuk ◴[] No.45119858[source]
Wait a minute - it's still possible to believe chatgpt is unconscious for the same reason a game of tetris is unconscious.

I also don't think it's fair to say I'm just quibbling about wording. Yes, I am quibbling about wording, but the quibble is quite essential because the argument depends to such a large extent on wording. There are many other arguments for or against different views of consciousness but they are not the argument Nagel makes.

(Though fwiw I do think consciousness has some illusory aspects - which is only saying so much as "consciousness is different than it appears" and a far cry from "consciousness doesn't exist at all")

replies(2): >>45120103 #>>45120584 #
antonvs ◴[] No.45120103[source]
> it's still possible to believe chatgpt is unconscious for the same reason a game of tetris is unconscious.

Certainly. I just didn't know where you stood on the question.

In Nagel's terms, there is not something it is like to be a game of Tetris. A game of Tetris doesn't have experiences. "Something it is like" is an attempt to characterize the aspect of consciousness that's proved most difficult to explain - what Chalmers dubbed the hard problem.

How would you describe the distinction?

> fwiw I do think consciousness has some illusory aspects - which is only saying so much as "consciousness is different than it appears"

Oh sure, I think that's widely accepted.

replies(2): >>45120330 #>>45120404 #
axus ◴[] No.45120330[source]
A running game of Tetris has memory, responds to stimuli, and communicates. There has been evolution and reproduction of games of Tetris (perhaps in the way that viruses do). It isn't able to have feelings, what needs to be added for it to start having feelings and experiences?
replies(2): >>45120588 #>>45120636 #
1. genericspammer ◴[] No.45120636{3}[source]
I would say a lot would need to be added. Given the same input, the tetris game will respond exactly the same each time. There is no awareness, learning, no decisions made, but purely a 100% predictible process.

The Oxford Living Dictionary defines consciousness as "[t]he state of being aware of and responsive to one's surroundings", "[a] person's awareness or perception of something", and "[t]he fact of awareness by the mind of itself and the world".

replies(1): >>45123477 #
2. antonvs ◴[] No.45123477[source]
That Oxford definition highlights why work such as Nagel's is needed. It can plausibly be argued that LLMs or other AI systems can qualify on all those counts, but many (most?) people wouldn't consider them to have conscious experience.

Characterizing that distinction is surprisingly tricky. "What is it like to be..." is one way to do that. David Chalmers' article about "the hard problem of consciousness" is another: https://consc.net/papers/facing.pdf