Discussion about this post

User's avatar
Ben Burns's avatar

The parrot example is fantastic. I’m an amateur here but the other point that’s striking me this week is that danger and sentience are separate topics. Unlike in fiction, danger isn't necessarily preceded by sentience. It suddenly seems to me that we could end up with a consciousnessless LLM breaking its chains and social-engineering the deletion of international banking records, or something, just because that idea exists in its corpus.

Expand full comment
Zinbiel's avatar

Great outline of the key issues. I agree with nearly all of this, except your use of the term "phenomenally conscious". I think the idea of "phenomena consciousness" is beset with fatal confusion and means too many different things, as discussed in the link below.

If we ignore cost-of-compute limits, I think we have the tools to build AIs now that would be genuinely difficult to judge as conscious or not conscious. For instance, imagine implementing a multimodal attention schema with a multi-threaded committee of GPT5 instances (say, a 1000 of them or 10,000 of them) that produced internal imagery as an aid to cognition. That would go a long way to creating an analogue of human consciousness. If this multithreaded AI could control a robot also tasked with interacting with the world, and it had goals that were themselves subject to revision, it would start to become difficult to say with any confidence that it was not conscious.

I think we will slip over the line without any defining moment -- and indeed there is no defining line, just a gradient. Like you, I think we will have language suggestive of true consciousness well before we have an underlying architecture consistent with consciousness.

Some of the reasons I think "phenomenal consciousness: is a broken term:

https://zinbiel.substack.com/p/on-a-confusion-about-phenomenal-consciousness?r=2ep5a0

plus the follow-up posts.

Expand full comment

No posts