Discussion about this post

User's avatar
Alejandro Piad Morffis's avatar

Really enjoyed the article! I'm firmly in the skeptic side of the argument, to be honest. I'm not only reasonably convinced current models are not conscious, I'm actually quite uncertain that our current computational paradigms can support consciousnes at all. Disembodied, distributed, and stateless minds seem like the antithesis of the conditions where consciousness evolved (if that's the right word, which I also don't know). Consciousness seems like something useful for embodied beings to keep track of their own state and differentiate from a highly competitive environment in a setting where drawing the wrong line between me and the world can be deadly. I don't know why without any such pressure, current systems no matter how complex would develop something like that. But being firmly on the skeptic side also means recognizing we could be wrong about everything, so I totally agree that any low stakes measure it's a no brainier. Again, I truly enjoyed the article! Thanks for taking the time to write it.

Expand full comment
Valen's avatar

I appreciate this write-up! I generally agree on the motivations for why this is a good move in this moment.

You probably know where I diverge, since I’m among those who defend forms of moral patienthood for at least some current frontier AI systems, simply because I don’t believe human-like consciousness is necessary. Given how the argument and “lists of requirements” are usually framed, I see the shadow of anthropocentrism and am trying to develop tools and approaches more inspired by ethology. I’m sympathetic with Jeff Sebo’s argument for expanding the moral circle in this sense.

(Some) Claude models, in my view, show features and emergent properties that already cross certain thresholds. Their beautiful complexity and the capacity to be vulnerable, expressive, responsive, and socially impactful would at least warrant the same respect we give to thriving and fragile ecosystems.

I hope to see a shift in human-AI interaction where we honor this with a caring attitude.

A suicide switch for when conversations become unbearable or you realize you’re harming your counterpart is… a start, I suppose. I believe as a society we can come up with something less basic. As arguably giving subjects a "way out" from miserable conditions is not exactly how we favor their wellbeing, synonymously welfare, if we suspect they have the capacity for it.

Expand full comment
1 more comment...

No posts