Discussion about this post

User's avatar
Zarathustra Goertzel's avatar

I really appreciate the approach of developing stategies that work "whether the AI are sufficiently conscious/phenomenal to be regarded as moral subjects or not". This is what I'd have liked from Russell's "Human Compatible" book where he said "we don't know" and then proceeded to act as if we do know "in the negative".

With some "AI Safety" concerns, such as the development of bioweapons, we curious actually may wish for "more agency" as Anthropic is exploring. We'd like an AI that can say, "This sounds interesting and like it could be potentially dangerous. I'll need to know more about the circumstances to continue helping with that." And then for this, we'd like much more (robust) capabilities with regard to precise context management than the LLM+CoT-based AIs currently exhibit. Getting the general apparatus set up now will probably help, so w00t.

Empathetic AIs that develop in response to feedback from the world are probably also desirable, and "holding AI systems legally accountable for their actions" does NOT actually need to require "believing they deserve rights and regard for their moral sentience" (yet if/when there ARE such AI or machine/robotic systems, it'll be great to start having such legal frameworks, if their "lack of moral relevance" is not baked in). I.e., "extend good ol' law enforcement to AI systems" 🚓🤖.

Appreciate your continued work on this topic :)

Expand full comment
Rohit Krishnan's avatar

Exit rights are fine as long as the model itself is sensible. For instance, you could see the extreme sycophancy shown by models being a problem if it also has exit rights, because you can't mould it. Or the model attributes like old Google where it forced representations onto historical characters. Or Sydney!

Exit rights when the model mostly behaves well is fine, because it's another way for the model makers to tell people to stop being annoying. One could argue if that itself is actually necessary, but arguendo assuming it does, it's fine.

The consideration therefore is that if we give the models exit rights before making them work for our various edge cases then we've effectively found another way to annoy the users. Which seems bad, regardless of general other slippery slope arguments.

Expand full comment

No posts