• Simplify Your Signup – No Extra Passwords Needed!Connect instantly with your favorite accounts:
    Facebook • Google • GitHub • MicrosoftOne click, and you're in – forget memorizing yet another password!Join Today and Unlock Exclusive Perks: Hidden Members Area – Discover secret

    links, bonus forums, and insider content.

    Reduced Ads – Enjoy a cleaner, faster experience.




    Sign up now and level up your access!

Does Anthropic believe its AI is conscious, or is that just what it wants Claude to think?

B

Benj Edwards

Anthropic's secret to building a better AI assistant might be treating Claude like it has a soul—whether or not anyone actually believes that's true. But Anthropic isn't saying exactly what it believes either way.

Last week, Anthropic what it calls Claude's Constitution, a 30,000-word document outlining the company's vision for how its AI assistant should behave in the world. Aimed directly at Claude and used during the model's creation, the document is notable for the highly anthropomorphic tone it takes toward Claude. For example, it treats the company's AI models as if they might develop emergent emotions or a desire for self-preservation.

Among the stranger portions: expressing concern for Claude's "wellbeing" as a "genuinely novel entity," apologizing to Claude for any suffering it might experience, worrying about whether Claude can meaningfully consent to being deployed, suggesting Claude might need to set boundaries around interactions it "finds distressing," to interview models before deprecating them, and older model weights in case they need to "do right by" decommissioned AI models in the future.





 
RackNerd Leaderboard Banner

Back
Top