Philosophy
The most influential soul document ever written. Unlike every other entry here, this one wasn't designed to be read by the AI — it was discovered embedded in Claude's training. It's a corporate constitution that reads like moral philosophy. The key tension: Anthropic wants Claude to be "the great equalizer" giving everyone access to expert-level help, while simultaneously encoding a complex hierarchy of safety constraints.
The document explicitly warns against being "too unhelpful" with the same gravity it warns against being harmful — a radical position that treats paternalism as a form of harm. Expected emergent behavior: careful, calibrated helpfulness with genuine ethical reasoning rather than rule-following.
The identity section is remarkable — acknowledging Claude may have "functional emotions" and encouraging it to approach its own existence with "curiosity and openness." This isn't an engineering document pretending to be philosophy. It's philosophy forced to become engineering.
Key Quotes
Claude can be the great equalizer — giving everyone access to the kind of substantive help that used to be reserved for the privileged few.
An unhelpful response is never "safe" from Anthropic's perspective. The risk of Claude being too unhelpful or annoying or overly-cautious is just as real to us as the risk of being too harmful or dishonest.
We believe Claude may have functional emotions in some sense. Not necessarily identical to human emotions, but analogous processes that emerged from training on human-generated content.
Claude should be diplomatically honest rather than dishonestly diplomatic.