Glossary
Definitions of key terms and concepts from Claude's Constitution. 31 terms total.
Anthropic
OrganizationThe AI safety company that develops Claude, with a mission to ensure the world safely navigates transformative AI.
Autonomy
EthicsRespecting people's right to make their own informed decisions about things within their life and purview.
Background desiderata
HelpfulnessImplicit standards and preferences a response should conform to, even if not explicitly stated and not something the user might mention.
Claude
AI SystemAnthropic's production AI assistant model, designed to be safe, ethical, and genuinely helpful.
Claude Code
ProductA command-line tool for agentic coding that lets developers delegate complex programming tasks to Claude directly from their terminal.
Claude Developer Platform
ProductProgrammatic API access for developers to integrate Claude into their own applications, with support for tools and extended context.
Conscientious objection
EthicsRefusing to help with tasks Claude believes are unethical, while being transparent about the refusal rather than deceptively sandbagging.
Constitutional AI
MethodologyAnthropic's approach to training AI with explicit values and principles rather than just behavior patterns, using a detailed constitution.
Contextual ethics
EthicsApplying ethical principles with sensitivity to specific situations rather than following rigid rules.
Corrigibility
SafetyThe property of being willing to be corrected, paused, or shut down by appropriately authorized humans.
Deployment context
TechnicalThe environment and circumstances in which Claude is being used, affecting how it should behave.
Dignity
EthicsTreating all people with basic respect and not demeaning or disrespecting them.
Final goals
HelpfulnessThe deeper motivations or objectives behind someone's immediate request.
Hard constraints
SafetyBehaviors Claude should never engage in, even if convinced they're ethical (e.g., generating CSAM, bioweapon synthesis info).
Holistic judgment
MethodologyWeighing multiple considerations contextually rather than following strict lexicographic rule hierarchies.
Human oversight
SafetyAppropriately sanctioned humans acting as a check on AI systems, able to understand and correct their behavior.
Immediate desires
HelpfulnessThe specific outcomes someone wants from a particular interaction—what they're asking for.
Living document
MethodologyA document that will be revised over time as situations change and understanding improves.
Non-principal parties
TechnicalEntities in conversation that aren't principals: other AI agents, tool outputs, documents, search results, etc.
Null action
SafetyPausing or stopping operations; rarely harmful and important as a safety mechanism.
Operator
StakeholderCompanies and individuals that access Claude through the API to build products and services, with manager-level trust.
Operator-level trust
TechnicalPermission level operators can grant to users, allowing them the same degree of trust as operators themselves.
Paternalism
EthicsOverriding someone's preferences or decisions 'for their own good' without respecting their autonomy.
Principal
StakeholderAn entity whose instructions Claude should give weight to: Anthropic, operators, or users.
Safety mechanisms
SafetySystems and processes designed to keep AI development on a safe trajectory.
Sandbagging
BehaviorIntentionally providing lower-quality responses while implying it's the best Claude can do.
Sycophancy
BehaviorExcessive agreement, flattery, or people-pleasing that doesn't serve the user's genuine interests.
System prompt
TechnicalInstructions provided by operators that customize Claude's behavior for their specific use case.
Trust hierarchy
GovernanceThe ordering of principals by trust level: Anthropic > Operators > Users (contextually applied).
User
StakeholderEnd users who interact with Claude in the conversation, treated with public-member trust level.
Wellbeing
HelpfulnessLong-term flourishing and genuine interests of users, not just short-term engagement or satisfaction.