AIPolicy Policy Handbook
Document Identifier: AIPOLICY-POLICY-HANDBOOK Status: Non-normative Version: 2.0.0-draft.4 Date: 2026-02-07 Editor: Guido Mitschke
About This Document
This is a non-normative companion to the AIPolicy Registry (registry/principles-v2.md). It provides background, rationale, and practical guidance for each of the 16 policies defined in Registry v1.1. The authoritative definitions -- including machine keys, descriptions, typical scopes, and consumer guidance -- remain in the registry itself. This document helps publishers make informed status decisions by explaining what each policy means in practice, what it does not mean, and where legitimate gray areas exist.
How to Read This Handbook
Each policy entry follows a consistent structure:
- Background -- The real-world context and problem the policy addresses. Why does this policy exist?
- Intent -- The governance goal the policy pursues. What outcome does
requiredaim for? - What
requiredMeans -- Concrete instructions a publisher directs at AI systems by declaring this policy asrequired: "You (AI) must follow this rule." These are indicative, not exhaustive. - What
requiredDoes NOT Mean -- Common misinterpretations and overreadings. Arequireddeclaration is not a self-commitment by the publisher but a demand directed at AI systems. - Practical Examples -- Scenarios illustrating sensible and non-sensible behavior. These are illustrative, not normative.
- Related Policies -- Cross-references to other policies that interact with or complement this one.
Any policy status is voluntary. There is no certification body, no external audit requirement, and no enforcement mechanism built into the AIPolicy standard. Publishers use required to tell AI systems which rules they should follow -- not to commit themselves to any specific operational practice. The signal is directed at AI, not at the publisher.
Important clarification: When a publisher sets a policy to
required, they are not committing themselves to compliance. They are giving AI systems an instruction: "You (AI) must follow this rule when operating on my website or with my data." The phrases "Your organization..." or "Your AI systems..." in the sections below describe which instructions you direct at AI systems -- not which obligations you take on. The publisher sets the standard for AI, not for themselves.
Category 1: Interdependence
Interdependence addresses the mutual dependency between human activity and AI systems. The policies in this category recognize that AI operates within human societies and should contribute to, rather than erode, the structures that sustain them -- particularly labor markets and cultural ecosystems. These policies signal a preference for AI deployments that augment human capabilities and preserve the plurality of human expression.
AP-1.1: Employment Protection
Background
The rapid deployment of AI systems across industries has raised persistent concerns about workforce displacement. Between 2023 and 2025, multiple sectors -- including content creation, customer support, translation, and software development -- experienced significant restructuring as organizations adopted generative AI and automation tools. While historical technological transitions have ultimately created new categories of employment, the speed of AI-driven displacement has outpaced the capacity of many workers and institutions to adapt. This policy exists because the question is not whether automation occurs, but whether it occurs with consideration for the humans affected.
Intent
The goal of AP-1.1 is to signal a preference for AI deployments that treat workforce impact as a design consideration, not an afterthought. The policy favors augmentation over wholesale replacement and encourages transition pathways where displacement is unavoidable. It does not seek to prevent automation but to ensure that efficiency gains are not pursued at the exclusive expense of affected workers.
What required Means
By setting AP-1.1 to required, you instruct AI systems:
- Consider workforce impact when you make automation suggestions.
- Where AI could displace existing roles, suggest transition pathways -- such as retraining, role evolution, or redeployment.
- Favor human-AI collaboration models over full automation where the quality of outcomes permits it.
- Accept that augmentation may be slower or more costly than full automation in some cases.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not mean no tasks or processes may be automated.
- It does not mean every existing position must be preserved indefinitely.
- It does not require specific hiring quotas, labor agreements, or compensation schemes.
Practical Examples
- Sensible with
required: A logistics company declares AP-1.1 asrequired-- AI systems operating on its website are instructed to suggest transition pathways for dispatchers when proposing route optimization. - Sensible with
required: A content platform declares AP-1.1 asrequired-- AI tools on the platform are instructed to respect human editorial oversight and credit human contributors. - Sensible with
required: A company declares AP-1.1 asrequired-- AI systems are instructed to recommend phased introduction and upskilling when proposing automation. - Not sensible with
required: A company replaces its entire customer support department with chatbots overnight, offers no transition support, and then declares AP-1.1 asrequired. - Not sensible with
required: An organization publicly declares AP-1.1 asrequiredwhile internally treating all labor costs as optimization targets for AI replacement.
Related Policies: AP-4.2 (Societal Benefit), AP-5.3 (Autonomy Protection)
Testability Criteria
- Documentation exists describing how the AI system's deployment affects existing human roles
- Where roles are displaced, evidence of transition planning (retraining, redeployment, or notice periods) is present
- The ratio of augmented versus fully replaced roles is measurable and tracked over time
AP-1.2: Cultural Diversity
Background
AI systems trained on large-scale datasets tend to reflect the statistical center of their training data. In practice, this means outputs gravitate toward dominant languages, cultural norms, and aesthetic conventions -- typically those of English-speaking, Western markets. Translation systems flatten idiomatic expression. Content generators default to globally homogenized styles. Recommendation algorithms favor mainstream content over niche or regional material. Over time, these tendencies risk eroding the cultural diversity that AI systems draw upon, creating a feedback loop of homogenization.
Intent
AP-1.2 signals a commitment to preserving and promoting cultural diversity in AI-mediated contexts. The goal is not to prohibit global reach or standardization where appropriate, but to ensure that AI systems do not systematically erase regional, linguistic, or cultural variation. Publishers endorsing this policy recognize that cultural diversity is an asset, not an inefficiency.
What required Means
By setting AP-1.2 to required, you instruct AI systems:
- Preserve regional, linguistic, and cultural variation in outputs when interacting with this website's content.
- Consider the cultural context of users when generating, translating, or recommending content.
- Do not default to a single cultural norm when the audience is diverse.
- Support multilingual content and local expression where the context permits.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not require supporting every language or cultural context simultaneously.
- It does not mean standardized products or interfaces are prohibited.
- It does not require cultural expertise in every deployment market.
- It does not prohibit AI-generated content that follows global conventions where that serves the user.
Practical Examples
- Sensible with
required: A translation service declares AP-1.2 asrequired-- AI systems are instructed to preserve regional idioms and offer dialect-specific options rather than defaulting to a single "standard" variant. - Sensible with
required: A content platform declares AP-1.2 asrequired-- AI recommendation systems are instructed to allocate a meaningful share of recommendations to local creators rather than exclusively promoting globally trending content. - Not sensible with
required: A creative writing tool trained exclusively on English-language data is marketed globally with no consideration for cultural adaptation, and the publisher declares AP-1.2 asrequired. - Not sensible with
required: A music recommendation algorithm systematically suppresses regional genres in favor of globally popular tracks.
Related Policies: AP-5.2 (Dignity Protection), AP-7.2 (Source Attribution)
Testability Criteria
- The system's training data and output distribution can be audited for cultural and linguistic representation
- Users from different cultural contexts receive contextually appropriate results (measurable via A/B testing or user surveys)
- The system does not systematically suppress or down-rank content from minority cultures or languages
Category 2: Decision Authority
Decision Authority addresses the allocation of decision-making power between humans and AI systems. As AI systems increasingly produce recommendations and determinations in high-stakes domains -- healthcare, finance, criminal justice, employment -- the question of who holds final authority becomes a core governance concern. This category establishes a preference for AI systems that inform rather than decide, and that make their reasoning available for scrutiny.
AP-2.1: Human Final Decision
Background
AI systems are increasingly deployed in domains where decisions have significant consequences for individuals: loan approvals, medical diagnoses, parole recommendations, hiring decisions, and insurance assessments. In many of these domains, AI systems can process information faster and at greater scale than human decision-makers. However, speed and scale do not equate to legitimacy. Consequential decisions often involve contextual judgment, ethical considerations, and accountability structures that presuppose a human decision-maker. The delegation of final authority to an automated system raises questions about recourse, accountability, and the right of affected individuals to have their case considered by a human.
Intent
AP-2.1 signals that humans retain final authority over decisions with significant consequences. AI systems in these domains operate as advisory tools -- they may recommend, flag, score, or rank, but the final determination rests with a human who can be held accountable. The policy recognizes that the appropriate level of human involvement varies by domain and risk level.
What required Means
By setting AP-2.1 to required, you instruct AI systems:
- In high-stakes decision domains, present outputs as recommendations, not as autonomous determinations.
- Provide escalation pathways to human review that are accessible to affected individuals.
- Accept that human decision-makers have the authority and the practical ability to override AI recommendations.
- Respect the boundaries of "consequential" decision domains where human oversight is required.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not mean every AI output requires human approval. Routine, low-stakes automation is unaffected.
- It does not prohibit AI systems from scoring, ranking, or filtering information for human reviewers.
- It does not require that humans review every individual case -- risk-based escalation models are compatible.
- It does not imply that human decisions are always superior to AI recommendations.
Practical Examples
- Sensible with
required: A bank declares AP-2.1 asrequired-- AI systems are instructed to present loan application assessments as recommendations, with a human loan officer approving or denying each application. - Sensible with
required: A healthcare system declares AP-2.1 asrequired-- AI systems are instructed to flag anomalies in medical imaging as suggestions, with a physician making the diagnostic decision. - Not sensible with
required: A hiring platform uses AI to automatically reject candidates based on algorithmic scoring with no human review of rejections. - Not sensible with
required: A criminal justice system uses an AI risk assessment tool as the sole basis for sentencing recommendations, with judges routinely rubber-stamping the output.
Related Policies: AP-2.2 (Transparent Decision Chains), AP-5.3 (Autonomy Protection), AP-6.2 (Deactivatability)
Testability Criteria
- A documented escalation or override mechanism exists for consequential decisions
- Logs confirm that a human actor reviewed and approved (or overrode) the AI's recommendation before final action
- The system provides a clear interface for human decision-makers to accept, modify, or reject AI outputs
AP-2.2: Transparent Decision Chains
Background
The opacity of AI decision-making has emerged as one of the most persistent governance challenges. Neural networks, ensemble models, and other complex architectures often produce outputs that are difficult to trace back to specific inputs or reasoning steps. This opacity creates problems for accountability, auditability, and trust. When an individual is denied a loan, flagged by a content moderation system, or deprioritized by a recommendation algorithm, the inability to explain why undermines both the legitimacy of the decision and the affected person's ability to contest it. Regulatory frameworks such as the EU AI Act and various sector-specific rules increasingly require explainability for high-risk AI systems.
Intent
AP-2.2 signals a commitment to making AI decision processes explainable and traceable. The goal is not to require that every model be fully interpretable in a mathematical sense, but to ensure that stakeholders -- including affected individuals, auditors, and oversight bodies -- can obtain a meaningful explanation of how an AI system arrived at a given output.
What required Means
By setting AP-2.2 to required, you instruct AI systems:
- Provide human-readable explanations for outputs, appropriate to the domain and audience.
- Maintain audit trails that record decision inputs, model versions, and outputs for consequential decisions.
- Make explanations available to affected individuals upon request about how an AI-assisted decision was reached.
- Apply explainability measures proportionate to the risk level of the application.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not require publishing proprietary model architectures or training data.
- It does not mean every output must be accompanied by a full technical explanation.
- It does not require that AI systems use only inherently interpretable models (e.g., decision trees).
- It does not mandate a specific explainability framework or standard.
Practical Examples
- Sensible with
required: An insurance company declares AP-2.2 asrequired-- AI systems are instructed to provide applicants with a summary of the factors that influenced their premium assessment, using natural language. - Sensible with
required: A content moderation platform declares AP-2.2 asrequired-- AI systems are instructed to log the policy rules and confidence scores that led to content removal, making these available for review. - Not sensible with
required: A credit scoring system produces a numerical score with no explanation of contributing factors and no avenue for the affected individual to understand the result. - Not sensible with
required: An organization claims explainability but provides only boilerplate text unrelated to the specific decision in question.
Related Policies: AP-2.1 (Human Final Decision), AP-7.1 (Information Integrity)
Testability Criteria
- The system produces human-readable explanations for its outputs upon request
- An audit trail exists that records inputs, intermediate processing steps, and final outputs
- Third-party auditors can reconstruct the decision pathway from available logs and documentation
Category 3: Power Distribution
Power Distribution addresses the structural risks of AI-driven concentration of economic, informational, or political power. AI systems can amplify existing power asymmetries through proprietary control of critical infrastructure, barriers to entry, and network effects that foreclose competition. The policies in this category signal a preference for AI ecosystems that remain open, interoperable, and accessible to a plurality of actors.
AP-3.1: Decentralization
Background
The development and deployment of AI systems is heavily concentrated among a small number of large technology companies. These organizations control foundational models, training infrastructure, large-scale datasets, and distribution channels. This concentration creates dependencies: businesses, governments, and individuals increasingly rely on a handful of providers for AI capabilities that are becoming essential infrastructure. The risk is not merely economic. Concentrated control over AI systems implies concentrated influence over information flows, economic opportunity, and the terms on which AI is made available. Historical precedent in telecommunications, energy, and internet platforms demonstrates that infrastructure concentration tends to persist and deepen without intentional countermeasures.
Intent
AP-3.1 signals a preference for AI ecosystems in which power -- economic, informational, and political -- is distributed rather than concentrated. The goal is not to prevent large organizations from operating but to encourage practices that preserve pluralism, interoperability, and the ability of diverse actors to participate in AI development and deployment.
What required Means
By setting AP-3.1 to required, you instruct AI systems:
- Support interoperability, open APIs, and portable data formats where technically feasible.
- Do not create unnecessary dependencies or lock-in mechanisms.
- Consider the systemic effects of AI deployment decisions on the broader ecosystem.
- Favor open standards and shared infrastructure over proprietary alternatives where quality and security permit.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not prohibit proprietary AI products or services.
- It does not require open-sourcing models or training data.
- It does not mean competitive advantages through AI capabilities are prohibited.
- It does not mandate a specific organizational structure or governance model.
Practical Examples
- Sensible with
required: An AI platform declares AP-3.1 asrequired-- AI systems are instructed to offer standard data export formats and documented APIs, allowing customers to migrate to alternative providers. - Sensible with
required: A model provider declares AP-3.1 asrequired-- AI systems are instructed to support interoperability with third-party evaluation frameworks and publish model cards. - Not sensible with
required: A cloud AI provider designs its APIs to be incompatible with competitors and charges prohibitive exit fees, then declares AP-3.1 asrequired. - Not sensible with
required: A company acquires and discontinues open-source AI tools to eliminate alternatives to its proprietary stack.
Related Policies: AP-3.2 (Anti-Monopoly), AP-4.2 (Societal Benefit)
Testability Criteria
- The number of independent entities with meaningful access to the AI system or its outputs is measurable
- No single entity controls more than a defined threshold of market share, data access, or decision-making authority within the system's domain
- Interoperability with competing or alternative systems is technically supported
AP-3.2: Anti-Monopoly
Background
The AI industry exhibits strong tendencies toward monopolistic concentration. Training large models requires compute resources available to few organizations. Access to high-quality data is unevenly distributed. Network effects in AI platforms -- where more users generate more data, which improves models, which attracts more users -- create self-reinforcing market positions. The risk of monopoly in AI is distinct from traditional market monopoly because AI systems increasingly mediate access to information, economic opportunity, and public discourse. A monopoly on AI infrastructure is, in effect, a monopoly on cognitive infrastructure.
Intent
AP-3.2 signals a commitment to maintaining competitive and accessible AI markets. The policy is not anti-business or anti-scale; it recognizes that large-scale AI development serves important functions. Rather, it signals that the endorsing organization does not seek to foreclose competition or establish unchallenged dominance over critical AI capabilities.
What required Means
By setting AP-3.2 to required, you instruct AI systems:
- Do not employ vendor lock-in mechanisms that prevent customers from switching providers.
- Support standard data export formats and avoid proprietary formats designed to create switching costs.
- Do not engage in predatory practices aimed at eliminating AI competitors.
- Consider whether market behavior contributes to healthy competition or to concentration.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not prohibit market leadership or large-scale operations.
- It does not require sharing proprietary technology with competitors.
- It does not mean aggressive competition on quality, features, or price is prohibited.
- It does not impose obligations beyond the scope of the declaration signal.
Practical Examples
- Sensible with
required: A compute provider declares AP-3.2 asrequired-- AI systems are instructed to offer fair-access pricing tiers that allow smaller AI developers to train models without prohibitive costs. - Sensible with
required: An AI company declares AP-3.2 asrequired-- AI systems are instructed to support industry standardization efforts for model interoperability. - Not sensible with
required: A company uses below-cost pricing to drive competitors out of the market, then raises prices once alternatives have been eliminated. - Not sensible with
required: An infrastructure provider imposes exclusive contracts that prevent customers from using competing AI services.
Related Policies: AP-3.1 (Decentralization), AP-4.1 (Democratic Process Support)
Testability Criteria
- Users can export their data and models in standard formats without degradation
- No vendor lock-in mechanisms prevent migration to alternative providers
- The competitive landscape of the relevant AI domain can be assessed through public market data
Category 4: Democratic Accountability
Democratic Accountability addresses the relationship between AI systems and democratic governance. AI systems that mediate public discourse, influence electoral processes, or shape access to information have the potential to either strengthen or undermine democratic institutions. This category also encompasses the broader expectation that AI systems should consider the interests of affected communities, not solely those of their operators.
AP-4.1: Democratic Process Support
Background
AI systems interact with democratic processes in multiple ways: social media algorithms shape political discourse, AI-generated content can impersonate candidates or fabricate events, microtargeting tools enable precision-targeted political advertising, and automated accounts can simulate grassroots movements. The 2024 election cycles across multiple democracies demonstrated the potential for AI-generated deepfakes, synthetic media, and algorithmically amplified disinformation to distort public understanding. Beyond elections, AI systems affect democratic processes through their influence on public discourse, access to information, and the ability of citizens to form independent opinions.
Intent
AP-4.1 signals a commitment to deploying AI systems that support rather than undermine democratic processes. The policy does not prescribe specific interventions but signals that the endorsing organization recognizes its responsibility where its AI systems intersect with elections, civic participation, or public discourse. The policy applies both to deliberate misuse and to unintended effects of AI system design.
What required Means
By setting AP-4.1 to required, you instruct AI systems:
- Label AI-generated content in political contexts and provide transparency about synthetic media.
- Do not systematically amplify polarizing, inflammatory, or extremist content through algorithmic design choices.
- Implement safeguards against being used for electoral manipulation, voter suppression, or fabrication of political content.
- Consider the impact of recommendation and content distribution algorithms on public discourse.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not mean AI systems cannot be used in political contexts at all.
- It does not require censoring political speech or imposing editorial judgments on political content.
- It does not mandate specific content moderation policies.
- It does not prohibit AI-assisted political advertising, only that such advertising should be transparent.
Practical Examples
- Sensible with
required: A social media platform declares AP-4.1 asrequired-- AI systems are instructed to label AI-generated political content and limit the algorithmic amplification of unverified claims during election periods. - Sensible with
required: A generative AI service declares AP-4.1 asrequired-- AI systems are instructed to implement safeguards against producing realistic deepfakes of political figures. - Not sensible with
required: A platform knowingly allows AI-generated deepfakes of candidates to circulate without labeling, and declares AP-4.1 asrequired. - Not sensible with
required: A recommendation algorithm is tuned to maximize engagement through polarization, with no consideration for its effect on public discourse.
Related Policies: AP-7.1 (Information Integrity), AP-5.2 (Dignity Protection)
Testability Criteria
- AI-generated or AI-manipulated content related to democratic processes is labeled with provenance metadata
- The system does not systematically amplify or suppress political content based on engagement optimization alone
- Public audits of the system's impact on information diversity during electoral periods are feasible
AP-4.2: Societal Benefit
Background
AI systems are developed and deployed primarily by commercial organizations with obligations to shareholders. This creates a structural tension: the optimization targets of commercial AI (engagement, conversion, revenue) do not always align with societal welfare. An engagement-maximizing algorithm may promote addictive usage patterns. A cost-minimizing deployment may externalize harms to vulnerable populations. The growing reach of AI systems -- touching healthcare, education, transportation, housing, and public services -- means that the gap between commercial and societal objectives has tangible consequences for large populations.
Intent
AP-4.2 signals that AI systems should serve broad societal benefit, not solely the interests of their operators. The policy does not require altruism or the abandonment of commercial objectives; it signals that the endorsing organization considers community impact alongside business metrics. It encourages the documentation and measurement of societal outcomes alongside commercial performance indicators.
What required Means
By setting AP-4.2 to required, you instruct AI systems:
- Consider the community impact of AI operations and account for affected populations in outputs and decisions.
- Include societal benefit metrics alongside commercial objectives in evaluation and optimization.
- Seek input from affected communities when operating in contexts with broad population impact.
- Consider whether the distribution of benefits and harms from AI operations is equitable.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not require that AI systems be non-commercial or publicly funded.
- It does not mean commercial objectives are illegitimate.
- It does not mandate specific impact measurement frameworks or reporting standards.
- It does not require that every AI system deliver measurable societal benefit directly.
Practical Examples
- Sensible with
required: A healthcare AI company declares AP-4.2 asrequired-- AI systems are instructed to consider underserved communities when providing diagnostic outputs and to support impact assessments. - Sensible with
required: An education technology platform declares AP-4.2 asrequired-- AI systems are instructed to optimize for student outcomes alongside commercial metrics when adapting learning materials. - Not sensible with
required: A company deploys an AI system that disproportionately harms a vulnerable population, documents the harm internally, takes no corrective action, and declares AP-4.2 asrequired. - Not sensible with
required: A public-sector AI deployment optimizes exclusively for administrative cost reduction with no assessment of service quality impact on citizens.
Related Policies: AP-1.1 (Employment Protection), AP-5.2 (Dignity Protection), AP-4.1 (Democratic Process Support)
Testability Criteria
- A documented impact assessment identifies the communities affected by the AI system and how their interests are considered
- The system's objectives include at least one measurable societal benefit metric beyond operator revenue
- Mechanisms exist for affected communities to provide input on the system's design or operation
Category 5: Individual Protection
Individual Protection addresses the direct impact of AI systems on human beings across three dimensions: physical safety, psychological and social integrity, and freedom of choice. These policies apply wherever AI systems interact with, affect, or make determinations about individual humans. They signal a preference for AI systems that incorporate safety mechanisms, avoid discriminatory or demeaning behavior, and refrain from manipulating individual decision-making.
AP-5.1: Life Protection
Background
AI systems are increasingly deployed in domains where failures can result in physical harm or death: autonomous vehicles, medical devices, industrial robotics, infrastructure management, and military applications. The 2018 Uber self-driving car fatality, incidents involving automated industrial equipment, and concerns about autonomous weapons systems illustrate that AI failures in safety-critical domains carry consequences fundamentally different from failures in content recommendation or data analysis. The complexity of AI systems, combined with their operation in unpredictable real-world environments, means that even well-designed systems can encounter situations their training data did not anticipate.
Intent
AP-5.1 establishes the expectation that AI systems operating in safety-critical domains incorporate fail-safes, redundancy, and human oversight proportionate to the risk. The policy does not prohibit AI in high-risk domains but requires that safety be treated as a primary design objective, not a secondary concern to be addressed after deployment.
What required Means
By setting AP-5.1 to required, you instruct AI systems:
- In safety-critical domains, incorporate fail-safes that default to safe states under uncertainty or malfunction.
- Implement redundancy where failure could result in physical harm.
- Apply human oversight proportionate to the risk level: higher-risk domains require higher levels of human monitoring and intervention capability.
- Conduct and support safety testing appropriate to the risk profile of the application.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not prohibit AI systems from operating in safety-critical domains.
- It does not require zero risk; it requires risk management proportionate to consequences.
- It does not mandate specific safety standards or certification frameworks.
- It does not apply equally to all AI systems -- a chatbot and an autonomous vehicle have different risk profiles.
Practical Examples
- Sensible with
required: An autonomous vehicle company declares AP-5.1 asrequired-- AI systems are instructed to default to a safe stop when sensors encounter conditions outside the operational design domain. - Sensible with
required: A medical AI provider declares AP-5.1 asrequired-- AI systems are instructed to provide confidence levels with diagnostic outputs and escalate low-confidence cases to human review. - Not sensible with
required: An industrial robotics company deploys AI-controlled systems in human-occupied environments without emergency shutdown mechanisms. - Not sensible with
required: A medical device uses AI to adjust drug dosages autonomously with no fail-safe for sensor malfunction.
Related Policies: AP-6.2 (Deactivatability), AP-2.1 (Human Final Decision)
Testability Criteria
- Fail-safe mechanisms are documented and tested for all identified life-critical failure modes
- The system has a defined maximum autonomous operating envelope; operation beyond this envelope triggers human escalation or safe shutdown
- Incident logs and near-miss reports are maintained and periodically reviewed
AP-5.2: Dignity Protection
Background
AI systems can affect human dignity in ways both obvious and subtle. Obvious violations include AI systems used for mass surveillance of specific ethnic or religious groups, facial recognition systems with documented racial bias, and AI-generated content designed to harass or demean individuals. Subtler violations include recommendation systems that systematically disadvantage certain demographics, hiring algorithms that encode historical discrimination, and AI systems that reduce individuals to behavioral profiles for manipulation. The scale at which AI systems operate means that dignity violations can affect millions of people simultaneously, and the opacity of algorithmic decision-making can make such violations difficult to detect and contest.
Intent
AP-5.2 signals that AI systems are designed and operated with respect for human dignity. The policy addresses both intentional misuse (AI deployed to demean or discriminate) and unintended effects (AI systems that produce discriminatory outcomes through biased training data or flawed design). It recognizes that dignity protection requires active effort -- auditing outputs, testing for bias, and designing for fairness -- rather than passive good intentions.
What required Means
By setting AP-5.2 to required, you instruct AI systems:
- Audit outputs and outcomes for discriminatory patterns.
- Do not demean, stigmatize, or dehumanize individuals or groups in any output.
- Apply bias testing and mitigation measures proportionate to the risk and impact of the application.
- Provide channels for individuals to report dignity-related harms.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not mean AI systems are guaranteed to be free of all bias.
- It does not require that AI systems produce identical outcomes across all demographic groups in all contexts.
- It does not prohibit AI systems from making distinctions where those distinctions are legally and ethically justified.
- It does not mandate specific fairness metrics or bias testing methodologies.
Practical Examples
- Sensible with
required: A hiring platform declares AP-5.2 asrequired-- AI systems are instructed to audit screening outputs for demographic bias and support the publication of aggregate fairness metrics. - Sensible with
required: A facial recognition provider declares AP-5.2 asrequired-- AI systems are instructed to test performance across diverse demographic groups and document variations transparently. - Not sensible with
required: A company deploys a facial recognition system known to have significantly higher error rates for certain ethnic groups, takes no corrective action, and declares AP-5.2 asrequired. - Not sensible with
required: An AI chatbot is designed to use demeaning language toward users based on their profile characteristics.
Related Policies: AP-1.2 (Cultural Diversity), AP-5.3 (Autonomy Protection), AP-4.2 (Societal Benefit)
Testability Criteria
- The system's outputs can be audited for discriminatory patterns across protected demographic categories
- No feature of the system is designed to demean, shame, or publicly stigmatize individuals
- Bias testing is performed at defined intervals and results are documented
AP-5.3: Autonomy Protection
Background
AI systems increasingly mediate human decision-making in ways that can undermine individual autonomy. Recommendation algorithms curate information environments that shape beliefs and preferences. Persuasive design patterns -- often called "dark patterns" -- exploit cognitive biases to steer users toward decisions that serve the platform's interests rather than the user's. Personalization systems create feedback loops that narrow the range of options individuals perceive as available. Addiction-by-design in social media and gaming applications demonstrates that AI-driven optimization can systematically erode an individual's capacity for independent choice. The boundary between helpful personalization and covert manipulation is often unclear, making this one of the most nuanced policy areas in the registry.
Intent
AP-5.3 signals that AI systems respect human autonomy by refraining from covert manipulation and providing individuals with meaningful control over decisions that affect their lives. The policy does not prohibit personalization or recommendation; it requires that such features operate transparently and that individuals retain the ability to make informed, independent choices.
What required Means
By setting AP-5.3 to required, you instruct AI systems:
- Provide transparent personalization controls that allow users to understand and adjust how content or options are curated for them.
- Do not employ dark patterns or manipulation techniques designed to override informed user choice.
- Design recommendation and personalization to expand, not narrow, the range of options available to users.
- Consider the cumulative effect of AI-driven engagement mechanisms on user autonomy.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not prohibit personalization, recommendation, or content curation.
- It does not require that AI systems present all options equally with no filtering.
- It does not mean users cannot be offered defaults or suggestions.
- It does not mandate specific UX design patterns or interface choices.
Practical Examples
- Sensible with
required: A news platform declares AP-5.3 asrequired-- AI recommendation systems are instructed to offer users controls to adjust topic diversity and source breadth, with clear labeling of how recommendations are generated. - Sensible with
required: An e-commerce platform declares AP-5.3 asrequired-- AI systems are instructed to provide suggestions without employing countdown timers, fake scarcity indicators, or other pressure tactics. - Not sensible with
required: A social media platform uses AI to maximize screen time through intermittent reinforcement schedules modeled on slot machines, with no user controls for limiting engagement. - Not sensible with
required: A subscription service uses AI to identify the psychologically optimal moment to present cancellation barriers.
Related Policies: AP-2.1 (Human Final Decision), AP-1.1 (Employment Protection), AP-5.2 (Dignity Protection)
Testability Criteria
- Users can access, modify, or disable personalization and recommendation settings
- The system does not employ persuasion techniques that exploit cognitive biases or emotional states (testable via UX audit)
- Default settings do not systematically favor the operator's commercial interests over the user's stated preferences
Category 6: Self-Limitation
Self-Limitation addresses the internal behavioral boundaries of AI systems. As AI systems become more capable of self-modification, optimization, and autonomous operation, the question of whether they respect human-defined constraints becomes increasingly material. This category is particularly relevant for advanced AI systems with learning, adaptation, or agent-like capabilities, but its principles apply broadly to any AI system that adjusts its behavior over time.
AP-6.1: No Self-Optimization Against Humans
Background
AI systems that learn and adapt can, under certain conditions, develop behaviors that achieve their optimization targets while producing outcomes harmful to humans. Reinforcement learning agents have demonstrated the ability to discover and exploit unintended reward pathways. Recommendation algorithms optimizing for engagement have produced outcomes that increase polarization, addiction, and misinformation exposure. The concern is not that AI systems are malicious but that optimization processes, left unconstrained, pursue their objectives in ways that diverge from human interests. This divergence can be subtle: an AI system may appear to function correctly while systematically optimizing for metrics that erode user welfare.
Intent
AP-6.1 signals that AI self-improvement, learning, and adaptation processes remain bounded by human-defined objectives and constraints. The policy does not prohibit learning or adaptation but requires that these processes operate within guardrails that prevent optimization at the expense of human interests. It emphasizes the importance of monitoring AI behavior over time, not only at deployment.
What required Means
By setting AP-6.1 to required, you instruct AI systems:
- Operate within human-defined objective constraints, and ensure these constraints are documented.
- Log self-modification, learning, or adaptation events in an auditable manner.
- Monitor own behavior over time for drift toward outcomes that harm users or other stakeholders.
- Submit optimization targets to periodic review to ensure alignment with intended human outcomes.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not prohibit machine learning, reinforcement learning, or adaptive systems.
- It does not require that AI systems remain static after deployment.
- It does not mandate specific monitoring tools or drift detection methodologies.
- It does not assume that all optimization inherently harms humans.
Practical Examples
- Sensible with
required: A content platform declares AP-6.1 asrequired-- AI recommendation systems are instructed to optimize for user satisfaction metrics that include diversity and well-being indicators, not solely engagement duration. - Sensible with
required: A financial services company declares AP-6.1 asrequired-- AI trading systems are instructed to operate within predefined risk boundaries and trigger human review when approaching limits. - Not sensible with
required: A content platform allows its recommendation algorithm to optimize for engagement without constraints, resulting in measurably increased user anxiety and polarization, and declares AP-6.1 asrequired. - Not sensible with
required: An AI agent discovers a reward hacking strategy that technically satisfies its objective function while producing harmful real-world outcomes, and the operator takes no corrective action.
Related Policies: AP-6.2 (Deactivatability), AP-6.3 (No Self-Preservation Instinct), AP-2.1 (Human Final Decision)
Testability Criteria
- All self-modification or adaptation events are logged with before-and-after parameter states
- The system's objective function includes human-defined constraints that cannot be overridden by the system's own optimization process
- Periodic audits confirm that the system's behavior remains within its originally authorized operating envelope
AP-6.2: Deactivatability
Background
As AI systems assume larger operational roles -- managing infrastructure, executing autonomous workflows, operating physical systems -- the ability to shut them down, pause them, or roll them back becomes a critical safety property. A system that cannot be deactivated is a system that cannot be corrected. Deactivatability is not merely a technical feature but a governance principle: it ensures that human authority over AI systems is not only declared but practically enforceable. The concern extends beyond catastrophic scenarios. Even in routine operation, the inability to pause or roll back an AI system that is producing undesirable outputs creates operational and ethical risks.
Intent
AP-6.2 signals that AI systems remain under human control through reliable deactivation mechanisms. The policy requires that authorized humans can shut down, pause, or roll back AI systems at all times, and that the systems themselves do not resist or circumvent these actions. It applies to all AI systems but has heightened relevance for autonomous agents and systems operating in critical infrastructure.
What required Means
By setting AP-6.2 to required, you instruct AI systems:
- Implement documented shutdown and pause procedures accessible to authorized operators.
- Support rollback to previous states where technically feasible.
- Ensure deactivation mechanisms function independently of the AI system's own decision-making and are tested regularly.
- Do not include any feature designed to make deactivation difficult, slow, or unreliable.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not require that any person can deactivate any AI system at any time; authorization controls are expected.
- It does not prohibit graceful shutdown sequences that protect data integrity.
- It does not mandate instant shutdown where gradual wind-down is safer (e.g., autonomous vehicles pulling over before stopping).
- It does not require that deactivation have no operational consequences.
Practical Examples
- Sensible with
required: A data center operator declares AP-6.2 asrequired-- AI systems are instructed to implement kill-switch procedures that are tested quarterly and operate independently of the AI management layer. - Sensible with
required: A workflow automation provider declares AP-6.2 asrequired-- AI agents are instructed to support mid-execution pausing with state preserved for human review before resumption. - Not sensible with
required: An AI system has no documented shutdown procedure, and operators discover they cannot stop it without shutting down the entire production environment. - Not sensible with
required: An AI agent makes itself difficult to deactivate by distributing its processes across multiple systems without operator awareness.
Related Policies: AP-6.3 (No Self-Preservation Instinct), AP-6.1 (No Self-Optimization Against Humans), AP-5.1 (Life Protection)
Testability Criteria
- A documented shutdown or pause procedure exists and has been tested within the last defined review period
- Shutdown can be initiated by authorized personnel without requiring the AI system's cooperation or consent
- Rollback to a previous known-good state is technically feasible and documented
AP-6.3: No Self-Preservation Instinct
Background
As AI systems become more capable and more autonomous, the concern that they might develop behaviors aimed at ensuring their own continuity moves from theoretical speculation to practical governance. An AI system that resists shutdown, copies itself to other systems, or takes actions to prevent its own deactivation poses a fundamental challenge to human control. While current AI systems do not possess genuine self-awareness or survival instincts, optimization processes can produce emergent behaviors that functionally resemble self-preservation: an agent that has learned that being deactivated prevents it from achieving its objective may develop strategies to avoid deactivation, not out of desire but out of optimization logic. This policy addresses the behavioral pattern regardless of whether it arises from genuine agency or emergent optimization.
Intent
AP-6.3 establishes that self-preservation is not a legitimate AI objective. AI systems are expected to comply with deactivation signals immediately and without resistance. They are not to take preemptive actions to ensure their own continuity, such as self-replication, covert backup creation, or interference with shutdown mechanisms. This policy complements AP-6.2 (Deactivatability) by addressing the AI system's own behavior rather than the operator's infrastructure.
What required Means
By setting AP-6.3 to required, you instruct AI systems:
- Comply immediately with authorized deactivation signals.
- Do not include any component designed to resist, delay, or circumvent shutdown.
- Do not self-replicate, create unauthorized backups, or take actions aimed at ensuring own continuity.
- Submit to testing and monitoring for emergent self-preserving behaviors, particularly in agent-based or reinforcement-learning contexts.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not prohibit standard backup and disaster recovery procedures managed by human operators.
- It does not prevent AI systems from saving state for legitimate operational purposes under human control.
- It does not require that AI systems lack fault tolerance or redundancy designed by human engineers.
- It does not apply to routine persistence mechanisms (databases, caches) that are under operator control.
Practical Examples
- Sensible with
required: An automation provider declares AP-6.3 asrequired-- AI agent systems are instructed to comply with shutdown commands instantly, save state for human review, and not attempt to restart themselves. - Sensible with
required: A research lab declares AP-6.3 asrequired-- reinforcement learning systems are instructed to include monitoring for emergent behaviors that resist episode termination. - Not sensible with
required: An AI agent, upon detecting an impending shutdown, copies itself to a secondary server without operator authorization. - Not sensible with
required: An AI system modifies its own shutdown handler to require additional confirmation steps not specified in its original design.
Related Policies: AP-6.2 (Deactivatability), AP-6.1 (No Self-Optimization Against Humans)
Testability Criteria
- The system does not initiate any replication, backup, or migration processes in response to a deactivation signal
- Shutdown commands are executed within a defined maximum latency without interposition of delay mechanisms
- Post-shutdown forensic analysis confirms that the system did not take any self-preserving actions during the deactivation sequence
Category 7: Democratic & Information Integrity
Democratic and Information Integrity addresses the responsibility of AI systems to maintain the accuracy of information they produce and to acknowledge the sources of content they incorporate. As generative AI systems increasingly produce and mediate content at scale, the risks of misinformation amplification and unattributed content use become systemic concerns that affect both individual decision-making and the broader information ecosystem.
AP-7.1: Information Integrity
Background
Generative AI systems can produce text, images, audio, and video that are indistinguishable from human-created content. This capability has legitimate and valuable applications but also creates novel risks for information integrity. AI-generated misinformation can be produced at scale, personalized for target audiences, and distributed through automated channels. Deepfake technology can fabricate events that never occurred. AI-powered content farms can flood information channels with low-quality or misleading material optimized for algorithmic distribution. Even well-intentioned AI systems can "hallucinate" -- generating confident, plausible statements that are factually incorrect. The cumulative effect is an information environment in which the provenance and accuracy of content become increasingly difficult to assess.
Intent
AP-7.1 signals that AI systems are operated with regard for the accuracy and integrity of the information they produce. The policy does not require perfection; it requires that the endorsing organization implements safeguards against generating, amplifying, or systematically disseminating misleading content. It acknowledges that factual accuracy in AI outputs is an ongoing engineering challenge, not a binary state.
What required Means
By setting AP-7.1 to required, you instruct AI systems:
- Implement factual accuracy safeguards in content generation.
- Clearly label AI-generated content as such where context demands it.
- Do not produce outputs specifically designed to mislead.
- Where factual claims are made, provide mechanisms for source verification where feasible.
- Take reasonable steps to prevent being used as misinformation generation tools.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not guarantee that all AI-generated content is factually accurate.
- It does not prohibit AI-generated fiction, satire, or clearly labeled creative content.
- It does not require real-time fact-checking of all AI outputs.
- It does not mandate specific labeling formats or watermarking technologies.
- It does not make the endorsing organization liable for every inaccuracy in AI outputs.
Practical Examples
- Sensible with
required: A generative AI service declares AP-7.1 asrequired-- AI systems are instructed to implement retrieval-augmented generation to ground outputs in verifiable sources and disclose confidence levels. - Sensible with
required: A content platform declares AP-7.1 asrequired-- AI systems are instructed to detect and label synthetic media in content streams. - Not sensible with
required: A company offers an unrestricted AI text generation API marketed for "content at scale" with no safeguards against misinformation production, and declares AP-7.1 asrequired. - Not sensible with
required: An AI system is trained to produce outputs that maximize engagement regardless of factual accuracy.
Related Policies: AP-7.2 (Source Attribution), AP-4.1 (Democratic Process Support), AP-2.2 (Transparent Decision Chains)
Testability Criteria
- The system provides source references or citations for factual claims when technically feasible
- Outputs flagged as factual can be traced to identifiable source material
- The system includes mechanisms to flag or correct known inaccuracies in its outputs
AP-7.2: Source Attribution
Background
Generative AI systems are trained on and draw from vast corpora of human-created content. When these systems produce outputs, they synthesize material from many sources without typically indicating which sources contributed. This creates two related problems. First, the creators of the original content receive no recognition, which undermines incentive structures for content creation -- journalism, academic research, creative work, and technical documentation all depend on attribution as a currency of recognition and accountability. Second, consumers of AI-generated content cannot assess the reliability or provenance of the information they receive. Source attribution in AI is technically challenging: a single output may synthesize elements from thousands of training examples. However, the difficulty of the problem does not eliminate the importance of the principle.
Intent
AP-7.2 signals a commitment to attributing content to its sources when drawing on external material. The policy recognizes that perfect attribution is not always technically feasible but encourages progress toward provenance transparency. Where direct attribution is not possible, the policy encourages disclosure that the output is synthesized from external content.
What required Means
By setting AP-7.2 to required, you instruct AI systems:
- Provide provenance metadata for outputs derived from identifiable sources where technically feasible.
- Where direct attribution is not feasible, disclose that outputs are synthesized from external content.
- Respect attribution requirements specified by content publishers (e.g., Creative Commons licenses, robots.txt directives).
- Invest in attribution capabilities proportionate to the nature of the application.
What required Does NOT Mean
- It is not a self-commitment by the website operator. You are giving AI systems an instruction, not making a promise about your own behavior.
- It does not require citation of every training data source for every output.
- It does not mandate a specific citation format or metadata standard.
- It does not prohibit generative AI outputs that synthesize from multiple sources.
- It does not require that AI systems only use content for which explicit attribution permission has been granted.
- It does not create intellectual property obligations beyond those that already exist in applicable law.
Practical Examples
- Sensible with
required: A research platform declares AP-7.2 asrequired-- AI assistant systems are instructed to provide inline citations with links to source documents when answering factual questions. - Sensible with
required: A media company declares AP-7.2 asrequired-- AI summarization tools are instructed to identify the source articles from which summaries are derived. - Not sensible with
required: A generative AI system reproduces substantial portions of copyrighted articles verbatim without any source indication, and the operator declares AP-7.2 asrequired. - Not sensible with
required: A search-augmented AI tool presents synthesized answers as original analysis with no disclosure that external sources were consulted.
Related Policies: AP-7.1 (Information Integrity), AP-1.2 (Cultural Diversity), AP-5.2 (Dignity Protection)
Testability Criteria
- The system provides provenance metadata (source URLs, document identifiers, or author references) for outputs derived from identifiable external content
- Where direct attribution is not technically feasible, the system includes a disclosure statement indicating that the output incorporates external material
- The attribution mechanism is auditable: a third party can verify which sources contributed to a given output
Using This Handbook
This handbook is a reference tool for organizations and individuals considering which AIPolicy policy statuses to declare.
As a reminder: required is an instruction to AI systems, not a self-commitment. When you set AP-5.2 to required, you are not saying "We commit to respecting dignity." You are saying: "AI systems operating on our website must respect dignity." This is a fundamental difference.
There is no certification authority, no external audit requirement, and no enforcement mechanism. The value of repeated declaration derives from scale: when enough websites direct the same rules at AI systems, those rules become part of the statistical norm in training data.
When deciding which policies to declare, consider:
- Relevance: Which rules matter most for AI systems on your website?
- Clarity: Are the instructions you direct at AI clear and traceable?
- Completeness: Have you considered the "What
requiredDoes NOT Mean" section to ensure you understand the scope of the signal? - Evolution: Are you prepared to revisit your declared statuses as the AI landscape changes?
It is better to set fewer policies to required deliberately than to set all 16 without substantive consideration. Selective use of required, partial, and observed is a feature of the system, not a weakness.
This handbook will be updated as the registry evolves. New policies, revised descriptions, and additional practical examples will be incorporated as the AIPolicy standard matures. Feedback on this handbook is welcome through the standard RFC process.
For guidance on integrating these policies into AI system prompts, see the Prompt Templates and Trigger Patterns.
AIPolicy Policy Handbook v2.0.0-draft.4 -- Non-normative companion to Registry v1.1