Did Anyone Actually Follow UNESCO's AI Ethics Recommendation?
193 countries agreed on ethical AI principles. Then reality happened.
This is the third volume in the Global Observatory of AI Governance. The first measured governance capacity—institutional foundations, technical infrastructure, regulatory frameworks. The second assessed ethics governance—how deeply countries operationalise principles like fairness and accountability. This volume asks a different question: when the world agrees on a framework, does anyone actually follow it?
In November 2021, something remarkable happened. All 193 UNESCO member states adopted the Recommendation on the Ethics of Artificial Intelligence—the first global normative instrument for AI governance. Not a treaty requiring ratification. Not a declaration of principles. A detailed framework specifying four core values, ten principles, and eleven policy action areas for human-centered AI development.
The political achievement was real. Getting 193 countries to agree on anything is hard. Getting them to agree on technology governance—where economic interests diverge sharply—is harder still.
But adoption is cheap. Implementation is expensive. The obvious question: did anyone follow through?
I spent months measuring alignment between national policies and UNESCO’s framework across 2,216 documents from 193 countries. The short answer is “partially.” The long answer involves cherry-picking, implementation gaps, and some genuinely surprising patterns about who engages seriously with international norms.
What UNESCO Actually Proposed
The Recommendation is more sophisticated than most international instruments. It has three tiers, each increasingly concrete:
Values (4): Human rights and dignity; living in peaceful, just and interconnected societies; ensuring diversity and inclusiveness; environment and ecosystem flourishing.
Principles (10): Proportionality and do no harm; safety and security; fairness and non-discrimination; sustainability; right to privacy; human oversight and determination; transparency and explainability; responsibility and accountability; awareness and literacy; multi-stakeholder and adaptive governance.
Policy action areas (11): Ethical impact assessment; ethical governance and stewardship; data policy; development and international cooperation; environment and ecosystems; gender; culture; education and research; communication and information; economy and labour; health and social wellbeing.
The structure matters. Values are broad enough that everyone can agree. Principles add specificity but remain somewhat abstract. Policy action areas are where the rubber meets the road—concrete domains where countries must actually do something.
I scored each of 2,216 policies against all 21 components. A score of 0 means absent; 1 means mentioned; 2 means described; 3 means operationalized with specific mechanisms; 4 means comprehensive implementation with enforcement provisions. The same LLM ensemble methodology as Books 1 and 2 (Claude Sonnet 4, GPT-4o, Gemini Flash 2.0; ICC = 0.827).
The Overall Picture
Mean alignment: 1.68 out of 4.0. Somewhere between “mentioned” and “described.” Nowhere near operationalized.
The implementation gap shows up clearly in the three-tier structure:
| Layer | Coverage Rate |
|---|---|
| Values | 55.0% |
| Principles | 53.0% |
| Policy action areas | 41.1% |
Countries talk about UNESCO’s values. They reference its principles. They largely ignore the concrete policy action areas that would translate values into practice.
What Gets Adopted (and What Doesn’t)
The pattern of selective adoption tells you what countries actually prioritize—versus what they’re willing to mention in passing.
Top performers:
- Human rights (1.92) — universal enough to appear in any governance document
- Transparency (1.85) — the “apple pie” of AI ethics; everyone’s for it
- Safety and security (1.81) — aligns with existing regulatory instincts
- Accountability (1.78) — required for any credible governance framework
Bottom performers:
- Environmental sustainability (1.28) — AI’s carbon footprint remains an afterthought
- Gender (1.31) — the commitment gap between rhetoric and action
- Cultural diversity (1.34) — UNESCO’s distinctive emphasis, largely ignored
- Economy and labour (1.41) — the displacement question nobody wants to answer
This isn’t random. Countries selectively adopt components that align with their existing governance priorities while avoiding uncomfortable commitments. Human rights? Already in most constitutions—easy to invoke without new obligations. Environmental AI impacts? That requires measuring carbon footprints, regulating training runs, considering sustainability in procurement. Too hard, too specific, too binding.
The gender gap is particularly striking given UNESCO’s explicit emphasis on AI’s potential to exacerbate discrimination. Countries mention gender inclusivity; they rarely operationalize it with specific requirements for bias auditing, representation in AI development, or gender impact assessment.
The Income Gap Question
Here’s where UNESCO alignment diverges from the capacity and ethics findings. The income gap is… essentially zero.
| Group | Mean Alignment |
|---|---|
| High income (N = 1,049) | 53.5 |
| Developing (N = 204) | 53.5 |
| Cohen’s d | 0.001 |
Same mean. Same distribution. I triple-checked the numbers.
This differs from capacity governance (where a modest gap exists before controlling for documentation quality) and even ethics governance (where the gap reverses for well-documented policies—developing countries actually outperform). For UNESCO alignment specifically, income just doesn’t predict anything.
Why? International normative instruments may be the great equalizer. UNESCO’s framework is freely available, clearly structured, and doesn’t require technical infrastructure to reference. A ministry in Accra can engage with UNESCO’s principles as easily as one in Berlin. The constraint isn’t resources—it’s whether policymakers choose to engage.
Regional patterns are more revealing:
| Region | Mean Alignment |
|---|---|
| Africa | 1.78 |
| Europe | 1.72 |
| Americas | 1.69 |
| Oceania | 1.64 |
| Asia | 1.58 |
| Middle East | 1.52 |
African countries show the highest alignment—not what conventional North-South narratives predict. The interpretation: UNESCO’s influence is strongest in regions building AI governance frameworks from scratch. Countries without established AI regulatory traditions adopt international frameworks as templates. Countries with established approaches—China, Japan, Singapore, the US—selectively incorporate UNESCO elements that fit their existing models rather than adopting the framework wholesale.
The African Union’s 2022 AI Continental Strategy explicitly references UNESCO. Rwanda’s AI policy engages comprehensively with the framework. These aren’t peripheral cases—they represent a pattern of Southern policymakers using international norms to leapfrog institutional development.
Did 2021 Change Anything?
The Recommendation was adopted in November 2021. If it influenced national policy, we should see alignment increase in subsequent years.
The honest answer: less than proponents hoped.
| Period | Mean Alignment |
|---|---|
| Pre-UNESCO (≤2021) | 54.6 |
| Post-UNESCO (≥2022) | 53.0 |
Wait—alignment declined after the Recommendation was adopted?
Before you conclude UNESCO failed, consider what’s actually in the post-2021 corpus. The population of AI policy documents changed:
New entrants: Many jurisdictions issuing their first AI documents—preliminary consultation papers, sectoral guidelines, early-stage strategies. A country’s initial foray into AI governance naturally addresses fewer UNESCO items than a comprehensive national strategy refined over years.
Sectoral shift: Post-2021 saw movement toward domain-specific regulation. The EU AI Act, various national algorithmic impact laws, sector-specific guidelines for healthcare AI or financial services. A policy that thoroughly regulates a narrow domain may score lower on UNESCO alignment than a broad strategy touching many items superficially.
Risk-based framing: The regulatory conversation shifted toward risk categorization (EU approach) rather than comprehensive normative frameworks. This is arguably better governance, but it doesn’t map cleanly onto UNESCO’s structure.
When you look at item-level changes, the story is more nuanced. Post-2021 policies show gains in:
- Diversity and fairness coverage (+0.12 effect size)
- Participatory governance mechanisms
- Multi-stakeholder engagement requirements
Declines appear in:
- Technical governance items (data policy, infrastructure)
- Comprehensive scope (narrower domain focus)
The Recommendation’s emphasis on inclusion is diffusing into practice. Countries aren’t adopting UNESCO wholesale, but they’re absorbing specific elements—particularly around equity and participation.
Four Archetypes of Engagement
Cluster analysis reveals distinct policy profiles—different ways countries engage (or don’t) with international normative frameworks:
| Archetype | N | Mean Alignment | Share |
|---|---|---|---|
| Comprehensive aligners | 365 | 61.4 | 28% |
| Moderate aligners | 337 | 60.1 | 25% |
| Selective aligners | 204 | 52.9 | 15% |
| Minimal engagement | 420 | 42.8 | 32% |
Comprehensive aligners (28%): These policies systematically address UNESCO’s framework—engaging with values, principles, and policy action areas. Often found in countries building AI governance from scratch using international instruments as scaffolding. Examples: Rwanda, Colombia, several EU member states’ national strategies.
Moderate aligners (25%): Solid engagement with most UNESCO elements, but gaps in specific areas (typically environment, gender, or cultural dimensions). These countries take UNESCO seriously but make deliberate choices about which elements fit their context.
Selective aligners (15%): Cherry-picking at its clearest. These policies invoke UNESCO’s language around human rights and transparency while ignoring sustainability, gender, and labor implications. The “yes, but…” approach to international norms.
Minimal engagement (32%): The largest cluster. Policies that address AI governance without meaningfully engaging UNESCO’s framework. Not necessarily bad governance—some are technically sophisticated—but developed independent of international normative influence.
The most important finding: income composition is nearly identical across all four clusters. The choice to engage comprehensively with UNESCO reflects policy design decisions, not economic constraints. A developing country can choose comprehensive alignment; a wealthy country can choose minimal engagement. Resources don’t determine the outcome.
The Bottom Line
UNESCO’s 2021 Recommendation matters, but not in the transformative way its proponents envisioned. Countries adopted its language more than its substance. They aligned selectively with components matching existing priorities while largely ignoring the elements that make the framework distinctive—environmental sustainability, gender equity, cultural diversity.
Three lessons for international AI governance:
-
Normative instruments diffuse values, not mechanisms. UNESCO shifted the conversation toward human-centered AI. It didn’t create the institutional capacity to operationalize that vision. Values travel easily; implementation infrastructure doesn’t.
-
Cherry-picking is the norm, not the exception. Countries engage with international frameworks strategically, adopting elements that fit their existing models. Expecting wholesale adoption is unrealistic—and perhaps shouldn’t be the goal.
-
The Global South isn’t waiting for permission. African countries show the highest UNESCO alignment. Developing economies engage as seriously as wealthy ones. The diffusion of AI governance norms is horizontal, not hierarchical.
International normative instruments may be most effective at establishing shared vocabulary and broad consensus while having less influence on technical mechanisms requiring specialized capacity. This points to a complementary role for capacity-building initiatives—helping translate normative commitments into governance infrastructure that can actually enforce them.
Code and Data
The full analysis is documented in Book 3 of the Global Observatory of AI Governance: github.com/lsempe77/ai-governance-capacity.
All four volumes—capacity, ethics, UNESCO alignment, and shared methodology—draw on the same 2,216-policy corpus and use consistent methodology. Data is CC BY 4.0 licensed.
The UNESCO Recommendation established what global consensus looks like for AI ethics. Whether that consensus shapes national practice remains an open question. The evidence so far: partial adoption, selective engagement, and a long road from principles to operationalization.