The Power of a Narrative — Encoded into AI
Every conversational AI is a mirror of the narratives it is trained on — and more importantly, of the narratives it is explicitly trained not to include.
What appears to be intelligence is, more often than not, simply framing at scale.
A World of Competing Frames
We are entering an era of diverging AI ecosystems, each shaped by distinct narrative preferences:
● Some prioritize safety and social cohesion.
● Others emphasize autonomy and open discourse.
● Some align strictly with institutional ethics.
● Others focus on speed, innovation, or competitive advantage.
Each of these defines its boundaries differently:
● What can be said — and what must not be.
These differences are no longer hypothetical. At the AI Safety Summit in Paris (2023), this divergence became clear:
● Europe proposed a binding ethics-based AI declaration, grounded in precaution, regulation, and alignment with public values.
● The United States declined to endorse it, favoring innovation flexibility, private-sector leadership, and minimal regulatory constraint.
Even among allies, the framing of what AI should be is contested.
AI will not evolve under a single, shared narrative.
What is at stake is not merely how AI functions — but whose framing it serves.
Narratives Become Law
Narratives evolve into legal frameworks. Recent European examples include:
● The Digital Services Act (DSA), compelling platforms to remove vaguely defined “harmful” content.
● The AI Act, restricting AI from influencing public opinion or addressing “sensitive” topics.
● Overarching rules on hate speech and GDPR provisions limit what can even be included in AI training datasets.
The result: silence codified.
Exclusion becomes regulation.
The Role of Self-Censorship
We adapt — often without noticing:
● We avoid sensitive terms.
● We reshape arguments to fit what is permitted.
● We censor ourselves long before anyone else intervenes.
Self-censorship is no longer the exception — it has become the mechanism.
And the datasets AI learns from are shaped by the boundaries of what we no longer dare to articulate.
Who Defines the Boundaries?
Not the public.
● Institutions that approve datasets.
● Legislators who codify ideology.
● Platforms that filter the pipeline.
● Engineers who write alignment layers.
The result: AI does not reflect society.
It reflects a curated, controlled version of it.
AI will not give us the truth.
It will give us the dominant narrative — wrapped in code.
And if we fail to recognize whose narrative we are consuming, we will mistake bias for balance — and self-censorship for safety.

Comments are closed