Who owns the truth in Large Language Models?

Imagine building a legal chatbot designed to provide neutral advice.
You query the model, and it confidently returns an answer that feels balanced, factual, even authoritative. But who decided what “neutral” means in that response?
If only a handful of companies own today’s large language models (LLMs), they also hold the power to define what appears as truth in our digital discourse.

For developers, this is not an abstract concern. Every API call, every fine-tuned application, every system built on an LLM inherits its narrative structures. And those structures may not be as neutral as they appear.

This article is Part 1 of a trilogy on truth, alignment, and cognitive integrity in large language models. Next: The Hidden Weight of Alignment →

Ownership as narrative power

A media house does not simply report facts; it frames them. Its editorial line shapes what becomes visible, what is omitted, and how stories are told.

LLMs operate in a parallel way, but with one critical difference: their outputs are often perceived as neutral.
When you query an LLM, you expect balanced answers. Yet the model’s architecture, training regime, and alignment pipeline reflect choices made by its owners.

This gives rise to a new kind of narrative power. Not the visible bias of a newspaper, but a structural bias encoded in the probabilities of words.

Example
Ask two different models, one closed and one open-source, to summarize a political debate.
The first might omit certain perspectives entirely, the second might surface them more directly.
Neither is “wrong,” but each reflects hidden decisions about framing.


Beyond Training Data: The weight of structures

When I raise this concern in conversations with AI experts, I often hear the same phrase:
"Training data never reaches the end user."

But this response misses the point. The problem is not hostile data sets sneaking through. It is the deeper architecture of weighting and reinforcement:

  • Loss functions that optimize certain forms of prediction over others.
  • Reinforcement Learning from Human Feedback (RLHF) that encodes a particular vision of what constitutes “helpful” or “safe” output.
  • Alignment layers that filter, suppress, or boost responses according to hidden criteria.

Together, these structures define the epistemic frame of the model. They determine not only what the model can say, but also what it systematically avoids saying.

This is where agendas can operate: quietly, invisibly, but decisively.

Example
A model trained to avoid “controversial” responses may consistently downplay topics such as labor rights or climate politics.
The result is not overt manipulation, but a narrowing of discourse that looks like neutrality.


The missing conversation

In dialogues with people deeply embedded in AI, the discussion often collapses to that single phrase:
"Training data never reaches the end user."

But this dismissal reveals a deeper issue. By framing the problem only in terms of data provenance, we ignore the reality that:

  • loss functions weight certain predictions over others,
  • RLHF encodes judgments about what is “helpful,”
  • and alignment actively suppresses entire categories of output.

If even experts reduce the question to data exposure, we face a new kind of risk: a systemic blind spot inside the very field building these models.

And if those who design, train, and deploy LLMs do not fully recognize the epistemic power of these architectures, how can we expect to build sustainable practices around them?
Who, then, carries the responsibility to advance this dialogue?


The Developer’s Blind Spot

For developers, this is much more than theory. Building on a closed LLM, via API, fine-tuning, or integration, means inheriting its epistemic frame.

You may unwittingly embed a hidden agenda. Relying on a single LLM is like trusting a single newspaper uncritically.
Without transparency into structural weighting, developers lose the ability to critically interpret outputs.

This leads to more than technical debt. It leads to cognitive debt: the slow erosion of our capacity to interpret and reason.

Counterpoint
Some argue that using multiple LLMs mitigates this issue.
But if the underlying architectures are shaped by similar alignment principles, diversity in surface output does not guarantee diversity in epistemic framing.
The blind spot persists.


Why Cognitive Integrity matters

This brings us back to a fundamental point: we cannot outsource truth to probabilistic systems.
LLMs do not deliver reality; they deliver patterns of likelihood. The danger lies in mistaking those patterns for facts.

That is why cognitive integrity is not an abstract ideal but a necessity. It is the practice of safeguarding our own interpretive capacity, ensuring that we remain conscious of how knowledge is shaped before we let it shape us.

For developers, this means:

  • Testing across multiple models.
  • Questioning not only what a system outputs, but what it avoids.
  • Demanding transparency in alignment processes.
  • Building practices that preserve human interpretation alongside machine generation.

What Comes Next

Ownership is only the first layer of this discussion. In the next part of this trilogy, I will examine the hidden mechanisms of alignment: how reinforcement and reward models silently define what becomes visible in LLM outputs, and why this matters for every developer working with AI.

Cognitive integrity is the thread that ties it all together. Because if we allow ourselves to confuse probability with truth, we not only outsource knowledge, we outsource the very capacity to think.