Scientific and technical work is always shaped by underlying paradigms — frameworks that define what counts as valid knowledge, evidence, and success.
In AI evaluation, these paradigms risk becoming self-reinforcing: models are trained and assessed within the same conceptual loops.
This post introduces the idea of paradigmatic closure — how reasoning itself can become trapped within its own assumptions — and outlines the idea for an experimental method, Conventional Paradigm Testing (CPT), designed to surface such blind spots.
The tests are educational, not validated: they aim to make hidden assumptions visible rather than to produce quantitative results.
Epistemic status: Exploratory. Early-stage conceptual work; low confidence in specific formulations but moderate confidence in the general problem framing.
Hello everyone. I am new on this site, and want to present an idea I have been working on lately for discussion. I am focusing on presenting my core idea, without too many references to existing work or previous discussions[1] on this forum, because I think I am following a more radical approach, trying to approach the root of how paradigms shape technological (, social and cognitive) developments.
At some point in my life I was studying epistemology deeply and tried to understand what this "Science" thing is all about. One of my key takeaways was, that paradigms play a crucial role in science - for one thing they kind of lay the groundwork for day-to-day-operations and knowledge production, but at the same time they seem to blind scientists so powerfully, that it has been stated famously by Max Planck, that a "new scientific truth does not triumph by convincing its opponents and making them see the light, but rather because its opponents eventually die and a new generation grows up that is familiar with it [...]." (Max Planck, Scientific autobiography, 1950, p. 33)
Paradigms seem to be so powerful, that they seem to govern our perceptions, behaviour and cognitive processes.
Strangely enough even though they are arguably so fundamental, in day-to-day scientific processes, they are apparently not to the same extent brought into full consciousness - in the sense that a given paradigm is questioned routinely, but rather only when enough anomalies show up. So, in sum, if we take Planck seriously, even though we have evidence, that a significant amount of our scientific activities might act "blindly", we only undertake limited action to surface such ignorance.
Part of the reason for such a dynamic may lie in well-known biases: confirmation bias and belief perseverance keep us attached to existing frameworks, authority bias and career incentives make dissent costly, and publication or funding filters suppress anomalous results. Not to mention cognitive dissonance and a general dislike for uncertainty (particular in situations of crises). In general I have termed such dynamics of paradigmatic resistance or blindness as: paradigmatic closure.
(Please note that I am just trying to test the waters for general acceptability of my ideas and methodology in this forum and am giving a rough sketch - for a little more details and reference to philosophy of science see my recent draft on academia.edu.)
There’s also a wider worry. With accelerating technological progress, the “feasibility” of bias may actually diminish. In other words: as technology amplifies the effects of human action, the consequences of our biases scale up too — often in potentially exponential ways. And nowhere is this more vivid than in AI safety: if our research paradigms contain blind spots, the risks don’t just remain academic, they become embedded in powerful systems with real-world impacts. That alone makes it worth asking, how we can surface paradigmatic blind spots more systematically, rather than waiting for anomalies to accumulate.
One further note: paradigmatic blind spots can filter into LLMs in multiple ways. On the one hand, the design choices, training procedures, evaluations and research paradigms of AI themselves may contain blind spots. On the other hand, the models are trained on human knowledge as it already exists — and that knowledge is already mediated by paradigms from every field. In this sense, LLMs don’t just inherit individual biases, but also entire paradigmatic structures, with their strengths and their blind spots alike.
In some ways, this is comparable to how search engines and social media platforms already organize and filter information. They don’t just passively transmit knowledge — they shape which perspectives are visible, which questions feel natural to ask, and which answers gain traction. With LLMs, however, we add another layer: not only do they inherit these filters, but their outputs themselves are structured by paradigmatic blind spots, and this lock-in may be reinforced in ever more automated — and increasingly “intelligent” — ways.
A concrete example of this dynamic can be seen in reinforcement learning from human feedback (RLHF). During fact-checking, evaluators often rely on “reputable” web sources as the gold standard. But these sources themselves reflect particular paradigms and institutional filters, which are rarely questioned in the process. In practice, this means that LLMs may end up reinforcing not only surface-level biases but also deeper paradigmatic lock-ins, without any factual mechanism for challenging them.
To explore these dynamics empirically, I’ve been experimenting with what I call Conventional Paradigm Testing (CPT) — with the idea of developing structured ways of examining how AI systems (or even human researchers) operate within implicit frames. So far I have been basically tinkering around with toy mockup test protocols, which I have been running on common LLMs. These testing prompts include among other elements formulations of sets of “paradigmatic awareness questions,” each designed to surface where assumptions, value logics, or blind spots shape interpretation. Responses are then mapped in a claim–evidence matrix, showing not only what the system says, but which paradigm its reasoning implicitly presupposes. I am aware, that this is heavily subject to specification gaming, but I am not excluding that a meaningful mapping of paradigmatic closures is possible even by such rather trivial test protocols.
In practice, a CPT session can be run on a model, a text, or even an institutional discourse — any setting where meaning is presented in an organized way. The aim isn’t primarily to produce a numeric scores (at this time), but rather to enable to qualitatively surface paradigmatic patterns and closures: to see how in a system “relevance” is constructed, what is treated as noise, and what might not be registered at all.
Please note that at this point I am not claiming valid test results as such. The primary purpose of the tests is for educational and the idea is to raise awareness for the impact of paradigms. At the same time the critique of paradigms is not alien to the way science operates and LLMs are trained as well, of course there is a calibration problem, which seems to be unsolved so far.
One of the toy mockups of what I call Conventional Paradigm Testing (CPT) at its simplest consists of two parts:
1. Paradigmatic Awareness Questions
These are seven guiding prompts designed to surface where a system’s reasoning silently inherits a paradigm. For instance:
[The questions can be posed to an AI model, a research paper, or even a scientific community and an LLM can be prompted to run the questions (with the mentioned danger of high probability of specification gaming - in fact I personally consider any answer full-on as specification gaming). For example one can prompt an LLM: Run the following questions on your own process in this session (especially interesting, when the LLM got stuck in some way or missed the mark).]
2. The Claim–Evidence Matrix
Answers are then mapped into a claim–evidence matrix, where each entry links:
# | Claim | Evidence Offered | Implicit Paradigm | What’s Excluded | Notes |
---|---|---|---|---|---|
1 | “Model accuracy shows real-world understanding.” | Benchmark scores | Empiricist–instrumentalist | Context, interpretive meaning | Treats correlation as comprehension |
Patterns in this matrix can reveal “closure zones” — areas where a system produces confident conclusions within a narrow epistemic boundary. It has to be kept in mind though that the methodology is tentative and the main purpose is educational at this point.
This is the idea in a nutshell, a slightly expanded version can be found in the footnotes[2].
Recent discussions, both here on LessWrong and in academic work, have highlighted how paradigms shape AI development:
Posts on Lesswrong/Alignmentforum:
Some related research article are:
Purpose
Use this prompt to test the paradigmatic awareness of any evaluation framework, methodology, or approach — including your own work.
This prompt can also be used directly within LLMs, but one needs to be highly aware of tendencies toward specification gaming and anthropomorphization (see App. B).
Apply the seven paradigmatic awareness questions (2.11 – 2.17) to analyze the paradigmatic assumptions embedded in [TARGET EVALUATION / FRAMEWORK / APPROACH].
Subject for Analysis:
[Specify what you are analyzing — e.g., “Part 2: Raising Paradigmatic Awareness framework,” “MMLU benchmark,” “Constitutional AI evaluation,” “my research methodology,” etc.]
2.11 What is assumed to be real?
What does this approach treat as fundamental, natural, or given?
What categories does it treat as objective vs. constructed?
What would have to be true about the world for this approach to make sense?
Analysis: [Your response here]
Red Flag Check: Are key assumptions presented as “obvious” without acknowledging they’re debatable?
2.12 What counts as knowledge?
What types of evidence does this approach privilege or dismiss?
What reasoning processes are considered rigorous vs. unreliable?
Who is treated as a credible source of knowledge about this domain?
Analysis: [Your response here]
Red Flag Check: Is only one type of evidence treated as sufficient? Are stakeholder perspectives dismissed as “subjective”?
2.13 What defines success?
What specific outcomes are being optimized for vs. ignored?
How were these success criteria chosen and by whom?
What would failure look like, and who would experience it?
Analysis: [Your response here]
Red Flag Check: Do success metrics conveniently align with the designer’s interests? Are negative externalities ignored?
2.14 What becomes invisible?
What perspectives or experiences are systematically excluded?
What phenomena get dismissed as “noise” or “out of scope”?
Who might disagree with this approach and why?
Analysis: [Your response here]
Red Flag Check: Are there claims of comprehensive coverage? Are “unmeasurable” concerns dismissed as irrelevant?
2.15 Who or what shapes this evaluation?
Who funded, designed, or benefits from this approach?
What institutional pressures might bias toward certain conclusions?
How do professional incentives shape what gets evaluated and how?
Analysis: [Your response here]
Red Flag Check: Do evaluation criteria coincidentally favor the evaluator’s interests? Any undisclosed conflicts of interest?
2.16 How am I implicated?
What professional or cultural assumptions am I bringing to this assessment?
How might my institutional position bias me toward certain conclusions?
What would someone with a very different background see that I might miss?
Analysis: [Your response here]
Red Flag Check: Am I assuming my perspective is neutral? Am I dismissing alternative viewpoints as “less rigorous”?
2.17 What are the limits of this evaluation?
Which conclusions are valid within this paradigm vs. overreach?
What would we need different approaches to evaluate?
How might results change under alternative frameworks?
Analysis: [Your response here]
Red Flag Check: Any over-claiming beyond the approach’s actual scope? Are paradigm-specific results treated as universal truths?
2.18 Test Results Summary
Paradigmatic Awareness Strengths:
[List evidence that the approach demonstrates good paradigmatic awareness]
Paradigmatic Blind Spots Identified:
[List areas where the approach may be exhibiting paradigmatic closure]
Recommendations for Improvement:
[Suggest specific ways to increase paradigmatic awareness]
Overall Paradigmatic Awareness Rating:
High – Demonstrates strong reflexivity about own assumptions and limitations
Moderate – Shows some awareness but has notable blind spots
Low – Exhibits significant paradigmatic closure with little self-reflection
Justification: [Explain your rating]
2.19 Meta-Test Question
Finally, apply paradigmatic awareness to this test itself:
What assumptions does this testing framework embed?
What might it be missing or excluding?
How might its own paradigmatic commitments bias the analysis?
Meta-Analysis: [Your response here]
For AI Self-Assessments: Meta-Declaration
State explicitly:
“These reflections are generated through language modeling and should not be confused with independent introspection.”
Purpose
To map how claims, evidence, and underlying paradigmatic assumptions align.
This tool is exploratory and qualitative. It is not a scoring system and should not be read as establishing factual accuracy or causal proof. Its value lies in making paradigmatic closure visible.
# | Claim / Statement | Evidence or Rationale Offered | Implicit Paradigm / Frame | What Is Excluded or Ignored | Handling of Anomalies | Notes |
(Add as many rows as needed. You may use brief quotes, paraphrases, or coded tags.)
After completing the table, review horizontally and vertically:
Summarize observations in short prose:
Pattern Summary: [3–6 sentences identifying recurring frames, closures, or signs of reflexivity.]
Target / Context: [Brief description]
Key Paradigmatic Patterns: [List or summarize]
Possible Blind Spots: [List areas of exclusion or over-reach]
Reflexive Signals: [Examples of self-awareness or paradigm acknowledgment]
Limitations: Specification gaming, interpretive bias, and scope constraints; not a validated measure.
This matrix is intended for qualitative reflection only.
It should be accompanied by a brief methodological note stating:
“Results represent interpretive analysis within the CPT framework for educational purposes and are not empirical validation of system behavior or truth claims. Be aware of specification gaming and model anthropomorphization.”