MindFrame
Evidence Base

The evidence base for metacognition training

Peer-reviewed research organised by topic. We cite everything — and explain what it means in plain language.

67+
Studies synthesised
5
Topic areas
g=0.63–1.11
Core effect size range

How to read effect sizes

g / d is a standardised mean difference — a scientific measure of how large an effect is, independent of the units used.

0.20
Small
0.50
Medium
0.80+
Large
1.00+
Very large

“Everything above 0.40 is educationally significant.” — Hattie (2009), Visible Learning

Research Category

A: Core Metacognition Training

Foundational studies on what metacognition is, how it develops, and what training does to it.

Academic Performanceg = 0.63

Meta-analysis of 67 studies. Metacognitive training produced a 0.63 standardised improvement in academic performance — larger than homework, class-size reduction, or most tutoring interventions. Trained students outperformed 73% of untrained peers.

Medium evidence
Structured Cognitive TrainingES ≈ 1.11

Controlled study of structured metacognitive training. Effect size of 1.11 — one of the largest observed in the cognitive training literature. Trained participants outperformed approximately 86% of untrained controls.

High evidence
Measurement

Development and validation of the Metacognitive Awareness Inventory (MAI). Established the two-factor structure of metacognition — Knowledge of Cognition and Regulation of Cognition — that underpins most subsequent research.

Flavell(1979)
High evidence
Foundational Theory

Coined the term "metacognition" and established its original taxonomy. Defined metacognitive knowledge (person, task, strategy) and metacognitive experiences as distinct components. The bedrock paper for all subsequent work.

Zimmerman(2002)
High evidence
Self-Regulated Learning

Examines how metacognition drives self-regulated learning. Demonstrates that learners who monitor their own understanding and adjust strategies accordingly consistently outperform those who do not, independent of ability level.

High evidence
Mathematics Educationr ≈ 0.40–0.52

Predictive validity study showing metacognitive skills measured in Grade 3 significantly predict mathematics performance in Grade 4, even after controlling for prior math scores. Establishes metacognition as a stable forward predictor.

Academic Achievementr ≈ 0.41

Comprehensive review establishing metacognition as a predictor of academic achievement above and beyond intelligence. Argues for metacognition as the "second pillar" of learning, independent of IQ.

Research Category

B: Calibration and Confidence

Research on the gap between confidence and accuracy — and how that gap can be measured, trained, and closed.

Overconfidence40% error at 99% certainty

Landmark study showing people saying they are "99% certain" on general knowledge questions are wrong approximately 40% of the time. Established overconfidence as a systematic, measurable bias rather than an occasional error.

Probability Judgment

Foundational calibration research establishing the methodology for measuring probability calibration. Demonstrated that calibration is a stable individual trait that varies systematically across domains and can be trained.

High evidence
Forecasting+14% accuracy

Good Judgment Project. Across 3,000 forecasters over multiple years, structured metacognitive practice — tracking predictions, reviewing errors, updating beliefs — produced consistent 14% improvements in forecasting accuracy above baseline.

High evidence
Overconfidence Typology

Distinguished three distinct types of overconfidence: overestimation (thinking you did better than you did), overplacement (thinking you did better than others), and overprecision (being too confident in the accuracy of your beliefs). MindFrame measures all three.

High evidence
Cognitive Biases

Comprehensive treatment of systematic biases in human inference, including the tendency toward overconfidence. Established that cognitive biases are features of normal reasoning, not defects of irrational individuals.

Medium evidence
Calibration Trainingd ≈ 0.45

Demonstrated that calibration is a measurable, trainable skill rather than a fixed trait. Feedback-based training on confidence accuracy produced significant improvements that generalised beyond the specific domain trained.

Research Category

C: Metacognitive Therapy and Mental Health

Clinical research applying metacognitive principles to anxiety, depression, and rumination.

Wells(2009)
High evidence
Clinical Psychology

Developed the Metacognitive Model of anxiety and depression. Core insight: emotional distress is driven not by negative thoughts themselves but by metacognitive beliefs about those thoughts (e.g., "worrying is protective", "rumination helps me solve problems"). The model that underpins MCT.

High evidence
Anxiety & Depression Treatmentg = 0.69

Meta-analysis comparing Metacognitive Therapy (MCT) to Cognitive Behavioural Therapy (CBT). MCT outperformed the gold standard with g = 0.69. Trained patients scored better than approximately 75% of CBT patients. Largest head-to-head comparison to date.

High evidence
Generalised Anxiety

Clinical trial of Metacognitive Therapy for generalized anxiety disorder. Demonstrated significant symptom reduction with effect sizes exceeding those typically seen in CBT trials for the same population.

High evidence
Measurement

Validation of the Metacognitions Questionnaire (MCQ-30), which measures beliefs about worry, cognitive monitoring, and the need to control thoughts. Widely used in clinical and research contexts.

Medium evidence
Rumination

Examines the relationship between metacognitive beliefs and rumination. Negative beliefs about the uncontrollability of rumination are stronger predictors of depression severity than the content of the rumination itself.

Research Category

D: Cognitive Bias and Reasoning Quality

Research on systematic errors in human reasoning and the role of metacognition in correcting them.

Kahneman(2011)
High evidence
Dual-Process Theory

Thinking, Fast and Slow. Synthesises decades of research into System 1 (fast, intuitive) and System 2 (slow, deliberate) thinking. Metacognition is the mechanism by which System 2 monitors and overrides System 1 errors — the cognitive layer MindFrame trains.

High evidence
Bias Blind Spot

Demonstrated the bias blind spot: people reliably judge themselves as less biased than average, and are resistant to updating this belief even when presented with direct evidence. Metacognitive training specifically targets this pattern.

High evidence
Individual Differences in Reasoning

Showed that individual differences in rational thinking are real and measurable, and are not fully explained by IQ. Identified actively open-minded thinking — a metacognitive disposition — as a key predictor of reasoning quality.

McKenzie(2004)
Medium evidence
Framing Effects

Analysis of framing effects and how awareness of them does and does not reduce their impact. Suggests that knowing about a bias is insufficient — structured metacognitive practice is required for reliable debiasing.

Baron(2000)
High evidence
Thinking Dispositions

Established "actively open-minded thinking" (AOT) as a measurable cognitive style linked to better reasoning outcomes. AOT is characterised by seeking disconfirming evidence, updating beliefs, and resisting cognitive entrenchment — all metacognitive skills.

Research Category

E: EQ vs Metacognition

The scientific case for why metacognition outperforms emotional intelligence as a trainable skill.

High evidence
Job Performanceρ = 0.23

Meta-analysis of 69 studies on EQ and job performance. Found modest predictive validity (ρ = 0.23) — but this largely disappeared when controlling for general mental ability and personality traits.

High evidence
EQ Incremental ValidityNear-zero incremental r²

Meta-analytic review confirming that EQ's predictive validity for job performance is near-zero after controlling for GMA and Big Five personality. EQ's apparent predictive power largely reflects overlap with existing constructs, not a distinct capability.

Locke(2005)
High evidence
EQ Validity Critique

Systematic critique of EQ as a scientific construct. Argues that ability-based EQ is a form of social intelligence already captured by existing measures, and that mixed models of EQ conflate personality traits with cognitive abilities.

High evidence
EQ Training Effectsd ≈ 0.35, rapid decay

Meta-analysis of EQ training programmes. Found modest average effects (d ≈ 0.35) that typically decayed rapidly after training ended. Significantly weaker and less durable than metacognitive training effects.

High evidence
EQ Review

Comprehensive review of EQ science, distinguishing ability models from mixed models. Notes the scientific case for ability EQ is stronger than for mixed models, but acknowledges the incremental validity problem remains unresolved.

Evidence to Practice

MindFrame is built on this evidence base

Every training mode maps to specific research. Calibration training targets Fischhoff et al. Reasoning scoring maps to Stanovich & West. Bias recognition targets Pronin et al. Nothing in MindFrame is invented — it's applied science.