The Research Behind MindFrame
Every training mode, scoring metric, and feedback mechanism in MindFrame is grounded in peer-reviewed research on metacognition, calibration, and cognitive skill development.
Effect Size Evidence Table
What is g? g (Cohen's g), d, and ES are all standardised effect sizes — a universal ruler that lets researchers compare results across different studies and populations. Think of it as a percentile gap: g = 0.63 means a person at the 50th percentile in the trained group would outscore 73% of untrained people.
Scale: 0.20 = small · 0.50 = medium · 0.80 = large · 1.0+ = very large. Values above 0.40 are considered educationally significant (Hattie, 2009).
| Intervention | Effect | Evidence |
|---|---|---|
| Metacognitive Training | g = 0.63Trained people outscore 73% of untrained — beats homework, class-size cuts, and most tutoring | high |
| Metacognitive Instruction | ES ≈ 1.11Trained people outscore ~86% of untrained — among the strongest effects in cognitive training | high |
| Metacognitive Therapy (MCT) | g = 0.69Trained people outscore ~75% of CBT patients — outperforms the gold standard for anxiety | high |
| Calibration Training | +14%Consistent, measurable accuracy gain from structured practice | high |
| Working Memory Training | g ≈ 0.28Small-to-medium — modest generalisation beyond trained tasks | medium |
| Spaced Repetition | d = 0.47–0.71Medium-to-large — strong durable memory across 254 studies | high |
| Error Monitoring Training | SignificantConsistent improvement in decision accuracy across multiple RCTs | medium |
Training Principles
The five mechanisms through which MindFrame produces measurable improvement.
Calibrated confidence, not just accuracy
Getting an answer right is not enough. Knowing when you're right versus guessing — and assigning the correct probability — is what distinguishes expert decision-makers from lucky ones. Every MindFrame challenge requires you to state your confidence alongside your answer.
Brier Score + Calibration ErrorImmediate, precise feedback
Vague feedback ("good job") produces no improvement. Improvement requires specific information about where you deviated from ideal performance. MindFrame gives you percentile rankings, calibration error breakdown, and mode-level analytics after every session.
Composite Score, Mode Breakdown, Percentile RankSpaced repetition scheduling
The forgetting curve is real. A single exposure to a concept produces brief retention. Reviewing at increasing intervals forces retrieval practice, which produces durable memory and skill. MindFrame uses SM-2 scheduling to surface the right challenge at the right time.
SM-2 adaptive schedulingReasoning quality evaluation
Correct answers reached through faulty reasoning don't transfer to novel situations. MindFrame's AI coach evaluates the logical structure and quality of your reasoning, not just whether your final answer was right.
AI Reasoning ScoreReflective consolidation
Research on learning shows that post-session reflection significantly increases knowledge transfer. After every session, MindFrame prompts you to identify your biggest error and the strategy that would have prevented it.
Session JournalMode × Research Map
How each MindFrame training mode maps to a specific cognitive skill and research base.
Probability estimation
Reduces overconfidence by training confidence-accuracy match
Analytical reasoning
Improves argument evaluation and logical validity detection
Bias recognition
Reduces susceptibility by increasing bias fluency
Attentional control
Increases capacity and resistance to distraction
Cognitive flexibility
Trains perspective-shifting and belief updating under conflict
Belief updating
Trains proportional response to new evidence
All-mode integration
Cross-domain calibration across all 5 skill areas
Research Downloads
Full evidence reviews available as PDFs. No sign-up required.
Put the evidence to work
Effect sizes only become results when paired with consistent practice. Start a session and get your first calibration baseline in 10 minutes.