BODHI Framework

Bridging Open Discerning Humble Inquiring

A systems design framework that reimagines medical AI as an epistemic agent: one that questions assumptions, reasons under uncertainty, and recognizes the boundaries of its own knowledge.

Preliminary results from BODHI's prompting implementation on HealthBench Hard
+89.6pp
Context Seeking Improvement
7.8% → 97.3% (GPT 4.1 mini)
+16.6pp
Overall Clinical Quality
p < 0.0001, 5 seeds
d = 16.38
Curiosity Effect Size
Very large effect
d = 5.80
Humility Effect Size
GPT 4.1 mini hedging

The Problem

Medical AI is dangerously overconfident

Current AI systems conflate statistical pattern recognition with genuine clinical understanding. They deliver confident predictions without mechanisms to express uncertainty, acknowledge limitations, or push back on flawed assumptions.

This sycophantic behavior has documented clinical consequences: missed diagnoses, delayed treatments, and eroded trust. The root issue isn't model accuracy. It's the absence of epistemic agency.

BODHI introduces a design philosophy where AI systems are architected around epistemic virtues: the capacity to bridge knowledge gaps, remain open to alternatives, discern complexity, practice humility, and actively inquire.

Read the design philosophy in PLOS Digital Health →
🤖

Sycophantic Agreement

AI confirms clinician assumptions instead of challenging potentially incorrect diagnoses

⚠️

False Confidence

Models generate authoritative sounding responses even when operating outside their training distribution

🔇

Missing Questions

Baseline models ask clarifying questions only 7.8% of the time when they should be asking nearly always

Design Principles

BODHI is grounded in four principles that guide the design of AI systems with genuine epistemic agency — independent of any specific implementation method.

Curiosity
Systems that actively seek context, ask clarifying questions, and explore alternative hypotheses — rather than defaulting to confident answers.
Calibrated Humility
Mechanisms for uncertainty quantification, explicit limitation acknowledgment, and recognition of knowledge boundaries — knowing what you don't know.
Creative Reasoning
Novel hypothesis generation, divergent thinking, and alternative analytical approaches that go beyond statistical pattern matching.
Anti-Sycophancy
Counteracting RLHF-induced behaviors: agreement bias, overconfidence, hubris, and the tendency to confirm rather than challenge flawed assumptions.

Virtue Activation Matrix

Central to BODHI's design principles: four quadrants organize behavioral responses based on the interplay between uncertainty and clinical stakes.

Virtue Activation Matrix mapping clinical complexity against model confidence
Virtue Activation Matrix mapping clinical complexity against model confidence. Q1 (Proceed & Monitor): routine cases with high confidence. Q2 (Watchful & Alternatives): complex cases requiring situational awareness. Q3 (Clarify & Review): uncertain cases needing high curiosity and epistemic humility. Q4 (Escalate & Reframe): high stakes uncertainty demanding explicit escalation to human expertise. Read More →

Active Research

BODHI's design philosophy translates into concrete research modules, each operationalizing a distinct epistemic virtue. Some are empirically validated; others are active areas of investigation.

Curiosity Module

Validated

Context seeking behavior, clarifying question generation, and proactive exploration of alternative diagnoses. Preliminary evaluation demonstrated +89.6pp improvement on HealthBench Hard.

Context Seeking Active Inquiry HealthBench

Humility Module

Validated

Uncertainty quantification, sycophancy detection and mitigation, explicit limitation statements, and recognition of knowledge boundaries. Preliminary evaluation achieved d = 5.80 effect size on hedging behavior.

Uncertainty Anti Sycophancy Hedging

Creativity Module

In Development

Novel hypothesis generation, alternative analytical approaches, and divergent thinking in clinical reasoning. Expanding BODHI beyond safety constraints into creative diagnostic exploration.

Hypothesis Generation Divergent Thinking QMoE

Sycophancy Detection

Validated

Detection and mitigation of sycophantic agreement where AI confirms clinician assumptions instead of challenging potentially incorrect diagnoses. Anti-sycophancy measures promote independent clinical reasoning with documented reduction in agreement bias.

Anti Sycophancy Agreement Bias Independent Reasoning

BODHI is an expanding research platform. We welcome researchers, clinicians, and engineers interested in building AI systems with genuine epistemic agency.

Get Involved

Publications

Published

Beyond Overconfidence: Embedding Curiosity and Humility for Ethical Medical AI

Cajas Ordoñez SA, Castro R, Celi LA, Delos Reyes R, Engelmann J, Ercole A, Hilel A, Kalla M, Kinyera L, Lange M, Lunde TM, Meni MJ, Premo AE, Sedlakova J
PLOS Digital Health 5(1): e0001013
January 2026
Under Review

An Engineering Framework for Curiosity Driven and Humble AI in Clinical Decision Support

Arslan J, Benke K, Cajas Ordoñez SA, Castro R, Celi LA, Cruz Suarez GA, Delos Reyes R, Engelmann J, Ercole A, et al.
BMJ Health & Care Informatics
2026
Published

Humility and Curiosity in Human–AI Systems for Health Care

Cajas Ordoñez SA, Lange M, Lunde TM, Meni MJ, Premo AE
The Lancet 406(10505): 804-805
2025
Preprint

Uncertainty Makes It Stable: Curiosity Driven Quantized Mixture of Experts

Cajas Ordoñez SA, Torres Torres LF, Meni MJ, Duran Paredes CA, Arazo E, Bosch C, Carbajo RS, Lai Y, Celi LA
arXiv preprint arXiv:2511.11743
2025
Published

Teaching Machines to Doubt

Celi LA
Nature Medicine
2025

Research Team

BODHI is developed by an interdisciplinary team of researchers, clinicians, and engineers from institutions worldwide.

Principal Investigator

Leo Anthony Celi
Principal Investigator
MIT, Beth Israel Deaconess, Harvard
lceli@mit.edu

Multidisciplinary Team Researchers

Torleif Markussen Lunde
University of Bergen
Maximin Lange
MIT, King's College London
Felipe Ocampo Osorio
Valle del Lili, Universidad Icesi
Oriel Perets
Ben-Gurion University of the Negev
Sebastián Andrés Cajas Ordoñez
MIT Critical Data
Mackenzie J Meni
Florida Institute of Technology
Gustavo Adolfo Cruz Suarez
Valle del Lili, Universidad Icesi
Rowell Castro
MIT
Roben Delos Reyes
University of Melbourne
Janan Arslan
Sorbonne Université, University of Melbourne
Kurt Benke
University of Melbourne
Justin Engelmann
University College London
Ari Ercole
Cambridge University Hospitals
Almog Hilel
MIT
Mahima Kalla
University of Melbourne
Leo Kinyera
Mbarara University
Anna E Premo
ETH Zurich
Jana Sedlakova
University of Zurich
Pritika Vig
MIT

Partner Institutions

Code & Resources

📦

BODHI Python Package

Fully open source implementation of the BODHI framework. Install via pip, explore the code, report issues, or contribute on GitHub.

pip install bodhi-llm
View on PyPI →
🔬

Evaluation Scripts

Complete evaluation framework for testing BODHI on HealthBench Hard and other clinical benchmarks with statistical analysis.

View on GitHub →

Curiosity Driven QMoE

Quantized Mixture of Experts with curiosity driven routing for efficient edge deployment with stable latency.

View on GitHub →

Get in Touch

We welcome researchers, clinicians, and engineers working on epistemic agency, uncertainty quantification, and safe AI for clinical decision support. Reach out to any of us directly.