SATIR
Precision Patient–Trial Matching
Scalable High-Recall Constraint-Satisfaction-Based
Information Retrieval for Clinical Trials Matching
1Stanford 2UCLA 3Emory 4Mayo Clinic
*Equal contribution
We propose SatIR, a scalable clinical trial retrieval method based on constraint satisfaction. Our approach uses formal methods — Satisfiability Modulo Theories (SMT) and relational algebra — to efficiently represent and match key constraints from clinical trials and patient records. We use Large Language Models (LLMs) to convert informal reasoning — regarding ambiguity, implicit clinical assumptions, and incomplete patient records — into explicit, precise, controllable, and interpretable formal constraints.
Evaluated on 59 patients and 3,621 trials, SatIR outperforms TrialGPT on all three retrieval objectives. It retrieves 32%–72% more relevant-and-eligible trials per patient, improves recall by 22–38 points, and serves more patients with at least one useful trial — in just 2.95 seconds per patient.
Ontology-Grounded Representation
An SMT-based representation grounded in SNOMED CT captures Boolean logic, numeric thresholds, temporal windows, and ontology-aware concept identity and entailment.
Accurate Semantic Parsing
LLM-based parsers translate complex clinical text — handling underspecified criteria, negation, temporality, and implicit assumptions — into executable formal constraints.
Salience-Based Missingness
SatIR introduces salience to resolve ambiguity from incomplete records — using LLMs to assess whether missing data is clinically significant enough to affect eligibility, with explicit auditable judgments.
Scalable Database Retrieval
SMT constraints are projected into relational algebra and stored in SQL, enabling fast recall-preserving retrieval across hundreds of thousands of trials in under 3 seconds.
Constraint Semantic Parsing
LLM-based pipelines that translate free-text trial criteria and patient records into formal SMT constraint programs.
→ ExploreConstraint Augmentation
Enriching constraints with ontology-grounded subsumption and salience-based handling of incomplete patient records.
→ ExploreConstraint Indexing
Projecting SMT constraints into recall-safe relational algebra and storing them as first-class data in a SQL database.
→ ExploreConstraint Execution
Executing patient constraints against the pre-built trial database via SQL queries in ~2.95 seconds per patient.
→ ExploreThe core assumption: any clinically salient fact about a patient will be documented in their medical record. If it is absent, that absence is itself informative.
SatIR uses targeted LLM queries to assess the salience of each missing concept — whether it is the kind of thing a physician would always record if present. These judgments are compiled into explicit, auditable policy decisions rather than left as implicit LLM reasoning.
| Objective | Retrieved/Patient | BMRetriever | ClinBest | BGE-large | PubMedBERT | TrialGPT | SatIR |
|---|---|---|---|---|---|---|---|
| Treat-Chief | 36 | 1.44 | 1.75 | 2.07 | 2.00 | 2.17 | 3.25 |
| Treat-Any | 64 | 2.00 | 2.54 | 2.85 | 2.68 | 2.98 | 5.12 |
| Relevant-to-Any | 121 | 6.39 | 7.24 | 8.68 | 8.92 | 8.93 | 11.76 |
| Objective | SatIR wins | Tie | TrialGPT wins | Served by SatIR | Served by TrialGPT |
|---|---|---|---|---|---|
| Treat-Chief | 20 | 33 | 6 | 39 | 35 |
| Treat-Any | 29 | 28 | 2 | 42 | 39 |
| Relevant-to-Any | 35 | 17 | 7 | 54 | 50 |
Interpretable, auditable, and permanently correctable. Because every eligibility decision traces back to explicit symbolic constraints — not opaque model weights — failures are diagnosable. When a parsing error or an ontology gap is identified and fixed, that fix propagates immediately and permanently across every patient and every trial that touches the corrected predicate. There is no need to retrain, re-embed, or re-evaluate a model. This is the core advantage of SatIR's formal approach: errors are local, fixes are global.
@article{zhou2026satir, title = {Scalable High-Recall Constraint-Satisfaction-Based Information Retrieval for Clinical Trials Matching}, author = {Cyrus Zhou and Yufei Jin and Yilin Xu and Yu-Chiang Wang and Chieh-Ju Chao and Monica S. Lam}, note = {Under review}, year = {2026}, }
We gratefully acknowledge support from the Verdant Foundation, the Hasso Plattner Institute, Microsoft Azure AI credits, Itaú Unibanco, BMO Financial Group, and the Stanford HAI Institute. Cyrus Zhou is partially supported by the Stanford School of Engineering Fellowship.
We thank Yucheng Jiang for the clinician annotation interface, and our clinical collaborators at Stanford and Mayo Clinic for their expert guidance.








