Our research team publishes in top-tier medical and AI journals, conducts multi-site clinical validation trials, and presents at leading conferences worldwide.
Selected publications from our research team across clinical AI, medical informatics, and computational medicine.
Rigorous validation through prospective clinical trials across leading health systems, conducted in partnership with independent academic researchers.
Prospective, multi-site observational study across 6 ICUs (n=42,318 admissions). Primary endpoint: AUROC for sepsis prediction 48-72 hours before clinical recognition. 18-month enrollment period with 6-month follow-up.
94.1% sensitivity, 89.3% specificity. Median early detection advantage of 52 hours. False positive rate of 4.7%, significantly below the 12.3% rate of existing scoring systems.
First prospective multi-site validation of a multimodal AI system for sepsis prediction achieving >0.94 AUROC. Results published in Nature Medicine and adopted into clinical practice at all 3 participating institutions.
Cluster-randomized controlled trial across 14 ICUs in 4 health systems (n=8,724 patients). Primary endpoint: reduction in non-actionable alarms. Secondary endpoints: critical event detection rate, nursing satisfaction, and response time.
Non-actionable alarms reduced by 68% (p<0.001). Critical event detection maintained at 99.7%. Nursing satisfaction improved by 34 NPS points. Mean response time to genuine alerts decreased by 22%.
Largest randomized trial of AI-based alarm management in critical care. Demonstrated that intelligent filtering can dramatically reduce fatigue without compromising safety. Published in The Lancet Digital Health.
12-month prospective comparison study at 3 health systems (n=156,832 medication orders). Parallel evaluation of AI-enhanced pharmacogenomic checking vs. standard formulary-based checking. Independent pharmacy adjudication of all flagged interactions.
3.2 times more clinically significant interactions detected (p<0.001). 847 potentially life-threatening interactions caught that standard checking missed. Precision of 91.4%, reducing pharmacist verification burden.
Demonstrated that patient-specific pharmacogenomic and organ function data substantially improves drug interaction detection. The 847 prevented life-threatening interactions represent direct patient safety impact. Published in JAMA Internal Medicine.
Stepped-wedge cluster randomized trial across 8 emergency departments (n=31,562 encounters). Primary endpoint: diagnostic error rate. Sequential rollout over 12 months with concurrent control analysis.
Diagnostic error rate reduced from 5.7% to 2.1% (p<0.001). Time to correct diagnosis decreased by 41%. 30-day unplanned return visits decreased by 28%. Clinician acceptance rate of 89%.
First large-scale randomized evidence that AI-assisted decision support meaningfully reduces diagnostic error in emergency settings. The 89% acceptance rate indicates strong clinician trust. Published in Annals of Internal Medicine.
Multi-national prospective cohort study (n=94,216 admissions enrolled to date). Primary endpoint: time to HAI detection vs. standard clinical identification. Phase 2 expansion to 5 additional countries in progress.
Median 36 hours earlier detection (IQR: 18-52 hours). 23% shorter infection duration. 18% lower antibiotic utilization. Estimated $4,200 cost saving per detected case. Phase 1 results published in The Lancet.
Largest international trial of AI-based HAI detection. Demonstrates cross-border generalizability of clinical AI models and meaningful impact on antibiotic stewardship. Phase 2 enrollment targeting 200,000 admissions.
Our research team combines deep expertise in machine learning, clinical medicine, biomedical informatics, and health system operations.
Every model we deploy follows a rigorous development and validation lifecycle designed specifically for clinical-grade AI.
All training data undergoes rigorous curation: automated de-identification exceeding Safe Harbor standards, quality validation with <0.1% error tolerance, demographic bias auditing, and IRB approval verification. We maintain data provenance records for full reproducibility.
Architecture selection is hypothesis-driven and based on the clinical task. We use ablation studies, cross-validation, and held-out test sets from different institutions than training data. Hyperparameter optimization uses Bayesian search with clinical performance metrics as objectives.
Before external validation, all models undergo internal testing across multiple dimensions: clinical accuracy (AUROC, sensitivity, specificity, PPV, NPV), fairness (performance parity across demographics), robustness (adversarial testing), and latency (real-time inference benchmarks).
Multi-site prospective clinical trials with independent academic partners. We use pre-registered study protocols, independent data safety monitoring boards, and follow CONSORT-AI and SPIRIT-AI reporting guidelines for transparent and reproducible trial design.
Post-deployment, every model is continuously monitored for performance drift, calibration degradation, and fairness regression. Automated retraining pipelines trigger when performance drops below pre-defined thresholds. Monthly model performance reports are shared with all partner institutions.
We publish all clinical validation results in peer-reviewed journals, regardless of outcome. Model cards documenting intended use, limitations, and performance characteristics are publicly available. We adhere to the TRIPOD-AI framework for transparent reporting.