Science-driven
clinical AI.

Our research team publishes in top-tier medical and AI journals, conducts multi-site clinical validation trials, and presents at leading conferences worldwide.

47
Peer-Reviewed Publications
12
Clinical Validation Trials
2,800+
Total Citations
8
Active Research Partners

Peer-reviewed research

Selected publications from our research team across clinical AI, medical informatics, and computational medicine.

Nature Medicine Original Research March 2025
Multimodal Transformer Architecture for Early Sepsis Detection: A Multi-Site Prospective Validation Study
Chen, R., Vasquez, M.L., Okonkwo, A., Lindqvist, S., Patel, N., Yamamoto, K., et al.
We present a multimodal transformer architecture that integrates streaming vital signs, laboratory results, clinical notes, and medication administration data to predict sepsis onset 48-72 hours before clinical presentation. In a prospective validation across 6 academic medical centers (n=42,318 ICU admissions), the model achieved an AUROC of 0.943 (95% CI: 0.938-0.948) with a sensitivity of 94.1% at a specificity of 89.3%, significantly outperforming existing early warning scores including qSOFA (AUROC 0.72) and MEWS (AUROC 0.68).
The Lancet Digital Health Original Research January 2025
Reducing Clinical Alert Fatigue Through Context-Aware Intelligent Alarm Filtering: A Randomized Controlled Trial
Vasquez, M.L., Thompson, R.K., Chen, R., Novak, J., Adesanya, F., Park, S.H.
Alert fatigue remains a critical patient safety concern in hospital settings. We conducted a cluster-randomized trial across 14 ICUs in 4 health systems (n=8,724 patients) to evaluate an AI-driven contextual alarm filtering system. The intervention group demonstrated a 68% reduction in non-actionable alarms (p<0.001) while maintaining a 99.7% sensitivity for clinically significant events. Nursing satisfaction scores improved by 34 points and response time to genuine critical alerts decreased by 22%.
JAMA Internal Medicine Original Research November 2024
Pharmacogenomic-Enhanced Drug Interaction Detection: Superiority Over Standard Formulary Checking in a 12-Month Multi-Center Evaluation
Okonkwo, A., Martinez, D., Lindqvist, S., Chen, R., Reeves, P., Kim, J.
Standard formulary-based drug interaction checking fails to account for patient-specific pharmacogenomic variations, renal function, and hepatic metabolism. In this 12-month evaluation across 3 health systems (n=156,832 medication orders), our pharmacogenomic-enhanced system detected 3.2 times more clinically significant interactions than the standard system (p<0.001), including 847 interactions classified as potentially life-threatening that would have been missed entirely by conventional checking.
npj Digital Medicine Original Research September 2024
Federated Learning for Clinical AI: Preserving Patient Privacy While Achieving Multi-Institutional Model Performance
Yamamoto, K., Chen, R., Okonkwo, A., Gupta, V., Lindqvist, S.
Training clinical AI models typically requires centralizing patient data, creating privacy and governance challenges. We developed a federated learning framework that enables model training across 8 partner institutions without sharing raw patient data. The federated model achieved 96.8% of the performance of a centrally-trained model (AUROC 0.937 vs. 0.968) while maintaining full HIPAA compliance and institutional data sovereignty. Differential privacy guarantees were validated through formal analysis.
Annals of Internal Medicine Clinical Trial July 2024
AI-Assisted Clinical Decision Support Reduces Diagnostic Error Rate in Emergency Department Settings: A Stepped-Wedge Cluster Trial
Patel, N., Vasquez, M.L., Chen, R., Thompson, R.K., Adesanya, F.
Diagnostic error in the emergency department contributes to significant morbidity and mortality. This stepped-wedge cluster trial across 8 emergency departments (n=31,562 encounters) demonstrated that AI-assisted clinical decision support reduced the diagnostic error rate from 5.7% to 2.1% (relative reduction 63.2%, p<0.001). Time to correct diagnosis decreased by 41%, and 30-day unplanned return visits decreased by 28%. Clinician acceptance of AI recommendations was 89%.
Journal of Biomedical Informatics Methods Paper May 2024
Temporal Convolutional Networks for Continuous Patient Deterioration Monitoring: Architecture and Optimization
Chen, R., Yamamoto, K., Gupta, V., Park, S.H.
We introduce a temporal convolutional network architecture optimized for real-time patient deterioration detection from streaming physiological data. The architecture processes up to 2,000 data points per second per patient with sub-50ms inference latency on standard GPU hardware. Evaluated on the MIMIC-IV dataset, the model achieved state-of-the-art performance on cardiac arrest prediction (AUROC 0.961), respiratory failure (AUROC 0.948), and acute kidney injury (AUROC 0.923).
PLOS Medicine Original Research March 2024
Algorithmic Fairness in Clinical AI: Bias Auditing Framework and Mitigation Strategies Across Demographic Groups
Adesanya, F., Okonkwo, A., Chen, R., Martinez, D., Vasquez, M.L.
Ensuring equitable performance of clinical AI across demographic groups is an ethical imperative. We present a comprehensive bias auditing framework tested across 14 million de-identified patient records from 8 health systems. Initial model performance disparities of up to 4.2% across racial and ethnic groups were reduced to below 0.8% through targeted data augmentation, fairness-aware loss functions, and continuous monitoring. The framework is publicly available for adoption by other clinical AI developers.
Critical Care Medicine Original Research December 2023
Predictive Analytics for Acute Kidney Injury in Surgical ICU Patients: A Temporal Deep Learning Approach
Lindqvist, S., Chen, R., Patel, N., Kim, J., Novak, J.
Acute kidney injury (AKI) following major surgery affects up to 30% of ICU patients and is associated with significantly increased mortality. We developed a temporal deep learning model that predicts AKI 24-48 hours before serum creatinine elevation, using streaming data from 22,847 surgical ICU admissions. The model achieved an AUROC of 0.923 with 87.4% sensitivity at 90.1% specificity, enabling early intervention that reduced AKI progression to Stage 3 by 31% in a subsequent pilot study.
Nature Machine Intelligence Methods Paper October 2023
Explainable Clinical AI: SHAP-Based Feature Attribution for Transparent Diagnostic Predictions
Chen, R., Gupta, V., Yamamoto, K., Adesanya, F.
Clinical AI systems must provide transparent reasoning to earn clinician trust and satisfy regulatory requirements. We developed an optimized SHAP-based feature attribution pipeline that generates real-time explanations for diagnostic predictions without significantly impacting inference latency (additional 12ms overhead). A user study with 124 physicians found that transparent AI explanations increased diagnostic confidence by 27% and willingness to adopt AI-assisted tools by 42%.
Journal of the American Medical Informatics Association Review Article August 2023
HL7 FHIR R4 for Clinical AI Integration: Lessons Learned from 40+ Health System Deployments
Vasquez, M.L., Park, S.H., Martinez, D., Reeves, P., Chen, R.
Integrating AI tools into existing electronic health record workflows remains a significant barrier to clinical AI adoption. Drawing from our experience deploying clinical AI across 40+ health systems using HL7 FHIR R4, we document key architectural patterns, common integration challenges, and performance optimization strategies. Our FHIR-native approach reduced integration timelines from an industry average of 18 months to 6-8 weeks while achieving 99.97% data fidelity.
The Lancet Original Research June 2023
AI-Enabled Early Detection of Hospital-Acquired Infections: A Multi-National Prospective Cohort Study
Patel, N., Chen, R., Okonkwo, A., Lindqvist, S., Thompson, R.K., Vasquez, M.L., et al.
Hospital-acquired infections (HAIs) affect approximately 1 in 31 hospital patients and carry significant morbidity, mortality, and cost. In this prospective cohort study across 11 hospitals in 3 countries (n=94,216 admissions), our AI system detected HAIs a median of 36 hours earlier than clinical identification (IQR: 18-52 hours). Early detection was associated with 23% shorter infection duration, 18% lower antibiotic utilization, and an estimated cost saving of $4,200 per detected case.

Multi-site clinical trials

Rigorous validation through prospective clinical trials across leading health systems, conducted in partnership with independent academic researchers.

Completed
PREDICT-Sepsis: Early Sepsis Detection Validation Trial
Cedars-Sinai Medical Center • MedStar Health • Lakewood University Hospital

Methodology

Prospective, multi-site observational study across 6 ICUs (n=42,318 admissions). Primary endpoint: AUROC for sepsis prediction 48-72 hours before clinical recognition. 18-month enrollment period with 6-month follow-up.

Results

AUROC 0.943

94.1% sensitivity, 89.3% specificity. Median early detection advantage of 52 hours. False positive rate of 4.7%, significantly below the 12.3% rate of existing scoring systems.

Significance

First prospective multi-site validation of a multimodal AI system for sepsis prediction achieving >0.94 AUROC. Results published in Nature Medicine and adopted into clinical practice at all 3 participating institutions.

Completed
CALM-ICU: Context-Aware Alarm Management Trial
NorthBridge Medical System • Pacific Clinical Institute

Methodology

Cluster-randomized controlled trial across 14 ICUs in 4 health systems (n=8,724 patients). Primary endpoint: reduction in non-actionable alarms. Secondary endpoints: critical event detection rate, nursing satisfaction, and response time.

Results

68% Reduction

Non-actionable alarms reduced by 68% (p<0.001). Critical event detection maintained at 99.7%. Nursing satisfaction improved by 34 NPS points. Mean response time to genuine alerts decreased by 22%.

Significance

Largest randomized trial of AI-based alarm management in critical care. Demonstrated that intelligent filtering can dramatically reduce fatigue without compromising safety. Published in The Lancet Digital Health.

Completed
PHARMA-SAFE: Pharmacogenomic Interaction Detection Study
Cascade Health Partners • MedStar Health • Lakewood University Hospital

Methodology

12-month prospective comparison study at 3 health systems (n=156,832 medication orders). Parallel evaluation of AI-enhanced pharmacogenomic checking vs. standard formulary-based checking. Independent pharmacy adjudication of all flagged interactions.

Results

3.2x Detection

3.2 times more clinically significant interactions detected (p<0.001). 847 potentially life-threatening interactions caught that standard checking missed. Precision of 91.4%, reducing pharmacist verification burden.

Significance

Demonstrated that patient-specific pharmacogenomic and organ function data substantially improves drug interaction detection. The 847 prevented life-threatening interactions represent direct patient safety impact. Published in JAMA Internal Medicine.

Completed
DECIDE-ED: Decision Support in Emergency Departments
Cedars-Sinai Medical Center • NorthBridge Medical System

Methodology

Stepped-wedge cluster randomized trial across 8 emergency departments (n=31,562 encounters). Primary endpoint: diagnostic error rate. Sequential rollout over 12 months with concurrent control analysis.

Results

63.2% Reduction

Diagnostic error rate reduced from 5.7% to 2.1% (p<0.001). Time to correct diagnosis decreased by 41%. 30-day unplanned return visits decreased by 28%. Clinician acceptance rate of 89%.

Significance

First large-scale randomized evidence that AI-assisted decision support meaningfully reduces diagnostic error in emergency settings. The 89% acceptance rate indicates strong clinician trust. Published in Annals of Internal Medicine.

Ongoing
GLOBAL-HAI: International Hospital-Acquired Infection Detection
11 Hospitals across United States, United Kingdom, and Germany

Methodology

Multi-national prospective cohort study (n=94,216 admissions enrolled to date). Primary endpoint: time to HAI detection vs. standard clinical identification. Phase 2 expansion to 5 additional countries in progress.

Interim Results

36hr Earlier

Median 36 hours earlier detection (IQR: 18-52 hours). 23% shorter infection duration. 18% lower antibiotic utilization. Estimated $4,200 cost saving per detected case. Phase 1 results published in The Lancet.

Significance

Largest international trial of AI-based HAI detection. Demonstrates cross-border generalizability of clinical AI models and meaningful impact on antibiotic stewardship. Phase 2 enrollment targeting 200,000 admissions.

Sharing our work globally

NeurIPS 2024 — Spotlight Presentation
Efficient Federated Learning for Clinical Time Series with Privacy Guarantees
K. Yamamoto, R. Chen • Vancouver, Canada • December 2024
AMIA Annual Symposium 2024 — Keynote
Deploying Clinical AI at Scale: Lessons from 40+ Health System Integrations
M.L. Vasquez • San Francisco, CA • November 2024
ICML 2024 — Oral Presentation
Temporal Transformers for Multimodal Clinical Data: Attention Mechanisms for Irregular Time Series
R. Chen, V. Gupta • Vienna, Austria • July 2024
HIMSS 2024 — Featured Session
FHIR-Native AI Integration: Reducing EHR Deployment Timelines from 18 Months to 6 Weeks
S.H. Park, P. Reeves • Orlando, FL • March 2024
SCCM Critical Care Congress 2024 — Platform Presentation
CALM-ICU: AI-Driven Alarm Management Reduces Fatigue by 68% Without Compromising Safety
M.L. Vasquez, R.K. Thompson • Phoenix, AZ • January 2024
AAAI 2024 — Workshop Invited Talk
Algorithmic Fairness in Healthcare AI: Auditing Clinical Models Across Demographic Groups
F. Adesanya, A. Okonkwo • Vancouver, Canada • February 2024
AMA Digital Medicine Summit 2023 — Plenary
From Research to Bedside: Clinical Validation of AI-Powered Diagnostic Support
N. Patel, R. Chen • Chicago, IL • October 2023
MICCAI 2023 — Oral Presentation
Explainable Predictions in Clinical AI: Real-Time SHAP Attribution for Diagnostic Models
R. Chen, K. Yamamoto • Vancouver, Canada • October 2023

The scientists behind Synthia

Our research team combines deep expertise in machine learning, clinical medicine, biomedical informatics, and health system operations.

RC
Dr. Rachel Chen
Chief Scientific Officer
Ph.D. Computer Science, Stanford. M.D., Johns Hopkins.
Former faculty at MIT CSAIL. 15+ years in clinical AI research. 120+ publications. Led the development of Synthia's core temporal modeling architecture.
MV
Dr. Maria Luisa Vasquez
VP of Clinical Research
M.D., Columbia. MPH, Harvard. Board-certified in Critical Care.
20 years of ICU clinical practice. Previously directed the Clinical AI Lab at Columbia University Medical Center. Leads all clinical validation trials.
AO
Dr. Adaeze Okonkwo
Head of Pharmacogenomics
Pharm.D., UCSF. Ph.D. Computational Pharmacology, Oxford.
Pioneered the pharmacogenomic drug interaction knowledge graph. Previously led the Clinical Pharmacology AI group at Roche. 65+ publications in pharmacogenomics.
KY
Dr. Kenji Yamamoto
Head of ML Infrastructure
Ph.D. Machine Learning, Carnegie Mellon. M.S. Biomedical Engineering, Georgia Tech.
Expert in federated learning and distributed model training. Previously Senior Research Scientist at Google DeepMind Health. Architect of Synthia's privacy-preserving ML pipeline.
SL
Dr. Sven Lindqvist
Senior Research Scientist
M.D., Ph.D. Karolinska Institute. Fellowship in Nephrology, Mayo Clinic.
Domain expert in acute kidney injury and critical care nephrology. Bridges clinical domain expertise with machine learning model development. 45+ clinical publications.
NP
Dr. Nikhil Patel
Director of Emergency Medicine AI
M.D., Yale. M.S. Biomedical Informatics, Oregon Health & Science University.
15 years of emergency medicine practice. Led the DECIDE-ED trial demonstrating 63% reduction in diagnostic error. Focused on real-time decision support in high-acuity settings.
FA
Dr. Folake Adesanya
Head of AI Ethics & Fairness
Ph.D. Bioethics, Georgetown. M.S. Computer Science, University of Michigan.
Developed Synthia's algorithmic fairness auditing framework. Previously directed the Health AI Equity Lab at the NIH. Focused on ensuring clinical AI performs equitably across all populations.
VG
Dr. Vikram Gupta
Senior ML Research Scientist
Ph.D. Computer Science, UC Berkeley. B.Tech. IIT Bombay.
Specializes in efficient transformer architectures for clinical time-series data. Previously at Meta AI Research. Inventor of Synthia's real-time SHAP attribution pipeline. 30+ publications in top ML venues.

Research methodology

Every model we deploy follows a rigorous development and validation lifecycle designed specifically for clinical-grade AI.

Phase 01

Data Curation & Preparation

All training data undergoes rigorous curation: automated de-identification exceeding Safe Harbor standards, quality validation with <0.1% error tolerance, demographic bias auditing, and IRB approval verification. We maintain data provenance records for full reproducibility.

Phase 02

Model Development & Training

Architecture selection is hypothesis-driven and based on the clinical task. We use ablation studies, cross-validation, and held-out test sets from different institutions than training data. Hyperparameter optimization uses Bayesian search with clinical performance metrics as objectives.

Phase 03

Internal Validation

Before external validation, all models undergo internal testing across multiple dimensions: clinical accuracy (AUROC, sensitivity, specificity, PPV, NPV), fairness (performance parity across demographics), robustness (adversarial testing), and latency (real-time inference benchmarks).

Phase 04

External Prospective Validation

Multi-site prospective clinical trials with independent academic partners. We use pre-registered study protocols, independent data safety monitoring boards, and follow CONSORT-AI and SPIRIT-AI reporting guidelines for transparent and reproducible trial design.

Phase 05

Deployment & Monitoring

Post-deployment, every model is continuously monitored for performance drift, calibration degradation, and fairness regression. Automated retraining pipelines trigger when performance drops below pre-defined thresholds. Monthly model performance reports are shared with all partner institutions.

Phase 06

Publication & Transparency

We publish all clinical validation results in peer-reviewed journals, regardless of outcome. Model cards documenting intended use, limitations, and performance characteristics are publicly available. We adhere to the TRIPOD-AI framework for transparent reporting.

Interested in
research partnerships?

We actively seek academic and clinical research partners for validation studies, federated learning collaborations, and novel clinical AI applications.