Healthcare AI Stumbles: Blood Pressure Prediction Hits a Wall
A recent study highlights the challenges of generalizing AI models across healthcare institutions. Despite promising internal results, external validation reveals significant hurdles.
Healthcare AI is often hailed as the next frontier, but a recent study throws cold water on those ambitions. Researchers developed an ensemble framework to predict blood pressure from electronic health records. While it showed moderate success internally, external validation painted a different picture.
Internal vs. External: A Stark Contrast
The model's internal validation using the MIMIC-III dataset showed room for optimism. For systolic blood pressure, it achieved an R2of 0.248 and a root mean square error (RMSE) of 14.84 mmHg. Similarly, diastolic predictions weren't far off, with an R2of 0.297 and an RMSE of 8.27 mmHg. However, when tested on the eICU dataset, the model's performance plummeted. Systolic prediction accuracy dropped from an R2of 0.248 to a shocking -0.024, while RMSE increased to 18.69 mmHg.
Why should you care? These results underscore a fundamental issue in deploying AI in healthcare: what works in one place may not work in another. The capital isn't leaving AI. it's wrestling with jurisdictional variability.
Imputation and Adaptation: A Band-Aid, Not a Cure
Researchers explored potential causes, like feature imputation, and attempted various fixes. They tried an intersection-only experiment with 16 common features. Still, the results worsened, with an R2of -0.115 and an RMSE of 17.32 mmHg. Post-hoc corrections like linear recalibration and covariate shift reweighting were tested. Yet, gains were minimal, with R2values ranging from -0.170 to 0.024.
Isn't it time we face reality? Strong internal performance doesn't promise external success. Asia moves first, but it seems the West isn't alone in facing these challenges.
Root Causes: More Than Just Data
The study identified three core issues: site-specific feature distributions, varying patient populations, and different measurement protocols. These barriers highlight the need for transparent reporting of validation failures. Without understanding these obstacles, predictive models will fall short of real-world deployment.
The licensing race in Hong Kong is accelerating. Yet, as we rush to integrate AI into healthcare, we must acknowledge these foundational issues. Could this be a wake-up call for the industry?
In the end, this research serves as a reminder. While AI in healthcare holds immense promise, we must confront the harsh reality of its limitations. Without overcoming cross-institutional barriers, the journey from lab to clinic will remain fraught with challenges.
Get AI news in your inbox
Daily digest of what matters in AI.