Published March 2025
In 2019, UPMC quietly rolled out what it called a breakthrough. The Pittsburgh-based health giant had spent years building a machine learning algorithm, trained on records from more than 1.25 million patients, designed to predict which surgical patients were most likely to die or suffer a serious complication. Deployed across 20 of its hospitals, the tool would scan patient charts each morning and flag the highest-risk cases before surgeons ever made the first incision. The goal, according to Dr. Aman Mahajan, the algorithm's lead architect, was to give clinicians "a more objective, data-driven assessment of who might run into trouble after surgery."
The data suggests that assessment was not enough.
The measure, PSI-04, tracked by the Centers for Medicare and Medicaid Services, is designed specifically to capture deaths that should not have happened. It counts surgical patients who develop a serious complication and then die, with the underlying assumption that these are deaths medicine had a chance to prevent. Dr. Mahajan, in an interview, acknowledged the weight of that framing. "PSI-04 focuses on patients who developed serious complications and then died, cases where medicine, theoretically, had an opportunity to intervene," he said. "When that number increases, it raises a fundamental question: Were high-risk patients identified but not effectively managed? Or were the interventions insufficient? Prediction is only the first step. What matters is what happens after the risk is identified."
In theory, the tool was designed to trigger a chain of action. Each morning, surgical teams would receive a list of flagged patients, prompting, as Dr. Mahajan described it, "heightened vigilance, perhaps additional monitoring, consultations, or changes to surgical planning." But he acknowledged the gap between alert and outcome is not automatic. "An algorithm can demonstrate strong predictive accuracy, but that does not automatically improve outcomes. There has to be a clear workflow: Who responds to the alert? What resources are mobilized? Is there adequate staffing to act on those warnings? Without an effective response system, prediction alone does not save lives."
His broader conclusion was pointed. "AI is a tool, not a solution," he said. "If a hospital system promotes AI as a safety breakthrough, it also assumes responsibility for demonstrating that it improves real-world outcomes. Otherwise, we risk mistaking predictive sophistication for clinical effectiveness."
For the patients counted in rising mortality rates, that distinction is not academic.