On April 28, 2026, researchers at Mayo Clinic and UT MD Anderson Cancer Center published a finding in Gut that quietly reframes how medicine could approach one of its most dangerous cancers. Their AI model detected signs of pancreatic cancer in routine CT scans up to three years before the tumor was clinically diagnosed. The scans looked normal at the time. They weren't.

Old scans contain warnings nobody knew to read

The model is called REDMOD, short for Radiomics-based Early Detection MODel. It does not look for visible tumors. It reads radiomic texture, fine-grained patterns in tissue density that sit below the threshold of what a trained human eye can detect. Every CT scanner captures these patterns. REDMOD's job is to recognize which ones matter.

The retrospective design is the detail that changes what this means. Those scans were not ordered to look for cancer. They were routine abdominal CTs done for other reasons, the kind hospitals perform thousands of times a year. Each patient had already been scanned. The warning was already recorded. It just was not extracted.

The numbers close the gap

On an independent test set of 63 pre-diagnostic cancer scans and 430 healthy controls, REDMOD caught 73% of the cancer cases. Radiologists reading the same scans caught 39%. That is not a marginal difference.

For scans taken more than two years before diagnosis, the gap widens sharply. REDMOD detected 68% of those cases; radiologists caught 23%. Nearly three times more, at the exact window where early detection has the most potential to change the outcome.

The specificity figure is the one most coverage omits. REDMOD has 81.1% specificity on the test set, meaning roughly one in five healthy patients would receive a false positive flag. That number does not undercut the result. It contextualizes it. A clinical tool is not evaluated by sensitivity alone. The trade-off between catching more cancers and generating more unnecessary follow-up investigations is the core question any health system deploying this model would have to answer explicitly. Mayo Clinic's own press release describes the study as a validation, not a screening program. That distinction matters.

475 days is a window, not a plan

The median lead time was 475 days, about 16 months before clinical diagnosis. The maximum observed was three years. For a cancer with a 13% overall five-year survival rate, driven almost entirely by how late the disease is caught, a 16-month head start is significant on paper.

But "significant on paper" is not the same as "changes outcomes." This study is retrospective. There is no prospective evidence yet that an earlier REDMOD flag translates to better survival. A prospective trial named AI-PACED (Artificial Intelligence for Pancreatic Cancer Early Detection) is underway as part of Mayo's Precure initiative, but results are years away.

The clinical workflow gap is not a technical problem. It is an institutional one. If REDMOD flags a patient today, a patient with no symptoms and no visible tumor, what happens next? Who orders the follow-up? What does it look like? When does watching become testing? Those are not questions the model answers. They are questions health systems would have to design protocols for before any broad deployment becomes defensible.

Why it matters

The story most outlets will miss is not the performance gap over radiologists. It is what the retrospective architecture implies.

Hospitals already hold millions of CT scans in archived storage. Many of those patients have since received diagnoses for conditions that showed no visible sign at scan time. REDMOD was built to run on exactly those images. The data already exists. The warning may already be in there. The question is whether medicine has the infrastructure, the workflows, and the institutional will to ask new questions of old data.

This is a different model of early detection. Historically, early detection meant new tests, new appointments, new costs, new asks of patients and clinicians. What REDMOD points toward is passive retrospective review: no new imaging required, no new patient encounter, just a different computation run on data already in the room.

The obstacles are real. Broader population validation is needed, particularly across hospitals that are not research centers, on different scanner hardware, and in patient populations that do not look like the study cohort. Regulatory clearance, integration infrastructure, and defined follow-up protocols all come before meaningful deployment. A commenter in the r/artificial thread where the study surfaced framed it well: "If AI can shift that detection window by even a year or two the downstream effects on mortality would be enormous. The tricky part is validating these models across diverse patient populations."

That is where the work actually is. Not in the headline comparison number. In the decade of prospective trials, workflow design, and population-level validation between a promising result and a tool that reliably changes whether someone lives or dies.

Pancreatic cancer's brutality is partly a timing problem. If the signal was in the scan three years ago, and medicine is only now developing tools to read it, the question is not whether AI can find the warning. It is whether health systems can build the infrastructure to act on it before the window closes.

Originally published as an Instagram carousel on @recul.ai.