By Kaveh Waddell, Axios | May 30, 2019

Facing increasingly overworked doctors and labyrinthine insurance systems, hospitals are searching for a lifeline in AI systems that promises to ease hard diagnoses and treatment decisions.

Reality check: The data underpinning the very first systems is often spotty, volatile and completely lacking in critical context, leading to a poor early record in the field.

The big picture: Basic clinical decision support (CDS) systems have been around for decades, but a skepticism of technology leads many doctors to ignore or override them. Now, experts say a nascent generation of CDS — infused with AI in academic labs and startups — may reduce the estimated 40,000–80,000 deaths a year that result from medical errors.

  • The grand vision: Researchers hope AI programs can point doctors toward the best medications, lab tests or treatment plans based on minute patterns discovered in huge numbers of patients’ past experiences.
  • Last week, we reported on the promise of combining pools of private data to strengthen AI systems, feeding them with ever more examples of past outcomes.
  • This helps solve the quantity issue. But data quality — a constant struggle in health care — remains an enormous threat to medical AI.

The big problem: Record keeping is so bad that doctors laugh when you ask about it.

  • Electronic medical records, central to CDS predictions, are notoriously error-ridden. Doctors fill them with generic diagnosis codes, and 82% of any given record is likely to have been copied or imported from elsewhere.
  • “Filling out documentation right is something that most physicians don’t care about,” says Jonathan Chen, a doctor and professor of biomedical informatics at Stanford.

Other quirks of health data make more problems for CDS systems:

  • Records reaching back even a year or two become useless for predicting future trends because of how quickly treatments shift. A study led by Chen found that medical data has a usefulness half-life of just 4 months.
  • Hospital readmission rates are a gold standard for measuring whether a particular treatment worked, but the statistic misses more important outcomes like patient happiness and long-term health.
  • Together, all this means that if doctors and researchers are not careful, “you’ll end up learning patterns that are either obvious or you’ll completely misinterpret what they mean in potentially dangerous ways,” says Chen.

It’s the oldest problem in data science: garbage in, garbage out.

  • “We are at a decided disadvantage because our core electronic record is so pitiful,” says Eric Topol, director of the Scripps Research Translational Institute.
  • The messy reality of medical records tripped up IBM’s much-ballyhooed Watson AI system when it was deployed at a Texas cancer hospital. “[T]he acronyms, human errors, shorthand phrases, and different styles of writing” were too much to handle, Stat News reported in 2017. An IBM executive today blamed data quality for past AI failures.

What’s next: The Food and Drug Administration, which currently doesn’t review most CDS systems, is considering policy changes that could head off some data issues. Scientists are pushing the agency to impose strict benchmarks and audits to prevent mistakes.

Go deeper: What your hospital knows about you