Within a few weeks’ time I read “How Algorithms Could Bring Empathy Back To Medicine,” “How AI Is Humanizing Healthcare,” and “Making Health Care Human Again.”Each came from somewhat varied publishing venues—Nature, Mission, and Fortune. I think such diversity in and of itself sends a clear message that while technology is cool, we are missing something in our experience of the provision of healthcare—both as doctors and as patients.

First off, full disclosure, I’m a longtime fan of Eric Topol, MD’s work. I had the chance to meet up with him at a South-By a few years back, having just finished enjoying his then new book, “The Patient Will See You Now.” I more recently read his piece on high-performance medicine in Nature Medicine and his great new book, “Deep Medicine,” both looking at what artificial intelligence can and cannot (yet) do in healthcare and medical practice. I like his informed, albeit ironic, outlook that something like machine learning can act to make healthcare more human.

It is also interesting to have learned from his exhaustive review, that Artificial Intelligence (AI) and Machine Learning (ML) while Jetson-like in our homes, and promising equal parts of being awesome/terrifying via the movies, it is a bit far from being a panacea in medicine.

Bottom-line and spoiler alert, AI is great for diagnostic imaging/radiology as it “sees” better than we can, and in data processing of medical histories and the voluminous evidence-based scientific, clinical literature, because it can remember and pattern-synthesize better and faster than us. Without question, neither of those are trivial successes, as it is a humanly impossible task to keep current with medical progress. But much of what’s what in AI and healthcare are just currently promising ideas making their way up the hype-scale, as many are not yet operational.

Tech to the Rescue

Many years ago, with my brand new license proudly mounted on my wall,I was actually a bit terrified in my first few years of post-doctoral practice that I would misdiagnose someone (Is it depression or an endocrine disorder...?) and thus mistreat them. In response to my anxiety and imposter feelings, I called upon my prior computer science training and did a mash-up differential diagnostics algorithm(that would certainly NOT qualify as being AI) to suss-out the probabilistic difference between a true positive psychiatric diagnosis and a mimicking physical one.

Even further back in time, 1955 to be precise, John McCarthy, coined the term “Artificial Intelligence” and four years later, Arthur Samuel described what we think of as machine learning. In a clinical realm, ELIZA was the world’s first psychotherapist chatbot in 1964, which was capable of passing the Turing Test. ELIZA was light years beyond what I wrote. But even with this decades long existence of AI and ML, and even clinical uses, it was not until 2016 that the term machine learning first made its appearance in the (arguably) top two American medical journals—The New England Journal of Medicine and the Journal of the American Medical Association.

Side-effects? What side-effects?

I’d like to think AI is agnostic, bias-free and data-centric, But in AI—just as with humans—you are what you eat. In one of my podcast episodes with Heather Dewey-Hagborg, we discussed the programmed-in biases of coders based on the availability of data/images’ subsequent impact on facial recognition based on what they built from. Topol noted Cathy O’Neil’s finding in her book Weapons of Math Destruction that “many of these models encoded human prejudice, misunderstanding, and bias.” Uh-oh.

"Ai is great for diagnostic imaging/ radiology as it “sees” better than we can"

So, it seems as with most procedures, medications and treatments, there may be an iatrogenic-bad that comes along with the therapeutic-good. This is particularly difficult to spot in considering the often “black box” aspect of AI algorithm development in everything, let alone medicine.

IBM’s Dr. Watson’s patients are not faring well either, as it was found to have “recommended ‘unsafe and incorrect’ cancer treatments.” And in our currently app-happy world, there may also be some worries. For example, Saeb and colleagues note that the standard approach to evaluate predictive algorithm accuracy is via cross-validation. However, not all are statistically meaningful. They noted that “… record-wise, cross-validation often massively overestimates the prediction accuracy of the algorithms… (and) …that this erroneous method is used by almost half of the retrieved studies… (to) predict clinical outcomes.”Uh-oh, again.

And as for the Future

Given my poor forecasting abilities when it comes to technology and medicine as I have previously written as a LinkedIn Influencer, perhaps a better approach would be to play a role in inventing the future rather than trying to predict it (a la Dennis Gabor’s observation). So, here are some of the projects afoot that we’re working on to help invent a better future in healthcare:

• Developing ways to measure and report on Clinician Performance to ensure consistent, quality of care in all clinics, better manage staff and help tailor training needs and resource deployment,
• Synthetic integration of Treatment Guidelines to be easily available to clinicians via the electronic medical record,
• Measurement of Patient Reported Outcomes and use in

o Optimizing Medicare reimbursements based on value of care provided
o Contracting with third party payers
o Contracting with self-insured employers
o Contracting with Unions and other guilds and associations
o National Outcomes Registry

• Scaling expansion of specialized treatment approaches along with clinical outcomes and concomitant economic savings aimed at combatting over utilization of expensive procedures and opioids,
• Use of Machine Learning and Artificial Intelligenceto more specifically tailor “bespoke” treatment guidelines,
• Expanding involvement in Bundled Payments and sophistication of managing risk and measuring clinical outcome performance,
• Development of Risk Adjusted Treatment Outcomes that enhance all programs as well as argue for rational reimbursement from less sophisticated payers and create an empirical foundation for Valued Based Care,
• Predicting Patient No shows and testing optimal responses to mitigation,
• Expansion of apps for home exercise program deployment and monitoring,
• Incorporating new Patient Reported Outcome measures for more specialty services (e.g., occupational therapy, women’s health, neurological conditions, etc.),
• Exploration of Telehealth/ Telerehab applications, and
• Exploration of Virtual Reality incare provision.

I’m also working with a group of colleagues to develop what may be the first integrative work-up that is inclusive of the psychological/ emotional health aspects of a person along with their genetic and microbiome data as well. The findings would be interpreted and provide treatment and/or lifestyle recommendations, powered by a proprietary algorithm.

I do believe we are really moving towards personalized medicine, but perhaps not leading with the biological, but more so behavioral and social. The point is that AI does indeed hold the promise of making healthcare more human by making it easier to get back to being human in making the space for our humanity and empathy in the exam room.