Today’s Solutions: December 15, 2025

At the Optimist Daily, we’re always on the lookout for solutions—whether they may be for lighter issues like gardening, or for much heavier topics, like how we can best discuss how to prepare for death. Today, we’ll be touching upon the latter.

The question of how and when to prepare for death is among the most difficult and human of conversations — one which centers around our (perhaps unique) ability to grasp, turn, and examine each facet of our mortality, like a diamond under a loupe. Yet, surprisingly, these important conversations are increasingly being guided by very non-human advice: artificial intelligence. For doctors and patients, crucial but difficult decisions about end-of-life care cannot be made until a conversation about dying begins. But the taboo around death and fear of discouraging patients often delay such conversations until it is too late.

Writing in STAT, Rebecca Robbins interviewed over a dozen clinicians, researchers, and AI developers and experts on the role of machine learning in addressing patient’s mortal concerns. “A lot of times, we think about it too late — and we think about it when the patient is decompensating, or they’re really, really struggling, or they need some kind of urgent intervention to turn them around,” said Stanford inpatient medical physician Samantha Wang.

The nudge provided by AI may help doctors and patients have this difficult talk before it’s too late.

Multiple artificial intelligence models are being applied to palliative care. The models use various machine learning techniques to analyze the medical records of patients, availing themselves to the vast troves of data to generate mortality probabilities. These AI actuaries are trained with, and then tested on, data of patients who have already been treated, including diagnoses, treatments, and outcomes, discharge or death; some also include socioeconomic data and insurance information, Robbins writes.

From there, clinicians receive notifications about those whom the algorithm feels are at the highest risk of death — and prompts that difficult discussion. Those messages have to be considered and curated carefully; at UPenn, clinicians never receive more than six at a time, to avoid overwhelming docs and generating alarm fatigue. At Stanford, the notifications do not include the patient’s probabilities.

“We don’t think the probability is accurate enough, nor do we think human beings — clinicians — are able to really appropriately interpret the meaning of that number,” Stanford physician Ron Li, per STAT.

It’s odd to think about relinquishing such a heavy human burden onto artificial intelligence, but perhaps that is the appeal of it too. Even for highly experienced doctors, having a conversation about death with patients is an incredibly difficult decision. With accurate, highly selective AI, doctors can feel more certain they are making the right choice.

Solutions News Source Print this article
More of Today's Solutions

Vision board ideas for adults: how to create one that inspires real change

BY THE OPTIMIST DAILY EDITORIAL TEAM A vision board might look like a crafty throwback to childhood afternoons spent collaging. But don’t write it ...

Read More

India’s social experiment: how paying women directly reshapes welfare, autono...

BY THE OPTIMIST DAILY EDITORIAL TEAM Across India, millions of women now receive a modest but unwavering deposit each month into their bank accounts. ...

Read More

New Zealand’s groundbreaking shift to renewables promises massive emiss...

New Zealand launched its most ambitious emissions reduction initiative to date in an incredible undertaking. The government announced a historic switch from coal to ...

Read More

Going for the goal: the impact of team sports on boosting young girls’ ...

In a pioneering study, the Here for Every Goal report demonstrates that team sports, particularly elite women's soccer (referenced from here on in this ...

Read More