A few months ago, a woman in her mid-50s—let’s call her Sophie—experienced a hemorrhagic stroke. Her brain started to bleed. She underwent brain surgery, but her heart stopped beating.
Sophie’s ordeal left her with significant brain damage. She was unresponsive; she couldn’t squeeze her fingers or open her eyes when asked, and she didn’t flinch when her skin was pinched. She needed a tracheostomy tube in her neck to breathe and a feeding tube to deliver nutrition directly to her stomach, because she couldn’t swallow. Where should her medical care go from there?
This difficult question was left, as it usually is in these kinds of situations, to Sophie’s family members, recalls Holland Kaplan, an internal-medicine physician at Baylor College of Medicine who was involved in Sophie’s care. But the family couldn’t agree. Sophie’s daughter was adamant that her mother would want to stop having medical treatments and be left to die in peace. Another family member vehemently disagreed and insisted that Sophie was “a fighter.” The situation was distressing for everyone involved, including Sophie’s doctors.
End-of-life decisions can be extremely upsetting for surrogates, the people who have to make those calls on behalf of another person, says David Wendler, a bioethicist at the US National Institutes of Health. Wendler and his colleagues have been working on an idea for something that could make things easier: an artificial-intelligence-based tool that can help surrogates predict what patients themselves would want in any given situation.
The tool hasn’t been built yet. But Wendler plans to train it on a person’s own medical data, personal messages, and social media posts. He hopes it could not only be more accurate at working out what the patient would want, but also alleviate the stress and emotional burden of difficult decision-making for family members.
Wendler, along with bioethicist Brian Earp at the University of Oxford and their colleagues, hopes to start building the tool as soon as they secure funding for it, potentially in the coming months. But rolling it out won’t be simple. Critics wonder how such a tool can ethically be trained on a person’s data, and whether life-or-death decisions should ever be entrusted to AI.
Live or die
Around 34% of people in a medical setting are considered to be unable to make decisions about their own care for various reasons. They may be unconscious, for example, or unable to reason or communicate. This figure is higher among older individuals—one study of people over 60 in the US found that 70% of those faced with important decisions about their care lacked the capacity to make those decisions themselves. “It’s not just a lot of decisions—it’s a lot of really important decisions,” says Wendler. “The kinds of decisions that basically decide whether the person is going to live or die in the near future.”
Chest compressions administered to a failing heart might extend a person’s life. But the treatment might lead to a broken sternum and ribs, and by the time the person comes around—if ever—significant brain damage may have developed. Keeping the heart and lungs functioning with a machine might maintain a supply of oxygenated blood to the other organs—but recovery is no guarantee, and the person could develop numerous infections in the meantime. A terminally ill person might want to continue trying hospital-administered medications and procedures that could offer a few more weeks or months. But someone else might want to forgo those interventions and be more comfortable at home.
Only around one in three adults in the US completes any kind of advance directive—a legal document that specifies the end-of-life care they might want to receive. Wendler estimates that over 90% of end-of-life decisions end up being made by someone other than the patient. The role of a surrogate is to make that decision based on beliefs about how the patient would want to be treated. But people are generally not very good at making these kinds of predictions. Studies suggest that surrogates accurately predict a patient’s end-of-life decisions around 68% of the time.
The decisions themselves can also be extremely distressing, Wendler adds. While some surrogates feel a sense of satisfaction from having supported their loved ones, others struggle with the emotional burden and can feel guilty for months or even years afterwards. Some fear they ended the life of their loved ones too early. Others worry they unnecessarily prolonged their suffering. “It’s really bad for a lot of people,” says Wendler. “People will describe this as one of the worst things they’ve ever had to do.”
Wendler has been working on ways to help surrogates make these kinds of decisions. Over 10 years ago, he developed the idea for a tool that would predict a patient’s preferences on the basis of characteristics such as age, gender, and insurance status. That tool would have been based on a computer algorithm trained on survey results from the general population. It may seem crude, but these characteristics do seem to influence how people feel about medical care. A teenager is more likely to opt for aggressive treatment than a 90-year-old, for example. And research suggests that predictions based on averages can be more accurate than the guesses made by family members.
In 2007, Wendler and his colleagues built a “very basic,” preliminary version of this tool based on a small amount of data. That simplistic tool did “at least as well as next-of-kin surrogates” in predicting what kind of care people would want, says Wendler.
Now Wendler, Earp and their colleagues are working on a new idea. Instead of being based on crude characteristics, the new tool the researchers plan to build will be personalized. The team proposes using AI and machine learning to predict a patient’s treatment preferences on the basis of personal data such as medical history, along with emails, personal messages, web browsing history, social media posts, or even Facebook likes. The result would be a “digital psychological twin” of a person—a tool that doctors and family members could consult to guide a person’s medical care. It’s not yet clear what this would look like in practice, but the team hopes to build and test the tool before refining it.
The researchers call their tool a personalized patient preference predictor, or P4 for short. In theory, if it works as they hope, it could be more accurate than the previous version of the tool—and more accurate than human surrogates, says Wendler. It could be more reflective of a patient’s current thinking than an advance directive, which might have been signed a decade beforehand, says Earp.
A better bet?
A tool like the P4 could also help relieve the emotional burden surrogates feel in making such significant life-or-death decisions about their family members, which can sometimes leave people with symptoms of post-traumatic stress disorder, says Jennifer Blumenthal-Barby, a medical ethicist at Baylor College of Medicine in Texas.
Some surrogates experience “decisional paralysis” and might opt to use the tool to help steer them through a decision-making process, says Kaplan. In cases like these, the P4 could help ease some of the burden surrogates might be experiencing, without necessarily giving them a black-and-white answer. It might, for example, suggest that a person was “likely” or “unlikely” to feel a certain way about a treatment, or give a percentage score indicating how likely the answer is to be right or wrong.
Kaplan can imagine a tool like the P4 being helpful in cases like Sophie’s, where various family members might have different opinions on a person’s medical care. In those cases, the tool could be offered to these family members, ideally to help them reach a decision together.
It could also help guide decisions about care for people who don’t have surrogates. Kaplan is an internal-medicine physician at Ben Taub Hospital in Houston, a “safety net” hospital that treats patients whether or not they have health insurance. “A lot of our patients are undocumented, incarcerated, homeless,” she says. “We take care of patients who basically can’t get their care anywhere else.”
These patients are often in dire straits and at the end stages of diseases by the time Kaplan sees them. Many of them aren’t able to discuss their care, and some don’t have family members to speak on their behalf. Kaplan says she could imagine a tool like the P4 being used in situations like these, to give doctors a little more insight into what the patient might want. In such cases, it might be difficult to find the person’s social media profile, for example. But other information might prove useful. “If something turns out to be a predictor, I would want it in the model,” says Wendler. “If it turns out that people’s hair color or where they went to elementary school or the first letter of their last name turns out to [predict a person’s wishes], then I’d want to add them in.”
This approach is backed by preliminary research from Earp and his colleagues, who have started running surveys to find out how individuals might feel about using the P4. This research is ongoing, but early responses suggest that people would be willing to try the model if there were no human surrogates available. Earp says he feels the same way. He also says that if the P4 and a surrogate were to give different predictions, “I’d probably defer to the human that knows me, rather than the model.”
Not a human
Earp’s feelings betray a gut instinct many others will share: that these huge decisions should ideally be made by a human. “The question is: How do we want end-of-life decisions to be made, and by whom?” says Georg Starke, a researcher at the Swiss Federal Institute of Technology Lausanne. He worries about the potential of taking a techno-solutionist approach and turning intimate, complex, personal decisions into “an engineering issue.”
Bryanna Moore, an ethicist at the University of Rochester, says her first reaction to hearing about the P4 was: “Oh, no.” Moore is a clinical ethicist who offers consultations for patients, family members, and hospital staff at two hospitals. “So much of our work is really just sitting with people who are facing terrible decisions … they have no good options,” she says. “What surrogates really need is just for you to sit with them and hear their story and support them through active listening and validating [their] role … I don’t know how much of a need there is for something like this, to be honest.”
Moore accepts that surrogates won’t always get it right when deciding on the care of their loved ones. Even if we were able to ask the patients themselves, their answers would probably change over time. Moore calls this the “then self, now self” problem.
And she doesn’t think a tool like the P4 will necessarily solve it. Even if a person’s wishes were made clear in previous notes, messages, and social media posts, it can be very difficult to know how you’ll feel about a medical situation until you’re in it. Kaplan recalls treating an 80-year-old man with osteoporosis who had been adamant that he wanted to receive chest compressions if his heart were to stop beating. But when the moment arrived, his bones were too thin and brittle to withstand the compressions. Kaplan remembers hearing his bones cracking “like a toothpick,” and the man’s sternum detaching from his ribs. “And then it’s like, what are we doing? Who are we helping? Could anyone really want this?” says Kaplan.
There are other concerns. For a start, an AI trained on a person’s social media posts may not end up being all that much of a “psychological twin.” “Any of us who have a social media presence know that often what we put on our social media profile doesn’t really represent what we truly believe or value or want,” says Blumenthal-Barby. And even if we did, it’s hard to know how these posts might reflect our feelings about end-of-life care—many people find it hard enough to have these discussions with their family members, let alone on public platforms.
As things stand, AI doesn’t always do a great job of coming up with answers to human questions. Even subtly altering the prompt given to an AI model can leave you with an entirely different response. “Imagine this happening for a fine-tuned large language model that’s supposed to tell you what a patient wants at the end of their life,” says Starke. “That’s scary.”
On the other hand, humans are fallible, too. Vasiliki Rahimzadeh, a bioethicist at Baylor College of Medicine, thinks the P4 is a good idea, provided it is rigorously tested. “We shouldn’t hold these technologies to a higher standard than we hold ourselves,” she says.
Earp and Wendler acknowledge the challenges ahead of them. They hope the tool they build can capture useful information that might reflect a person’s wishes without violating privacy. They want it to be a helpful guide that patients and surrogates can choose to use, but not a default way to give black-and-white final answers on a person’s care.
Even if they do succeed on those fronts, they might not be able to control how such a tool is ultimately used. Take a case like Sophie’s, for example. If the P4 were used, its prediction might only serve to further fracture family relationships that are already under pressure. And if it is presented as the closest indicator of a patient’s own wishes, there’s a chance that a patient’s doctors might feel legally obliged to follow the output of the P4 over the opinions of family members, says Blumenthal-Barby. “That could just be very messy, and also very distressing, for the family members,” she says.
“What I’m most worried about is who controls it,” says Wendler. He fears that hospitals could misuse tools like the P4 to avoid undertaking costly procedures, for example. “There could be all kinds of financial incentives,” he says.
Everyone contacted by MIT Technology Review agrees that the use of a tool like the P4 should be optional, and that it won’t appeal to everyone. “I think it has the potential to be helpful for some people,” says Earp. “I think there are lots of people who will be uncomfortable with the idea that an artificial system should be involved in any way with their decision making with the stakes being what they are.”