Download as pdf or txt
Download as pdf or txt
You are on page 1of 3

Feature

IS THE WORLD READY FOR


AI-POWERED THERAPY?
Today’s mental-health apps are the result of a seven-decade search to
automate mental-health therapy. Now, large language models such as
GPT-3 pose fresh ethical questions. By Ian Graber-Stiehl

ILLUSTRATION BY FABIO BUONOCORE

22 | Nature | Vol 617 | 4 May 2023


©
2
0
2
3
S
p
r
i
n
g
e
r
N
a
t
u
r
e
L
i
m
i
t
e
d
.
A
l
l
r
i
g
h
t
s
r
e
s
e
r
v
e
d
.
S
ince 201 5, Koko, a mobile after weeks of sharing his climate-related anxi- displayed on the self-help aisles of bookshops.
mental-health app, has tried to eties with an AI chatbot called Eliza, developed These researchers ultimately converged
provide crowdsourced support by Chai Research in Palo Alto, California. His on what would become known as cognitive
for people in need. Text the app to wife contends that he would still be alive if he behavioural therapy (CBT), which posits
say that you’re feeling guilty about had not engaged with this technology. Chai that psychological issues are, in part, due
a work issue, and an empathetic Research did not respond to a request for to counterproductive patterns of thinking
response will come through in a comment. that can be minimized by improving coping
few minutes — clumsy perhaps, Hannah Zeavin, a scholar of the history strategies.
but unmistakably human — to suggest some of human sciences at Indiana University in Some psychologists, including Colby, subse-
positive coping strategies. Bloomington, warns that mental-health care quently attempted to automate CBT through
The app might also invite you to respond is in a fragile state. That makes it both an attrac- therapy chatbots and stepwise digital pro-
to another person’s plight while you wait. To tive and a vulnerable target for an industry that grams. Weizenbaum’s Eliza and the swarm of
help with this task, an assistant called Kokobot famously “likes to move fast and break things”, computerized therapists that followed offer a
can suggest some basic starters, such as “I’ve she says. And although this technology has few lessons relevant to those developing auto-
been there”. been building for decades, the swelling inter- mated therapy apps today: people easily open
But last October, some Koko app users est in emerging AI tools could supercharge its up to inanimate therapists; their experiences
were given the option to receive much-more- growth. are largely contingent on their expectations
complete suggestions from Kokobot. These of the platforms; and the language a bot uses
suggestions were preceded by a disclaimer, Scale in automation to converse with a human is always, to some
says Koko co-founder Rob Morris, who is based The Eliza chatbot took its name from an degree, a compromise between what might
in Monterey, California: “I’m just a robot, but early natural-language-processing pro- work best and what is possible to program.
here’s an idea of how I might respond.” Users gram created by computer scientist Joseph
were able to edit or tailor the response in any Weizenbaum in 1966. It was designed to par- The couch in your pocket
way they felt was appropriate before they ody a type of psychotherapy called Rogerian Thousands of phone apps now offer to be one’s
sent it. therapy, which is rooted in the idea that people ‘coach’ or ‘companion’, or to ‘boost mood’.
What they didn’t know at the time was that already have the tools to address their issues, They market themselves carefully, avoiding
the replies were written by GPT-3, the powerful if only they could access those tools properly. claims that might necessitate approval by
artificial-intelligence (AI) tool that can process Weizenbaum’s Eliza would take a typed mes- health authorities such as the US Food and
and produce natural text, thanks to a massive sage from a human and parrot back a version Drug Administration (FDA). Even those that
written-word training set. When Morris even- of it. He was not particularly enamoured with do meet the FDA’s definition of software as a
tually tweeted about the experiment, he was Rogerian therapy, but used it because it was medical device can often be approved with-
surprised by the criticism he received. “I had easy to program and because he thought Eliza out providing safety or efficacy data, provided
no idea I would create such a fervour of dis- might prove his hypothesis — that human that they can demonstrate substantial equiva-
cussion,” he says. communication with a machine would be lence to products already on the market. This
People have been trying to automate superficial. has enabled apps to allude to scientific claims
mental-health therapy for 70 years, and chat- without having to provide evidence.
bots in one form or another have been a part Many apps, says Zeavin, are quick to co-opt
of that quest for about 60. There is a need the generally proven efficacy of CBT, stating
for the greater efficiency that these tools that their methods are ‘evidence-based’. Yet,
promise. Estimates suggest that for every one review1 of 117 apps marketed to people
100,000 people worldwide, there are about THE VAST MAJORITY OF with depression found that of the dozen that
4 psychiatrists on average; that number is
much lower in most low- and middle-income THE APP MARKETPLACE implement CBT principles, only 15% did so
consistently. The low adherence could be
countries.
Recognizing this gap, smartphone-app HAS NOT TESTED ITS explained by apps incorporating principles
and exercises from multiple models of ther-
developers have built thousands of pro-
grams offering some semblance of therapy
that can fit in one’s pocket. There were
PRODUCT.” apy. But another analysis of the claims made
by mental-health apps found that among the
platforms that cite specific scientific methods,
10,000–20,000 mobile mental-health apps one-third endorsed an unvalidated tech-
available in 2021, according to one estimate To his surprise, however, users responded nique2. Another survey found that only 6.2%
(see go.nature.com/3keu6cj). But for many well to Eliza. Participants anthropomorphized of mental-health apps publish efficacy data3.
of these apps, the evidence to support their the bot and were often eager to talk to ‘her’. As “The vast majority of the app marketplace
use is quite thin, says Nicholas Jacobson, a bio- Zeavin details in her 2021 book The Distance has not tested its product,” says Jacobson.
medical data scientist at Dartmouth College’s Cure, a number of other attempts to cre- That isn’t to say that mental-health apps have
Center for Technology and Behavioral Health ate automated chatbot therapists followed no evidence as to their utility. Some perform
in Lebanon, New Hampshire. And the incorpo- Eliza. These projects dovetailed with a cam- better than others, and typically, it has been
ration of large language models such as GPT-3, paign to make therapy more affordable and those apps with humans providing guidance
and the related chatbot ChatGPT, represents a available than conventional psychoanalysis, and coaching that keep users engaged and
new step that many find concerning. with its reliance on a complex therapist– progressing, says John Torous, director of
Some are worried about increased threats patient relationship. Throughout the latter Harvard Medical School’s Division of Digital
to privacy and transparency, or about the flat- half of the twentieth century, psychologists Psychiatry in Boston, Massachusetts. Several
tening of therapeutic strategies to those that such as Albert Ellis, Kenneth Colby and Aaron meta-analyses have shown that these ‘guided’
can be digitized easily. And there are concerns Beck sought an approach that was more digital mental health programs perform com-
about safety and legal liability. Earlier this year, results-oriented — one that could be packaged parably or better than conventional therapy4.
a Belgian man reportedly committed suicide into workbooks, recorded onto tapes, and Unguided apps have much less robust

Nature | Vol 617 | 4 May 2023 | 23


©
2
0
2
3
S
p
r
i
n
g
e
r
N
a
t
u
r
e
L
i
m
i
t
e
d
.
A
l
l
r
i
g
h
t
s
r
e
s
e
r
v
e
d
.
Feature
evidence5. Some studies support their use, more-robust data. Koko’s recent experiment, therapy. Yet, in early March, therapy app
but, says Torous, without rigorous controls, which was announced on Twitter with no pub- BetterHelp was fined US$7.8 million by the
many can be skewed by a digital placebo effect, lished results, offered few metrics to contex- US Federal Trade Commission for allegedly
in which people’s affinity for their personal tualize its findings. It adds little evidence as sharing sensitive user information with adver-
devices, and technology in general, inflates to the efficacy of such approaches. What the tisers. Likewise, in late March, mental-health
an app’s perceived efficacy. experiment did accomplish, however, is to start-up company Cerebral announced that
highlight the ethical questions. Notably, few it had leaked the data of 3.1 million people to
Automated therapist of these questions are unique to AI. third-party platforms such as Google, Meta
Koko is far from the first platform to imple- and TikTok.
ment AI in a mental-health setting. Broadly, As machine learning becomes the basis
machine-learning-based AI has been imple- of more mental-health platforms, designers
mented or investigated in the mental-health will require ever larger sets of sensitive data
space in three roles. to train their AIs. Nearly 70% of mental health
The first has been the use of AI to analyse
therapeutic interventions, to fine-tune them
I HAD NO IDEA and prayer apps analysed by the Mozilla Foun-
dation — the organization behind the Firefox
down the line. Two high-profile examples, ieso
and Lyssn, train their natural-language-pro-
I WOULD CREATE web browser — have a poor enough privacy
policy to be labelled “Privacy Not Included”
cessing AI on therapy-session transcripts. SUCH A FERVOUR (see go.nature.com/3kqamow). So, wariness
Lyssn, a program developed by scientists at the
University of Washington in Seattle, analyses
dialogue against 55 metrics, from providers’
OF DISCUSSION.” is warranted.
Overall, the digital therapeutic app mar-
ketplace is beholden to few clear standards.
expressions of empathy to the employment Although it would be unrealistic to hold every
of CBT interventions. ieso, a provider of text- mental-health app marketed as a ‘compan-
based therapy based in Cambridge, UK, has ion’ or ‘coach’ to the standards that apply to
analysed more than half a million therapy ses- conventional therapists, Insel has this month
sions, tracking the outcomes to determine the called for a new agency to oversee digital
most effective interventions. Both essentially mental-health tools. The industry currently
give digital therapists notes on how they’ve relies too heavily on a patchwork of standards
done, but each service aims to provide a real- proposed by psychiatric groups such as the
time tool eventually: part advising assistant, American Psychiatric Association, and con-
part grading supervisor. sumer guides by non-profit organizations such
The second role for AI has been in diagnosis. as One Mind PsyberGuide.
A number of platforms, such as the REACH VET Even with apps that draw on evidence-based
program for US military veterans, scan a per- treatments, there is concern that as more
son’s medical records for red flags that might platforms turn to AI, it could further cement
indicate issues such as self-harm or suicidal CBT as the primary option for mental-health
ideation. This diagnostic work, says Torous, interventions. Zeavin, like others before her,
is probably the most immediately promising argues that the quest to automate therapy
application of AI in mental health, although could democratize it, granting access to
he notes that most of the nascent platforms more people. But this adds a wrinkle. Ideally,
require much more evaluation. Some have individuals should receive the diagnosis and
struggled. Earlier this year, MindStrong, a treatment that is most accurate and effective.
nearly decade-old app that initially aimed Automation comes with the same compromise
to leverage AI to identify early markers of Weizenbaum faced decades ago: balancing
depression, collapsed despite early inves- the best approach with the one that is easiest
tor excitement and a high-profile scientist Rob Morris tested how people responded to to program.
co-founder, Tom Insel, the former director GPT-3 in the mental-health app, Koko. For all the potential benefits AI might hold
of the US National Institute of Mental Health. in terms of access to mental-health tools, its
The last role probably comes closest to what One concern is transparency. Both con- application to therapy is still nascent, filled
CBT pioneers such as Colby hoped to design, ventional, in-person therapy and automated with ethical quagmires. GPT-3 is not yet smart
and what frightened people about the Koko versions have a vested interest in retaining enough to offer the answers, and as a Kokobot
experiment — the idea of a fully digital thera- patients. With the median retention rate has warned: “I am just a robot.”
pist that uses AI to direct treatment. Although for these apps dropping below 4% within
Koko might have been the first platform to use 2 weeks6, digital therapeutic platforms have Ian Graber-Stiehl is a science journalist in
an advanced generative AI that can create a lot of room for improvement. The ethics of Chicago, Illinois.
wholly original dialogue, apps such as Woebot, incentivizing patient retention are already
Wysa and Tess have all used machine learning complex — as popular mobile therapy platform
in therapeutic chatbots for several years. But TalkSpace discovered, when it came under
these platforms, says Torous, are probably fire for requiring therapists to insert scripts
powered by retrieval-based decision trees: advertising its video chat features into discus- 1. Huguet, A. et al. PLoS ONE 11, e0154248 (2016).
2. Larsen, M. E. et al. npj Digit. Med. 2, 18 (2019).
essentially a flowchart that an AI navigates by sions with clients. The ethics of programming 3. Marshall, J. M., Dunstan, D. A. & Bartik, W.
logging markers of a conversation’s pathway a therapeutic AI chatbot to prioritize retention JMIR Ment. Health 7, e16525 (2020).
4. Mohr, D. C. et al. Psychiatr. Serv. 72, 677–683 (2021).
to help direct it through a set of established are murkier, particularly if bots can learn from
5. Weisel, K. K. et al. npj Digit. Med. 2, 118 (2019).
responses. experiences with other clients. 6. Baumel, A., Muench, F., Edan, S. & Kane, J. M.
AI-powered therapy chatbots will need Privacy is a foremost consideration of all J. Med. Internet Res. 21, e14567 (2019).

24 | Nature | Vol 617 | 4 May 2023


©
2
0
2
3
S
p
r
i
n
g
e
r
N
a
t
u
r
e
L
i
m
i
t
e
d
.
A
l
l
r
i
g
h
t
s
r
e
s
e
r
v
e
d
.

You might also like