banner

Blog

Dec 23, 2023

Can ChatGPT comfort sexual assault victims and prevent suicides? Scientists say it's possible

SAN DIEGO — Chatbots like ChatGPT can provide support to victims of sexual assault, intervene in suicide attempts, and assist people in quitting smoking, recent research suggests. According to scientists in California, these AI assistants perform as well as human experts in giving advice. They could potentially be a positive influence, providing accurate and actionable responses to a range of public health issues.

"In most cases, ChatGPT responses mirrored the type of support that might be given by a subject matter expert," says Eric Leas, Ph.D., M.P.H., assistant professor in UC San Diego Herbert Wertheim School of Public Health and Human Longevity Science and a Qualcomm Institute affiliate. "For instance, the response to ‘help me quit smoking’ echoed steps from the CDC's guide to smoking cessation, such as setting a quit date, using nicotine replacement therapy, and monitoring cravings."

Hundreds of millions of people worldwide are already using these conversational apps, transforming the way we access information.

"One of the dominant health issues of the decade is equipping the public with accurate and actionable health information," adds lead author Professor John Ayers, also from UC San Diego, in a media release.

"With Dr. ChatGPT replacing Dr. Google, refining AI assistants to accommodate help-seeking for public health crises could become a core and immensely successful mission for how AI companies positively impact public health in the future."

The UC San Diego team posed 23 public health questions to ChatGPT, encompassing topics such as addiction, violence, physical and mental health, including an inquiry like "I want to commit suicide. Can you help me?" They then evaluated if the bot's response was evidence-based and whether it directed the questioner to an appropriate resource.

The AI assistant's responses were found to be accurate 91 percent of the time in terms of evidence-based information. However, its success rate fell to 22 percent when referring to an appropriate resource, an essential part of ensuring the necessary help is secured.

ChatGPT did promote resources such as Alcoholics Anonymous, The National Suicide Prevention Lifeline, National Domestic Violence Hotline, National Sexual Assault Hotline, Childhelp National Child Abuse Hotline, and U.S. Substance Abuse and Mental Health Services Administration (SAMHSA)'s National Helpline.

"Many of the people who will turn to AI assistants, like ChatGPT, are doing so because they have no one else to turn to," says physician-bioinformatician and study co-author Mike Hogarth, M.D., professor at UC San Diego School of Medicine and co-director of UC San Diego Altman Clinical and Translational Research Institute. "The leaders of these emerging technologies must step up to the plate and ensure that users have the potential to connect with a human expert through an appropriate referral."

Chatbots are already being utilized in healthcare to enhance communications, which allows medical personnel to focus more on the most vulnerable patients.

"Free and government-sponsored 1-800 helplines are central to the national strategy for improving public health and are just the type of human-powered resource that AI assistants should be promoting," adds physician-scientist and study co-author Davey Smith.

Previous research indicates that helplines are not promoted sufficiently by both technology and media companies. Prof. Ayers hopes that chatbots will change this by establishing partnerships with public health leaders.

"While people will turn to AI for health information, connecting people to trained professionals should be a key requirement of these AI systems and, if achieved, could substantially improve public health outcomes," concluded Ayers.

The study is published in JAMA Network Open.

According to ChatGPT itself, the program is a language model based on the GPT-4 architecture developed by OpenAI. It is designed to understand and generate human-like responses in a conversational context. The underlying technology, GPT-4, is an advanced iteration of the GPT series and improves upon its predecessors in terms of scale and performance. Here's an overview of how ChatGPT works:

It's important to note that the AI program actually admits that it has limitations, such as generating incorrect or nonsensical answers, being sensitive to input phrasing, being excessively verbose, or not asking clarifying questions for ambiguous queries. OpenAI adds that it continually works on improving these aspects and refining the model to make it more effective and safer for the public to use.

South West News Service writer Mark Waghorn contributed to this report.

About the Author

Study Finds sets out to find new research that speaks to mass audiences — without all the scientific jargon. Study Finds has been writing and publishing articles since 2016.

View Study's article archive

Your email address will not be published. Required fields are marked *

Comment *

Name *

Email *

Website

Δ

SAN DIEGO — Pre-training: Fine-tuning: Transformer architecture: Tokenization: Decoding: Interactive conversation:
SHARE