AI-Mediated Health Communication: Evaluating Large Language Model-Based FAQ Rewriting to Foster Clinical Trial Participation via A Randomized Controlled Study
ABSTRACT
Background:
Effective communication about clinical trials is essential, as low enrollment undermines scientific validity and contributes to healthcare inequities. However, recruitment remains a persistent challenge, particularly among older adults, minority populations, and individuals with limited health literacy. Although large language models (LLMs) show promise in understanding and generating health information, it is unclear whether these generative AI tools can improve the content of hospitals’ frequently asked questions (FAQ) pages to enhance public attitudes and intentions toward clinical trial participation.
Objective:
This study aimed to compare clinical trial FAQs from health organizations and hospitals with versions rewritten by LLMs to examine whether the generated content improves public attitudes and intentions toward clinical trial participation, and to identify the mechanisms underlying these effects.
Methods:
308 question–answer pairs were collected from the FAQ pages of 38 health organizations and hospitals, categorizing them into 52 types and selecting the 11 most frequent for testing. In a quasi-experiment with 440 online participants, the control group viewed the original FAQs, while the experimental group read GPT-4o-generated answers emphasizing comprehension and empathy. The study compared the impact of AI-generated versus standard FAQ content on attitudes toward clinical trials and examined Theory of Planned Behavior constructs to determine for whom and how AI information is most effective.
Results:
Participants were recruited through CloudResearch, yielding a 96.94% completion rate, resulting in 440 valid responses across the two conditions. Participants who viewed GPT-generated information (M = 0.26, SD = 0.65) showed a marginally greater positive change in outcome evaluation attitudes than those who viewed standard FAQs (M = 0.13, SD = 0.70, p = .056, 95% CI [0.00, 0.25]). Follow-up linear regression analyses revealed that several individual factors significantly moderated the effect of the information type (FAQ vs. GPT) on attitude change, including age (MDiffer = 0.87, SE = 0.33, t(394) = 2.62, p = .009), race (MDiffer = 0.36, SE = 0.15, t(383) = 2.47, p = .014), risk aversion (B = 0.12, SE = 0.06, t = 2.23, p = .026), fear of ineffective treatment (B = 0.11, SE = 0.05, t = 2.03, p = .043), and fear of unknown treatment effects (B = 0.21, SE = 0.07, t = 3.10, p = .002).
Conclusions:
This study is the first to apply the Theory of Planned Behavior (TPB) to compare LLM-rewritten versus original FAQ content for clinical trial communication. While GPT-generated responses did not directly increase participation intentions, they improved attitudes among traditionally underrepresented groups, including older adults, Black participants, and those with higher uncertainty avoidance or treatment concerns. These attitude gains were positively linked to participation intentions, suggesting that AI-generated language can enhance public attitudes, perceptions, and engagement with clinical research.
Citation
The author of this paper has made a PDF available, but requires the user to login, or create an account.
Copyright
© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.