Primary care is facing multiple crises, including an increase in health misinformation. Digital health messaging by primary care providers has been shown to reach a diverse patient population. With the uptake of Generative Artificial Intelligence (GenAI) usage in healthcare, there is an important opportunity to rapidly create messages that are tailored to different populations and conditions. However, thoroughly assessing artificial intelligence (AI)-generated content is essential, as GenAI raises concerns regarding its accuracy, understandability, actionability and bias perpetuation. We aim to investigate whether digital health messages created by GenAI are evaluated as non-inferior compared with those created by human experts.
The AI-CARE (AI to Create Accessible and Reliable patient Education materials) study is a double-blind, crossover, non-inferiority randomised controlled trial. Data collection began on 30 May 2025, and is expected to be completed at the end of May 2026. Over 12 months, 192 messages on 48 topics will be written: half by primary care and public health experts and half by a GenAI tool (OpenAI’s ChatGPT). Review Panels composed of 24 primary care providers and 24 patients will evaluate these messages using an Evaluation Grid developed to assess the messages’ quality of information, adaptation to the target audience, relevance and usefulness, and readiness to be shared with patients. Evaluations will be completed via online REDCap (Research Electronic Data Capture) surveys and the order in which the 192 messages appear will be randomised and will vary between individuals. Participants and analysts will be blinded to the generation source. The primary outcome will be the Clarity and Understandability score.
The Research Ethics Boards of the Hôpital Montfort (24-25-11-038) and the University of Ottawa (S-12-24-11153) formally approved this study in December 2024. Reported data will be grouped and anonymised for dissemination in peer-reviewed scientific journals and conferences.