Artificial intelligence-powered chatbots in search engines: a cross-sectional study on the quality and risks of drug information for patients.

Wahram Andrikyan, Sophie Marie Sametinger, Frithjof Kosfeld, Lea Jung-Poppe, Martin F Fromm, Renke Maas, Hagen F Nicolaus
Author Information
  1. Wahram Andrikyan: Institute of Experimental and Clinical Pharmacology and Toxicology, Friedrich-Alexander-Universit��t Erlangen-N��rnberg, Erlangen, Germany wahram.andrikyan@fau.de. ORCID
  2. Sophie Marie Sametinger: Institute of Experimental and Clinical Pharmacology and Toxicology, Friedrich-Alexander-Universit��t Erlangen-N��rnberg, Erlangen, Germany.
  3. Frithjof Kosfeld: Institute of Experimental and Clinical Pharmacology and Toxicology, Friedrich-Alexander-Universit��t Erlangen-N��rnberg, Erlangen, Germany.
  4. Lea Jung-Poppe: Institute of Experimental and Clinical Pharmacology and Toxicology, Friedrich-Alexander-Universit��t Erlangen-N��rnberg, Erlangen, Germany.
  5. Martin F Fromm: Institute of Experimental and Clinical Pharmacology and Toxicology, Friedrich-Alexander-Universit��t Erlangen-N��rnberg, Erlangen, Germany.
  6. Renke Maas: Institute of Experimental and Clinical Pharmacology and Toxicology, Friedrich-Alexander-Universit��t Erlangen-N��rnberg, Erlangen, Germany.
  7. Hagen F Nicolaus: Institute of Experimental and Clinical Pharmacology and Toxicology, Friedrich-Alexander-Universit��t Erlangen-N��rnberg, Erlangen, Germany.

Abstract

BACKGROUND: Search engines often serve as a primary resource for patients to obtain drug information. However, the search engine market is rapidly changing due to the introduction of artificial intelligence (AI)-powered chatbots. The consequences for medication safety when patients interact with chatbots remain largely unexplored.
OBJECTIVE: To explore the quality and potential safety concerns of answers provided by an AI-powered chatbot integrated within a search engine.
METHODOLOGY: Bing copilot was queried on 10 frequently asked patient questions regarding the 50 most prescribed drugs in the US outpatient market. Patient questions covered drug indications, mechanisms of action, instructions for use, adverse drug reactions and contraindications. Readability of chatbot answers was assessed using the Flesch Reading Ease Score. Completeness and accuracy were evaluated based on corresponding patient drug information in the pharmaceutical encyclopaedia drugs.com. On a preselected subset of inaccurate chatbot answers, healthcare professionals evaluated likelihood and extent of possible harm if patients follow the chatbot's given recommendations.
RESULTS: Of 500 generated chatbot answers, overall readability implied that responses were difficult to read according to the Flesch Reading Ease Score. Overall median completeness and accuracy of chatbot answers were 100.0% (IQR 50.0-100.0%) and 100.0% (IQR 88.1-100.0%), respectively. Of the subset of 20 chatbot answers, experts found 66% (95% CI 50% to 85%) to be potentially harmful. 42% (95% CI 25% to 60%) of these 20 chatbot answers were found to potentially cause moderate to mild harm, and 22% (95% CI 10% to 40%) to cause severe harm or even death if patients follow the chatbot's advice.
CONCLUSIONS: AI-powered chatbots are capable of providing overall complete and accurate patient drug information. Yet, experts deemed a considerable number of answers incorrect or potentially harmful. Furthermore, complexity of chatbot answers may limit patient understanding. Hence, healthcare professionals should be cautious in recommending AI-powered search engines until more precise and reliable alternatives are available.

Keywords

References

  1. J Telemed Telecare. 2025 Feb;31(2):306-308 [PMID: 37350055]
  2. Eur J Hosp Pharm. 2024 Oct 25;31(6):491-497 [PMID: 37263772]
  3. PLoS Med. 2007 Oct 16;4(10):e296 [PMID: 17941714]
  4. Ann Pharm Fr. 2024 May;82(3):507-513 [PMID: 37992892]
  5. Int J Clin Pract. 2010 Dec;64(13):1824-31 [PMID: 21070533]
  6. Radiology. 2023 Jun;307(5):e230922 [PMID: 37310252]
  7. Br J Clin Pharmacol. 2024 Jan;90(1):232-238 [PMID: 37626010]
  8. J Am Pharm Assoc (2003). 2024 Mar-Apr;64(2):422-428.e8 [PMID: 38049066]
  9. BMJ Qual Saf. 2024 Oct 18;33(11):748-754 [PMID: 39048292]
  10. Educ Health (Abingdon). 2017 Jan-Apr;30(1):84-88 [PMID: 28707643]
  11. Qual Health Care. 2001 Sep;10(3):135-40 [PMID: 11533420]
  12. Dtsch Arztebl Int. 2023 May 19;120(20):355-361 [PMID: 37070271]
  13. Am J Med Qual. 2018 Sep/Oct;33(5):487-492 [PMID: 29345143]
  14. J Med Libr Assoc. 2021 Oct 01;109(4):613-617 [PMID: 34858091]
  15. Eur J Clin Pharmacol. 2023 Oct;79(10):1375-1384 [PMID: 37566133]
  16. Nature. 2023 Aug;620(7972):172-180 [PMID: 37438534]
  17. Drug Healthc Patient Saf. 2023 Sep 20;15:137-147 [PMID: 37750052]
  18. J Patient Saf. 2015 Mar;11(1):52-9 [PMID: 24080718]
  19. Res Social Adm Pharm. 2023 Dec;19(12):1590-1594 [PMID: 37696742]

MeSH Term

Generative Artificial Intelligence
Search Engine
Patient Safety
Information Seeking Behavior
Humans
Surveys and Questionnaires
Comprehension

Word Cloud

Created with Highcharts 10.0.0answerschatbotdrugpatientsinformationsearchchatbotssafetypatient0%AI-poweredharm95%CIpotentiallyenginesenginemarketqualityquestions50drugsPatientFleschReadingEaseScoreaccuracyevaluatedsubsethealthcareprofessionalsfollowchatbot'soverall100IQR20expertsfoundharmfulcauseBACKGROUND:SearchoftenserveprimaryresourceobtainHoweverrapidlychangingdueintroductionartificialintelligenceAI-poweredconsequencesmedicationinteractremainlargelyunexploredOBJECTIVE:explorepotentialconcernsprovidedintegratedwithinMETHODOLOGY:Bingcopilotqueried10frequentlyaskedregardingprescribedUSoutpatientcoveredindicationsmechanismsactioninstructionsuseadversereactionscontraindicationsReadabilityassessedusingCompletenessbasedcorrespondingpharmaceuticalencyclopaediacompreselectedinaccuratelikelihoodextentpossiblegivenrecommendationsRESULTS:500generatedreadabilityimpliedresponsesdifficultreadaccordingOverallmediancompleteness0-100881-100respectively66%50%85%42%25%60%moderatemild22%10%40%severeevendeathadviceCONCLUSIONS:capableprovidingcompleteaccurateYetdeemedconsiderablenumberincorrectFurthermorecomplexitymaylimitunderstandingHencecautiousrecommendingprecisereliablealternativesavailableArtificialintelligence-poweredengines:cross-sectionalstudyrisksClinicalpharmacologyInformationtechnologyMedicationPolypharmacy

Similar Articles

Cited By