AI Prompt Engineering

This area of my work investigates how large language models (LLMs) like ChatGPT can be applied to healthcare and public health challenges, with particular focus on understanding how prompt engineering, which is the strategic design of requests to AI systems, affects performance and output.
My recent work demonstrated that incorporating social identity cues (such as political beliefs and education levels) into prompts significantly reduced misinformation detection accuracy, with political identity alone causing the most dramatic decline. Upcoming projects examine LLMs in specialized medical domains—including orthopedic surgery and pharmacy practice.
Related Publications
Haupt, M. R.*, Yang, L.*, Purnat, T., & Mackey, T. (2024). Evaluating the Influence of Role-Playing Prompts on ChatGPT’s Misinformation Detection Accuracy: Quantitative Study. JMIR Infodemiology, 4(1), e60678. *= Co-first authors [LINK] [FILE]
Haupt, M. R., Yang, L., Dinesh, D.,...& Mackey, T. (Under Review). Investigating the Impact of Prompt Engineering on ChatGPT’s Ability to Generate and Evaluate Pharmaceutical Medicines Advice
Haupt, M. R., Massillon, D., Yang, L., Chen, Y., Natarajan, A., Ward, S. R., & Mackey, T. (Under Review). Prompt Engineering Experiment on ChatGPT’s Ability to Recommend Orthopedic Surgeons