Qualitative Literature Reviews: A Comparison of Researcher and AI Screening of Articles to Inform Conceptual Model Development

Author(s)

Burbridge C1, Lloyd-Price L2, Hudgens S3, Thorlund K4
1Clinical Outcomes Solutions, Ltd., Folkestone , UK, 2Clinical Outcomes Solutions Ltd, Folkestone, Kent, UK, 3Clinical Outcomes Solutions, Tucson, AZ, USA, 4McMaster University, Hamilton, ON, Canada

OBJECTIVES: Artificial intelligence (AI) models are being used in systematic literature reviews, reducing researcher burden and improving efficiency. However, in structured (not fully systematic) literature reviews when terminology and reporting is not formal or standardized within the literature, as in reviews to identify qualitative research and insights exploring the patient lived experience, it can be challenging to develop focused yet comprehensive search strategies and screening criteria without compromising results. A novel AI model is being developed and trained specifically to facilitate the expert researcher in screening literature for qualitative reviews.

METHODS: Data from 27 medical literature database reviews (5671 citations overall, ranging from 20 to 942 per review) was used to compare researcher and AI title/abstract screening decisions. The reviews were previously conducted to identify qualitative research across a number of conditions to inform the development of conceptual models of the patient experience. Screening decisions were annotated using researcher developed eligibility screening criteria based on PICO principles adapted for the specific context of a qualitative review: Population (search dependent), study design (qualitative research), and outcomes/concepts of interest (patient experience).

RESULTS: Level of agreement between researcher and AI screening decisions was 86% overall, ranging from 44% to 100% across individual screening criteria. For all but 4 searches, agreement was 75% or above on all screening criteria. The main criteria on which there was discrepancy was population, which comprised individual of interest and condition. Time to screen reduced from a skilled researcher screening approximately 30-40 citations/hour to the AI software screening approximately 1000 citations/hour.

CONCLUSIONS: There is a high level of agreement between expert researcher and the AI model in title/abstract screening, highlighting the potential of AI to facilitate the researcher in efficient screening for qualitative literature reviews, supporting research informing the development of conceptual models of the patient experience in the context of COA research.

Conference/Value in Health Info

2024-11, ISPOR Europe 2024, Barcelona, Spain

Value in Health, Volume 27, Issue 12, S2 (December 2024)

Code

MSR194

Topic

Clinical Outcomes, Methodological & Statistical Research, Patient-Centered Research, Study Approaches

Topic Subcategory

Artificial Intelligence, Machine Learning, Predictive Analytics, Clinical Outcomes Assessment, Literature Review & Synthesis, Patient-reported Outcomes & Quality of Life Outcomes

Disease

No Additional Disease & Conditions/Specialized Treatment Areas

Your browser is out-of-date

ISPOR recommends that you update your browser for more security, speed and the best experience on ispor.org. Update my browser now

×