A NEXT-GENERATION FRAMEWORK FOR AI-AUGMENTED, SUBMISSION-GRADE LITERATURE REVIEWS: A MULTI-PROJECT ANALYSIS OF SPEED, ACCURACY, AND WORKFLOW EVOLUTION

Author(s)

Angeline Babitha Dhas, BS1, Revanth M, B.E2, Diwyashri Govindarajaperumal, B.Pharm2, Siva Karthick Bagavathiappan, B.E.3, Vince Salerno, PharmD, RPh1, Viji Queen, PharmD1.
1MadeAi, Cambridge, MA, USA, 2MadeAi, Nagercoil, India, 3MadeAi, Etobicoke, ON, Canada.
OBJECTIVES: Although AI has been applied to literature reviews for several years, adoption within regulatory-oriented evidence workflows remains limited due to transparency, reproducibility, traceability, and methodological concerns. This study evaluates a standardized, submission-grade framework for AI-augmented literature reviews, assessing its impact on speed, accuracy, workflow efficiency, and ROI across multiple real-world regulatory projects.
METHODS: A comparative analysis was conducted across five regulatory literature reviews, supporting CERs and PSURs. One project was completed using a fully manual workflow as a baseline, while four projects applied an AI-augmented methodology using a structured framework. All followed an identical process, including protocol development, database searching, deduplication, title/abstract screening, full-text screening, data appraisal, evidence table creation, and PRISMA generation. In AI-augmented projects, AI was incorporated across all stages except the baseline. Human reviewers adjudicated screening decisions, verified extracted data, and performed final quality control. Metrics included time spent at each stage, AI-output accuracy, workflow efficiency, traceability, and ROI.
RESULTS: The AI-augmented framework demonstrated progressively increasing time savings, achieving a 32% time savings (initial adoption project), which increased to 33-51% in subsequent projects, highlighting sustained gains as workflows matured. Time savings increased even as article volume expanded nearly 19 times (120 to 2,300 articles), demonstrating strong scalability and hours-per-article efficiency. AI reduced the manual effort in screening, data extraction, and evidence table drafting, while human reviewers maintained regulatory accuracy and oversight. AI accuracy across protocol development, screening, and data extraction remained consistently high (84-90%), enabling reliable outputs with reduced manual effort and shorter quality-control cycles. Human reviewer oversight ensured regulatory accuracy and methodological defensibility throughout all projects.
CONCLUSIONS: This multi-project evaluation demonstrates that a structured, transparent AI-augmented framework can deliver substantial efficiency gains while maintaining accuracy and regulatory rigor. The approach enables scalable, defensible adoption of AI in submission-grade literature reviews and supports its integration into regulatory evidence-generation workflows.

Conference/Value in Health Info

2026-05, ISPOR 2026, Philadelphia, PA, USA

Value in Health, Volume 29, Issue S6

Code

SA38

Topic

Study Approaches

Topic Subcategory

Literature Review & Synthesis

Disease

No Additional Disease & Conditions/Specialized Treatment Areas, STA: Multiple/Other Specialized Treatments

Your browser is out-of-date

ISPOR recommends that you update your browser for more security, speed and the best experience on ispor.org. Update my browser now

×