THE FUTURE OF EVIDENCE SYNTHESIS AND GENERATION IN HEOR: A SECURE FRAMEWORK FOR DOMAIN SPECIFIC ADAPTATION AND FINE-TUNING OF LARGE LANGUAGE MODELS
Author(s)
Rajdeep Kaur, PhD, Mrinal Mayank, B.Tech, Ruhi j, B.Tech, Shubhram Pandey, MSc, Gagandeep Kaur, M Pharma, Barinder Singh, RPh;
Pharmacoevidence Pvt. Ltd., Mohali, India
Pharmacoevidence Pvt. Ltd., Mohali, India
OBJECTIVES: Large language models enable new applications of artificial intelligence (AI) in healthcare; however, they require domain adaptation and fine-tuning to meet the quality, tone, and domain-specific medical writing tasks. The objective of this proof of concept was to validate a secure, cloud-based framework for fine-tuning of large language models to support writing-intensive HEOR workflows
METHODS: A modular framework was implemented as a proof of concept to enable domain adaptation and fine-tuning of large language models within a secure, cloud-based environment using AWS SageMaker AI to support HEOR workflows. Phase 1: Domain-specific source documents (e.g., published literature, HTA guidelines, dossiers, and study reports) were collected, normalized, chunked, tokenized, and securely stored with encryption. The curated data were then partitioned into training 80% and testing 20% datasets. Phase 2: Fine-tuning was performed on an open-source large language model (LLaMA-2-7B) using multiple adaptation strategies, including QLoRA, LoRA, prefix tuning, adapter tuning. Phase 3: Outputs generated on the testing dataset were assessed for writing quality, tone, style adaptation, and cost considerations, followed by subject matter expert validation to ensure alignment with HTA and regulatory
RESULTS: The framework was validated as a proof of concept using medical writing prompts for dossier, report, and protocol development. Training cost increased with data size across fine-tuning approaches. QLoRA achieved greater cost efficiency but comparatively lower output refinement, whereas LoRA produced higher-quality outputs with improved tonal consistency, clearer structure, and stronger adaptation to medical writing style. Fine-tuned models were generated and deployed securely, supporting the feasibility of the framework for HEOR evidence-generation workflows
CONCLUSIONS: This study demonstrates the feasibility of a secure, cloud-based framework for fine-tuning large language models and presents a novel application of such a framework within HEOR workflows. Future work will focus on enterprise-scale deployment and the integration of domain-adapted models into multi-agent HEOR workflows
METHODS: A modular framework was implemented as a proof of concept to enable domain adaptation and fine-tuning of large language models within a secure, cloud-based environment using AWS SageMaker AI to support HEOR workflows. Phase 1: Domain-specific source documents (e.g., published literature, HTA guidelines, dossiers, and study reports) were collected, normalized, chunked, tokenized, and securely stored with encryption. The curated data were then partitioned into training 80% and testing 20% datasets. Phase 2: Fine-tuning was performed on an open-source large language model (LLaMA-2-7B) using multiple adaptation strategies, including QLoRA, LoRA, prefix tuning, adapter tuning. Phase 3: Outputs generated on the testing dataset were assessed for writing quality, tone, style adaptation, and cost considerations, followed by subject matter expert validation to ensure alignment with HTA and regulatory
RESULTS: The framework was validated as a proof of concept using medical writing prompts for dossier, report, and protocol development. Training cost increased with data size across fine-tuning approaches. QLoRA achieved greater cost efficiency but comparatively lower output refinement, whereas LoRA produced higher-quality outputs with improved tonal consistency, clearer structure, and stronger adaptation to medical writing style. Fine-tuned models were generated and deployed securely, supporting the feasibility of the framework for HEOR evidence-generation workflows
CONCLUSIONS: This study demonstrates the feasibility of a secure, cloud-based framework for fine-tuning large language models and presents a novel application of such a framework within HEOR workflows. Future work will focus on enterprise-scale deployment and the integration of domain-adapted models into multi-agent HEOR workflows
Conference/Value in Health Info
2026-05, ISPOR 2026, Philadelphia, PA, USA
Value in Health, Volume 29, Issue S6
Code
MSR37
Topic
Methodological & Statistical Research
Topic Subcategory
Artificial Intelligence, Machine Learning, Predictive Analytics
Disease
No Additional Disease & Conditions/Specialized Treatment Areas