Chansarkar, Anupam (2025) LLM cross-validation frameworks: Mitigating hallucinations in enterprise content generation systems. World Journal of Advanced Engineering Technology and Sciences, 15 (2). pp. 1721-1728. ISSN 2582-8266
![WJAETS-2025-0722.pdf [thumbnail of WJAETS-2025-0722.pdf]](https://eprint.scholarsrepository.com/style/images/fileicons/text.png)
WJAETS-2025-0722.pdf - Published Version
Available under License Creative Commons Attribution Non-commercial Share Alike.
Abstract
This article examines the efficacy of using one language learning model (LLM) to validate the outputs of another as a quality assurance mechanism in content generation workflows. Drawing from a comprehensive experiment conducted during the Prime Video Project Remaster Launch, it demonstrates the implementation of a dual-LLM verification system designed to detect and reduce hallucinations in automatically generated book summaries. It also demonstrates that while LLM cross-validation significantly improves content accuracy through iterative prompt refinement and systematic error detection, it cannot completely eliminate hallucination issues inherent to generative AI systems. This article provides valuable insights for organizations seeking to balance the efficiency of automated content generation with the need for factual accuracy, particularly in customer-facing applications where trust and reliability are paramount.
Item Type: | Article |
---|---|
Official URL: | https://doi.org/10.30574/wjaets.2025.15.2.0722 |
Uncontrolled Keywords: | LLM Cross-Validation; Hallucination Mitigation; Prompt Engineering; Content Verification; Generative Ai Reliability |
Depositing User: | Editor Engineering Section |
Date Deposited: | 04 Aug 2025 16:30 |
Related URLs: | |
URI: | https://eprint.scholarsrepository.com/id/eprint/3881 |