Leveraging large language models (LLMs) in critical business processes, customer-facing agents, or compliance-driven scenarios requires accurate, contextual, and verifiable information to ensure accuracy. Establishing a reliable ground-truth dataset, which includes questions and validated answers representing the correct responses for a given domain, is key. However, generating such a dataset can be costly, complex, and labor-intensive. A new toolkit enables organizations to automate this process by harnessing the power of LLMs themselves and using an image-based workflow that preserves the original layout and structure of documents. This approach delivers more accurate and reliable ground truth datasets, faster than traditional methods, by considering every element, including table cells, images, captions, and layout nuances. By anchoring LLMs in authoritative sources, organizations can ensure answers are both domain-relevant and contextually precise, reducing hallucinations, fostering trust, and supporting compliance. The toolkit also provides a streamlined workflow for creating and refining ground truth datasets, making it easier to build production-grade language models.