Most companies are adopting AI features and putting them into production, with 66.2% having at least one internal use case live and 43.1% having at least one external use case live. However, building and productionizing AI apps is challenging due to issues such as model output accuracy, hallucinations, and prompt engineering. Companies are seeking ways to evaluate their AI apps, including how to integrate evaluations with Continuous Integration and Continuous Deployment (CICD) systems and how to score AI outputs using tools like GPT-4. Many companies are either manually checking their AI's work or not checking it at all, highlighting the need for efficient evaluation tools. Braintrust offers a platform that enables users to add evaluations to their AI apps in under 10 minutes, providing a convenient solution for companies looking to streamline their internal tool development.