Stream Query is a new API endpoint that reduces perceived latency by delivering search results first and then streaming summary responses in small chunks, eliminating the frustration of waiting for Large Language Models (LLMs) like GPT-4 to generate complete responses. This approach virtually eliminates perceived latency, offering a smooth and continuous interaction that keeps pace with the speed of thought. By returning search results first and then sending generative summary responses as they become available, Stream Query reduces the wait time for LLMs, enhancing overall responsiveness and enriching the user experience. The API endpoint is designed to be easy to use, with complementary concatenation tools to seamlessly integrate it into a fluid user experience.