/plushcap/analysis/cloudflare/http-analytics-for-6m-requests-per-second-using-clickhouse

HTTP Analytics for 6M requests per second using ClickHouse

What's this blog post about?

Cloudflare has evolved its analytics pipeline over the last year to provide HTTP traffic analytics to its customers. The old data pipeline was built in 2014 and faced challenges as the company grew, including a single point of failure (SPOF) for Postgres and Citus main, complex codebase, high maintenance cost, and many dependencies. To address these issues, Cloudflare replaced the old pipeline with a new one based on ClickHouse, an open-source column-oriented database management system. The new data pipeline is more scalable, fault-tolerant, and easier to operate, providing analytics for over 7 million customers' domains. Additionally, historical data was transferred from the old Citus cluster into ClickHouse during the switch. Looking forward, Cloudflare plans to develop products like Log Push and Logs SQL API to enhance its analytics offerings.

Company
Cloudflare

Date published
March 6, 2018

Author(s)
Alex Bocharov

Word count
3831

Hacker News points
3

Language
English


By Matt Makai. 2021-2024.