Learn how to configure Kerberos for Kafka and ClickHouse to benefit from a centralized authentication and authorization service. This article explains how to deploy Kafka, Zookeeper and Clickhouse and configure them to authenticate via Kerberos showcasing the process with a docker compose deployment.
Want fast MySQL analytics? Then check out the Altinity Sink Connector for ClickHouse. Merging the power of ClickHouse with MySQL, the sink connector lets you replicate data from MySQL to ClickHouse in real-time.
eBay depends on Kafka to solve the impedance mismatch between rapidly arriving messages in event streams and efficient block insert into ClickHouse clusters. Naïve loading procedures from Kafka to ClickHouse generate non-deterministic blocks, leading to data loss and incorrect results in applications. Learn how the eBay team solved this problem.
Our colleague Mikhail Filimonov just published an excellent ClickHouse Kafka Engine FAQ. It provides users with answers to common questions about using stable versions, configuration parameters, standard SQL definitions, and many other topics. Even experienced users are likely to learn something new.
But what if you are getting started and need help setting up Kafka and ClickHouse for the first time? Good news! This article is for you.
Kafka is a popular way to stream data into ClickHouse. ClickHouse has a built-in connector for this purpose — the Kafka engine. This article collects typical questions that we get in our support cases regarding the Kafka engine usage. We hope that our recommendations will help to avoid common problems.
Important notice for our beloved Apache Kafka users. We continue to improve Kafka engine reliability, performance and usability, and as a part of this entertaining process we have released 126.96.36.199 Altinity Stable ClickHouse release. This release supersedes the previous stable release.
Apache Kafka is a popular way to load large data volumes quickly to ClickHouse. In this webinar we will cover best practices for integrating Kafka and ClickHouse including setup of Kafka clusters, defining materialized views to pull data into ClickHouse, and organization of target tables.
Apache Kafka is a popular way to load large data volumes quickly to ClickHouse. These presentation slides will cover best practices for integrating Kafka and ClickHouse from our webinar hosted on January 21, 2020.