Analytical Databases in the Cluster: ClickHouse and TimescaleDB for High-Volume Data
In an industrial concept, millions of data points are generated daily. When these data flow into …

In modern manufacturing, data is generated not in batches, but as a continuous stream. Sensors on rolling mills, flow meters in chemical reactors, and logistics systems produce status messages every second. Those who analyze this data only in nightly batch runs miss the opportunity for immediate reaction—whether in the case of quality deviations or impending machine failures.
To harness this “data tsunami,” we rely on Apache Kafka within the Kubernetes cluster. Kafka acts as a highly available digital nervous system that ingests, stores, and distributes events in real-time to the appropriate analysis tools.
In traditional industrial architectures, machines are often directly connected to a database or a specific application, leading to rigid dependencies. Kafka breaks this pattern:
Running Kafka on Kubernetes (often supported by operators like Strimzi) provides the necessary elasticity for fluctuating production loads:
A concrete example from an industrial corporation:
Apache Kafka on Kubernetes is far more than just a data transport medium. It is the technological prerequisite for true Industry 4.0. By decoupling data sources and analysis applications, we create a flexible, highly scalable infrastructure that grows with the demands of production. This way, we transform fleeting sensor data into valuable, immediately usable knowledge.
Isn’t Kafka too complex for smaller data volumes? Kafka shows its full strength with large volumes but also offers the advantage of clean architectural separation in smaller setups. For very simple use cases, lighter brokers like NATS can be an alternative—in a Kubernetes environment, this can be flexibly decided.
How secure are the data streams in Kafka? We use end-to-end TLS encryption for transmission and strict authentication (e.g., via SASL or certificates). Within the corporate network, this ensures that only authorized systems have access to sensitive production data.
Can Kafka also deliver historical data? Yes. Kafka is not a transient storage. Depending on configuration, data can be retained for days, weeks, or months (retention). This allows new AI models to be “retrained” with real historical data streams.
What happens in the event of a complete network outage at the plant? Local gateways at the plants buffer the data on-site until the connection to the central Kafka cluster in the data center is restored. Kafka then ensures seamless synchronization of the buffered events.
How does ayedo support the setup of event streaming? We not only implement the Kafka cluster on Kubernetes, but also advise you on the design of topics and the integration of your existing SCADA or ERP systems. We ensure stable monitoring of data streams so that your real-time pipeline runs reliably 24/7.
In an industrial concept, millions of data points are generated daily. When these data flow into …
Scaling at the Limit: How Track & Trace Processes Millions of Events in Real-Time During the …
Data Warehouse as a Cloud Product or as an Open Analytical Platform Analytical data is no longer …