Video Processing
From Bare-Metal Tinkering to Elastic Video Infrastructure: How ayedo Made Streambase Scalable for …

Data-driven innovation rarely fails due to a lack of ideas. It fails due to infrastructure.
Many industrial companies invest in Data Science, AI models, and Advanced Analytics—only to find that their platform doesn’t scale. GPU resources are scarce, development environments are rigid, and ETL processes are hard to scale. Every new use case becomes an infrastructure project.
In this post, we demonstrate through an anonymized project how ayedo built a Kubernetes-based data engineering platform for a global industrial corporation—for ETL pipelines, event streaming, analytical databases, and GPU-supported AI workloads.
The client remains anonymous. The approach is reproducible—especially for companies looking to combine on-prem stability with cloud flexibility.
The client is an international manufacturer of industrial raw materials with around 10,000 employees. About 30 specialists work in the field of Data Engineering and Advanced Analytics. The goal is to optimize production processes based on data, reduce energy consumption, and use AI models for quality control.
Technically, there was already an in-house orchestration platform based on HashiCorp Nomad and Docker. It worked—but it wasn’t designed to scale dynamically or give data teams autonomy.
Resources had to be requested through a central infrastructure team. GPU capacities were scarce, and on-prem procurement took months. Individual development environments—such as customized Jupyter setups or special R stacks—could only be realized with significant coordination effort.
The result was not only technical friction but also organizational. Data engineers waited for infrastructure instead of training models. ETL pipelines were delayed in going live. Innovation cycles were extended—not for technical reasons, but due to a lack of elasticity.
Especially with AI workloads, time is a critical factor. When training jobs wait weeks for GPU resources, data-driven optimization quickly becomes a strategic bottleneck.
The actual problem wasn’t Nomad or Docker. It was the underlying operating model.
Infrastructure was a centrally controlled bottleneck. Resources were manually planned. Development environments weren’t reproducibly containerized but partially individually configured. GPU usage was tied to fixed hardware.
This led to three structural weaknesses:
First, there was a lack of true self-service capability for data teams.
Second, scaling was linearly tied to physical resources.
Third, reproducibility across projects wasn’t systematically ensured.
To sustainably operate AI, ETL, and BI applications, a platform is needed that decouples compute, storage, and orchestration and makes them declaratively controllable.
Together with the client, we migrated the entire data orchestration to Kubernetes—not as a mere infrastructure upgrade, but as a foundation for a modern, hybrid data platform.
The goal was to create an architecture that meets the following requirements:
Kubernetes is particularly well-suited for these requirements because it standardizes workloads, dynamically allocates resources, and seamlessly integrates into hybrid cloud scenarios.
A central element was the introduction of Coder on Kubernetes.
Data engineers can now start fully containerized development environments on-demand—via browser, RDP, or VS Code Extension. Each environment is versioned, reproducible, and isolated.
This solves several problems:
Development environments are no longer tied to individual workstations.
Configurations are defined as code.
Teams can share and standardize setups.
Instead of “It works on my machine,” there is now a consistent, containerized workspace.
For the orchestration of ETL processes, Apache Airflow was established on Kubernetes. Airflow jobs run containerized and can be horizontally scaled. Compute-intensive transformations can be dynamically distributed to additional workers.
Apache Kafka serves as the event streaming backbone for production data from plants and sensors. Data streams are ingested almost in real-time and distributed to downstream systems.
For analytical workloads, TimescaleDB and ClickHouse were integrated—optimized for time-series and high-volume analyses. Both benefit from Kubernetes resource management and scalable storage.
The platform requires not only compute but also scalable, highly available storage for training data, models, and artifacts.
CEPH was implemented as an S3-compatible backend. It combines high availability with horizontal scalability and allows the separation of performance and capacity requirements.
This enables efficient storage and processing of large data volumes—without proprietary dependencies.
One of the biggest bottlenecks was GPU availability.
Through an integrated cloud-layer architecture, GPU resources can now be dynamically provisioned from European cloud providers. Training and simulation jobs are outsourced to the cloud as needed, without having to adjust the on-prem architecture.
The key here is: The workloads remain Kubernetes-native. There is no separate “cloud version.” Only the location of the cluster changes.
This creates true elasticity without dependency on a single hyperscaler.
The entire platform is integrated into Azure Entra ID. Single sign-on and role-based access control ensure that the data platform remains compliant with the company’s security policies.
Harbor serves as a dedicated container registry with long-term artifact persistence. Models, ETL jobs, and container images are versioned and traceable.
This ensures reproducibility not only technically but also regulatorily.
After migrating to Kubernetes, the working method of the data engineering team has fundamentally changed.
GPU compute is available on-demand. Development environments can be started in minutes. New projects no longer require weeks of coordination with infrastructure teams.
ETL pipelines are containerized, orchestrated, and horizontally scalable. Training jobs can be flexibly shifted between on-prem and cloud.
Above all, the platform is reproducible. Models, pipelines, and environments are versioned. New teams can start immediately without historical legacy or setup issues.
What was previously an organizational bottleneck is now a strategic asset.
The new platform is more than a technical upgrade. It enables sustainable, economically scalable use of AI and big data workloads.
Innovations can be developed iteratively and put into production—without waiting for infrastructure. GPU bottlenecks no longer block roadmaps. New analysis projects can run in parallel without resource collisions.
Kubernetes acts as a universal orchestration layer for compute, storage, and network—independent of the physical location.
Complex ETL pipelines, BI applications, and AI workloads are inherently resource-intensive and dynamic. Rigid infrastructure models are not designed for this.
Kubernetes enables:
Especially in corporate structures, this creates a balance between governance and speed.
If your data engineering team is being slowed down by infrastructure bottlenecks, GPU resources are scarce, or ETL workloads only scale with ticket processes, it’s time for a new platform model.
ayedo supports the development of Kubernetes-based data engineering platforms—with hybrid GPU usage, scalable storage, orchestrated ETL pipelines, and reproducible development environments.
This way, data is not just collected but strategically utilized.
Wir helfen Ihnen, diesen Use Case auf Ihrer Infrastruktur zu realisieren – skalierbar, sicher und DSGVO-konform.
From Bare-Metal Tinkering to Elastic Video Infrastructure: How ayedo Made Streambase Scalable for …
From VM Operation to Platform: How ayedo’s Planwerk Led to Scalable, Auditable SaaS …
From GPU Bottlenecks to Industrial-Scale MLOps: How ayedo Led Sensoriq to a Kubernetes-Based ML …