120 GB from Kafka to ClickHouse in 4 minutes
Recently, we conducted a performance test of Megaladata using a realistic ETL workflow – without synthetic benchmarks, simplified datasets, or artificial optimizations.
Test setup
- Data source: Apache Kafka
- Data volume: 120 GB
- Message format: JSON
- Target database: ClickHouse
- Operating system: Linux
- Megaladata Server: 24 CPU cores | 64 GB RAM
- Platform version: Megaladata 7.3.0
Workflow
The workflow is as close to a real production scenario as possible:
- JSON messages consumed from Apache Kafka
- Parsing into a flat table
- Lightweight data transformations
- Statistics calculation
- Loading results into ClickHouse
In other words, a typical stream → transform → analytical storage (ETL) pipeline that many teams run every day.
Result
- Total execution time: 4 minutes 2 seconds
120 GB of data was processed from Kafka to ClickHouse – including parsing, transformations, and calculations.
No Spark cluster. No manual JVM tuning. No complex code.
A video demonstrating the full execution of the test is available below.
Why this matters
ETL platform performance is often tested:
- On small datasets,
- Without real transformations,
- Or in laboratory conditions far from production.
Here, this is a realistic, honest scenario that can be easily transferred to real systems:
- Streaming analytics
- Logging
- Telemetry
- Event processing
- Transactional data streams
- IoT
- Finance
What this shows in practice
- Megaladata comfortably processes tens and hundreds of gigabytes of data
- Suitable not only for analytics, but also for heavy ETL / streaming workloads
- Low-code does not mean “slow”
- The platform scales vertically and uses hardware efficiently
Conclusion
Megaladata is not just a “visual Excel for analysts.” It is a full-featured, high-performance data processing platform that can compete with traditional ETL stacks while remaining simple to use.
If you need to quickly ingest data from Kafka, transform it, calculate metrics, and load it into ClickHouse or another analytical database, this test speaks for itself.
A video of the test execution:
See also